AN INVESTIGATION INTO THE EFFICIENT MARKET HYPOTHESIS A CANONICAL CORRELATION ANALYSIS APPROACH by DAREN MCCROSSAN SMITH B.Sc, University of Western Ontario, 1993 A THESIS SUBMITTED IN PARTIAL FULFILLMENT OF THE REQUIREMENTS FOR THE DEGREE OF MASTER OF SCIENCE in THE FACULTY OF GRADUATE STUDIES Department of Statistics We accept this thesis as conforming to the required standard THE UNIVERSITY OF BRITISH COLUMBIA October 1995 ©Daren Smith, 1995 In presenting this thesis in partial fulfilment of the requirements for an advanced degree at the University of British Columbia, I agree that the Library shall make it freely available for reference and study. I further agree that permission for extensive copying of this thesis for scholarly purposes may be granted by the head of my department or by his or her representatives. It is understood that copying or publication of this thesis for financial gain shall not be allowed without my written permission. Department The University of British Columbia Vancouver, Canada Date Qof- /J/fr DE-6 (2788) ABSTRACT .1 In this thesis we will consider the Efficient Market Hypothesis (EMH). Fama (1970) defined three levels in which to test market efficiency: weak, semi-strong, and strong, each level depending on the particular set of information being used to assess efficiency. We will mainly address weak level efficiency in which the information set is past security data. Before the mid 1980's it was widely believed that the EMH was true at the weak and semi-strong levels. It was not until the pioneering work of Shiller (1984) and Summers (1986) that some doubt was cast on the EMH. They proposed an inefficient model in which prices consist of a sum of a random walk component and a stationary (predictable) component which represents the market valuation error. Since their initial conjecture about a stationary component in stock prices much effort has been spent in trying to determine if it exists and if it does, determining how much of the variations in stock prices it accounts for. To investigate this problem we will use a combination of data filtering, canonical correlation analysis, simulations and bootstrapping. Using industry price data obtained from the Toronto Stock Exchange over the period January 1956 to June 1995, we find some evidence against the EMH. ii Table of Contents Abstract ii Table of Contents iiList of Tables vi List of Figures viList of Nomenclature viii Acknowledgements ix 1 Introduction 1 1.1 Investigations into the Efficient Market Hypothesis (EMH) 1 1.1.1 The EMH universally accepted 3 1.1.2 A digression: Terminology in the EMH debate explained 5 1.1.3 The EMH questioned 7 1.1.4 Evidence against the validity of the EMH 11 1.1.5 A re-evaluation of the F&F (1988) Results 2 1.1.6 Mounting evidence that the EMH does not hold 13 1.1.7 The EMH re-considered 11.2 Our proposal 14 1.3 Layout of the thesis 5 1.4 The Datain 2 The Mean-Reverting Price Component 21 2.1 Our Method and Rationale 22.2 On the use of returns to avoid the confounding problem 22 2.3 Details of the Canonical Correlation Analysis 23 2.3.1 Theoretical Canonical Correlation Analysis 22.3.2 Sample Canonical Correlation Analysis 4 2.4 A New Look at Canonical Correlation Analysis 26 2.4.1 Historical Overview 22.4.2 Filters 22.4.3 Uses of Filtering 8 2.4.4 Filtering in a time-series setting . 29 2.4.5 Filtering in an investment setting 30 2.4.6 Filtering used to solve an identifiability problem 31 2.5 ARIMA modeling and diagnostics 32 2.5.1 ARIMA modeling 32.5.2 ARIMA diagnostics 2 2.5.3 Using the acf to check for serial dependence in residuals 32 2.5.4 Using a runs test to check for serial dependence in residuals ... 34 2.5.5 The Ljung-Box test for lack of fit 32.6 The Dickey Fuller Unit Root Test 5 2.7 Variation in prices explained by stationary components 36 2.8 Procedure that was followed 32.9 The Data Revisited 7 2.10 Results 38 2.10.1 Results using a linear lag-1 prediction filter 41 2.10.2 Results using a linear lag-2 prediction filter 2 iv 2.10.3 Results using a quadratic lag-1 prediction filter 43 2.10.4 Results using a quadratic lag-2 prediction filter 44 2.10.5 Results using a square Root lag-1 prediction filter 45 2.10.6 Results using a square Root lag-2 prediction filter 46 2.11 Testing the significance of the i?2's 47 2.12 Bootstrapping 42.12.1 Bootstrapping without standardization 48 2.12.2 Bootstrapping with standardization 9 2.13 Simulations 50 2.13.1 Simulations with no standardization 51 2.13.2 Simulations with standardization 2 3 Discussion 53 3.1 Comparison of the three predictive information filters 53 3.2 The significance of the R2,s and their relevance to the EMH . 55 3.3 Why the quadratic information filter results in the highest R2,s 56 3.4 Comparison with Tsay's (1990) results 53.5 Future Developments 57 3.6 Conclusions 8 3.7 Final Comments 59 Bibliography 60 Appendix 2 v List of Tables 1.1 The 14 industries of the TSE 17 2.1 The industry weights of each portfolio 39 2.2 R2,s under the RW Null using bootstrapping vs actual .......... 48 2.3 R2,s under the RW Null using bootstrapping vs actual .......... 49 2.4 i?2's under the RW Null using simulations vs actual 51 2.5 R2,s under the RW Null using simulations vs actual 52 3.1 Summary of predictive information filter R2,s without standardization . . 53 3.2 Summary of predictive information filter R2,s with standardization ... 55 A.l ARIMA models using linear lag-1 prediction 71 A.2 Portfolio R2,s using linear lag-1 predictionA.3 ARIMA models using linear lag-2 prediction 72 A.4 Portfolio R2,s using linear lag-2 prediction . 72 A.5 ARIMA models using quadratic lag-1 prediction 73 A.6 Portfolio industry R2,s using quadratic lag-1 prediction 73 A.7 ARIMA models using quadratic lag-2 prediction 74 A.8 Portfolio industry i?2's using quadratic lag-2 prediction .......... 74 A.9 ARIMA models using square-root lag-1 prediction 75 A. 10 Portfolio industry R2,s using square-root lag-1 prediction 75 A. 11 ARIMA models using square-root lag-2 prediction ............. 76 A.12 Portfolio industry i?2's using square-root lag-2 prediction 76 vi List of Figures 1.1 Simulated AR(1) series with varying coefficients 9 1.2 Plots of monthly TSE industry portfolio indexes . 18 1.3 Plots of monthly TSE industry portfolio indexes 9 1.4 Plots of monthly TSE industry portfolio indexes 20 A.T Quantiles of R2 using bootstrapping with original data 77 A.2 Quantiles of R2 using bootstrapping with standardized data 78 A.3 Quantiles of R2 using simulations with original data 79 A.4 Quantiles of R2 using simulations with standardized data. ........ 80 A.5 Q-Q plots of monthly log(index) for TSE industry portfolios 81 A.6 Q-Q plots of monthly log(index) for TSE industry portfolios 82 A.7 Q-Q plots of monthly log(index) for TSE industry portfolios ....... 83 vii List of Nomenclature vec(-) - The column stacking operator. V - Extracts the upper diagonal portion of the matrix it is applied to B - is the backshift operator, for example BYt — Yt — Yt_\. ~ - this is read "distributed as". Np(fj,, S) - this represents a p-dimensional normal vector. £(•) - this denotes a linear portfolio filter. - this denotes an information filter. - this is read "converges in distribution". viii Acknowledgements I would like to thank my supervisor, Dr. Jian Liu, for providing me with a very interesting topic to investigate in this thesis. I would also like to thank Dr. Ruben Zamar for his comments and suggestions on improving the manuscript. Thanks go out also to Wade Blanchard, who provided much of the data that I worked with and also valuable latex scripts. I would like to thank the Statistics Department as a whole for making my two years in Vancouver enjoyable and rewarding. ix Chapter 1 Introduction 1.1 Investigations into the Efficient Market Hypothesis (EMH) The capital market's primary role is the allocation of ownership of the economy's capital stock. Its secondary role is to determine the current price of a stock. The ideal market is one in which the prices determined by the market are accurate indicators of the real value of the stock price. Such an ideal market in which prices always fully reflect all available information is called efficient. It has been debated for many years whether the capital markets are efficient or not. Generally speaking a believer in the EMH presumes the following two statements to be valid. 1) The market has a very efficient pricing mechanism which is capable of interpreting fundamental economic realities and arrive at a price which comes close to reflecting the true worth of the security. It is able to do this because the market investors are acting rationally. 2) Except for a very small number of highly skilled professionals the time and effort spent in trying to locate mis-priced stocks is not worth the reward. This means that the market may not reflect the "true" price, but the difference is so small that almost no gain can be achieved by revealing the discrepancy. On the other hand, a non-EMH believer would say the following: 1) The prices determined by the market are not reflective of fundamental economic realities, since they are determined in part by irrational investors. 2) It is highly probable that with a little effort abnormally high returns can be earned. This is because the market price and the "true" price may 1 differ by a large amount and revealing the difference can lead to a large financial gain. As can be seen, these two views are diametrically opposed to one another. At the heart of the difference is the EMH belief in rational investors and little or no potential for abnormally high returns (market exploitability), and the non-EMH belief that some investors act irrationally and that this can lead to prices which deviate from "true" values enough that revealing the difference can lead to large financial rewards. Rationality of investors and non-exploitability of the market have in the past been considered as equiv alent, but in fact they should be considered as separate entities. For example, irrational investing is a necessary condition for market exploitability but it is not sufficient. This implies that even if the market value is substantially different from the "true" value it may not be possible for investors to tell the difference (for example West (1988)). In the early stages of research into the EMH most of the effort was spent addressing the possi bility of earning abnormally high returns and only recently have researchers considered the rationality of investors. If either irrationality in market prices or exploitability of the market can be shown than the EMH has been proven invalid. The implications of whether the EMH can be considered as a reasonable depiction of the market are very far-reaching. Much of the modern theory of finance and economics has been built up on the assumption that the EMH is valid, in particular the assumption of a rational investor (e.g the Capital Asset Pricing Model (CAPM)). This debate also concerns the average investor, since if the market can be considered as efficient than the optimal portfolio investing strategy in light of transaction costs is to buy and hold, whereas if the market is not efficient than it would be more profitable to pursue an active investment strategy. Much of the early work on the EMH debate is summarized by Fama (1970) in a survey paper entitled "Efficient Capital Markets: A review of theory and empirical work" [7]. 2 1.1.1 The EMH universally accepted Fama begins by dividing "information" into three subsets, which allows the EMH to be tested at three levels of efficiency. • The first level is what he calls weak efficiency, in which the information set is simple historical prices. At this level the market is efficient if tomorrow's price moves independently of any past prices. This implies that any "trading rules" or attempts to determine the future price by studying charts of past price data is useless. • The second level is semi-strong efficiency, where the information set is all publicly available information (e.g., quarterly reports, announcements of annual earnings, stock splits, etc.). At this level it is of interest to determine if prices respond instantaneously and in an unbiased manner to new information. Efficiency at this level would imply that the market is able to immediately arrive at the best interpretation of new information. • The third level is strong efficiency, which is concerned with whether certain investors or groups who have monopolistic access to information can use this knowledge to predict future prices. In other words this level is concerned with whether the market incorporates all published and unpublished information in its determination of the price. If the market is efficient at this level it implies that insider trading would not be successful. Each of the three levels of efficiency has its own collection of statistical tests, called weak form level tests, semi-strong form tests, and strong form tests. It should be noted that the above three levels of efficiency are not independent. If the market is inefficient at the weak level than it will be inefficient at both of the other two levels. As well, inefficiency at the semi-strong level implies inefficiency at the strong level. Fama perceptively notes that almost all of the empirical literature up to 1970 is based on the assumption that the conditions of market equilibrium can be stated in terms of expected returns. Notationally speaking, this can be summarized in the following 3 equation: E{pj,t+1\$t) = [l + E(rj,t+i\$t)]pjt where E is the expected value operator: pjt is the price of security j at time t; Pj,t+i is its price at t+1; r,it+1 is the one-period percentage return {pj,t+i — Pjt)/Pjt', ®t 1S a general symbol for whatever set of information is assumed to be "fully reflected" in the price at time t; and the tildes indicate that Pjtt+i and rjit+i are random variables at t (because they are unknown at time t and will not be known until they are realized at time t+1). For all three sets of "information", theories were developed and tested. In the class of weak form tests almost all of the literature up to 1970 supported the EMH. In fact, so much support was found for the EMH at this level that most researchers stopped testing for it and switched to testing at the semi-strong form level, assuming that all work at the weak forms level was completed. At the semi-strong level as well, the consensus was that the market was efficient. Finally, at the strong form test level a few instances where individuals were able to exploit inside information for profit were documented. The first example concerned specialists on major security exchanges who had monopolistic access to information on unexecuted limit orders and who were able to utilize their information to generate trading profits. The second example concerned corporate insiders who had monopolistic access to information concerning their firms. Besides these two cases there was little or no evidence against the EMH at any of the three levels of "information". Fama's paper was published in 1970 and the views expressed in it were wide spread among statisticians and econometricians alike. This is what the great economist John Maynard Keynes had to say about the subject. Professional investment may be likened to those newspaper competitions in which the competitors have to pick out the six prettiest faces from a hundred photographs, the prize being awarded to the competitor whose choice most nearly corresponds to 4 the average preferences of the competitors as a whole; so that each competitor has to pick not those faces which he himself finds prettiest, but those which he thinks likeliest to catch the fancy of the other competitors, all of whom are looking at the problem from the same point of view. It is not a case of choosing those which, to the best of one's judgement, are really the prettiest, nor even those which average opinion genuinely thinks the prettiest. We have reached the third degree where we devote our intelligences to anticipating what average opinion excepts the average opinion to be. And there are some, I believe, who practice the fourth, five and higher degrees. [13] This may not be the case, but what it does indicate is a belief that investors are merely guessing at future price movements and therefore that the EMH is true. This was a belief that many people held. 1.1.2 A digression: Terminology in the EMH debate explained In the sections that follow certain statistical concepts relating to time series analysis will be used in the discussion of approaches to resolving the EMH debate. Some of these concepts will be explained now. Let Yt represent a general time series. Definition 1.1 (A white noise time series) Yt = et where, E(et) = 0 Cov(ejtk) — 0 Definition 1.2 (A random walk (RW) time series) Yt = n + Yt^ + et Var{Yt) = ta2 where et follows a white noise process. Definition 1.3 (An autoregressive process of order p, AR(p)) Yt = 7 + + a2Yt-2 + ... + apYt-p + et where et follows a white noise process. When p = 1 we have an AR(1) time series which has the following variance: Definition 1.4 (Stationary time series) A stationary time series Yt must satisfy the following two conditions. If it does not satisfy them it is referred to as nonstationary. E(Yt) = constant, does not depend on t Cov(Yt,Yt-s) — a2, does not depend on t Of the time series we have considered above, the AR(1) process with a < 1 and the white noise process are stationary. The random walk is nonstationary. Definition 1.5 (Predictability of a time series) This refers to the noise to signal variance ratio of a stationary time series given by: Var{et) Var{Yt) Greater predictability implies a lower noise to signal variance ratio. This ratio is 1 — cx\ for an AR(1) process and 1 for a white noise process indicating that it is stationary but unpredictable. As a increases in the AR(1) process, but is less than 1, greater predictabil ity is obtained. 6 Much of the early work in support of the EMH claimed that prices followed a RW. This implies that prices and their first differences returns (which are white noise if prices follow a RW), do not have any predictable components. 1.1.3 The EMH questioned It was not until 1986 that someone seriously challenged the EMH. In his seminal paper "Does the Stock Market Rationally Reflect Fundamental Values" [19] Summers (1986) submitted that the current statistical rationale for regarding capital markets as efficient was somewhat flawed. He noted that most of the early work on market efficiency was a result of analysing the auto-correlations of daily and weekly stock returns. The common consensus of the studies was that the estimated auto-correlations were very close to zero. This implied that there was no significant predictability of returns and hence the capital markets were efficient. His main contribution was to demonstrate that the auto-correlations of short-horizon returns can give the impression that the predictable components of prices are of no aid in prediction when in fact they can actually explain a moderate amount of return variations. As Summers says, the existing evidence does not establish that financial markets are efficient in the sense of rationally reflecting fundamentals. It demonstrates that the types of sta tistical tests which have been used to date have essentially no power against at least one interesting hypothesis to market efficiency [19]. Summers hypothesized the following inefficient model for prices: Pt = Pt* + ut (1.1) ut = aut-i + vt where Pt is the market price, P* is the market efficient price which follows a random walk (RW), {ut} are the valuation errors which follow a first-order autoregressive 7 process (AR(1)), with errors given by {vt}. This model is commonly referred to as the "simple fads" model. The reasoning behind this name is explained well by Shiller (1984), who says: Investing in speculative assets is a social activity. Investors spend a substantial part of their leisure time discussing investments, reading about investments, or gossiping about others' successes or failures in investing. It is thus plausible that investors' behavior (and hence prices of speculative assests) would be influenced by social movements. Model (1.1) is inefficient and partially predictable because the valuation errors {ut} are predictable as they represent persistent errors with a defined structure to them. If the {ut} were white noise than Model (1.1) would still be inefficient but there would be no predictable component to it. The random walk component P* is commonly referred to as the permanent component of prices and ut as the temporary component. The AR(1) component is called temporary because it is assumed that it is stationary and that even though it may differ from its constant expected value it will eventually revert to its mean. The larger the AR(1) coefficient the longer on average the process will take to return to its expected value. This is exemplified by the plots in Figure 1.1. Six AR(1) processes were simulated all with expected value zero. The process with the smallest a is highly variable and returns to its expected value very quickly; as a increases the {ut} become less variable but take longer swings away from their mean. Summers assumed an a of 0.98 for the {ut} process, which allows prices to take long temporary (since it will return to its mean eventually unlike a nonstationary series) swings away from fundamental values given by P*. This has led some authors to refer to the investigation of a temporary component in prices as a search for mean reversion in prices. 8 Simulated AR(1) series, alpha = .1 Simulated AR(1) series, alpha = .3 Simulated AR(1) series, alpha = .8 0 100 200 300 400 Time Simulated AR(1) series, alpha = .975 Simulated AR(1) series, alpha = .9 100 200 300 400 Time Simulated AR(1) series, alpha = .99 3 O 100 200 300 400 Time Figure 1.1: Simulated AR(1) series with varying coefficients 9 Summers argued along the lines of Shiller (1984) that Model (1.1) is a reasonable hypothesis and also that it allows many instances of a market failing to rationally reflect fundamental values. Using Model (1.1) and a combination of assumed and empirically derived parameter values he conducted a weak form efficiency test. It is important to note that he is not saying his price model is the true one, only that it is reasonable and that the current statistical techniques would not be able to identify it if it was true. This basis for this is a belief that there may be many alternative models to efficiency which can also not be identified. In his analysis using monthly stock market returns over a 50-year period he showed that the standard statistical model could not reject the hypothesis of market efficiency. This was despite the fact that the market price frequently differed from the market efficient price by more than 30 percent. He also conducted a semi-strong form test of market efficiency. The main conclusions of which were again that the common statistical tests of semi-strong efficiency had very little power against detecting his inefficient model for prices. Summers makes a point of noting that he is not implying that abnormally high returns should be easy to make if prices actually follow his model. Instead he argues that for the same reasons it is very difficult for statisticians to uncover deviations of market prices from fundamental values it will be even more so for the average investor. The main point of his paper is to say that yes it may be exceedingly difficult to make abnormally high returns but that this does not imply that market prices reflect rational assessments of fundamental values which many people heretofore had assumed. Thus Summers was able to demonstrate that at the weak form test level and the semi-strong form test level the EMH did not necessarily hold. 10 1.1.4 Evidence against the validity of the EMH The additive components model for prices proposed by Summers spawned a flurry of research to determine if prices actually followed such a model. The most famous study was by Fama and French (hereafter F&F) (1988) entitled "Permanent and Temporary Components of Stock Prices" [8]. They proposed to investigate Summer's model by analysing the auto-correlations of long-horizon returns. Their rationale is given below: A slowly decaying component of prices (the AR(1) component in Summers' model) induces negative auto-correlation in returns that is weak for the daily and weekly holding periods common in market efficiency tests. But such a temporary compo nent of prices can induce strong negative auto-correlation in long-horizon returns. [8] Most of their work centered around the continuously compounded return from time t to t+T, rtj+T, which can be obtained from prices as follows: rt,t+T — Pt+T - Pt where T is the holding period in years and ranges from 1 to 8. Their tests involved calculating the slopes fir in the following regressions: rt,t+T = <*T + &Trt-T,t + £t,t+T (1-2) The slopes fir had to be biased adjusted because of the use of overlapping price data to determine T period returns. Under the assumption that prices follow Summers' "fads model", F&F were able to show that the biased adjusted slopes fir in (1.2) represented the fraction of variation of returns that could be predicted. Using this hypothesis they reached some startling conclusions. The data they used consisted of the ten size-based deciles (decile ten contains the largest firms) plus the value-weighted and equal-weighted 11 portfolios from the New York Stock Exchange (NYSE). Examining the slopes of the regressions in (1.2) for varying amounts of holding periods (T) for the time interval 1926-1985, they found that the predictable component of industry portfolio variances ranged from 25% to 40% of total 3-5 year return variances. This conclusion was in stark contrast to the expected near 0% under the efficient market hypothesis. 1.1.5 A re-evaluation of the F&F (1988) Results Following the work of F&F (1988) were Eckbo and Liu (1993) in a paper entitled "Tem porary Components of Stock Prices: New Univariate Results" [6]. In this paper they showed that if the model of stock prices that F&F used was modified slightly, their re sults were no longer valid. The slight modification involved allowing the {zt} to be a more general stationary process than the AR(1) that F&F had assumed. This modifica tion was reasonable as Daniel and Torous (1991) [4] and Eckbo and Liu (1993) had both demonstrated that the "simple fads" model does not hold for the decile price portfolios of the NYSE. The use of a more general stationary process implied that exact results for the frac tion of variation of returns that can be explained by the stationary component was no longer possible. With this in mind Eckbo and Liu proposed a finite-sample lower bound estimator of the predictable variance proportion. Using this estimator they found that their lower bounds of the predictable component of 3-5 year return variances ranged from 10% to 17%. Though this was a conservative estimate, it was much different from the 25% to 40% that F&F obtained. The reason for the large difference in results is that the optimistic estimates of F&F are sensitive to model specification, whereas the conser vative estimates of Eckbo and Liu allow for all possible model specifications subject to additivity of the price components. 12 1.1.6 Mounting evidence that the EMH does not hold Tsay (1990) attacked the EMH question from a different angle than the studies previ ously mentioned. Tsay reasoned that under the EMH, the price today contains all of the available information on the price tomorrow. Hence, all the expected information lies in the relationship between prices today and prices tomorrow. Using this observation led Tsay, in his paper entitled "Correlation Transformation and Components of Stock Prices" [21], to consider a canonical correlation analysis between prices today and prices tomorrow. The idea to use a canonical correlation analysis to determine the structure of a time series was an old one first proposed by Box and Tiao (1977) in their paper "A canonical analysis of multiple time series" [2]. In this paper they showed that a k-dimensional autoregressive process of order p could be transformed into a process whose components were ordered from least to most predictable. Tsay used the same NYSE portfolio data as F&F (1988). His results were that for the period 1926 to 1989 the sta tionary (predictable) components of each decile explained from 10-15% of the variability in prices. He also subdivided the period into a smaller segment, 1941 to 1989, and found that the stationary price components explained 0-1.2% of the variation in stock prices. This indicated that the importance of the stationary components may have diminished greatly over the last 50 years. 1.1.7 The EMH re-considered For a time after the work by Summers (1986) and F&F (1988) many researchers strongly believed that the EMH had been disproved and that all that remained was to determine by how much the market deviated from efficiency . Recently however, some disbelievers have appeared. Kim, Nelson, and Startz (1991) in their paper "Mean Reversion in Stock Prices? A Reappraisal of the Empirical Evidence" [14] discuss the results by F&F (1988). 13 They take issue with among other things, the claim issed by F&F that up to 30% of stock returns are predictable. This is an important consideration since what F&F demonstrated was a high degree of in-sample predictability, they mentioned nothing about an out-of-sample forecast. Kim et al considered such forecasting and showed that estimating the fix coefficients of F&F in real-time, leds to very poor correlations between predicted returns and actual returns. Another disenter is Richardson (1993) with his paper "Temporary Components of Stock Prices: A Skeptic's View" [18]. He demonstrates that the large coefficients of fix that F&F found could be consistent with prices that follow a RW. His work considers the distribution of the largest 0T , i.e the first order statistic, in the work by F&F. He shows via Monte Carlo simulation and bootstrapping that conditional on the largest value of /3j the results obtained by FSzF can be explained by prices which follow a RW. 1.2 Our proposal Obviously the EMH debate has not been settled yet. We propose to investigate it at the weak level, via price filtering and canonical correlation analysis. If inefficiency is found at this level it implies inefficiency at the other two levels. One of the implications of an inefficient market is that prices today can no longer be assumed to have digested all of the information available up to today. At best, prices today can provide a subset of information for price movement tomorrow. Consequently, we are led to find a better indicator for the expected price tomorrow based on all available price information up to today than merely the price today. This will be accomplished by searching for a (relatively) optimal predictive information filter of stock prices from the class of linear, quadratic and square-root filters. Once a predictive information filter has been found, its association with tomorrow's price will be investigated. This will be done using a 14 canonical correlation analysis. The resulting transformed series will be analysed and used to infer results about the original stock price series. The primary goal of this thesis is to find the predictive filter that produces a stationary (predictable) component in prices which can explain the largest (if any) percentage of variations in stock prices, and to determine if this percentage is statistically different from those generated by a RW model for prices. One reason why this is important will be explained by an example. Suppose that we are a fund manager and have several techniques at our disposal with which we can measure the amount of stock price variations which can be explained by a stationary (predictable) component of the stock price. We are primarily interested in finding the technique which offers the lowest amount of risk. In this case lower risk is equivalent to better prediction of future price movements. Thus we would like to obtain the method that allows us to predict the largest possible amount of variation in stock prices. Since the predictive filters are simply a re-arrangement of the past data, we are interested in finding the filter which allows us to explain the largest possible amount (if any) of variation in stock prices. 1.3 Layout of the thesis Chapter 2 contains a description of our techniques and results. Chapter 3 contains a summary of the findings of this thesis, interpretations and indicates future directions for development. 1.4 The Data The data consists of 14 monthly industry index values taken from the Toronto Stock Exchange, see Table 1.1 for a description. Each of the Toronto Stock Exchange indices measures the current aggregate market value (i.e. number of presently outstanding shares 15 x current price) of the stocks included in the index as a proportion of an average base aggregate market value (number of base outstanding shares x average base price ± changes proportional to changes made in the current aggregate market value figure) for such stocks. The starting level of the base value has been set equal to 1000. Notationally this corresponds to: »»r ~ .-, „ Current aqqreqate market value INDEX = ——• ^— — x 1000 Adjusted average base aggregate market value The data was obtained from the Toronto Stock Exchange Review [20]. The Appendix contains a detailed description of the method used to calculate the index. Time series plots of the industry index values are displayed in Figures 1.2, 1.3 and 1.4 on pages 18,19 and 20. The indexes exhibit the same generals trends. They all started low and steadily rose to a local maximum around 1980 after which time there was a short period of decline. Following this decline they generally rose sharply and had another local maximum just before 1989, after this there was a period of decline and then a gradual increase until the present. 16 Table 1.1: The 14 industries of the TSE. Industry First Last Percentage of TSE market Date Date value as of June, 1995 Communications and Media Jan, 1956 June, 1995 3.50 Financial Services Jan, 1956 June, 1995 15.42 Gold and Silver Jan, 1956 June, 1995 11.20 Industrial Products Jan, 1956 June, 1995 15.95 Conglomerates Jan, 1956 June, 1995 4.25 Merchandising Jan, 1956 June, 1995 3.89 Metals and Minerals Jan, 1956 June, 1995 8.36 Oil and Gas Jan, 1956 June, 1995 11.20 Paper and Forest Products Jan, 1956 June, 1995 5.47 Pipelines Jan, 1956 June, 1995 2.32 Trans and Env. Services Jan, 1956 June, 1995 1.70 Utilities Jan, 1956 June, 1995 8.83 Consumer Products Jan, 1956 June, 1995 7.67 Real Estate and Construction Jan, 1968 June, 1995 0.26 17 Monthly Index for Communications 1960 1970 1980 1990 Time Monthly Index for Gold and Silver 1960 1970 1980 Time 1990 Monthly Index for Conglomerates 1960 1970 1980 1990 Time Monthly Index for Financial Services 1960 1970 1980 1990 Time Monthly Index for Industrial Products 1960 1970 1980 Time 1990 Monthly Index for Merchandising .1960 1970 1980 1990 Time Figure 1.2: Plots of monthly TSE industry portfolio indexes 18 Monthly Index for Metals 1960 1970 1980 1990 Time Monthly Index for Real Estate o 1970 1980 1990 Time Monthly Index for Oil and Gas 1960 1970 1980 1990 Time Monthly Index for Pipelines 1960 1970 1980 1990 Time Monthly Index for Transportation o 1960 1970 1980 1990 Time Figure 1.3: Plots of monthly TSE industry portfolio indexes 19 Figure 1.4: Plots of monthly TSE industry portfolio indexes 20 Chapter 2 The Mean-Reverting Price Component 2.1 Our Method and Rationale We will consider price models of the following form, Pt = qt + zt where pt is the log stock price at time t, zt is the stationary or mean-reverting component of the stock price at time t, and qt is the nonstationary component of the stock price at time t. Often {qt} and {zt} are assumed to be independent or at least uncorrelated to avoid trivial identifiability problems. This is a generalization of the Summers (1986) "fads" model for prices, in which qt and Zt can be any arbitrary nonstationary and stationary series. We will propose a method to obtain this decomposition. We will then then test via simulations and bootstrapping whether the the percentage of price variation explained by the stationary components is significantly different from results obtained under the EMH view that prices follow soley a RW. Many early studies approached the EMH problem by simply using a linear regression of prices at time t on prices at time t-1, time t-2, and so on. This approach failed to uncover any relationship beyond a random walk for prices. The reason we believe it failed is because of a confounding effect of the stationary and nonstationary components of prices. The stationary component of prices represents the mean-reverting price com ponent, and it is this component which may allow some of the variation in stock prices to 21 be predicted. One approach to alleviating this problem of confounding is to first separate the stationary and nonstationary components of prices. We propose to do this through the use of predictive information filtering and canonical correlation analysis. Once this has been done a regression with stock prices at time t as the dependent variable and the stationary components of the canonical variate at time t as the independent variable can be performed. The R2 of this regression represents the percentage of stock price variation that can be accounted for by the stationary zt component. 2.2 On the use of returns to avoid the confounding problem At this point it seems natural to ask if returns can be used to avoid the issue of confound ing. To answer this question it is appropriate to recall the main goal of this thesis: to determine the amount of variation in stock prices that is accounted for by the stationary (predictable) component zt in (2.1). The following notation will help us out. For sim plicity, we assume that {qt — qt-i} is stationary. This would be the case if {qt} followed a random walk. Pt = qt + zt (2.3) var(pt) = var(qt) + var(zt) = a2 (2.4) varies in t const in t varies in t n = pt- pt-i = (qt - qt-i) + (zt - zt-i) (2.5) var(rt) = var(qt - qt-i) + var(zt - zt-x) = a2 (2.6) v. const in t const in t const m 1 Model (2.3) represents a general decomposition of prices into a stationary zt compo nent and a nonstationary qt component. The price variance is given in Equation (2.4). The stationary component has a variance which is constant in time and the nonstationary component has a time-varying variance, which together imply that prices have a time-varying variance. Equation (2.5) represents the decomposition of returns. If returns are 22 stationary then the variance of returns will be time independent and we have an identi-fiability problem. The reason for this is that we will not be able to separate the amount of variance attributed to (qt — qt-i) and [zt — zt-\) since both of these variances would also be constant. To avoid such problems we will use prices to resolve the confounding problem. In the following three sections we will discuss the technical aspects of the canoni cal correlation analysis, Filtering, Auto-regressive integrated moving average (ARIMA) modeling and diagnostics, and how to determine the amount of variation in prices that can be attributed to stationary components. 2.3 Details of the Canonical Correlation Analysis Before we carry out the canonical correlation analysis the theoretical and sample results will be discussed: Notation: Let X_ = (XLX2 .. • Xk)' represent a(ixl) vector random variable. The Jfcxl sample realizations of the jth component of this vector will be denoted by {xjt}. 2.3.1 Theoretical Canonical Correlation Analysis Canonical Correlation analysis seeks a measure of association between two groups of variables. Let the first group of p variables be represented by the (pxl) random vector 2£^ and the second group of q variables be represented by the (qxl) random vector X}2\ In the development which follows we assume that X}1^ represents the smaller set, so that p<q. The following theorem will be used to calculate the theoretical canonical variates Y_ and Z, both (pxl) random vectors: Theorem 2.1 (Canonical Variates and Correlation) Let the random vectors 2L^ andX™ have Cov{X(l)) = En, Cov(X(2)) = £22, and Cov(X(1\X(2)) = E12, 23 where E has full rank. For coefficient vectors a and b, form the linear combinations Yi = o-Z(1) and Zi = b\X{2). Then the maximum Corr(Yi, Zi) = p\ is attained by the following linear combinations (called the first canonical variate pair): Fx =£;E1-11/a2C(1) and Z^f^X^ ' The kth pair of canonical variates j=2,3, ..., p is given by: Yu = cLEu1/a2C(1) and Zk = f^X^ maximizes, Carr(Yk,Zk) = pl among those linear combinations uncorrelated with the preceding 1,2, ..., k-1 canonical variables. Here p*2 > p*2 > ... p*2 are the p ordered eigenvalues o/E111/'2Ei2E2~21E2iS1~11^2 and el5£2,..., ep are the associated (pxl) orthonormal eigenvectors. (The quantities p\2, p*2 ... p*2 are also the p largest eigenvalues of the matrix EJ21/'2E2iS1_11Ei2EJ21/'2 with corresponding (qxl) orthonormal eigenvectors f1, /2,... ,f^) 2.3.2 Sample Canonical Correlation Analysis The following theorem will be used to obtain the sample canonical variates {yjt} and [zjt] J = 1>2, • •. ,p. The notation {yjt} denotes the vector of observations for the jth canonical variate. The two groups in which we are interested in assessing the association between are represented by x^1' and x^2) (bold faces are used to indicate matrices). Each group is assumed to have n observations per variable. 24 Definition 2.1 (Notation for the two groups) X = ' r(l) x12 x21 2-22 • X2n x(l)' r(l) xp2 • xpn X(2) r(2) xll r(2) • x\n r(2) x2\ ^•22 r(2) • x2n r(2) ?1 T(2) x.j2 • xqn = x.\,x2,... ,xn) where, Xj Definition 2.2 (Sample Covariance) r(2) -3 s = Su S12 S21 S22 Now that we have defined the notation above we can use the following theorem to obtain the p sample canonical variate pairs {yjt} and Theorem 2.2 (Sample Canonical Variates and Correlation) Let pi2 > p*2 > ... p*2 be the p ordered eigenvalues of S-^^2S\2S22S2\S11X^2 with corre sponding orthonormal eigenvectors el5 e2,..., ep, where the Ski art defined in Definition (2.2). Let f1,f2, • • - if v be the p largest orthonormal eigenvectors of S22 ^2S2\S 11 S\2S22 ^• The kth canonical variate pair is: {Vkt} = JlSn'V1) and {zkt} = /^V2) where x^1^ and x^2^ are as in Definition (2.1). See Johnson and Wichern [11] for a more detailed treatment of canonical correlation analysis. 25 2.4 A New Look at Canonical Correlation Analysis 2.4.1 Historical Overview The classical canonical correlation analysis was developed by Hotelling (1936) primarily in the context of analyzing I.I.D. (independent identically distributed) data sampled from populations with multivariate normal distributions. Recently it has also been used to partition a general multivariate time series into stationary and nonstationary components (Tsay (1990)). One of the drawbacks of canonical correlation analysis is that it only uses the first two moments of the empirical distribution of the data. Since the normal distribution is typically characterized by its first two moments and is well known and easy to use, such an analysis is usually confined to the class of population distributions which are at least approximately normal. However, since most data is only approximately normal at best we are led to consider the use of Filters which allow the use of higher order moment information. 2.4.2 Filters Let X(1) be Jfei x 1 and 2—2^ be ^2 x 1 random vectors. Definition 2.3 (Portfolio Filter) A portfolio filter £(•) is a (measurable) transformation given by £(•) : dikl —> . Two examples of a portfolio filter are considered below. Example #1: A linear portfolio filter £(•) is given by: c(xw) = AX^ where A is a ki x k\ matrix of constants. 26 Example #2: A quadratic portfolio filter £(•) is given by: C(X^) = AX™ + £vec(VX(1)Z(1)') where A is a &i x k\ matrix of constants, B is a ki x M*^1) matrix of constants, vec(-) is the column stacking operator and V is an operator that returns the upper diagonal portion of a matrix. We need to use the V operator because the matrix it operates on is symmetric and we only want the unique elements. Definition 2.4 (Information Filter) An information filter !F(-) is a (measurable) transformation given by Jr{-) • dtkl+lt2 —• 3Jfcl. There are two main uses of an information filter. The first is as a smoother or recalibration technique in which both 2L^ and XJ2^ are observed. For example, let 2L^ and XJ2^ represent repeated observations from the same phenomena in which we would like to estimate a parameter. Information filtering may lead to a more accurate measurement of this parameter than either of them could provide individually. Two examples of an information filter are considered below. Example #3: A Linear Smoothing Information Filter is given by: HX(l\X(2)) = MXP + A2X^ where Ai is a k\ x kx matrix of constants and A2 is a k\ x k2 matrix of constants. The second use of information filtering is as a prediction scheme. In this scenario it is assumed that X^ is unobserved and X}2^ is observed. We also assume that there is some association between X}1^ and X}2^ that we are interested in exploiting to predict X}1\ In the analysis which follows in the rest of the thesis X}1^ represents the vector of 27 tomorrow's prices, X_2^ represents the vector of all available past price information and T(X_2^) represents our prediction of tomorrow's prices, which we will call X_t. Example #4: A Linear Predictive Information Filter T(-) is given by: where A3 is a k\ x hi matrix of constants. More generally, for the purpose of predicting X_X^ using XS*\ the choice of the Pre dictive Information Filter is often given by ^LY(2)) = E(X(1)\X_2)). 2.4.3 Uses of Filtering Our goal in using filtering is to restructure the components of XJ1^ and J-(X^) such that the new portfolios C\(2—^) an(l £2 0 J-(X^) have their respective pairwise correlations in decreasing order of magnitude and all pairs of the new portfolios are uncorrelated with previous pairs. When we restrict our attention to the class of linear portfolio filters, we are led to examine the canonical correlation analysis between and the current information F(X(2)). From the point of view of portfolio prediction, £2 0 T(X_2^) can be viewed as the best linear predictor of tomorrow's portfolios £1 (X^). The utility of using information filtering for predictive purposes will be illustrated in the following example: Example #5: Let X™ „ jV(0,1) and X^ = X^2. Since COTT(X^,X^) = 0 the only canonical variates possible in the usual canonical correlation analysis setting are and X^2\ We must then conclude falsely that the canonical variates X^ and X^ are unrelated due to their zero correlation and hence that X^ has no predictive power for 28 X^. However, with the use of information filtering it is possible to achieve maximum correlation if one uses the canonical variates X^ and T{X^), where J-'(X^) = X^2. 2.4.4 Filtering in a time-series setting As mentioned previously the other main use of canonical correlation analysis is to sepa rate the stationary and nonstationary components of a general multivariate time-series. Since most time-series are only approximately normally distributed the usual canonical correlation analysis will give sub-optimal results, as measured by the regression R2 in a regression of the time-series on its stationary components. However, if we utilize higher order moment information contained in the data by selecting appropriate portfolio and information filters we will obtain a more precise R2. In general, the use of information filtering and canonical correlation analysis go hand in hand. The first step is to determine a (relatively) optimal predictive information filter within a class of filters {J7}. In this thesis we will consider the class of linear, quadratic and square-root predictive information filters. More generally, we may con sider J-{X}2^) = E(X}X1\X}2}). The next step is to choose an optimal portfolio filter by maximizing the correlation between C\(X}^) and £-? oJ7(X^), subject to both £i(-) and £2(") being in a certain class of filters. Using this procedure will result in a constantly changing portfolio given by £i(2£^). It changes because as each month goes by and another data point is added to the data the entire analysis is re-done and a new & (X^) results. In this thesis we will consider the class of linear filters which is equivalent to a canonical correlation analysis. It is the £i(2L^) portfolio that we are interested in. Using it we will be able to infer properties of X}1\ For ease of notation the }th series of the canonical variate £i(2C^) will be denoted by {yjt}-29 2.4.5 Filtering in an investment setting If we were to consider long term investments with relatively unchanging portfolio's, we would require our portfolio filter C\ to be the same each time we repeated the analysis. To simplify our discussion, we shall in the following consider only linear portfolio filters. The following equations will be discussed: - H2L(2))+i (2.7) pxl a'X(1), aeW (2.8) y_ Va _F(XW) + a_ (2.9Var(ya) = Var(a >(X(2))) + Var(a_) (2.10) > * v ' signal noise Varjae) Varja^X^)) C^ = V^)=1 Var(ya) (211) = _Var(^))_ {2l2Equation (2.7) represents the prediction equation for X_l\ Equation (2.8) represents a portfolio of stocks. If we substitute Equation (2.7) into (2.8) we obtain Equation (2.9), the variance of which is given in Equation (2.10). We now define the noise-to-signal variance ratio as Equation (2.11). We want to minimize this value, which is equivalent to maximizing Equation (2.12). Now let p\ > p2 _ • • • pp and aL, a2 • • • Q.p he the eigenvalues and linearly independent right eigenvectors of [Var(2£(1))]-1 • VW[JF(X(2))]- Then the portfolio's a[Xw,..., a^(1) are arranged from the least predictable to most predictable, as measured by the noise-to-signal variance ratio. However, we will not consider such an analysis, the interested reader is referred to Box and Tiao (1977) [2] for a detailed discussion in the linear time series setting. 30 2.4.6 Filtering used to solve an identifiability problem Sometimes, an intelligent choice of a portfolio filter will help to solve the problem of identifiability of the additive stationary and nonstationary price components. For example, suppose prices follow a k-dimensional vector RW model driven by an I.I.D. noise sequence given by: Pt = Et-i+^ * = 1,2,... {et} I.I.D. /Yfc(0,E) (2.13) If the components of et are highly correlated, so are those of p . This means that some portfolio filters £, could result in components of C{pt) that are approximately stationary. For the stationary series of C(pt) it is no longer possible to to separate the stationary components from the nonstationary ones. However, if we left-multiply pt in Equation (2.13) by £-1/2 to create p*, which has independent components, we will be free of this particular identifiability problem. This is illustrated in the following equations: £(£) = AR; Etl a2iP*t Since the components of p* are independent RW's then the variance of the bth component of £(p*) is given by: k k Var(J2 abip*t) = £ a2bita^ i=l i=l This variance depends on t for all b, hence all of the components of C(p*) are nonstation ary. This means that it will still be possible to separate the stationary and nonstationary components. 31 2.5 ARIMA modeling and diagnostics In the following two sections we will discuss the techniques used to fit the auto-regressive integrated moving average (ARIMA) models and also the diagnostic tools that were used to evaluate the goodness of fit. 2.5.1 ARIMA modeling The basic procedure followed was as in Box and Jenkins [3]. Suppose our series is represented by {yjt}- The procedure is as follows: 1. Difference {yjt} until it is stationary. 2. Identify the appropriate auto-regressive moving average (ARMA) process. The main tools used in step 2 are the auto-correlation function (acf) and the partial auto-correlation functions (pacf) as defined in Box and Jenkins. The behavior of these functions will indicate which class of ARMA models is appropriate for the series under consideration. The model(s) indicated by the acf and pacf were fitted using the S-plus arima.mle function. If more than one model was appropriate the one with the better goodness of fit based on the Ljung-Box statistic was chosen. The Ljung-Box statistic is discussed in the next section. 2.5.2 ARIMA diagnostics In this section we will discuss the diagnostic tools that were used to check the adequacy of the fitted ARIMA models. 2.5.3 Using the acf to check for serial dependence in residuals To check for serial dependence in the residuals, we define the following: 32 Definition 2.5 The auto-covariance function 7T of a series , {zt}i , where N is the number of observations, is given by: ^T = Cov(Zt,Zt+r) and the auto-correlation function (acf) pT is given by: IT PT = — 7o Definition 2.6 The sample analogue of ~fT, denoted by jT, is given by: ir = ^Ni_{zt-z){zt+T-z) ly z'=l and the sample estimate of pT, denoted by pT, is 7r PT = — 7o For more details on these functions see Box and Jenkins [3]. We will also need the following result due to Bartlett: Definition 2.7 Var(pT) « : — q— T> q. We are sometimes interested in testing for zero auto-correlation after a specified lag q (for example, a moving average process of order 2 (MA(2)) should have an acf which is zero for any lag greater than 2, in Definition (2.7) r would represent a general lag > 2). To do this we first compute the estimated variance of the sample auto-correlation given by Definition (2.7). We can then use the result due to Andersen that for reasonably large N, pT is approximately distributed as a standard normal random variable. The last step is to make a plot of the sample auto-correlation function and determine how many points lie outside the 95% confidence interval of zero auto-correlation. If many points are outside this limit it is an indication that the residuals are serially correlated and the ARIMA model may be inadequate. 33 2.5.4 Using a runs test to check for serial dependence in residuals Another method of checking for serial dependence is given by using a runs test [15]. A runs test is usually used to test data to see if the order is "random". A run is one or more observations in a row greater than the median or one or more observations in a row less than or equal to the median. Wald and Wolfowitz (1940) showed that as m and n tend infinity with m/n tending to 7 then: Theorem 2.3 R-*n/(l + l)±mi) where R is the total number of runs, m is the number of observations greater than the median, and 7 is the ratio of m and the number of observations less than or equal to the median. A simple check for independence is given by calculating the above ratio and comparing it to the appropriate critical value from the standard normal distribution. 2.5.5 The Ljung-Box test for lack of fit Ljung and Box [16] have suggested the following measure of lack of fit for time series models: Definition 2.8 m Q(m) = n(n + 2)J2(n-k)-1pl k=i where m is the number of lags used, n is the length of the series, and pk is the fitted residual auto-correlation lag-k. For large n, Q(m) is distributed as x2 with m-p-q degrees of freedom, where p is the number of AR components in the model and q is the number of MA components in the model. 34 A check on the adequacy of any ARIMA model is given by computing Q(m) in Definition (2.8) and comparing it with the 95% critical value from the appropriate Xm-p-q distribution. 2.6 The Dickey Fuller Unit Root Test We will be interested in determining whether a given series is stationary or nonstationary. Due to the fact that existing studies have primarily centered around the RW, we will consider testing for this type of nonstationarity. We use the test developed by Dickey and Fuller (1979) [5]. Consider the following autoregressive model: Yt = pYt-1+eu i = l,2,...,n It is assumed Vo = 0, p is a real number and {et} is a sequence of I.I.D. random variables with mean zero, variance cr2 and finite fourth moment. If p = 1 the time series is a random walk (RW), and it has a unit root. Dickey and Fuller developed tests for a unit root based on the following two statistics. P = (E^r'tYtYt-i (2.14) t=l t=2 r = (/> - i)s;\£??-i)* where> (2-15) t=2 S2e = (n-2)-*±(Yt-pYt_ir t=2 The empirical distribution of these two statistics is given in Fuller [9] (1976, pp. 371,373). To test for a unit root one simply calculates either of the above two statistics and com pares it to a critical value. We choose to use p to classify the series. The series is nonstationary if n(p — 1) is between the critical values of -10.4 and 1.61 (this corresponds to a two sided test region with probability of falsely rejecting the null of a unit root equal to 5%). 35 A more recent unit root test given by Phillips (1987) [17] was also used in our subse quent analysis. The initial results obtained by both methods were very similar but since the Phillips test took much longer computationally to evaluate it was decided to use only the Dickey Fuller unit root test. 2.7 Variation in prices explained by stationary components Each of the k series of the canonical variate £I(2L^)> which we have denoted by {yjt} j — 1,2... k, were analysed to determine if they were stationary or nonstationary. When the stationary {yjt} were identified we performed k separate regressions. For each regression one of the original logarithm of stock price series was the dependent variable and the stationary {yjt} series were the independent variables. To estimate the percentage of stock price variation that is attributed to stationary (predicatable) components we used the R2 from the following regression: Definition 2.9 The percentage of stock price variation in each industry that is attributed to stationary (predictable) components is given by the R2 in the following regression: Xjt = To + Jv+iVv+u + • • • + JkVk,t + £jt where {xjt} is the jth original series, {yjt} is the )th stationary canonical variate series, k is the number of stock price series in the data, v is the number of nonstationary {yjt} series and {ejt} is white noise. 2.8 Procedure that was followed We started out with k logarithm of stock price series. We first considered filters within the class of linear, quadratic and square root predictive information filters. We then performed a canonical correlation analysis between the original logarithm of stock price 36 series 2L^ and a predictive information filter F{XS^). The resulting k canonical variate series {yjt} j — 1,2,... k were modeled using the standard ARIMA modeling techniques discussed in Section 2.3. We were primarily interested in determining which of the k {yjt} series were stationary and which were nonstationary. We then performed k separate regressions with each of the original logarithm of stock price series as the dependent variable and the stationary components of {yjt} as the independent variables, as discussed in Section 2.4. The J?2,s of these regressions represent the amount of variation in the stock price series that can be explained using the particular predictive information1 filter under consideration. Finally, we consider the question of whether the R2,s obtained in these regressions are statistically different than what is expected under a RW model for prices. 2.9 The Data Revisited The data that we used are the industry index values taken from the Toronto Stock Exchange (TSE), see Table 1.1 for a description. As can be seen from this table all of the series except real-estate start on Jan, 1956. Real-estate commences on Jan, 1968. In our analysis we performed a canonical correlation analysis using these industry index values. To do such an analysis, all of the series had to be the same length. We could have either truncated all of the series so that they started on Jan, 1968, or simply deleted the real-estate index. Since the real-estate industry only consists of 0.26% of the total TSE market value, we chose to proceed by deleting the real-estate index from the data. Thus the data that we used consisted of 13 industries of the TSE each with 474 monthly observations from Jan, 1956 to June, 1995. 37 2.10 Results We present results for the original data set and also a standardized version. The standard ized set is considered because of the possible confounding problem of serial correlation and correlation between variables. We standardized the price data using the following formulas: Kt - Kt-i = £*, if Prices follow a RW (2.16) Var(et) = E ' (2.17) E = AA' (2.18Kt = A~xKt (2.19) where Kt is the vector of pries at time t, et is the error if prices follow a RW and E is the variance-covariance matrix of et. The original industry data is highly correlated, to lessen this problem we can simply form a new standardized version of X_t which is given by Kt- As can be seen by Equation (2.18) the variance of the original industry data can be broken into two parts , A and A'. When A'1 premultiplies the original data the new variable Kt 1S created. The new data Kt is a linear combination of the industries from the TSE, and is referred to as a portfolio. The weights of the original industries in each portfolio are given in Table 2.1. Canonical correlation analysis seeks a linear combination of variables from one group that have maximum correlation with a linear combination of another set of variables such that each combination is independent of all others. Since standardization merely involves linear combinations of the original data this does not change the resulting canonical variates. 38 Table 2.1: The industry weights of each portfolio # Industry PI P2 P3 P4 P5 P6 P7 1 Communications Sz Media 28.09 -1.78 -0.28 -1.51 -2.20 -4.25 -0.29 2 Financial Services -1.78 34.79 -0.03 -4.65 -1.06 -3.72 -0.69 3 Gold & Silver -0.28 -0.03 11.83 0.40 -0.09 0.18 -2.53 4 Industrial Products -1.51 -4.65 0.40 38.15 -1.35 -5.24 -4.40 5 Conglomerates -2.20 -1.06 -0.09 -1.35 24.86 -1.69 -2.08 6 Merchandising -4.25 -3.72 0.18 -5.24 -1.69 33.59 -0.88 7 Metals &; Minerals -0.29 -0.69 -2.53 -4.40 -2.08 -0.88 25.19 8 Oil k Gas 0.35 0.38 -1.60 -1.21 -2.82 -1.90 -2.26 9 Paper & Forest Products -1.50 -2.10 -0.84 -3.90 -2.74 -1.48 -4.88 10 Pipelines -1.04 -3.38 -0.51 -2.03 -1.12 -0.84 -1.66 11 Trans. & Environmental -0.48 -0.87 -0.21 -2.76 -2.48 -0.58 -1.37 12 Utilities -2.89 -6.04 -0.20 -2.00 -0.34 -0.44 1.67 13 Consumer Products -4.20 -3.55 0.18 -5.05 -1.61 -4.20 -1.58 # Industry P8 P9 P10 Pll P12 P13 1 Communications & Media 0.35 -1.50 -1.04 -0.48 -2.89 -4.20 2 Financial Services 0.38 -2.10 -3.38 -0.87 -6.04 -3.55 3 Gold & Silver -1.60 -0.84 -0.51 -0.21 -0.20 0.18 4 Industrial Products -1.21 -3.90 -2.03 -2.76 -2.00 -5.05 5 Conglomerates -2.82 -2.74 -1.12 -2.48 -0.34 -1.61 6 Merchandising -1.90 -1.48 -0.84 -0.58 -0.44 -4.20 7 Metals & Minerals -2.26 -4.88 -1.66 -1.37 1.67 -1.58 8 Oil and Gas 20.57 0.61 -5.11 -1.02 0.01 -2.64 9 Paper & Forest Products 0.61 25.29 0.46 -2.33 0.77 -3.11 10 Pipelines -5.11 0.46 27.94 -1.22 -3.90 0.07 11 Trans. & Environmental -1.02 -2.33 -1.22 20.76 -1.29 -1.52 12 Utilities 0.01 0.77 -3.90 -1.29 42.15 -4.46 13 Consumer Products -2.64 -3.11 0.07 -1.52 -4.46 40.37 where P indicates a standardized portfolio, for example PI is standardized portfolio 1. Note that portfolio i is primarily composed of industry i. The negative weights for many industries may be taken to represent certain buying or selling strategys such as selling short. 39 Notation used in the results section: k - refers to the dimension of the price vector, which in our case is 13. vec(-) - this is the column stacking operator. V - extracts the upper diagonal portion of the matrix it is applied to (this is needed for any matrix of the form XtXt since it will be symmetric and when the vec(-) operator is used on it, k2 — fc(fc+1) elements in this new vector will be identical). %R2 - this is simply 100 xR2. R2W - this is a weighted average R2, with weights given by the percentage of TSE market value for each industry as of June, 1995 (see Table 1.1). To calculate this statistic each industry R2 is multiplied by its share of TSE market value given in Table 1.1 and the resulting values are added together. This statistic will be used for the original data set. R2 - this is an unweighted average R2. This statistic will be used for the standardized data where no direct correspondence with an industry is possible (each portfolio is a linear combination of industries). B - this is the usual backshift operator, i.e BYt — Yt — Yt-\. This operator is used in the ARIMA models in the Appendix. Note: All Tables in the results section with an A preceeding a number are located in the Appendix. The critical values for the Dickey-Fuller unit root test statistic are -10.4 and 1.61. If the Dickey-Fuller statistic is between these two values the series is classified as nonstationary. 40 2.10.1 Results using a linear lag-1 prediction filter The linear lag-1 predictor was obtained using the following regression model: Kt = ^AX^ + et, * = 2,3,...,474. (2.20) (fcxl) (fcxfc) Here Kt is the vector of stock prices at time t, A is a constant coefficient matrix, and {et} is vector white noise. Our predictor X_t is obtained by estimating A and setting et to zero in (2.20). A canonical correlation analysis between Kt and Kt was performed, and this resulted in 13 transformed series, {yjt} j=l,2,...,13. The ARIMA models for the 13 transformed {yjt} series, and the Q(12)-statistics of Ljung-Box for residuals of the fitted model are given in Table A.l. Referring to Table A.l we can see that all of the transformed series are AR(l)'s, except series 1, 2 and 4. The Dickey-Fuller unit root test was performed on the transformed series and indicated that the first 5 series were nonstationary. Carrying out the regressions given in Definition (2.9) with the number of nonstation ary series equal to 5, we obtained the results summarized in Table A.2. Original Data: The R2,s range from a low of 0.50% for industry 1 and 2 to a high of 9.49% for industry 7. is 3.76% which indicates that very little variation in industry prices is accounted for by stationary components. Standardized data: The i?2's range from a low of 0.63% for portfolio 1 to a high of 61.66% for portfolio 10. R2a is 17.57% which indicates that a large amount of variation in standardized prices is accounted for by stationary components. 41 2.10.2 Results using a linear lag-2 prediction filter The linear predictor of lag-2 was obtained using the following regression model: Zt, = ^AJLt-i + <BJ£*-2 + * = 3,4,...,474. (2.21) (kxl) (*xfc) (fcxfc) Here X_t is the vector of stock prices at time t, X_t_x is the vector of stock prices at time 2 — 1, K.t-2 is the vector of stock prices at time t — 2, A and B are constant coefficient matrices, and {et} is vector white noise. Our predictor X_t is obtained by estimating A and B and setting tt to zero in (2.21). A canonical correlation analysis between X_t and X_t was performed, and this resulted in 13 transformed series, {yjt} j = l,2...13. The ARIMA models for the 13 transformed {yjt} series, and the Q(12)-statistics of Ljung-Box for residuals of the fitted model are given in Table A.3. Referring to Table A.3 we can see that all of the transformed series are AR(l)'s, except series 1, 2 and 4. The Dickey-Fuller unit root test was performed on the transformed series and indicated that the first 5 series were nonstationary. Carrying out the regressions given in Definition (2.9) with the number of nonstation ary series equal to 5, we obtained the results summarized in Table A.4. Original Data: The R2,s range from a low of 0.48% for industry 1 to a high of 9.52% for industry 7. R2, is 3.77% which indicates that very little variation in industry prices is accounted for by stationary components. Standardized data: The i?2,s range from a low of 0.60% for portfolio 1 to a high of 61.06% for portfolio 10 . R2 is 17.55% which indicates that a large amount of variation in standardized prices is accounted for by stationary components. 42 2.10.3 Results using a quadratic lag-1 prediction filter The quadratic predictor of lag-1 was obtained using the following regression model: 2G=^X,-i+ ,^vec(V2G-i2£Li)+^ * = 2,3,...,474. (2.22) Here 2£< is the vector of prices at time t, 2Lt-i is the vector of prices at time t-1, A and B are constant coefficient matrices, vec(-) is the column stacking operator, V extracts the upper diagonal portion of the matrix it is applied to (this is needed because the matrix J—t-iK-t-i 1S symmetric and we only want the unique elements), and {et} is vector white noise. Our predictor X\t is obtained by estimating A and B and setting et to zero in (2.22). A canonical correlation analysis between X_t and X_t was performed, and this resulted in 13 transformed series, {yjt} j=l,2...13. The ARIMA models for the 13 transformed {yjt} series, and the Q(12)-statistics of Ljung-Box for residuals of the fitted model are given in Table A.5. Referring to Table A.5 we can see that all of the transformed series are AR(l)'s, except series 1, 2 and 4. The Dickey-Fuller unit root test was performed on the transformed series and indicated that the first 5 series were nonstationary. Carrying out the regressions given in Definition (2.9) with the number of nonstation ary series equal to 5, we obtained the results summarized in Table A.6. Original Data: The R2,s range from a low of 0.43% for industry 6 to a high of 9.43% for industry 7. R_ is 3.74% which indicates that very little variation in industry prices is accounted for by stationary components. Standardized data: The R2,s range from a low of 0.65% for portfolio 1 to a high of 64.22% for portfolio 10 . R2a is 17.81% which indicates that a large amount of variation in standardized prices is accounted for by stationary components. 43 2.10.4 Results using a quadratic lag-2 prediction filter The quadratic predictor of lag-2 was obtained using the following regression model: 2Lt = ^AJLt-i+ ^vec(VX(_1X;_1)+ ^vec(VXt_1X;_2) + M,.2 -I- ^ vec(VXt_2X!t_2) + et t = 3,4,... 474 (2.23) (*xfc) (fcxM^tll) Here 2£t is the vector of prices at time 2, 2Lt-i is the vector of prices at time 2 — 1, Xt_2 is the vector of prices at time 2 — 2, A, B, C, D and E are constant coefficient matrices and {e(} is vector white noise. Our predictor X_t is obtained by estimating A, B, C, D and E and setting et to zero in (2.23). A canonical correlation analysis between X_t and X_t was performed, and this resulted in 13 transformed series, {yjt} j = l,2...13. The ARIMA models for the 13 transformed {yjt} series and the Q(12)-statistics of Ljung-Box for the residuals of the fitted model are given in Table A.7. Referring to Table A.7 we can see that all of the transformed series are AR(l)'s, except series 1, 2,3 and 4. The Dickey-Fuller unit root test was performed on the transformed series and indicated that the first 5 series were nonstationary. Carrying out the regressions given in Definition (2.9) with the number of nonstation ary series equal to 5, we obtained the results summarized in Table A.8. Original Data: The i?2's range from a low of 0.51% for industry 6 to a high of 9.84% for industry 7. R?w is 4.46% which indicates that very little variation in industry prices is accounted for by stationary components. Standardized data: The J?2,s range from a low of 0.99% for portfolio 1 to a high of 63.36% for portfolio 10 . i?2 is 18.68% which indicates that a large amount of variation in standardized prices is accounted for by stationary components. 44 2.10.5 Results using a square Root lag-1 prediction filter The square-root predictor of lag-1 was obtained using the following regression model: = ^ Xt-i + ^vec(V^Zi_1Z;_1) + et, t = 2,3,..., 474. (2.24) Here X_t is the vector of prices at time t, X_t_i is the vector of prices at time t — 1, A and B are coefficient matrices and {et} is vector white noise. Our predictor X_t is obtained by estimating A and B and setting et to zero in (2.24). A canonical correlation analysis between X_t and X_t was performed, and this resulted in 13 transformed series, {yjt} j=l,2...13. The ARIMA models for the 13 transformed {yjt} series, and the Q(12)-statistics of Ljung-Box for the residuals of the fitted model are given in Table A.9. Referring to Table A.9 we can see that all of the transformed series are AR(l)'s, except series 1,2 and 4. The Dickey-Fuller unit root test was performed on the transformed series and indicated that the first 5 series were nonstationary. Carrying out the regressions given in Definition (2.9) with the number of nonstation ary series equal to 5, we obtained the results summarized in Table A.10. Original Data: The i?2,s range from a low of 0.43% for industry 6 to a high of 9.50% for industry 7. Pc^ is 3.75% which indicates that very little variation in industry prices is accounted for by stationary components. Standardized data: The JR2'S range from a low of 0.64% for portfolio 1 to a high of 64.32% for portfolio 10 . i?2 is 17.80% which indicates that a large amount of variation in standardized prices is accounted for by stationary components. 45 2.10.6 Results using a square Root lag-2 prediction filter The square-root predictor of lag-2 was obtained using the following regression model: Xt = ^AX^ + sBvec(V^-iZ;_1) + sC>ec(VV/2L4-i2Cl-2) + Jo) (fcxfc) (kxii^) (*x*2) ^DXt-2 + J^ec(V^Xt_22Ct_2) + et 2 = 3,4,..., 474 (2.25) (kxk) (fcx^tii) Here X_t is the vector of prices at time 2, X_t-i is the vector of prices at time 2 — 1, X.t-2 is the vector of prices at time 2 — 2, and A, B, C, D and E are constant coefficient matrices, vec(-) is the column stacking operator, V extracts the upper diagonal portion of the matrix it is applied to and {et} is vector white noise. Our predictor X_t is obtained by estimating A,BVC,D and E and setting et to zero in (2.25). A canonical correlation analysis between 2Lt and X_t was performed, and this resulted in 13 transformed series, {yjt} j = l,2...13. The ARIMA models for the 13 transformed {yjt} series, and the Q(12)-statistics of Ljung-Box for the residuals of the fitted model are given in Table A.ll. Referring to Table A.ll we can see that all of the transformed series are AR(l)'s, ex cept series 1, 2 and 4. The Dickey-Fuller unit root test was performed on the transformed series and indicated that the first 5 series were nonstationary. Carrying out the regressions given in Definition (2.9) with the number of nonstation ary series equal to 5, we obtained the results summarized in Table A. 12. Original Data: The i?2's range from a low of 0.49% for industry 6 to a high of 9.77% for industry 7. is 4.38% which indicates that very little variation in industry prices is accounted for by stationary components. Standardized data: The R2,s range from a low of 0.75% for portfolio 1 to a high of 65.43% for portfolio 10 . R\\ is 18.58 which indicates that a large amount of variation in standardized prices is accounted for by stationary components. 46 2.11 Testing the significance of the i?2's The last question we consider is whether the R2,s that we have obtained are statistically different from those that we could expect to obtain under a RW model for prices. To answer this question we used bootstrapping and simulation techniques with the linear lag-1 prediction filter. In all cases the sample size was 1000. The linear lag-1 prediction filter was used because it took the least amount of computational time. See Figures A.l, A.2, A.3 and A.4 in the Appendix for plots of the empirical R2 distribution for all cases. The y-axis of these Figures represents the range of numerical quantiles (0,1) and the x-axis represents the corresponding empirical R2 values. 2.12 Bootstrapping Bootstrapping is a technique used when the exact values for parameters of a distribution are unknown. It involves randomly sampling elements from the original data set. In our case we will difference the price data to obtain returns and sample randomly from this set, then reconstruct prices using the fact that Pt+i = Pt + Rt (the initial price was chosen to accord with the original data). If the prices actually follow a RW model then this should not make a difference in the R2 values (if prices follow a RW, returns are independent and identically distributed and their ordering should not matter). For both the original data set and the standardized data set we will generate 1000 samples and calculate the R2,s for each industry or portfolio according to the outlined scheme. We will then calculate the empirical quantiles of the R2 statistic and compare the R2 from the real price data to the empirical quantiles. If the real price data R2 is located outside of the empirical 95% quantile we can conclude that it is significantly different than would have occurred under a RW model for prices. 47 Table 2.2: R2,s under the RW Null using bootstrapping vs actual Empirical Quantiles Industry 0.85 0.90 0.95 Actual Significant at 5% (Y/N) Commercial 0 0197 0 0250 0 0376 0 0050 N Financial 0 0772 0 0942 0 1444 0 0050 N Gold 0 1185 0 1575 0 2364 0 0396 N Industrial 0 0999 0 1409 0 2002 0 0585 N Conglomerates 0 1411 0 1841 0 2633 0 0332 N Merchandising 0 0451 0 0613 0 1070 0 0051 N Metals 0 1642 0 2179 0 2984 0 0949 N Oil 0 1460 0 1730 0 2523 0 0425 N Paper 0 1706 0 2145 0 2927 0 0450 N Pipeline 0 1009 0 1388 0 2197 0 0570 N Transportation 0 1224 0 1669 0 2424 0 0196 N Utilities 0 0929 0 1121 0 1503 0 0384 N Consumer 0 0369 0 0475 0 0813 0 0146 N where 0.85, 0.90 and 0.95 are the empirical E? quantiles based on the RW Null. 2.12.1 Bootstrapping without standardization As Table 2.2 indicates, none of the i?2's obtained with the original data are significantly different from those generated under the hypothesis that prices follow a RW, at the 5% significance level. 48 Table 2.3: R2,s under the RW Null using bootstrapping vs actual Empirical Quantiles Portfolio 0.85 0.90 0.95 Actual Significant at 5% (Y/N) Portfolio #1 0 0285 0 0397 0 0661 0 0063 N Portfolio #2 0 2282 0 2860 0 4228 0 1334 N Portfolio #3 0 1421 0 1846 0 2734 0 0537 N Portfolio #4 0 2803 0 3416 0 4487 0 2748 N Portfolio #5 0 2372 0 2956 0 4110 0 3665 N Portfolio #6 0 1455 0 1944 0 2727 0 0392 N Portfolio #7 0 2476 0 3262 0 4460 0 2773 N Portfolio #8 0 2607 0 3200 0 4276 0 1310 N Portfolio #9 0 2430 0 3106 0 4160 0 1241 N Portfolio #10 0 2211 0 2743 0 3967 0 6166 Y Portfolio #11 0 2082 0 2684 0 3812 0 0429 N Portfolio #12 0 1950 0 2306 0 2868 0 1842 N Portfolio #13 0 0883 0 1321 0 1977 0 0340 N where 0.85, 0.90 and 0.95 are the empirical R2 quantiles based on the RW Null. 2.12.2 Bootstrapping with standardization As Table 2.3 indicates, the R2 obtained for portfolio #10 of the original standardized data is significantly different than the one generated under the hypothesis that standard ized prices follow a random walk. We take this as evidence that this portfolio does not follow a RW. The other portfolio's are adequately explained by a RW model, at the 5% significance level. 49 2.13 Simulations We also simulated 1000 vector random walk price data using normal errors with a mean and variance the same as the return data. The procedure used to simulate the data is briefly outlined below. If prices follow a vector RW then the following two equations are valid for prices and returns: Et = a + Et-i + tt Rt = Et-Pt-i = £t+!± If we assume et is distributed as a p-dimensional normal random vector, we can construct a simulated group of 13 prices which we can use to compare to thepriginal price series. We do this by generating observations from a p-dimensional normal distribution with mean and variance given by the original return data. We then construct simulated prices by letting the initial simulated price series be the same as the original price series and then recursively generate subsequent prices using the fact that Et+i = Et + Et-After each simulation has been completed it is subjected to the same canonical corre lation analysis as the real data. Each simulation generates a series of R2,s. After all the simulations and canonical correlation analyses have been completed the empirical distri bution of the simulated RW i?2,s are compared to those obtained by the real data. If any of the real data R2,s are larger than the simulated RW R2,s we take this as evidence that the real price series does not follow a RW. 50 Table 2.4: i?2's under the RW Null using simulations vs actual Empirical Quantiles Industry 0.85 0.90 0.95 Actual Significant at 5% (Y/N) Commercial 0 0188 0 0243 0 0329 0 0050 N Financial 0 0792 0 1035 0 1549 0 0050 N Gold 0 1081 0 1434 0 2319 0 0396 N Industrial 0 1006 0 1344 0 1826 0 0585 N Conglomerates 0 1375 0 1969 0 2846 0 0332 N Merchandising 0 0448 0 0570 0 0822 0 0051 N Metals 0 1564 0 1952 0 2688 0 0949 N Oil 0 1434 0 1863 0 2521 0 0425 N Paper 0 1775 0 2326 0 3103 0 0450 N Pipeline 0 1049 0 1572 0 2224 0 0570 N Transportation 0 1263 0 1659 0 2279 0 0196 N Utilities 0 0907 0 1174 0 1718 0 0384 N Consumer 0 0352 0 0447 0 0663 0 0146 N where 0.85, 0.90 and 0.95 are the empirical R? quantiles based on the RW Null. 2.13.1 Simulations with no standardization The procedure outlined above was completed with 1000 samples and the results are presented in Table 2.4. The first column describes the series being analysed, the sec ond through fourth columns are the upper tail quantiles of the i?2's obtained from the simulated vector random walk series, the fifth column contains the R2 obtained with the real data and the last column indicates whether the real data R2 is larger than the corresponding 95% quantile from the simulated vector RW series. As Table 2.4 indicates, none of the original price data R2,s seems to be significantly different from the R2,s generated under the hypothesis that prices follow a vector RW, at the 5% significance level. 51 Table 2.5: i?2's under the RW Null using simulations vs actual Empirical Quantiles Portfolio 0.85 0.90 0.95 Actual Significant at 5% (Y/N) Portfolio #1 0.2751 0 3376 0.4566 0 0063 N Portfolio #2 0.2726 0 3352 0.4366 0 1334 N Portfolio #3 0.2735 0 3392 0.4425 0 0537 N Portfolio #4 0.2735 0 3462 0.4821 0 2748 N Portfolio #5 0.2680 0 3327 0.4350 0 3665 N Portfolio #6 0.2488 0 3089 0.4002 0 0392 N Portfolio #7 0.2754 0 3321 0.4669 0 2773 N Portfolio #8 0.2708 0 3188 0.4093 0 1310 N Portfolio #9 0.2635 0 3124 0.3945 0 1241 N Portfolio #10 0.2739 0 3394 0.4591 0 6166 Y Portfolio #11 0.2641 0 3242 0.4133 0 0429 N Portfolio #12 0.2518 0 2931 0.4027 0 1842 N Portfolio #13 0.2546 0 3247 0.4268 0 0340 N where 0.85, 0.90 and 0.95 are the empirical R2 quantiles based on the RW Null. 2.13.2 Simulations with standardization All of the portfolio R2,s except one are not significantly different from those expected if prices followed a vector RW process, at the 5% significance level. However, as Table 2.5 indicates, the R2 obtained by Portfolio #10 is significantly different than we would expect if it followed a RW process. We take this as evidence that this portfolio does not follow a RW process. 52 Chapter 3 Discussion 3.1 Comparison of the three predictive information filters In this thesis we have considered three predictive information filters, linear, quadratic and square-root with lags 1 and 2, as well as standardized and unstandardized data. A summary of the performance of each predictive filter relative to the linear lag-1 predictive information filter for each data set is given in Tables 3.1 and 3.2. Table 3.1: Summary of predictive information filter R2,s without standardization L 1 L 2 Q i Q 2 SR 1 SR 2 I %Kf1 %Rf2 %A %KU %A %A %A %A 1 0.50 0.48 -4.23 0.51 2.34 0.88 74.74 0.51 2.56 0.82 63.23 2 0.50 0.52 3.94 0.58 17.42 1.07 114.45 0.59 18.33 1.02 104.60 3 3.96 3.83 -3.28 3.94 -0.42 5.34 34.87 3.94 -0.40 5.34 34.98 4 5.85 5.84 -0.25 5.86 0.06 6.79 16.05 5.88 0.37 6.69 14.27 5 3.32 3.31 -0.22 3.16 -4.83 3.61 8.86 3.15 -5.02 3.62 9.23 6 0.51 0.49 -4.02 0.43 - 16.04 0.51 -0.69 0.43 -16.07 0.49 -4.01 7 9.49 9.52 0.37 9.43 -0.63 9.84 3.77 9.50 0.18 9.77 3.00 8 4.25 4.35 2.23 4.20 -1.33 3.56 -16.43 4.21 -1.03 3.51 -17.60 9 4.50 4.75 5.46 4.53 0.64 7.82 73.71 4.57 1.44 7.76 72.32 10 5.70 5.61 -1.55 6.03 5.85 5.88 3.08 6.06 6.32 5.82 2.02 11 1.96 2.04 4.06 1.68 - 14.19 2.01 2.66 1.72 -12.12 2.12 8.26 12 3.84 3.79 -1.42 3.59 -6.70 4.93 28.37 3.56 -7.38 4.58 19.16 13 1.46 1.49 1.85 1.47 0.54 2.04 39.89 1.49 1.88 1.95 33.39 R* 1 LW 3.76 3.77 3.74 4.46 3.75 4.38 I is industry (see Table 2.1 for a description) , L 1 is linear lag-1 prediction, L 2 is linear lag-2 prediction, Q 1 is quadratic lag-1 prediction, Q 2 is quadratic lag-2 prediction, SR 1 is Square-root lag-1 prediction, SR 2 is Square-root lag-2 prediction, %A is the relative percentage change in R2 using each of the prediction methods as compared to the R2 obtained with linear lag-1 prediction. For example the %A for Quadratic lag-1 was calculated using 1lRi " . R^, is the weighted R . 53 Examining Table 3.1 we make the following observations: • The R2,s increase with lag for quadratic and square-root prediction for most industries but did not for linear prediction. • The weighted R2 for quadratic lag-2 prediction is the largest of any predictive filter, it is 18.6% larger than the weighted R2 for linear lag-1 prediction. • Based on the quadratic lag-2 predictor, the following industries have the most variance explained by stationary components: 1. Metals &; Minerals 2. Paper 3. Industrial . Each of these industries has an R2 > 6%. • Based on the quadratic lag-2 predictor, the following industries have the least amount of variance explained by stationary components: 1. Merchandising 2. Media & Communications 3. Financial. Each of these industries has an R2 < 2% Examining Table 3.2 we make the following observations: • The R2,s increase with lag for quadratic and square-root prediction for most industries but the same does not hold for linear prediction. • The average R2 for quadratic lag-2 prediction is the largest of any predictive filter, it is 6.3% larger than linear lag-1 prediction. • Based on the quadratic lag-2 predictor, the following portfolios have the largest amount of variance explained by stationary components: 1. portfolio 10, 2. portfolio 6, 3. portfolio 8. Each of these portfolios has an R2 > 30%. • Based on the quadratic lag-2 predictor, the following portfolios have the least amount of variance explained by stationary components: 1. portfolio 1, 2. portfolio 11, portfolio 6. Each of these portfolios has an R2 < 4%. 54 Table 3.2: Summary of predictive information filter R2,s with standardization L 1 L 2 Q i Q 2 SR. 1 SR. 2 p %n:f1 %Rh %A %A %A %A %Rir2 %A 1 0.63 0.60 -4.55 0.65 3.41 0.99 57.54 0.64 2.10 0.75 18.82 2 13.34 13.96 4.64 11.84 - 11.26 11.54 -13.50 11.56 -13.35 11.96 -10.36 3 5.37 5.18 -3.49 5.43 1.25 7.65 42.62 5.48 2.18 7.18 33.75 4 27.48 26.60 -3.21 30.93 12.57 23.03 -16.20 31.22 13.61 26.64 -3.06 5 36.65 35.84 -2.20 35.94 -1.93 41.91 14.35 35.82 -2.25 38.18 4.18 6 3.92 3.80 -3.30 3.61 -8.08 3.89 -0.76 3.48 -11.45 4.65 18.41 7 27.73 27.75 0.08 29.21 5.36 32.95 18.83 28.95 4.39 31.72 14.39 8 13.10 13.51 3.15 12.43 -5.08 8.82 -32.67 12.66 -3.34 8.44 -35.53 9 12.41 13.72 10.58 13.49 8.71 22.54 81.66 13.74 10.75 22.17 78.67 10 61.66 61.06 -0.98 64.22 4.15 63.36 2.75 64.32 4.31 65.43 6.10 11 4.29 4.45 3.65 3.83 - 10.87 3.09 -28.05 3.87 -9.94 4.35 1.20 12 18.42 18.25 -0.92 16.52 - 10.32 18.98 3.02 16.34 -11.32 16.47 -10.60 13 3.40 3.41 0.21 3.49 2.45 4.09 20.14 3.41 0.32 3.65 7.18 R'i 17.57 17.55 17.81 18.68 17.80 18.58 P is portfolio (see Table 2.1 for a description), L 1 is linear lag-1 prediction, L 2 is linear lag-2 prediction, Q 1 is quadratic lag-1 prediction, Q 2 is quadratic lag-2 prediction, SR 1 is Square-root lag-1 prediction, SR 2 is Square-root lag-2 prediction, %A is the relative percentage change in R2 using each of the prediction methods as compared to the R2 obtained with linear lag-1 prediction. For example the %A R3 —R2 for Quadratic lag-1 was calculated using *lRi " . R2a is the average R2 (unweighted). 3.2 The significance of the R2is and their relevance to the EMH Tables such as 3.1 and 3.2 alone do not address the question of whether the R2,s obtained are significantly different from those that could occur under the hypothesis that prices follow a RW. We have answered this question by using simulations and bootstrapping to construct hypothetical RW price data, which we then performed a canonical correlation analysis on. These results which are summarized in Tables 2.2, 2.3, 2.4, 2.5 indicate that for the original industries the R2,s that we have obtained can be expected under the hypothesis that prices follow a RW. However, when the data is standardized (which results in 13 portfolios) we find evidence that portfolio 10 can not be explained by a RW. Since this portfolio is a linear combination of each of the industries we view this as evidence that market prices as a whole do not follow a RW. 55 3.3 Why the quadratic information filter results in the highest R2ys We believe that one of the main reasons the quadratic predictive information filters works better than the linear predictive information filters is because the data is non-normal. Linear prediction uses the first two order moments of the sampling distribution of the data to perform its prediction whereas the use of quadratic prediction allows the use of up to the fourth order sample moment. This means that quadratic prediction can more accurately reflect the empirical distribution of the data. As shown in Figures A.5, A.6 and A.7 in the Appendix, the industry data is in fact non-normal with very heavy tails. 3.4 Comparison with Tsay's (1990) results The linear lag-1 predictive information filter is essentially the same as developed by Tsay (1990). In fact, the results obtained by both of these methods are identical. Using this predictor implies an implicit belief that today's prices contain all of the available information on tomorrow's prices. Since he also uses linear prediction, he would like the data to be normally distributed. We have been able to show that both of these assumptions are in error. On the basis of this we used non-linear predictive filters and obtained R2,s that are higher than with Tsay's method for most portfolios (10 out of 13 with the original data and 8 out of 13 with the standardized data). This meets the main objective of this thesis, namely to to be able to account for a larger percentage of stock price variations using stationary components than previously possible. On average our quadratic lag-2 predictor beat Tsay's linear lag-1 predictor by 18.6% with the original data and 6.3% with the standardized data. 56 3.5 Future Developments The choice of our predictive information filter depends on the type of filter and also on the number of time lags used. A stopping criteria for the number of lags used would be helpful. One possible solution is to calculate the following statistic of weighted R2,s for each lag used, £ W{R2 = Rf (3.26) t=i where Wi is a weight that could reflect the portfolios market value relative to the total, k is the number of portfolios used in the analysis and / is the lag used by the predictive filter. When and if R2 does not change much or reaches a maximum, the lag where this occurs is the best lag to choose. We decided to stop at lag-2 because of two reasons. The first was that at lags greater than two computational difficulties begin to arise for the quadratic and square-root predictive filters. The design matrix becomes singular and prevents the regression from being done, this is a direct result of a very large number of regressors (to be exact there are 3(13-f 132 +13(134-1)) = 1092). The second reason was that we felt two months was enough time for the market to absorb any information released by companies in their quarterly report. A second consideration is that we have only searched the class of linear, quadratic and square-root filters. It is very possible that a better filter exists which we did not investigate, such as E(Xt\Xi_1. Xt_2). Further studies may be to implement some para metric, semiparametric or non-parametric numerical schemes for approximating E(Kt\X_t-\,£.t-2)-57 3.6 Conclusions In this thesis we have considered three different predictive information filters. In the usual canonical correlation analysis setting only the linear lag-1 filter has been extensively used. Based on Tables 3.1 and 3.2 it is apparent that using different filters can substantially increase the i?2,s that we have considered in this thesis. In particular, the quadratic lag-2 filter resulted in a higher average R2 for both data sets that we considered. This indicates that filtering may lead to better results in the general canonical correlation setting. We have not derived any formal results to support this conclusion, but based on our empirical results we believe that filtering may play an important role in canonical correlation analysis. At the individual industry level we can not reject the EMH view that prices follow a RW. Some industries such as metals & minerals and paper & forest products indicate that stationary components in their prices can account for around 8% of their total variation in prices. However, when compared to the results from bootstrapping and simulations it indicates that such results are expected from a RW model for prices. When we consider portfolios of stocks, we find that portfolio ten has an R2 which lies beyond any of the 95% R2 quantiles calculated via bootstrapping or simulation. Since this portfolio is a linear combination of stocks from each of the industries, we take this as evidence that market prices as a whole may not be adequately be described as a RW. Our evidence is not overwhelmingly strong however, considering that the other 12 portfolios did not show any departure from the results obtained under the hypothesis that prices follow a RW. If the market is inefficient, than many new and creative models for prices are possible. Some authors (for example Arrow (1982)) have suggested incorporating psychological models of "irrational decision making" to try to explain the behavior of speculative asset 58 prices. Such models have in the past received little attention because they are based on an inefficient market. However, with the recent doubt cast on the EMH such models should receive more attention in the future. 3.7 Final Comments The debate surrounding the validity of the EMH is far from over. It is very interesting to note that many people vehemently believed the EMH was true before Summers (1986) proposed his "fads" model for prices . Jensen (1968) referred to the EMH as "best established empirical fact in economics" [10] and Keane (1980) in his monograph "The Efficient Market Hypothesis" [12] said: It is perhaps unfortunate that the efficient markets phenomenon should con tinue conventionally to be described as a "hypothesis", as if it were little more than academic speculation, when the fact is that it is a proposition which has received significant support of two decades of elaborate and rigorous testing Currently, many researchers strongly believe that the EMH hypothesis is not valid. The Wall Street Journal (October 23, 1987) called the EMH "the most remarkable error in the history of economic theory". The moral of the story is that one can never be be too sure about anything. The French philosopher Voltaire realized this two hundred years ago (referring of course to a different subject than statistics) when he wrote: It is only charlatans who are certain. 59 Bibliography [1] Blanchard, W., (1993), "Forecasting value-weighted real returns of TSE portfolios using dividend yields", UBC Press. [2] Box, G.E.P. and Tiao, G.C., (1977), "A canonical analysis of multiple time series", Biometrika, 64, 355-365. [3] Box, G.E.P. and Jenkins, G.M., (1976), Time Series Analysis: Forecasting and Control, Rev. Ed., Holden-Day, Oakland. [4] Daniel, K., and Tourous, W., (1991), "Common Stock Returns and the Business Cycle", Working Paper, UBC. [5] Dickey, D.A. and Fuller, W.A., (1979), "Distribution of the Estimators for Autore-gressive Time Series With a Unit Root", JASA, 74, 427-431. [6] Eckbo, E. and Liu, J., (1993), "Temporary Components of Stock Prices: New Uni variate Results", Journal of Financial and Quantitative Analysis, 28, 161-176. [7] Fama, E.F., (1970), "Efficient Capital Markets: A Review of Theory and Empirical Work", Journal of Finance, 25, 383-416. [8] Fama, E.F. and French, K.R., (1988), "Permanent and Temporary Components of Stock Prices", Journal of Politcal Economy, 96, 246-273. [9] Fuller, W., (1976) Introduction to Statistical Time Series, 1st Ed., John Wiley & Sons, New York. [10] Jensen, M., (1968), "The performance of Mutual Funds in the Period 1945-64", Journal of Finance, 23, 389-416. [11] Johnson, R.A. and Wichern, D.W., (1992), Applied Multivariate Statistical analysis, 3rd Ed., Prentice-Hall, New Jersey. [12] Keane, S., (1980), The Efficient Market Hypothesis, Gee and Co, Oxon. [13] Keynes, J.M., (1936), The General Theory of Employment, Interest, and Money, Harcourt, New York. [14] Kim, M., Nelson, C, and Startz, R., (1991), "Mean Reversion in Stock Prices? A Reappraisal of the Empirical Evidence", Review of Economic Studies, 58, 515-528. 60 [15] Lehmann, E.L., (1975), Nonparametrics: Statistical -methods based on ranks, Holden-Day, San Fransico. [16] Ljung, G.M. and Box, G.E.P., (1978), "On a Measure of Lack of Fit in Time Series Models", Biometrika, No. 2, 297-303. [17] Phillips, P.C.B., (1987), "Time Series Regression With a Unit Root", Econometrica, 55, No. 2, 277-301. [18] Richardson, M., (1993), "Temporary Components of Stock Prices: A skeptic's View", Journal of Business and Economic Statistics, 11, No. 2, 199-207. [19] Summers, L.H., (1986), "Does the Stock Market Rationally Reflect Fundamental Values?", Journal of Finance, 41, 591-601. [20] Traynor, P., ed The Toronto Stock Exchange Review, The Toronto Stock Exchange. [21] Tsay, R.S., (1990), "Correlation Transformation and Components of Stock Prices", University of Chicago technical report. 61 Appendix A TSE Index Formula and Rules [1] Each of the Toronto Stock Exchange indices measure the current aggregate market value (i.e. number of presently outstanding shares x current price) of the stocks included in the index as a proportion of an average base aggregate market value (number of base outstanding shares x average base price ± changes proportional to changes made in the current aggregate market value figure) for such stocks. The starting level of the base value has been set equal to 1000. Expressed more briefly this is: r*T-—~ Current aqqreqate market value INDEX = — ^— — x 1000 Adjusted average base aggregate market value Essentially, there are two stages in the production of indices: (1) establishment of an initial base and initial calculation of the indices; and (2) subsequent calculation of the indices taking into account recurring shifts of the market. Following is a detailed description of how The Toronto Stock Exchange indices are produced. The following formula is the basis for initial calculation of each of the indices of The Toronto Stock Exchange: INDEX = (K*Q*) + fr*QB) + --- + lKxQN) x 1000 (PAB x QAB) + (P_B x QBB) + .-. + (PNB x QNB) A,B, ... N: the various stocks in the index portfolio. PA,PB, • • • PN- the current board-lot market prices of each stock in the index. QAIQB, ...QN'- the numbers of currently outstanding shares of each stock in the index less any individual and/or related control blocks of 20% or more. 62 PAB,PBBI • • -PNB: the trade weighted average board-lot prices of each stock in the index during the base period QABIQBBI • • - QNB- the number of shares of each stock in the index outstanding in the base period less any individual and/or related control blocks of 20% or less. The base period is 1975. Calculation of the 1975 average base aggregate market value i.e. (PAB x QAB) + (PBB x QBB) + ... + (PNB x QNB) for each index was accomplished by multiplying the trade-weighted average board-lot price for each stock for the 1975 base period by the number of shares (share weight) of each stock outstanding at the beginning of the base period i.e. January 1, 1975 less any individual and/or control blocks of 20% or more. The current aggregate market value is determined using closing prices for each period for which the index is calculated multiplied by the number of shares then outstanding, less any individual and/or related control blocks of 20% or more, as at that period. As an example of these calculations, assume there are only two stocks in a hypothetical index. The problem is to calculate the level of the index as of January 31, 1975. Company 1 The current price (January 31, 1975) is $10 and the number of shares currently outstanding is 18,000. The average base aggregate market value in 1975 is $162,000. Company 2 The current price (January 31, 1975) is $25 and the number of shares currently outstanding is 30,000. The average base aggregate market value in 1975 is $690,000. Computation of the index would be as follows: INDEX = (IQx 18,000)+ (25 x 30,000) 162,000 +690,000 63 INDEX = _m x 1000 852,000 INDEX = 1091.55 ADJUSTMENT TO INDEX To calculate the indices subsequent to the establishment of the average base aggregate market value, recurring capital changes must be taken into account. Adjustments to the indices resulting from these changes must normally be introduced without altering the level of the index(see Bankruptcy Rule (7) for exception). In other words, continuity of the index must be preserved. To accomplish this, certain procedures are followed. These vary according to whether the adjustments result from: (1) the issuance of additional shares of a stock in the indices; or the addition to, withdrawal from, or substitution of stocks in the indices; (2) stock rights; (3) stock dividends and stock splits; (4) a liquidation of the company; (5) an asset spin-off; (6) takeover bid, amalgamation or merger; (7) a bankruptcy; or (8) a control block adjustment. Addition or Withdrawal of Shares or Changes in Number of Stocks Two steps are necessary to make adjustments for additions or withdrawals of shares to or from the index calculations: (1) Updating the current aggregate market value of the index. If additional shares of an index stock are issued, the current aggregate market value of the stocks in that index will be accordingly higher. Likewise, if a new stock is added to the index, or if a stock is removed, the current aggregate market value of that stock will be added to, or subtracted from the current aggregate market value of the other stocks in that index. 64 (2) Adjusting the average base aggregate market value of the index propor tional to the change in the current aggregate market value so that the index level will remain the same. The first step, therefore, towards making an adjustment is to calculate the new current aggregate market value as indicated in (1) above. The second step is to calculate the new average base aggregate market value. Ex pressed as a formula the second step would be as follows: Let the old average base aggregate market value = A. Let the un-adjusted current aggregate market value = C. Let the current aggregate market value of the capital to be added or withdrawn = D. The current adjusted aggregate market value will equal C±D. Therefore, to establish a new average base aggregate market value (B) for an index that formula is: To calculate the index on the new base, the formula for the hypothetical example given above would be: INDEX = (C:tD^ x 1000 B Continuing the example above, assume that Company 1 issued 2,000 new shares. This required an addition of $20,000 ($10 x 2,000) to the aggregate market value of the stocks in the index and therefore the new current aggregate market value resulting from the change is: 930,000 + 20,000 = 950,000. The average base aggregate market value of the index also has to be changed propor tionately. Here the formula B = A x ^CQD^ is used. B = 852,000 x<"+2°-00°> 930,00 65 B = 870,323 The index level remains unchanged as shown below: INDEX = x 1000 870,323 INDEX = 1091.55 Stock Rights The day the stock sells ex-rights, the additional shares resulting from the rights are included in the calculations to establish the current aggregate market value of the indices. The average base aggregate market value, however, is adjusted by taking into account both the market price and the subscription price because on ex-rights day the current market price, and accordingly aggregate market value, discounts the rights. The formula to calculate the new base aggregate market value following subscription to stock rights would be: B~bX C + D-S where S = the total capital subscribed for the newly issued shares. A concrete example of how a stock rights issue is incorporated into the index is the December 5, 1975 Bank of Nova Scotia offer. The Bank, with an outstanding capital of 18,562,500 shares, offered the shareholders of record at the close of business on December 5, 1975 rights to buy one new share at $36 per share of each 9 shares held. As a result, 2,062,500 new shares were issued. Ex-rights date was December 3, 1975 and from that date additional capitalization for the Bank of Nova Scotia used in the bank index was 66 2,062,500 shares times the current price (theoretically, at this opening on the "ex" date, $411 adjusted for the value of the right) amounting to $85,335,938. Actual subscription price was 2,062,500 shares times $36, amounting to $74,250,000. Calculations for the proportionately adjusting the base were as follows: Bank Index Un-adjusted current aggregate market value:* $4,193,109,375 Un-adjusted base ag gregate market value:* $1,337,840,000 New current aggregate market value after allowing for rights: (4,194,109,375 + 85,335,938) = $4,278,445,313 New base aggregate market value after allowing for rights offering: 1,337,840,000 x _n ' ' V » » nnn = $1,361,467,499 4,278,445,313 - 74,250,000 , As at the close on the day prior to the ex-date. Adjustments are made after the close and before the market opens the following day. Bank of Nova Scotia closed at $42 on December 2, 1975. Stock Dividends, Splits, and Consolidations On the ex-dividend day the outstanding share total is increased by the number of shares issued in the form of dividends. Theoretically, the price of the stock should drop by the extent of the worth of the dividend. The current aggregate market value, therefore, will not change. Hence the base figure is not adjusted. Similarly, in the case of share splits, the increased number of shares times the lower price should equal the old number of shares times the higher price. Thus, the current aggregate market value is theoretically unchanged, and the base figure is not adjusted. The same reasoning holds in the case of stock consolidations, except that the higher price time the smaller number of shares leaves the current aggregate market value unchanged. 67 Liquidation of A Company Effective January, 1979, where a capital distribution is announced as being a liquidation of a company whose stock is included in the index, that stock will be removed from the index effective the ex-distribution date. Asset Spin-off Effective January, 1979, adjustments necessary to leave the level of an index unchanged when a stock in that index has its per share value decreased through an asset spin-off are made at the opening of the ex-distribution day or as soon thereafter as the value of the asset being spun-off is known by the Exchange staff. Thus the staff may have to recalculate index values if a stock trades "ex-asset spin-off" without the index being stabilized. Takeover Bid, Amalgamation or Merger Effective January, 1979, changes in share weight or control blocks resulting from takeover bids, amalgamations or mergers are incorporated into the index as soon as is administra tively possible after the fact. This procedure replaces the former procedure of incorpo rating such changes at the next quarterly update made just after the end of the calendar quarter to which they relate. Bankruptcy of Stock in Index System If and when any company, whose stock is included within the TSE "300" indices, has made an assignment in bankruptcy or been placed in receivership, its stock will be re moved as soon as possible at the lowest possible price per share (one-half cent under the present computer programmes) rather than at the last board-lot price before trading was 68 suspended. If, as, and when the company recovers in any form, it will only be eligible to be included in the index system again after fully complying with and meeting all criteria; that is, after qualifying in the normal fashion. Control Blocks (a) All known individual and related control blocks equal to 20% or more of the share capital of any stock included in the indices is removed in order to reflect, as nearly as may be practical, the market float or stock normally available to portfolio investors. (b) If at any time more than 90% of the outstanding shares which are included in the TSE 300 index is held by a controlling group; as defined by the methods of computing control group holdings for index weighting purposes, or if the shares in public hands of the same class are so reduced that the value calculated by multiplying the most recent share price by the number of shares held by parties other than the control group is insufficient to meet the market capitalization criterion for admission to the index, then each such class of equity security shall be removed from the index as soon as is conveniently practicable. (c) If an individual control block of 20% or more, or a related group of control blocks which in aggregate total 20% or more of the relevant shares outstanding, are initially removed from the total of such shares then outstanding for purposes of computing the share weight of the stock in the index portfolio, and (1) the holder or holders of such stock subsequently sell stock from their position to reduce the amount of such stock holding(s) below 20%, then the holding(s) will be added back to the float at the first practical time subsequent to such sale; (2) if the 20% or more block(s) subsequently falls below 20% as a result of an increase or increases in the total of such share capital outstanding, then such block(s) will not be added back to the share weight until such time as the holding falls or is reduced to 15% or less and as soon thereafter as is practical for it to be added 69 back. Frequency of Adjusting the Index Stock rights, stock dividends, splits, consolidations, and liquidations are reflected in the calculations of the indices immediately as they become affective, i.e. on the "ex" date. Asset spin-offs are reflected effective the "ex" date or as soon thereafter as the value of the asset being spun-off is known by the Exchange staff. Takeovers, amalgamations and mergers are reflected as soon as possible after the fact. Bankruptcy and receivership sit uations are reflected as soon as possible after they are announced. Any changes resulting from the annual post-year-end revision as noted in the section entitled "Stock Eligibility Criteria" are made at the end of the first calendar quarter. Other changes (such as those related to control blocks or to addition or withdrawal of shares) are usually made on a quarterly basis. Additions or deletions of stocks are usually made on a quarterly basis but may be necessary at other times due to delistings caused by takeovers, amalgamations, or mergers or to normal delistings. 70 Results using linear lag-1 prediction Table A.1: ARIMA models using linear lag-1 prediction Model for Transformed series Q(12) DF (1 - .1045 + .11552)(1 - B)y1>t = 17.0 0.2 (1 - .9135)(1 - 5)2/2,, = (1 + MB)e2,t 12.3 -1.0 (l-.993B)jfcit=e3,t 3.9 -2.9 (l-.12B)(l-B)y4,«=C4,« 14.9 -8.5 (l-.985B)ifcit=cB>« 29.7 -7.0 (l-.977B)ite,t=C6,t 18.7 -10.9 (l-.967B)j/7,t=e7,4 9.7 -15.2 (l-.952B)ite,,=e8i, 16.7 -22.7 (l-.944B)ite,,=£9.t 14.4 -26.0 (l-.928B)i/io,i=eio,t 10.8 -33.6 (l-.913B)i/n,t=cii,» 9.8 -41.0 (l-.871B)yia,i=ci2,t 10.5 -60.7 (l-.854B)yi3,J=£i3,t 12.1 -68.8 All parameters are greater than their two standard errors. Series 5 has a significant Q(12) statistic. A runs test was performed on the residuals of series 5 and indicated no serial dependence. Table A.2: Portfolio R2,s using linear lag-1 prediction Portfolio %R2 O %R2 S Portfolio %R2 O %R2 S 1 0.50 0.63 8 4.25 13.10 2 0.50 13.34 9 4.50 12.41 3 3.95 5.37 10 5.70 61.66 4 5.85 27.48 11 1.96 4.29 5 3.31 36.65 12 3.84 18.42 6 0.51 3.92 13 1.46 3.40 7 9.49 27.73 where O is the original data and S is the standardized data. Portfolio i for the original data corresponds to industry i. 71 Results using linear lag-2 prediction Table A.3: ARIMA models using linear lag-2 prediction Model for Transformed Series Q(12) DF (1 + .475£)(1 - B)yltt = (1 - .6QB)e1<t 19.8 0.2 (1 - .911J3)(1 - 5)2/2,t = (1 + .835)e2,t 11.6 -1.1 (l-.993B)2/3,«=e3,< 3.5 -3.1 (1-.12B)(1-B)y4lt=£4,« 13.2 -7.8 (l-.984B)ite,j=C5,« 28.2 -7.1 (l-.976B)iteit=e6,t 18.4 -10.9 (l-.967B)y7,t=£7,t 9.3 -15.4 (l-.953B)ite,«=£8lt 16.2 -21.7 (l-.945B)ife,t=e9i4 13.6 -25.7 (l-.931B)»i0l«=cio.t 10.0 -32.4 (l-.913B)yii,t=ci1>t 10.1 -40.8 (l-.863B)yi2,t=ei2,« 12.0 -64.5 (l-.852B)yi3,t=ci3,t 11.55 -69.4 All parameters are greater than their two standard errors. Series 1 and 5 have a significant Q(12) statistic. A runs test was performed on the residuals of series 1 and 5 and indicated no serial dependence. Table A.4: Portfolio R2,s using linear lag-2 prediction Portfolio %R2 O %R2 S Portfolio %R2 O %R2 S 1 0.48 0.60 8 4.35 13.50 2 0.52 13.96 9 4.74 13.72 3 3.83 5.18 10 5.61 61.06 4 5.84 26.60 11 2.04 4.45 5 3.30 35.84 12 3.79 18.25 6 0.49 3.80 13 1.49 3.41 7 9.52 27.75 where O is the original data and S is the standardized data. 72 Results using quadratic lag-1 prediction Table A.5: ARIMA models using quadratic lag-1 prediction Model for Transformed Series Q(12) DF (l-.10B-r.llBa)(l-B)yM = eM 17.1 0.2 (1 - B2)y2,t = (1 + .96B)e3,t 16.1 -1.1 (l-.994B)y3,j=C3,t 3.4 -2.8 (1-.040B)(1-B)y4 t = (1 - .08B)c4,t 14.7 -8.7 (l-.984B)j/5,«=£5,t 29.5 -7.4 (l-.977B)y6,t=C6,t 15.4 -11.0 (l-.967B)y7|(=e7,« 10.4 -15.4 (l-.952B)j/8,,=e8,( 12.9 -22.6 (l-.945B)y9il=c9,1 12.8 -25.6 (l-.922B)yio,t=cio,t 7.1 -36.7 (l-.919B)yii,,=£ii,« 5.4 -38.4 (l-.869B)yi2,.=ei2,« 9.6 -61.9 (l-.859B)W3,t=£i3,t 12.6 -66.5 All parameters are greater than their two standard errors. The Q(12) statistic for series 5 is significant at the 5% level. A runs test was performed on the residuals of series 5 and indicated no serial dependence. Table A.6: Portfolio industry R2,s using quadratic lag-1 prediction Portfolio %R2 O %R2 S Portfolio %R? O %R2 S 1 0.51 0.65 8 4.20 12.43 2 0.58 11.84 9 4.53 13.49 3 3.94 5.43 10 6.03 64.22 4 5.86 30.93 11 1.68 3.83 5 3.16 35.94 12 3.59 16.52 6 0.43 3.61 13 1.47 3.49 7 9.43 29.21 where 0 is the original data and S is the standardized data. 73 Results using quadratic lag-2 prediction Table A.7: ARIMA models using quadratic lag-2 prediction Model for Transformed Series Q(12) DF (l-.999B)yM=eM 21.2 0.2 (l-.876B)(l-B)i/2,t=(l + .77B)e2,t 6.62 -1.2 (1+.002B)(1-B)y3,,=e3,< 2.6 -2.9 (1-.156B)(1-B)y4i<=e4,t 13.1 -10.0 (l-.986B)y6,i=c5,t 15.3 -6.3 (l-.977B)y6il=c6,« 15.3 -10.8 (l-.962B)y7,t=e7,( 7.9 -17.9 (l-.958B)y8i«=cg,t 18.2 -20.0 (l-.942B)jte,,=c9,t 16.2 -27.5 (l-.925B)yio,.=cio,« 15.9 -34.9 (l-.858B)y1M=en,t 9.9 -66.6 (l-.909B)yi2,«=ei2,t 8.6 -42.8 (l-.875B)y13,t=ei3,t 10.2 -58.8 All parameters are greater than their two standard errors. The Q(12) statistics for series 1 is significant at the 5% level. A runs test was performed on the residuals of series 1 and indicated no serial dependence. Table A.8: Portfolio industry i?2,s using quadratic lag-2 prediction Portfolio %R2 O %R2 S Portfolio %R2 O %R2 S 1 0.88 0.99 8 3.56 8.82 2 1.07 11.54 9 7.82 22.54 3 5.34 7.65 10 5.88 63.36 4 6.79 23.03 11 2.00 3.09 5 3.61 41.91 12 4.93 18.98 6 0.51 3.89 13 2.04 4.09 7 9.84 32.95 where 0 is the original data and S is the standardized data. 74 Results using square-root lag-1 prediction Table A.9: ARIMA models using square-root lag-1 prediction Model for Transformed Series Q(12) DF (1 - .1035 + .1105^(1 - B)yM=eM 17.3 0.2 (1-.909B)(1-B)yj,, = (1 + .8275)e2|1 10.4 -1.1 (l-.994B)y3.«=C3,t 3.1 -2.8 (l-.12B)(l-B)y4>1=C4,« 14.8 -8.7 (l-.984B)y5,(=€5,t 29.6 -7.4 (l-.976B)y6lt=C6,t 15.4 -11.0 (l-.967B)y7,t=e7,t 10.7 -15.5 (l-.952B)y8,,=e8,( 16.1 -22.6 (l-.946B)y9,,=£9,( 12.7 -25.6 (l-.923B)y10,«=cio1« 7.1 -36.6 (l-.918B)yii,1=en>, 5.6 -38.7 (l-.869B)y12,1=ci2.« 9.8 -61.7 (l-.859B)y13,t=ci3,i 12.8 -66.5 All parameters are greater than their two standard errors. The Q(12) statistic for series 5 is significant at the 5% level. A runs test was performed on the residuals of series 5 and indicated no serial dependence. Table A.10: Portfolio industry i?2's using square-root lag-1 prediction Portfolio %R2 O %R? S Portfolio %R2 O %R2 S 1 0.51 0.64 8 4.21 12.66 2 0.59 11.56 9 4.57 13.74 3 3.94 5.48 10 6.06 64.32 4 5.88 31.22 11 1.72 3.87 5 3.15 35.82 12 3.56 16.34 6 0.43 3.48 13 1.49 3.41 7 9.50 28.95 where O is the original data and S is the standardized data. 75 Results using square-root lag-2 prediction Table A. '. .1: ARIMA models using square-root lag-2 p Model for Transformed Series Q(12) DF (l-B)yM=(l - .05fl)eM 18.4 0.2 (1-.883B)(1-B)y2,t=(l + .785)e2,t 7.1 -1.2 (l-.993B)y3,,=e3,t 2.5 -2.9 (1-.158B)(1-B)y4l,=e4,< 13.8 -10.0 (l-.986B)ite,«=C5,t 15.8 -6.4 (l-.976B)y6>t=c6,t 13.1 -10.9 (l-.961B)y7,t=C7,« 7.9 -18.6 (l-.958B)y8,t=e8,« 19.7 -19.6 (l-.943B)y9,t=f9,« 15.8 -27.0 (l-.922B)yio,«=cio,t 15.9 -36.6 (l-.864B)yu,*=eii,t 8.4 -64.0 (l-.906B)yi2,i=ei2,t 8.8 -44.3 (l-.876B)yi3,t=ei3,t 10.3 -58.3 All parameters are greater than their two standard errors. The Q(12) statistic for series 5 is significant at the 5% level. A runs test was performed on the residuals of series 5 and indicated no serial dependence. Table A. 12: Portfolio industry R2,s using square-root lag-2 prediction Portfolio %Rl O %R2 S Portfolio %R2 O %R2 S 1 0.82 0.75 8 3.51 8.42 2 1.02 11.96 9 7.76 22.17 3 5.34 7.18 10 5.81 65.43 4 6.69 26.64 11 2.12 4.34 5 3.62 38.18 12 4.58 16.47 6 0.49 4.65 13 1.95 3.65 7 9.77 31.72 where 0 is the original data and S is the standardized data. 76 Bootstrapping from original data: Sample size = 1000 Communications Financial Gold Industrial 0.0 0.15 0.30 R-square o ci 0.0 0.2 0.4 R-square 0.0 0.4 R-square o 0.0 0.4 R-square Conglomerates Merchandising Metals Oil <8 o 0.0 0.4 R-square CO d 0.0 0.4 R-square a O 0.0 0.4 R-square o 0.0 0.4 R-square Paper Pipeline Transportation Utilities CO ci ci o ci 0.0 0.4 R-square 0.0 0.4 R-square CO ci ci o ci 0.0 0.4 R-square CO o o ci 0.0 0.2 0.4 R-square Consumer CD o 0.0 0.2 R-square Figure A.1: Quantiles of R2 using bootstrapping with original data. 77 Bootstrapping from standardized data: Sample size = 1000 Portfolio 1 Portfolio 2 Portfolio 3 Portfolio 4 0.0 0.2 0.4 R-square Portfolio 5 a o o d 0.0 0.4 0.8 R-square Portfolio 6 0.0 0.4 0.8 R-square Portfolio 7 a 00 o 0.0 0.4 0.8 R-square Portfolio 8 0.0 0.4 0.8 R-square Portfolio 9 0.0 0.4 R-square Portfolio 10 o.o 0.4 0.8 R-square Portfolio 11 O o d 0.0 0.4 0.8 R-square Portfolio 12 0.0 0.4 0.8 R-square Portfolio 13 o d 0.0 0.4 0.8 R-square o 0.0 0.4 0.8 R-square 0.0 0.3 0.6 R-square 0.0 0.4 0.8 R-square Figure A.2: Quantiles of R2 using bootstrapping with standardized data. 78 Simulated data without standardization: Sample size = 1000 Communications Financial Gold Industrial I 3 o co o 0.0 0.10 R-square Conglomerates 3 o o ci 0.0 0.2 0.4 R-square Merchandising 3 o o d 0.0 0.4 0.8 R-square Metals CO ci o ci 0.0 0.4 R-square Oil W 3 o 0.0 0.3 0.6 R-square Paper 0.0 0.2 0.4 R-square Pipeline n o 0.0 0.4 R-square Transportation o Utilities 3 o o ci 0.0 0.4 R-square 0.0 0.3 0.6 R-square 0.0 0.3 0.6 R-square 3 a o ci 0.0 0.2 0.4 R-square Consumer 18 3 o o d 0.0 0.2 R-square Figure A.3: Quantiles of R2 using simulations with original data. 79 Simulated data with standardization: Sample size = 1000 Portfolio 1 Portfolio 2 Portfolio 3 Portfolio 4 0.0 0.4 0.8 R-square Portfolio 5 o o d 0.0 0.4 0.8 R-square Portfolio 6 o d 0.0 0.4 0.8 R-square Portfolio 7 Portfolio 8 o d Portfolio 9 0.0 0.4 0.8 R-square Portfolio 10 o 0.0 0.4 0.8 R-square Portfolio 11 o d 0.0 0.4 0.8 R-square Portfolio 12 0.0 0.4 0.8 R-square =3 o o d 0.0 0.4 0.8 R-square a o d 0.0 0.4 R-square cd a 00 d o d 0.0 0.4 0.8 R-square Portfolio 13 d 0.0 0.4 0.8 R-square Figure A.4: Quantiles of R2 using simulations with standardized data. 80 q-q plot for Communications -3-2-10 1 2 3 Quantiles of Standard Normal q-q plot for Gold and Silver CO *>? CD _ (D -3-2-10 1 2 3 Quantiles of Standard Normal q-q plot for Conglomerates -3-2-10 1 2 3 Quantiles of Standard Normal q-q plot for Financial Services -3-2-10123 Quantiles of Standard Normal q-q plot for Industrial Products -3-2-10 1 2 3 Quantiles of Standard Normal q-q plot for Merchandising -3-2-10123 Quantiles of Standard Normal Figure A.5: Q-Q plots of monthly log(index) for TSE industry portfolios 81 q-q plot for Metals q-q plot for Oil and Gas 3 -2-10 1 2 Quantiles of Standard Normal q-q plot for Paper Products -2-10 1 2 Quantiles of Standard Normal q-q plot for Pipelines -3-2-10 1 2 Quantiles of Standard Normal q-q plot for Real Estate and Construction -3 -2-10 1 2 Quantiles of Standard Normal q-q plot for Transportation -3-2-10 1 2 3 Quantiles of Standard Normal -2-10 1 2 Quantiles of Standard Normal Figure A.6: Q-Q plots of monthly log(index) for TSE industry portfolios 82 Figure A.7: Q-Q plots of monthly log(index) for TSE industry portfolios 83
- Library Home /
- Search Collections /
- Open Collections /
- Browse Collections /
- UBC Theses and Dissertations /
- An investigation into the Efficient Market Hypothesis:...
Open Collections
UBC Theses and Dissertations
Featured Collection
UBC Theses and Dissertations
An investigation into the Efficient Market Hypothesis: a canonical correlation analysis approach Smith, Daren McCrossan 1995-12-31
pdf
Page Metadata
Item Metadata
Title | An investigation into the Efficient Market Hypothesis: a canonical correlation analysis approach |
Creator |
Smith, Daren McCrossan |
Date | 1995 |
Date Issued | 2009-01-31 |
Description | In this thesis we will consider the Efficient Market Hypothesis (EMH). Fama (1970) defined three levels in which to test market efficiency: weak, semi-strong, and strong, each level depending on the particular set of information being used to assess efficiency. We will mainly address weak level efficiency in which the information set is past security data. Before the mid 1980's it was widely believed that the E M H was true at the weak and semi-strong levels. It was not until the pioneering work of Shiller (1984) and Summers (1986) that some doubt was cast on the E M H . They proposed an inefficient model in which prices consist of a sum of a random walk component and a stationary (predictable) component which represents the market valuation error. Since their initial conjecture about a stationary component in stock prices much effort has been spent in trying to determine if it exists and if it does, determining how much of the variations in stock prices it accounts for. To investigate this problem we will use a combination of data filtering, canonical correlation analysis, simulations and bootstrapping. Using industry price data obtained from the Toronto Stock Exchange over the period January 1956 to June 1995, we find some evidence against the EMH. |
Extent | 3647425 bytes |
Genre |
Thesis/Dissertation |
Type |
Text |
File Format | application/pdf |
Language | eng |
Collection |
Retrospective Theses and Dissertations, 1919-2007 |
Series | UBC Retrospective Theses Digitization Project |
Date Available | 2009-01-31 |
Provider | Vancouver : University of British Columbia Library |
Rights | For non-commercial purposes only, such as research, private study and education. Additional conditions apply, see Terms of Use https://open.library.ubc.ca/terms_of_use. |
DOI | 10.14288/1.0087017 |
URI | http://hdl.handle.net/2429/4045 |
Degree |
Master of Science - MSc |
Program |
Statistics |
Affiliation |
Science, Faculty of Statistics, Department of |
Degree Grantor | University of British Columbia |
Graduation Date | 1995-11 |
Campus |
UBCV |
Scholarly Level | Graduate |
Aggregated Source Repository | DSpace |
Download
- Media
- ubc_1995-0645.pdf [ 3.48MB ]
- Metadata
- JSON: 1.0087017.json
- JSON-LD: 1.0087017+ld.json
- RDF/XML (Pretty): 1.0087017.xml
- RDF/JSON: 1.0087017+rdf.json
- Turtle: 1.0087017+rdf-turtle.txt
- N-Triples: 1.0087017+rdf-ntriples.txt
- Original Record: 1.0087017 +original-record.json
- Full Text
- 1.0087017.txt
- Citation
- 1.0087017.ris
Full Text
Cite
Citation Scheme:
Usage Statistics
Country | Views | Downloads |
---|---|---|
United States | 13 | 0 |
Russia | 10 | 0 |
China | 8 | 60 |
France | 4 | 0 |
Turkey | 1 | 0 |
Poland | 1 | 0 |
City | Views | Downloads |
---|---|---|
Saint Petersburg | 10 | 0 |
Shenzhen | 6 | 60 |
Unknown | 5 | 41 |
Ashburn | 3 | 0 |
Dallas | 3 | 0 |
Los Angeles | 2 | 0 |
Beijing | 2 | 0 |
Jacksonville | 2 | 0 |
Mountain View | 2 | 0 |
Kansas City | 1 | 0 |
Çorum | 1 | 0 |
{[{ mDataHeader[type] }]} | {[{ month[type] }]} | {[{ tData[type] }]} |
Share
Embed
Customize your widget with the following options, then copy and paste the code below into the HTML
of your page to embed this item in your website.
<div id="ubcOpenCollectionsWidgetDisplay">
<script id="ubcOpenCollectionsWidget"
src="{[{embed.src}]}"
data-item="{[{embed.item}]}"
data-collection="{[{embed.collection}]}"
data-metadata="{[{embed.showMetadata}]}"
data-width="{[{embed.width}]}"
async >
</script>
</div>
Our image viewer uses the IIIF 2.0 standard.
To load this item in other compatible viewers, use this url:
http://iiif.library.ubc.ca/presentation/dsp.831.1-0087017/manifest