The full text on this page is automatically extracted from the file linked above and may contain errors and inconsistencies.
1996 Quarter 1 Practical Issues in Monetary Policy Targeting 2 by Stephen G. Cecchetti The Reduced Form as an Empirical Tool: A Cautionary Tale from the Financial Veil 16 by Ben Craig and Christopher A. Richardson Predicting Real Growth Using the Yield Curve by Joseph G. Haubrich and Ann M. Dombrosky FEDERAL RESERVE BANK OF CLEVELAND 26 http://clevelandfed.org/research/review Economic Review 1996 Q1 1 ECONOMIC REVIEW 1996 Quarter 1 Vol. 32, No. 1 Practical Issues in Monetary Policy Targeting 2 by Stephen G. Cecchetti This paper outlines the considerable information requirements faced by monetary policymakers and looks at the data to see what we actually know and how well we know it. The author’s main conclusion is that our forecasting ability is very poor, which creates uncertainty that leads to cautious policymaking. At a more practical level, he finds that nominal-income targeting rules are more robust than price-targeting rules in the sense that someone who cares about the aggregate price path loses little by targeting nominal income, but someone who cares about nominal income is made much worse off by moving to a price-level target, which substantially destabilizes real output. The Reduced Form as an Empirical Tool: A Cautionary Tale from the Financial Veil 16 Economic Review is published quarterly by the Research Department of the Federal Reserve Bank of Cleveland. Copies of the Review are available through our Corporate Communications & Community Affairs Department. Call 1-800-543-3489 (OH, PA, WV) or 216-579-2001, then immediately key in 1-5-3 on your touch-tone phone to reach the publication request option. If you prefer to fax your order, the number is 216-579-2477. Economic Review is also available electronically through our home page on the World Wide Web: http:// www.clev.frb.org. by Ben Craig and Christopher A. Richardson The reduced-form empirical strategy has been used for more than 30 years to test the Modigliani–Miller model of corporate financial structure. Curiously, the early tests almost always accepted the model, whereas subsequent tests almost always reject it. This paper considers the limitations of the reduced-form strategy that led to the early, spurious results, and demonstrates why an empirical strategy that is not closely tied to an underlying economic theory of behavior will usually yield estimates that are too imprecise or too unreliable to form a basis for policy. Predicting Real Growth Using the Yield Curve 26 by Joseph G. Haubrich and Ann M. Dombrosky The yield curve, which relates interest rates to notes and bonds of various maturities, is often used by economists and business analysts to predict future economic growth. But how reliable is it? This article uses out-ofsample regressions to determine how well the 10-year, three-month yield spread predicts future real GDP growth. The authors show that although the yield curve is a good predictor over the entire 30-year sample period, it has become much less accurate over the last decade. Editorial Board: Charles T. Carlstrom Ben Craig Kevin J. Lansing William P. Osterberg Editors: Tess Ferg Michele Lachman Robin Ratliff Design: Michael Galka Typography: Liz Hanna Opinions stated in Economic Review are those of the authors and not necessarily those of the Federal Reserve Bank of Cleveland or of the Board of Governors of the Federal Reserve System. Material may be reprinted provided that the source is credited. Please send copies of reprinted material to the editors. ISSN 0013-0281 2 Practical Issues in Monetary Policy Targeting by Stephen G. Cecchetti Introduction What do monetary policymakers need to know and when do they need to know it? Textbook descriptions and academic discussions of policymaking usually ignore the practical problems faced by those who make the decisions and take the actions. While most economists would agree that monetary policy has real short-run effects and is most likely neutral in the long run, they could provide no more than informed speculation in helping decide at what level to set the target for a policy instrument and when to change it. This paper’s purpose is to outline the type of information monetary policymakers need in practice, and to examine the data to see what we actually know.1 Any policy rule must be formulated in several clearly defined steps. First, one must identify an operational instrument, best thought of as something policymakers can control precisely, like the federal funds rate or the monetary base. Next, there must be a target. Many central banks have stated that price stability is their goal, but an obvious alternative to targeting the aggregate price level is targeting nominal income.2 In addition to choosing the target variable itself, formulating policy necessi- Stephen G. Cecchetti is a professor of economics at Ohio State University and a research associate at the National Bureau of Economic Research and the Federal Reserve Bank of Cleveland. The author thanks Charles T. Carlstrom for his helpful comments on this article. tates specifying a loss function: What is the relative importance of large and small, or positive and negative, deviations of aggregate prices from their target path? One might also assign a cost to large movements in the target variable. For example, it might be important for the Federal Reserve to have a reputation for changing the federal funds rate target smoothly, without large movements or sudden reversals, to avoid creating uncertainty in financial markets. The next stage in devising a monetary rule is to link the operating instrument with the target. This requires specification and estimation of a macroeconomic model. One needs quantitative answers to questions of the form “If the federal funds rate is moved by one percentage point, what will be the path of the aggregate price level and real output over the following three years?’’ Not only do we require a point estimate of this response function, but it is also crucial that we know how precise our knowledge is in a statistical sense. ■ 1 This work is based on Cecchetti (1995). ■ 2 I will not discuss the difference between price-level and inflation targeting. While this is a potentially important practical distinction, it is beyond the scope of this paper. 3 Finally, policymakers need a timely estimate of their target variable’s future pathin the absence of any policy actions. In other words, they must know when external shocks hit the economy, how large they are, and what their impact on the time path of aggregate prices and real output will be. The next section offers a detailed discussion of the modeling issue: How do we formulate and estimate the necessary simple, dynamic, empirical macroeconomic model? The section’s first major part looks at econometric identification. What must we assume in order to disentangle the fluctuations in output and prices into their various components? How might we actually estimate the impact of monetary policy on macroeconomic quantities of interest? The section’s second major part discusses the issue of structural stability. Monetary policymakers change their emphasis fairly frequently, focusing on one indicator one year and another the next. How does this affect our attempt to estimate the relationship between the variables that policymakers control (like the federal funds rate) and the things we care about? Can we estimate a stable relationship between output, prices, and interest rates over any reasonable period? The methodological discussion of modeling issues is followed by section II, in which I present a particular model and examine its properties. Several different types of results are included. First, I look at the impact of different sources of shocks on the variables of interest. Besides allowing answers to questions like “If the federal funds rate were to rise by 100 basis points, how much would output change over the next three years?” this approach makes it possible to examine the sources of fluctuations in output and prices. For example, has monetary policy been responsible for a significant share of output variation over the past decade? Section III discusses how a policy rule can be formulated. The first step is to specify an objective function: What do policymakers actually want to stabilize? This discussion emphasizes the need for taking account of imprecision when forming a policy rule. We are uncertain how changes in the interest rate affect the size and timing of output and price movements. This means we cannot confidently predict policy actions’ impact on target variables, so that policy actions differ from what they would be if we had perfect knowledge. From the theoretical discussion, I move on to examine several possible objective functions of policy and the interest rate paths implied by the combination of each rule and the estimated model. I focus throughout on the importance of employing rules that recognize the imprecision of our knowledge regarding the size of the linkages we need to estimate. I reach three significant conclusions: First, since prices take time to respond to all types of economic shocks, the objective of price stability implies raising the federal funds rate immediately after a shock, instead of waiting for prices to rise. Second, and more important, comparing the results of price-level targeting with those of nominal-income targeting implies that the difficulties inherent in forecasting and controlling the former provide an argument for concentrating on the latter. Finally, it is possible to use policy rules to see how closely recent movements in the federal funds rate conform to those implied by either price-level or nominalincome targeting rules. The results show that the policy that is optimal in this limited sense involves faster, bigger movements than those shown by the actual federal funds rate path. This suggests that policymakers’ actions have been based on something akin to nominalincome targeting, but with costs attached to interest rate movements. I. Modeling Issues The single biggest problem in formulating monetary policy rules is how to construct an empirical macroeconomic model that describes the critical aspects of the economy. It is important that the model be dynamic, summarizing the impacts of shocks to the economy—as well as those of intended policy actions—over time. The standard response to this challenge has been to construct various forms of vector autoregressions (VAR). A VAR can answer a question of the following type: “If the federal funds rate moves, when and by how much does the price level change?” Policymakers require quantitative answers to exactly these kinds of questions. To construct any usable empirical model, a researcher must make a number of choices. I will describe four of these: 1) Which variables should be included in the model? 2) What is the appropriate measure of monetary policy? 3) How can the model be identified econometrically? and 4) Over what sample period should the model be estimated? 4 Variable Inclusion Identification When trying to discern the relationship between A model builder’s most complex decision is formulating a set of “identifying assumptions.” This is also the subtlest issue and the one that has generated the most discussion in the literature. It is like the textbook question about estimating supply and demand curves: There, if data on the price and quantity of a good in a market both move, we cannot tell whether the root cause of the change was a shift in supply or a shift in demand. Here, things are a bit less transparent, because there are no clearly defined supply and demand curves in the standard microeconomic sense. Instead, it is necessary to distinguish whether prices, output, and interest rates moved as a result of policy shifts, or because of factors like changes in the price of oil (an aggregate supply shock) or in the demand for money (an aggregate demand shock). To understand the problem and its solution more fully, we can begin by writing down a dynamic structural model in its moving-average form: inflation, output, and monetary policy, should we include other variables in the model? Our answer is guided by the findings of Sims (1992), who estimates a model with prices, output, and an interest rate for several countries. His robust overall conclusion is that with this specification, increases in the interest rate (which should signal policy contractions) lead to prices that are higher than otherwise expected, not lower. This problem, which came to be known as the “price puzzle,” can be eliminated by including commodity prices in the model. The reasoning is that the policymaker has additional knowledge about prices’ future path that the three-variable model does not adequately summarize. Policy contractions, being based on this omitted information, signal that these other indicators are pointing toward higher prices. More recent research, like that of Christiano, Eichenbaum, and Evans (1996a, 1996b), has shown that including commodity prices eliminates the puzzle. They suggest that higher commodity prices mean higher future overall prices, and that policymakers respond to this. In other words, an upward move in commodity prices precedes both a rise in the price level and a tightening of policy in the form of an increase in the federal funds rate. The omission of this information from the original Sims formulation led to a bias in which contractionary policy predicts higher aggregate prices. This is not a policy change, but simply a reaction to external events. The models of Christiano, Eichenbaum, and Evans do have the following property: Moving toward a more contractionary monetary policy drives prices down (relative to the trajectory they would follow without the policy change). Choice of Policy Instrument Beyond the question of which variables the model should include, it is necessary to specify a monetary policy instrument. Should one assume that policymakers are focusing on the federal funds rate itself (or behaving as if they were), or would it be more realistic to use nonborrowed reserves as the instrument? The literature takes up this issue in some detail.3 Because events of the past 15 years suggest that the primary focus has been on the federal funds rate, I will assume that it contains the information necessary to gauge policy actions.4 (1) pt = A11(L)ept + A12(L)ect + A13(L)eyt + A14(L)u t (2) p tc = A21(L)ept + A22(L)ect + A23(L)eyt + A24(L)u t (3) yt = A31(L)ept + A32(L)ect + A33(L)eyt + A34(L)u t (4) rt = A41(L)ept + A42(L)ect + A43(L)eyt + A44(L)u t , where pt , p tc, and yt are the logs of the aggregate price level, commodity prices, and output, respectively, rt is the policy indicator, the e’s are exogenous shocks, and u is the policy innovation. Equations (1) – (4) summarize the impact of all the shocks to the economy. The Aij (L)’s are lag polynomials in the lag operator L. For example, ¥ A11(L)ept = S a11i Li ept i=0 = a110 ept + a111ept – 1+ .... . ■ 3 See, for example, discussions in Christiano, Eichenbaum, and Evans (1996a, 1996b) and Bernanke and Mihov (1995). ■ 4 Most results are unaffected by the substitution of nonborrowed reserves, suggesting that the funds rate elasticity of reserve demand is relatively stable. 5 Because we do not observe the shocks, it is not possible to estimate the model (1)–(4) directly. Instead, we estimate the more familiar VAR form and place restrictions on the coefficients (the ai j k ’s) in order to recover estimates of the shocks. Identification entails determining the errors in this four-equation system, that is, the actual sources of disturbances that lead to variation in prices, output, and interest rates. As the appendix to this paper describes, when there are four endogenous variables, six restrictions are required for complete identification. All identification schemes involve assumptions about how these sources of variation are correlated. Researchers use two types of restrictions for this purpose. The first, based on the pioneering work of Sims (1980), is what I will call a “triangular identification,” which assumes that a shock does not affect a variable contemporaneously, and so one or more of the aij 0’s are zero. For example, it is commonly assumed that no variable other than policy itself responds to monetary shocks immediately, and so a 140 = a 240 = a 340 = 0. A more formal description of a triangular identification begins by writing the matrix A(0) that is composed of all the coefficients of the Aij (0)’s—that is, all the aij 0’s. Triangular identification means assuming that six of these aij 0’s are zero, and so (5) a 110 a A(0) = 210 a 310 a 410 0 a 220 a 320 a 420 0 0 a 330 a 430 0 0 . 0 a 440 In other words, triangular identification means that the monetary policy shock ut is identified by assuming that no variable other than the federal funds rate responds to it contemporaneously. The output shock, eyt , is identified by assuming that it is the portion of the error in the output equation that is orthogonal to the policy shock, while the commodity price shock, ect , is the portion of the error in the commodity price equation that is orthogonal to these. The final part of the residual in the aggregate price equation that is orthogonal to all three of these is the aggregate price shock, ept . There are many other ways to constrain the four-variable VAR and achieve identification. One, based on the work of Galí (1992), combines two types of restrictions. The first are contemporaneous and resemble those used in the triangular method. The second, following Blanchard and Quah (1989), assume that some shocks have temporary, but not permanent, effects on some variables. For example, we might claim that monetary shocks have no long-run effects on real output, and so the impact of u t on y t dies out. Formally, this involves assuming ¥ that the a34k ’s sum to zero: kS= 0 a 34k = 0. Recalling that we need six restrictions, the Galí-style procedure begins with two contemporaneous restrictions based on the logic of data availability and the time people in the economy take to act. The first constraint is that monetary policy does not affect real output contemporaneously (within the month). In the notation used above, the assumption is that a 340 = 0. This seems sensible, since production planning is unlikely to change suddenly after a policy innovation. The second constraint is that the aggregate price level does not enter the money supply rule. This also seems sensible, because the Bureau of Labor Statistics does not publicly release the Consumer Price Index (CPI) until the month following its price survey. The Galí-style, long-run restrictions, based on Blanchard and Quah (1989), amount to assumptions that neither monetary policy nor aggregate price (other aggregate demand) shocks permanently affect real output or the real commodity price level. Together, the two contemporaneous and four long-run restrictions allow us to estimate the impact of monetary policy shocks on prices and output. Structural Stability Variable inclusion and identification are related. The way in which we name various estimated shocks in a model obviously depends on the quantities being modeled in the first place. While connected to the other choices, the final, more general issue concerns the period over which the empirical model is estimated. The problem is that the reduced-form relationships in the data are unlikely to be stable over any reasonable sample.5 The problem, known widely as the Lucas (1976) critique, is that policy rule changes alter the relationship among endogenous variables in the economy. It is easy to see why this might happen. For the sake of discussion, assume that inflation is actually determined by the following structural model: (6) pt + 1 = art + b1X1t + b2 X2t + w t + 1, ■ 5 For a more detailed discussion, see section 4 of Cecchetti (1995). 6 where rt is policy, w t + 1 is a random variable, and X1t and X2t are measures of general economic conditions, like things that influence aggregate supply and money demand. Next, assume that we can write down a reaction function whereby policymakers automatically change their policy control variable when economic conditions change: (7) rt = g1X1t + g2 X2t + nt . The policymaker’s role is to choose g1 and g2, the reaction of rt to X1t and X2t . Since the g’s can be zero, a policy regime need not react to the X ’s. The term nt is a measure of the random component in the policy rule. Now, consider the reduced-form regression: (8) pt + 1 = f1X1t + f2 X2t + xt . Since fi = a gi + bi , changes in policy, which are changes in the g’s, will alter the correlation between the X’ s and p. In effect, the reducedform inflation regression subsumes the monetary-policy reaction function (7), so that a change in the monetary authorities’ policy rule—which may be a change in the relative weight placed on various indicators—will cause changes in (8). As a practical matter, there are several ways to deal with the instability that may be caused by changes in monetary policy rules. First, one can use institutional information to restrict the data to a period when there were no large changes in policy procedure. Second, one can try to estimate the timing of structural breaks.6 Alternatively, one can use time-varying parameter models, as Sims (1992) suggests. It is also possible to simply ignore the problem and use all of the available data. Following my earlier work, I use only the past decade’s data, beginning in 1984. Excepting the truncated sample period, I will ignore the problems created by the Lucas critique in all of the calculations that follow. This is an unfortunate necessity if any progress is to be made. Commerce index of industrial materials prices, and the federal funds rate, along with the triangular identification in equation (5), straightforward procedures yield estimates of the aijk ’s, as well as a covariance matrix for these estimates. These are the time path of the impact of innovations on the model’s endogenous variables. They tell us how any one of the four shocks will affect any of the four variables initially—and after several months. It is easiest to present these results in a series of figures. Figure 1 shows estimates of 16 impulse response functions, plotted with two standard-error bands.7 These are the response of output, aggregate prices, commodity prices, and the federal funds rate to a unit innovation to each of the four shocks. The impulse response functions are straightforward and easy to understand. Taking the policy innovation as an example, the last column of figure 1 shows the result of an unanticipated 100-basis-point change in the federal funds rate for one month on yt , pt , p ct , and rt over the next three years. For example, the fourth plot in the third row shows the impact of monetary policy shocks (ut ) on the aggregate price level (pt ). The estimates suggest that a one-time policy tightening—an increase in the federal funds rate—causes prices to rise slightly initially, then to fall below their original level after about six months. Over the next 30 months, the price level continues to fall. The standard-error bands on this figure imply that we are actually very unsure of the response. The data indicate a strong possibility that the policy tightening will result in a price-level increase. Several additional features of figure 1 are worth noting. First, in all cases, commodity prices (second row) respond more quickly and in the same direction as aggregate prices (third row). Second, for the three e shocks, the output response seems to be more precisely estimated ■ 6 This is the technique used in Cecchetti (1995). II. Results from Estimating the Model Impulse Response Functions Using monthly industrial production data for January 1984–November 1995, the CPI for urban wage earners (CPI-U), the Journal of ■ 7 The standard-error bands in the figure are constructed using the simple Taylor-series approximation: ^ ~ ^ – b), ~ F (b) + dF (b) F ( b) ^ (b db b= b ^ follows where F is any differentiable function. The variance of F ( b) immediately as ^ – F (b) 2 ~ ^ E [F ( b) ] ~ dFd b(b) b = b^ 2 Var ( b). ^ Here, we can think of the estimated impulse response functions, the Aij’ s, as functions of the estimated reduced-form VAR coefficients, the elements ^ of R(L). Given the estimated variance of these coefficient estimates, the ^ variance of the Aij ’s can be computed by numerical differentiation. | [ | ] 7 F I G U R E 1 Impulse Response Functions: Triangular Identificationa a. Estimated response, with two standard-error bands. NOTE: Horizontal axes are in months; vertical axes are in the change in the log per month. SOURCE: Author’s calculations. 8 than the aggregate price response. This second conclusion is consistent with Cochrane’s (1994) observation that real output is forecastable with high R 2 at horizons of several years, and with my finding (see Cecchetti [1995]) that inflation is difficult to forecast at horizons longer than a single quarter. It is very tempting to seek a correspondence between the shocks in this four-variable VAR and those discussed in macroeconomics textbooks. In a simple model, the basic result is that aggregate supply shocks move prices and output in opposite directions, while aggregate demand shocks move them in the same direction. With this categorization, the impulse responses shown in figure 1 suggest that all of the shocks in this model come from the demand side. While this makes intuitive sense for the monetary policy shock, it renders the other classifications unsatisfactory. One can either accept this at face value or ask whether it might result from the identification used to generate the estimates. Taking the second possibility seriously leads to examination of an alternative identification—the one proposed by Galí being a natural choice. Figure 2 plots the impulse response functions from such a model, estimated using exactly the same data. Because of the technical difficulty associated with their construction, I do not include standard-error bands. Here, the results differ markedly. It now appears that the output shock, eyt , behaves like an aggregate supply shock, while the three remaining shocks, representing the aggregate price-level shock, the raw material price shock, and the monetary policy shock, lead to reactions consistent with those expected from aggregate demand shocks. However, we can draw an important positive conclusion by comparing these two sets of identifying restrictions: The impulse response functions of the monetary policy shock are robust to changes in the identification procedure. Policy’s impact on output and prices seems fairly robust to the exact methods used in estimation. Since they are easier to compute, I will now proceed using only the estimates obtained with the simpler triangular identification. Historical Decompositions While the ultimate goal is to use the estimated dynamic model to construct policy rules, the impulse response functions and structural innovations also allow us to compute the quantities known as “historical forecast decompositions.” These allocate output and price movements into the portions accounted for by each of the structural shocks. It is easy to understand how these estimates are constructed from the structural model’s equations (1)–(4): Define the impact of the monetary policy shock on output as Hyu (t). From equation (3), this is just ¥ (9) Hyu (t) = Sa 34i ut – i , i and analogously for the other shocks. Its estimated value, constructed from the parameter estimates, is (10) ^ ¥ Hyu (t) = S a^ 34i u^ t – i , i ^ where ut is the estimated monetary policy innovation. Figure 3 plots the decomposition of the movements in real output and aggregate prices into the components attributable to monetary and nonmonetary shocks. In constructing these, I have truncated the sum in (9) at 60 months. Because of the difficulty in identifying innovations from nonmonetary sources, it seems prudent to simply sum them together. That is, I plot the fluctuations in yt and pt attributable to ut , ^ ^ Hyu (t), and Hpu (t), and the portion not attribut^ ^ able to policy, [yt – Hyu (t)] and [pt – Hpu (t)]. The results show that, for the past seven years, important movements in both output and prices are largely accounted for by innovations other than those coming from mone^ tary policy. The blue line representing Hyu (t) ^ and Hpu (t) in the figure’s two panels has much less variation than the green line representing the fluctuations in yt and pt that are attributable to nonmonetary policy shocks. This result is particularly striking for prices, where variation seems to be driven by innovations to output, raw materials prices, and the aggregate price level itself. Aggregate supply shocks and nonmonetary aggregate demand shocks account for most of the recent movements in key macroeconomic variables. III. Formulating a Policy Rule Issues The main use of the empirical model described in section I and estimated in section II is to provide quantitative answers to the questions required for implementing a policy rule. To see 9 F I G U R E 2 Impulse Response Functions: Galí Identificationa a. Estimated response. NOTE: Horizontal axes are in months; vertical axes are in the change in the log per month. SOURCE: Author’s calculations. 10 F I G U R E 3 Forecast Error Attributable to Various Innovations functions. Mankiw (1994) includes several papers that deal with this topic explicitly. There are two primary candidates: price-level targets and nominal-income targets. One version of these involves setting the policy instrument— the ut ’s in the model—to minimize the average expected mean square error (MSE) of either inflation or nominal-income growth over some future horizon. In the inflation case, the objective function can be written as (13) 1 min h {u i } where po is the log of the base-period price level and h is the policymaker’s horizon. The expectation in (13) is over the sampling distrib^ which is related to the covariance ution of p, matrix of the estimated coefficients in equation (11). Nominal-income targeting simply replaces the log price level in (13) with the sum of pt and yt . One important distinction between the objective function (13) and more standard formulations is the treatment of parameter uncertainty. As the results in figure 1 clearly show, we are very unsure about the size and timing of price movements following innovations to the federal funds rate. When constructing a policy rule, it seems prudent to account for this lack of knowledge. As Brainard (1967) originally pointed out, the presence of uncertainty has important implications. This is easily demonstrated in the present context. Consider a simplified version of the structural price and interest rate equations SOURCE: Author’s calculations. how this is done, first note that the model implies estimated values for the aggregate price level and real output: (11) h E ( p^i – po )2, S i=1 p^ t = A^11 (L) ^ept + A^12 (L) ^ect + A^13 (L) ^eyt + A^14 (L)ut , (14) pt = ept + gut (15) rt = ut , where g is a parameter. Next, take the horizon in (13) to be one period (h = 1), and the initial log price level to be zero, po = 0. The policy control problem then reduces to (16) ^ min E [p i2 ]. {u i } (12) y^t = A^31 (L) ^ept + A^32 (L) ^ect + A^33 (L) ^eyt + A^34 (L)ut . A policy rule is a sequence of ut ’s that is constructed to meet some objective. In other words, the policymaker is allowed to pick the path of the federal funds rate to meet a particular objective.8 The monetary policy literature includes many discussions of the efficacy of various objective Substituting in the expression for pt , this is simply (17) ^ min E [epi + g ui ]2. {u i } ■ 8 Feldstein and Stock (1994) examine an identical experiment, but without parameter uncertainty. 11 If we ignore that g is estimated, then it is trivial to generate the policy rule. It is just (18) ui* = – 1^ epi . g Taking account of uncertainty in the estimate of g, but continuing to assume that ept is known, the minimization problem yields (19) ui* = – ^ g ^2 ^ epi . [g + Var (g )] For a given ept , this leads to an unambiguously smaller response. In other words, imprecision creates caution, with policy reactions being muted in the face of uncertainty. Reactions are further attenuated if policymakers attach a cost to the movement in instrument. Taking the same simple setup, imagine the modified objective function ^ (20) min E [p i2 + a r^ 2t ]. {u i } This produces the reaction function (21) u i* = – ^ g e , ^2 ^ [g + Var (g ) + a ] pi which will yield an even smoother path for the interest rate than does (19). Results I examine results based on several policy objectives. It is worth noting that the exercise described here appears to be a gross violation of the Lucas critique. That is to say, contrary to the implications of the discussion in section I, I assume that the reduced-form correlations among output, prices, and interest rates described by equations (11) and (12) are unaffected by the change in the policymaker’s reaction function. There are two ways to defend the procedure. The first is to take the view of Sims (1982)— that parameters in these models evolve slowly enough to make Lucas-critique considerations quantitatively unimportant. The second defense is to reinterpret the exercise as an attempt to recover the objective function that policymakers were implicitly using, by trying to match the actual federal funds rate path with that implied by an optimal rule. I report results for three different policy rules. The first, which might be termed passive, holds the federal funds rate fixed in the face of the shock. (The model makes it clear that this is not really a passive policy, since it involves shocks to overcome the estimated reaction func- tion.) The other two, which I will call active, minimize the average MSE of either the log of the price level or the log of nominal output over a 36-month horizon (h = 36).9 For each rule, I examine three experiments—one for each structural shock. In each of the nine resulting cases, ejo = 1 and elk = 0 for l =/ j and k =/ 0. In other words, there is a unit innovation to one of the structural disturbances in the base period, and that is all. I then construct individual estimates for the optimal response of interest rates to each of the shocks. Figure 4 reports the implied path of the federal funds rate, aggregate prices, and industrial production for each policy objective in response to each of the three structural shocks. The fixed federal funds rate policy results in consistently higher output and prices than does either of the other two polices. The activist policies both have the same profile, whatever the source of the shock. Output and prices both rise initially, and then fall, with output dropping more than prices. Interestingly, both of the activist policies involve raising the funds rate immediately and then lowering it slowly. This follows directly from the fact that prices respond slowly to policy innovations (see the third row of figure 1). The implication is that a policymaker who wishes to stabilize prices must respond to exogenous shocks quickly, in order to ensure that future price movements are minimized. That is the argument for the Federal Reserve’s tightening up at the first sign of upward price pressure. Comparing Targeting Objectives These calculations have direct implications for the debate between advocates of price-level targeting and those who favor targeting nominal GDP. To see why, I have computed the implied root-mean-square error (RMSE) for inflation and nominal income for each policy. For the pricetargeting case, these are the square root of the minimized objective function (13). Table 1 shows the results. The computations suggest that nominal-income targeting has a certain robustness, since inclusion of real output in the objective function increases the RMSE for inflation only slightly. For the case of an output shock, the increase is from 0.24 to 0.61. However, when the output shock is the ■ 9 Because the model is estimated in logs, the minimum MSE of the nominal-income policy minimized the MSE of the sum of the log of industrial production and the log of the CPI. 12 F I G U R E 4 Interest Rate, Output, and Price Paths following Shocks, and the Policy Response ε ε ε ε ε ε ε ε ε . Min MSE (p) policy Fixed interest rate policy Min MSE (p + y ) NOTE: Horizontal axes are in months; vertical axes are in the change in the log per month. SOURCE: Author’s calculations. 13 T A B L E source of the instability, the move from pricelevel targeting to nominal-income targeting decreases the RMSE of nominal income substantially—from 4.12 to 0.69. In other words, the inability to estimate precisely either the impact of shocks on prices or prices’ response to policy innovations argues strongly for including real variables in the objective function. 1 Comparison of Policy Responses A v e r a g e R M S E of Inflation over a 36-Month Horizon Source of Shock Aggregate Commodity Price Price Policy Rule Fixed interest rate . Min MSE (p + y ) Min MSE (p) 2.35 2.15 0.99 1.98 1.50 0.51 Output 1.14 0.61 0.24 Comparing Actual and Implied Interest Rate Paths A v e r a g e R M S E of Nominal Income over a 36-Month Horizon Finally, one might ask how closely recent pol- Source of Shock Policy Rule Aggregate Price Commodity Price Output 1.86 0.32 0.99 4.89 0.35 10.85 6.19 0.69 4.12 Fixed interest rate . Min MSE (p + y ) Min MSE (p) SOURCE: Author’s calculations. F I G U R E 5 Comparison of Optimal and Actual Federal Funds Rate Pathsa 0 200 400 600 800 1000 1200 1400 1600 1800 π π a. Monthly data, June 1987 to November 1995. SOURCES: Author’s calculations; and Board of Governors of the Federal Reserve System. 2000 icy conforms to what would have been implied by either the price-level or nominal-income targeting rules plotted in figure 5. A simulated interest-rate path can be calculated by taking the estimated structural innovations, the ^ejt ’s, and then computing the optimal policy responses implied by each rule before substituting the result into the equation for the federal funds rate, which is the equivalent of (11).10 Figure 5 compares the actual path of the federal funds rate with that implied by the estimated price-level and nominal-income targeting policies. When we examine the figure, several findings emerge. First, targeting the price level alone yields larger swings, as the funds rate reaches both higher and lower extremes. The actual funds rate is the least variable, looking like a smoothed version of the two simulated paths, but the general character of the plot suggests that the optimal policy response simply involves faster, bigger movements than those on the actual path.11 Figure 5, however, allows an even more interesting conclusion. From its results, it is possible to infer something about the procedures policymakers were actually following. Such a calculation does not violate the Lucas critique, since it is an attempt to recover the loss function implicit in the policy actions we actually observed. The estimates imply that the actual fundsrate path was very similar to one that would ■ 10 Performing the calculations in this way ignores a number of elements. In particular, there is no guarantee that the policy rules generated from the artificial experiment of one unit shock in one ejk at a time will be robust to sequences of shocks in all the ejk ’s simultaneously. One clear reason for this is that it ignores the covariance of estimated coefficients both within and across the elements of the A^ij (L)’s. ■ 11 As one would expect, these large policy innovations result in less stable real output, highlighting that the ultimate issue in policymaking is still the relative weight of prices and output in the objective function. 14 have been implied by a nominal-income targeting procedure, only smoother. It is as if, over the past decade or so, the federal funds rate had been set to conform to a nominal-income targeting regime, but with policymakers attaching a cost to actually moving the funds rate. That is, the objective function that we can construct from the actual path of interest rates would minimize the sum of squared deviations in nominal income from a target path and squared movements in the federal funds rate, over a horizon of about three years. Appendix: Identification To understand the more general issues of identification, it is useful to rewrite the four-equation model [(1)–(4)] in a more compact form: (A1) where x t and e t are now vectors, and A (L ) is a matrix of lag polynomials. We can also write the model in its more familiar VAR reduced form as (A2) IV. Summary The information requirements for any policy rule are daunting. Not only do policymakers need timely information about current economic conditions, they also need forecasts of the future path of the variables they wish to control (aggregate prices and real output) and quantitative estimates of how their actions will affect these objectives. This paper’s purpose is to suggest that much of our knowledge is very inexact, and that our inability to precisely forecast the results of policy changes should make us cautious. Even more important, the fact that we have a much better understanding of the impact of our policies on real output than on prices suggests that nominal-income targeting rules are more robust than price targeting rules. From a purely pragmatic viewpoint, someone who cares about nominal income is made substantially worse off by moving to a price-level target, which destabilizes real output considerably. Thus, practical issues make a strong argument for nominalincome targeting. In addition, we have seen that the actual path of interest rates over the past decade is very similar to that implied by a nominalincome targeting rule, albeit one in which interest rate movements are viewed as costly. By comparing the actual interest-rate path with the path implied by the nominal-income targeting rule, we see that policymakers have smoothed interest rate movements more than the rule would have dictated, but not by much. x t = A (L )e t , R (L )x t = ht , where R (0) = I, the ht ’s are i.i.d. (implying that they are orthogonal to the lagged x t’ s), and E (hh ¢) = S. It immediately follows that A(L)et = R(L)–1ht . This allows us to write A(0)et = ht , and A(L) = R(L)–1A(0). As a result, given estimates of A(0), R(L), and h, we can recover estimates of both the structural innovations— the et ’s—and the structural parameters—the components of A(L). The issue of identification is the problem of estimating A(0). To show how this is done, note that A(0)E (ee ¢)A(0)¢= S, where E (ee ¢) is diagonal by construction. Normalizing E (ee ¢) = I, we obtain the result that A(0)A(0)¢= S. In a system with n variables, S has [n (n + 1) ] unique 2 elements, and so complete identification requires an additional [n (n – 1) ] restrictions. In a 2 four-variable model, six more restrictions are needed. This is a necessary but not sufficient condition for identification. Sufficiency can be established by proving that the restrictions lead to construction of an A(0) matrix that is invertible. The long-run restrictions of the Galí-style identification can be understood by defining A(1) as the matrix of long-run effects computed by summing the coefficients in A(L). That is, the (i,j ) element of A(1) is ¥ Aij (1) = S aijk . k=0 There are two long-run restrictions. The first is that the impact of ept and ut on yt is transitory, and so A31(1) = A34(1) = 0. The second is that ept and ut have no permanent impact on the relative price of commodities, ( pct – pt ), that is, A11(1) – A21(1) = A14(1) – A24(1) = 0. 15 References Bernanke, B. S., and I. Mihov. “Measuring Monetary Policy,” Federal Reserve Bank of San Francisco, Working Paper No. 95-09, March 1995. Blanchard, O. J., and D. Quah. “The Dynamic Effects of Aggregate Demand and Supply Disturbances,” American Economic Review, vol. 79, no. 4 (September 1989), pp. 655–73. Lucas, R. E., J r . “Econometric Policy Evaluation: A Critique,” Journal of Monetary Economics, vol. 1, no. 2 (Supplementary Series 1976), pp. 19–46. Mankiw, N. G., ed. Monetary Policy. Chicago: University of Chicago Press, 1994. S i m s , C. A. “Macroeconomics and Reality,” Econometrica, vol. 48, no. 1 (January 1980), pp. 1–48. Brainard, W. C. “Uncertainty and the Effectiveness of Policy,” American Economic Review, vol. 57, no. 2 (May 1967), pp. 411–25. . “Policy Analysis with Econometric Models,” Brookings Papers on Economic Activity, no. 1 (1982), pp. 107–52. Cecchetti, S. G. “Inflation Indicators and Inflation Policy,” in B.S. Bernanke and J.J. Rotemberg, eds., NBER Macroeconomics Annual 1995. Cambridge, Mass.: MIT Press, 1995, pp. 189–219. . “A Nine Variable Probabilistic Macroeconomic Forecasting Model,” Cowles Foundation for Research in Economics at Yale University, Discussion Paper No. 1034, June 1992. Christiano, L. J., M. Eichenbaum, and C . L . E v a n s . “The Effects of Monetary Policy Shocks: Evidence from the Flow of Funds,” Review of Economics and Statistics, vol. 78, no. 1 (February 1996a), pp. 16–34. , , and . “Identification and the Effects of Monetary Policy Shocks,” in M.I. Blejer et al., eds., Financial Factors in Economic Stabilization and Growth. New York: Cambridge University Press, 1996b, pp. 36–74. Cochrane, J. H. “Shocks,” Carnegie–Rochester Conference Series on Public Policy, vol. 41 (December 1994), pp. 295–364. Feldstein, M., and J. H. Stock. “The Use of a Monetary Aggregate to Target Nominal GDP,” in N.G. Mankiw, ed., Monetary Policy. Chicago: University of Chicago Press, 1994, pp. 7–70. G a l í, J . “How Well Does the IS-LM Model Fit Postwar U.S. Data?” Quarterly Journal of Economics, vol. 107, no. 2 (May 1992), pp. 709–38. 16 The Reduced Form as an Empirical Tool: A Cautionary Tale from the Financial Veil by Ben Craig and Christopher A. Richardson Introduction Economic data usually influence policy through a reduced-form analysis. Using such an analysis, the researcher generally poses an empirical relationship between an outcome variable, such as a firm’s total investment, and a policy variable, such as the design of a particular tax. This relationship serves as a point of departure in the analysis. Explicit assumptions about behavior that underlie the relationship are not emphasized; rather, the researcher asserts that the “data do the talking.” Policy implications, where they exist, are directly observed in the pattern estimated in the data. Most empirical analyses of policy questions follow a reducedform strategy.1 It is easy to understand why a reduced-form approach might, at first glance, appear to be the best way to analyze policy. It is a simple methodology, and thus can more easily keep track of what is happening during the complicated process of analyzing data. One does not need to specify a sophisticated and consistent model of behavior to use this approach. Further, the answers embodied in the model estimates may accord with a wide variety of behaviors that could be true of the firm. Ben Craig is an economist at the Federal Reserve Bank of Cleveland, and Christopher A. Richardson is a graduate student of economics at Indiana University. The authors have benefited from the generous help and advice of Charles Carlstrom. A different approach to estimating the effect of taxes would be to specify a model of optimizing behavior on the part of the firm and to model the tax policy as a set of constraints on this optimizing behavior. A simplistic reason for preferring the reduced-form approach is that economists are interested only in the overall effect of a proposed tax policy on investment. Why should we care about the intermediate steps by which a tax will affect the firm? How successful is the reduced-form approach at testing a behavior or measuring a policy effect? Given that we are never shown the truth behind the mystery, this paper will examine the history of an economic question that has been subjected to 35 years of intense scrutiny: Does corporate financial structure affect real investment? The empirical answer to this question, which lies at the heart of corporate financial economics, has heavily influenced every tax reform bill since the 1960s. ■ 1 Reduced form has a different meaning here than in simultaneous equations estimation, where a reduced form is estimated by regressing an endogenous variable on all of the exogenous variables in a system of equations. We use the term in a wider context, where the pattern in the data—not an assumed behavioral structure—forms the point of departure for estimation. 17 The initial econometric strategy was to follow a pure reduced-form approach. How well have the results of this research program held up to further scrutiny? Modigliani and Miller (1958) provide the first theoretical model showing the influence of corporate debt structure on investment. In the world they portray, perfect capital markets, coupled with symmetric information about the investment prospects of the firm, the investors, and the lenders, mean that the firm’s debt level is irrelevant to the amount of investment it undertakes. One reduced-form approach would be to directly examine the empirical relevance of the Modigliani–Miller (hereafter “MM”) hypothesis that with perfect capital markets, no taxes, and a given investment policy, capital structure is irrelevant to firm value. As a consequence, neither capital structure nor dividends should affect investment behavior. The MM propositions provide the following broad empirical prediction: In a properly specified regression of investment on the debt/equity ratio, dividends, and other covariates, the coefficients of debt/ equity and dividends should equal zero. A reduced-form estimating strategy uses this prediction as the point of departure. The following two sections discuss the history of tests of this hypothesis from both a cross-section and a time-series perspective. Next, we step back and explore the reasons for the pattern in the early reduced-form estimates through a simple structural model. We then look at what we have learned about whether a tax policy can affect investment through its influence on a company’s financial structure. We conclude with the object lessons that accompany 35 years of intensive research on this topic—lessons that could be applied to other situations where the reduced form is used to help shape policy. What did we first believe the data were telling us, and how did these beliefs change under close scrutiny? After all this time spent researching a single hypothesis, what limitations in our knowledge may be embedded in the reduced-form approach? I. Cross-Section Regression Tests of the MM Hypothesis The clear and simple MM hypothesis that there is no relationship between financing and real capital investment seems to lend itself easily to cross-section regressions. The early reducedform models assume away the importance of differential corporate and personal income taxes, which are a clear violation of the original MM statement. Thus, they jointly test the MM model and the hypothesis that the income tax structure is irrelevant to the effect of financial structure on real investment. We will treat the two tests separately later in this paper. The test of the joint hypothesis measures the statistical significance of financial variables in an investment equation where the dependent variable is capital investment and the independent variables are measures of a firm’s financial position, which may include its debt/equity ratio, cash flow, and dividends. The hypothesis of no relationship between financing and investment is rejected if the coefficients of the debt/equity ratio and dividends are statistically close to zero. A simple regression is not adequate here because both dividend payments and the firm’s debt are endogenous. Thus, absence of a correlation between the debt/equity ratio and investment is not necessarily evidence that the MM hypothesis holds. To alleviate this problem, early cross-section studies specified instruments in a system that estimated investment (I ), dividends (D), and new debt (ND) equations of the general form Iit = a0 + a1Dit + a2NDit + a3Xit + eI Dit = b0 + b1Iit + b2NDit + b3Yit + eD NDit = g0 + g1Iit + g2Dit + g3Zit + eND , where i and t are firm and time subscripts, the e’s are statistical error terms, and X, Y, and Z are vectors of exogenous explanatory variables. For the investment equation to be identified (so that we are estimating a separate equation for investment, not a hodgepodge of all three equations), the vectors Y and Z must contain variables that are not included in X. It is this process of identification that proved so problematic in the early reduced-form studies. What exogenous variable affects dividends and debt levels but does not influence investment behavior? It is important to note here that the MM hypothesis is not a theory of investment, but of why a firm’s financial structure does not influence investment. The estimating system of equations that test the MM hypothesis must include a theory of investment (even if it is implicit) to control for its endogeneity. Thus, the reduced form is a joint test of both the MM hypothesis and an underlying theory of investment. For example, if the researcher holds investment opportunities constant through using a measure of Tobin’s q, then the test of the MM 18 hypothesis also tests whether Tobin’s q is an empirically useful model of investment behavior.2 Hence, the test is only as good as the theory of investment. The early studies used identifying instruments 3 that included profits, proxies for firm size and taxes, and firm and industry dummy variables to allow for fixed firm and industry effects.4 (See, for example, Dhrymes and Kurz [1967], McDonald, Jacquillat, and Nussenbaum [1975], and McCabe [1979].) These studies, like so many modern consulting reports, argue by assertion—for example, the profit level should affect dividends but not investment. Unfortunately, a researcher’s assertion that a variable is an instrument does not necessarily make it so. A reduced form offers few checks as to whether the assertion reflects reality. Empirical tests of the MM irrelevance hypothesis during the 1970s and early 1980s, though more advanced econometrically, still came up short in modeling differences in the financial environment firms face. Although the studies varied in their conclusions, all suffered from the lack of a convincing instrument that would control for a firm’s investment opportunities. Again, the studies could not adequately address the fact that firms with better investment opportunities might choose higher levels of debt. McDonald, Jacquillat, and Nussenbaum estimate cross-section models using ordinary least squares (OLS) and two-stage least squares (2SLS), as does McCabe, but their conclusions differ. McDonald et al. find support for the MM propositions, while McCabe does not. Because investment opportunities surely vary across firms and certainly affect investment independently of financial structure, these early studies were never conclusive tests of the MM irrelevance hypothesis. Another reason for conflicting conclusions among the early empirical studies seems to lie in the differences in equation specification. McDonald et al., like many other researchers before McCabe, estimate investment as a function of contemporaneous variables only. Because it is likely that the decision to invest today will depend in part on financial decisions made previously, excluding lagged financing and dividend variables from an investment equation results in a misspecification. How was one to choose between these early studies? If they had been structural, a researcher could affirm that a particular study was the most convincing if it had more believable parameter estimates (for example, if it generated rate-ofreturn estimates of the same general magnitude as the interest rate). A classic indication that a system is identified improperly (that is, by false assumptions) is that estimates of the individual equations yield parameters that make little sense economically. One reason the earliest tests of the MM hypothesis seemed, on balance, to support the theory was that the estimates which rejected MM had the “wrong” expected sign for the dividend equation. This seemed to indicate that the studies which did not reject the hypothesis used more convincing instruments. One pitfall of a simple reduced-form strategy is that it yields so few checks of whether an estimated parameter makes economic sense. Subsequent cross-section studies made significant improvements over previous work. For example, Peterson and Benesh (1983) estimate a system of three equations similar to that used in earlier studies (adding a lagged profit variable to the investment equation and a lagged dividend variable to the dividend equation), but in addition to estimating the standard OLS, 2SLS, and 3SLS models, they also conduct MM hypothesis tests on the reduced-form equations by estimating a seemingly unrelated regressions (SUR) model. Their SUR results corroborate the 2SLS and 3SLS findings, which reject the null hypothesis that financing and investment decisions are independent. The lagged profit variable serves as a proxy (albeit an imperfect one) for investment opportunities, which makes the rejection of the MM irrelevance hypothesis somewhat more convincing. The use of lagged profits suffers from a problem common to all studies that rely on lagged variables for identifying instruments. Although it is true that lagged profits are approximate measures of investment opportunities, they may also affect both dividends and debt in the same ways that these variables were correlated with the original contemporaneous error term. It is not clear that simply including the lagged profit term will correct the original statistical bias. Most recent reduced-form cross-section models reject the MM hypothesis. (See, for example, Gilchrist and Himmelberg [1995].) However, ■ 2 Tobin’s q is defined as the ratio of the market value of capital to the replacement cost of capital. ■ 3 An instrument is a variable that is correlated with a variable on the right-hand side of the equation (in this case, corporate debt or dividends) without being correlated with the statistical error term. An identifying instrument in this case is one that is correlated with the right-hand variables without being included as a right-hand variable. Thus, it may be included in the equation where dividends or debt are left-side variables, but it must be excluded from the original investment equation. ■ 4 In some cases, X, Y, and Z contain the same variables. 19 these models often suffer a distressing lack of robustness to econometric specification. This makes precise determination of the estimated reduced-form parameters problematic. Further, the reduced-form approach does not present us with an easy point of departure for determining the correct econometric specification through convincing tests. We also lack a consensus on parameter estimates that are specific and precise enough to be more useful in a policy context than are cross-section reduced-form models. Can we glean additional evidence on the empirical validity of the MM irrelevance hypothesis from time-series patterns in the data? II. Granger Causality Tests Given the difficulty of pinning down a convincing set of instruments to tease out that part of the correlation of debt and investment stemming from debt’s possible impact on investment, some researchers have tried to determine the causality by studying the timing of debt and investment. Thus, if investment precedes debt, the correlation may be spurious because the firm, seizing its more potent investment opportunities, creates more debt, whereas the less fortunate firms do not have as much debt. This would be the case when the higher debt level was due to more investment opportunities for the high-debt firm. The test of a causal relationship between the variables proposed by Granger (1969) says that if a variable or event X (a change in a financial variable) causes another variable or event Y (a change in investment), then X should precede Y. The test involves measuring the power of lagged values of X in predicting Y. A test of whether debt affects investment is connected to whether corporate debt “Granger causes” investment. Smirlock and Marshall (1983) perform Granger causality tests on a sample of 194 firms from 1958 to 1977. Using annual data on dividends and investment, they fail to reject the null hypothesis of no Granger causality for the aggregate sample of firms. Causality tests on each of the 194 firms’ series do not reject the null any more often than would be expected by chance, leading the authors to conclude that their results support the MM irrelevance hypothesis. True to the pattern of cross-section tests of the MM hypothesis, the early test did not hold up to later scrutiny. It is imperative that enough variables be included in a Granger causality study so that nearly identical firms are being compared. For example, Smirlock and Marshall omit a financing variable, so that the analysis compares noncomparable firms that differ in precisely that dimension which the causality test requires to be the same. Mougoue and Mukherjee (1994) address this issue by including a long-term debt-financing variable in their causality tests. They find that dividend and investment growth rates Granger cause each other negatively, long-term debt and investment growth rates Granger cause each other positively, and debt and dividends Granger cause each other positively, thus rejecting the MM irrelevance principle.5 If the reduced-form test is to be appropriate, some sort of implicit structure must underlie it. In this case, the researchers had to have an idea about which financing variables were important so that the Granger causality could test comparable firms. Although Mougoue and Mukherjee’s Granger causality tests can detect significant interactions among investment, debt, and dividend variables, they do not tell us much else. That dividends and investment Granger cause each other simply means that a motion in one precedes a motion in the other. Which comes first, the investment chicken or the dividend egg? Moreover, it is somewhat ironic that Mougoue and Mukherjee’s causality tests may also suffer from a misspecification bias due to the omission of a proxy for investment opportunities, such as cash flow. If internal funds are omitted from the system of equations, the observed negative causality from dividends to investment may actually stem from a negative causality from dividends to retained earnings and a negative causality from retained earnings to investment. The MM irrelevance hypothesis would still be rejected, but for different reasons. Although more properly specified equation systems may be useful in illustrating the existence of causal relationships, it appears that Granger causality tests have only limited utility in distinguishing among the different hypotheses of how, why, and to what degree financing and real investment decisions interact. In addition, Granger causality tests suffer from a difficulty related to the forming of expectations. If debt Granger causes investment, the interpretation is that the corporate structure effects a change in investment behavior. However, expectations about investment ■ 5 It is assumed here that firms use borrowed funds to finance future investment or to increase dividend payments. Because the variables are expressed as changes in logarithms (growth rates), positive bidirectional causality between debt and investment does not support or refute the presence of financing constraints, as it might if debt and investment were expressed in levels. 20 opportunities could just as well influence both investment and corporate debt levels, but affect debt sooner because debt levels adjust more quickly. Tests that center on the timing of debt and investment thus provide weak evidence on the relevance of the MM hypothesis. Interestingly, the pattern of evidence in the Granger causality tests is the same as the pattern in the cross-section regression results. Initially, the evidence seemed to support the MM hypothesis. However, closer scrutiny and clearer identifying assumptions tend to reject the hypothesis. Even current studies are unable to provide more information than a crude rejection of the hypothesis. Precise parameter estimates needed for policy prediction seem to require a different estimation strategy. Why do we observe this pattern in the reduced-form estimates? It is not clear which part of the joint hypothesis—perfect capital markets or the empirical irrelevance of the personal and corporate income tax structure—is being rejected by the above tests. To further define the two hypotheses, a simple heuristic model is needed. III. A MM Structural Model In this section, we explore a model in which the underlying behavior of firms generates the data. A simple statement of the model will clarify the measurement problems inherent in testing the MM hypothesis with cross-section or time-series data. A MM firm chooses the levels of investment, I 0, in a project that will pay F (I 0) dollars for each period in the future, F ¢(I 0 ) > 0, F ²(I 0 ) < 0, so that the firm receives (1) ¥ S F (I 0) t = 1 (1 + r)t = F (I 0) r from the investment, where r represents the interest rate.6 The firm starts with a predetermined amount of cash, C, and must decide how much of this cash to pay out in dividends, Cd , and how much to reinvest, CI (C = Cd + CI ). The firm can also issue debt, D, to finance the investment. A tax rate is imposed on a corporation at rate tc and on individuals at rate tp. MM’s first observation is that the market value of the shares, S, is just a tax-adjusted value of the investment payoffs (including the corporate cash paid out today, Cd ) minus the value of debt, or (2) S = (1 – tp )Cd + (1 – tc ) [ F r(I ) – D]. 0 The firm maximizes S with respect to the amount of investment subject to the financing constraints I 0 = CI + D and C = Cd + CI. A simple substitution gives (3) S = (1 – tp ) (C – CI ) + (1 – tc ) [F (C r+ D) – D], I with first-order conditions (4) F ¢ = r and (5) F ¢ = r (1 – tp ) , (1 – tc ) corresponding to investment financed out of debt and cash, respectively. If the personal tax rate is equal to the corporate tax rate (or tp = tc , which nests the special case of MM’s no-tax scenario), the first-order conditions make it apparent that the expense of an additional dollar of investment is the same however it is financed, and that the firm finances from either debt or retained earnings until the marginal benefit from investment is equal to the outside rate of return (or F ¢[I 0 ] = r). It is important to note that the first-order condition in this case simultaneously says two things about the firm’s behavior. First, a firm’s decision to finance a given level of investment out of debt or retained earnings is irrelevant: The firm is indifferent between the two. Second, the investment level is determined by the rule that the firm invests until the marginal benefit of investment is equal to the interest rate.7 The level of debt says nothing about the value of the firm except that it has traded debt for dividends at a rate of one for one. Investors who are paying for a share of the company and who might prefer a higher level of debt in their portfolio can continue holding shares of this firm, but elect to borrow more on the outside market to increase the debt level within their own portfolios. This and other similar possible arbitrages force the share value to treat debt and retained earnings symmetrically. In a world where the corporate tax rate is higher than the personal tax rate, tc > tp , the firm’s rule is to finance until the return on ■ 6 The MM results do not depend on risk neutrality, a constant stream of benefits, or a constant discount rate. These are assumed here for expositional convenience. ■ 7 In the MM exposition, the firm invests until the marginal benefit equals the rate of return for the firm’s risk class. The theory also shows that investment financed out of new equity issues is equivalent to investment financed out of retained earnings or debt. 21 investment is equal to r. The investment is financed entirely out of debt and all of the cash appears as a dividend. Financing out of debt rather than retained earnings costs less because debt payments are fully deductible. MM (1963) makes the point that the tax advantages of debt financing (interest payments are deductible as a cost in calculating corporate taxes) imply that investment should never be financed out of retained earnings in a pure example of their model. Indeed, because of the tax advantages, the firm prefers to borrow more than its investment amount to finance a larger dividend. Clearly, a more complicated model is needed to explain why a corporation chooses one method of financing over another. However, the simpler model may still be adequate to explain the level of investment if the data show no relationship between that level and corporate financial structure. If the corporate tax rate is less than the personal tax rate, then cash is the relatively less expensive investment source. The firm will use only cash to finance investment unless the marginal return on investment after all of the cash is used (and the dividend is zero) exceeds the cost of financing additional investment out of debt (or F ¢[C ] > r). At this point, debt becomes the marginal investment source, with the first-order condition given by equation (4). If F ¢(C ) < r, then debt will equal zero, and only cash will be used to finance investment, until equation (5) is satisfied. The top federal individual tax rate and the top federal corporate tax rate are currently about the same (39.6 percent versus 38 percent). However, in many states, the top corporate rate is higher than the top individual rate (in Ohio, the respective figures are 8.1 percent and 4.1 percent). Based on the above analysis, one would expect to find corporations financing investment entirely out of debt and never using retained earnings for this purpose. Clearly, this is not the case, as firms use both debt and equity financing. One reason companies do not rely solely on debt is that outside credit constraints (or the costs of bankruptcy) may make the marginal cost of debt rise as the total level of debt increases. In other words, the market value of debt may decrease the firm’s value faster at higher debt levels because high debt may alert the capital markets that the firm is less likely to survive, or because lenders become less willing to lend to firms that could be hit with bankruptcy costs. In this world, the value of the shares might be written as (6) S = (1 – tp ) (C – CI ) + (1 – tc ) [F (Ir ) – [D + d(D)]] . 0 The parameter d is the increasing cost of debt not captured in the interest rate, where d(0) = 0, d¢ > 0, and d¢¢ > 0. The new rule for debt-financed investment is (7) F ¢(I 0 ) = r [1 + d¢(D)]. The first-order condition for investment financed out of equity is the same as in equation (5). The rule for investment explicitly makes the amount of investment a decreasing function of the firm’s debt level if the firm finances out of debt. Similar to the MM model, the first-order conditions can generate corner solutions in which the firm finances investment either completely out of debt or completely out of cash. For example, if tc < tp , the firm will finance up to its total cash holdings out of equity, then finance out of debt only if the marginal return on investment at that point is greater than or equal to r. If the corporate tax rate is greater than the personal tax rate (as it is for most U.S. corporations), then the investment rule is more complicated. The firm will finance out of debt only if [1 + d¢(D)] < (1 – tp )/(1 – tc ); that is, if the marginal cost of increasing the firm’s debt burden is small enough. If this is not the case, firms will use both debt and cash to finance their investment projects. First-order conditions for investment financed out of cash remain the same (equation [5]), so that the equation determining the debt level, if both debt and cash are used to finance investment, is [1 + d¢(D)] = (1 – tp ). (1 – tc ) It is important for policymakers to know whether a world represented by MM or an environment of substantially increasing marginal cost of debt, crudely represented by equation (6), best reflects investment behavior. One easy reduced-form test of the MM assumptions in the earlier cross-section studies was to examine whether investment was negatively correlated with debt. If the study detected no negative relationship, then the conclusion might be that equation (6) did not make empirical sense. However, as the following example illustrates, lack of correlation between debt and equity might occur in a world that is very non-MM. Suppose our sample consists of two types of firms that differ only in their set of investment opportunities. Each type faces an investment (8) 22 payoff of ai F (I ), where a1 > a 2. The empirical researcher observes only the debt and investment outcomes of the two firms. The firms face a very non-MM world, one in which increasing debt discourages investment, represented by equation (6). We further assume that both debt and equity are used to finance investment, so that equation (8) holds. First-order conditions for the two firms are F ¢(I 0) = ar [1 + d¢(D)] = i r(1 – tp )/ai (1 – tc ). The behavior rule gives the following outcome: I1 > I2 and, if both firms finance investment out of some of their retained earnings, D1 = D2.8 A simple regression of investment on the firm’s debt level would lend support to the MM hypothesis of the irrelevance of debt for the level of investment, even though the data are generated by a behavior where the debt level, ceteris paribus, discourages investment. Clearly, lack of a simple correlation between debt and investment is not enough to test the appropriateness of the theory. If the potential cash available to the individual firms, C, is unobserved by the empirical researcher (as is likely), then the dividend amount may also be uncorrelated with the investment level. The problem is that the corporate financial instruments are behavior variables chosen by the agents, not experimental variables applied by the researcher. Any estimation must take into account that both investment and corporate financial structure are caused by the environment facing the firm, and that the available data contain very little of the information needed to reconstruct the decision process for each firm’s investment and corporate financial structure, even if all of the correct variables are included. The key to the earlier studies lay in finding sufficient instruments to control for the different investment opportunities represented by ai and for the fact that debt was a behavioral variable chosen by the firm. Poor instrument choice was bound to lead to poor estimates. In this estimating context, the underlying structure of behavior and a clear notion of what is generating differences across observations are needed to formulate a useful reduced form. The problems with reduced-form analysis are clear from this example, yet researchers do not necessarily have the data to conduct a fullblown structural estimation. Despite these limitations, we can profit from structural models by using them to devise a test that can help uncover some of the important factors driving firms’ investment decisions. IV. Tax Effects and the Investment/ Financing Relationship Early reduced-form empirical work on the tax effects of the investment/financing relationship, such as Long and Malitz (1985), Titman and Wessels (1988), and Fischer, Heinkel, and Zechner (1989), failed to find economically or statistically significant effects, just as early reduced-form studies failed to find a link between corporate financial structure and investment. These early nonstructural studies had an important influence on the policy debate, particularly when federal tax reform was discussed. For example, when the Economic Recovery Tax Act of 1981 (ERTA) was being debated, it was theoretically understood that in a creditconstrained world, the investment tax credit might yield a strong substitution effect as firms changed their investment funding from debt financing to retained earnings. This was not considered important because the early reducedform estimates indicated that the effect of taxes on corporate financial structure was negligible.9 Subsequent, more careful work has generally found evidence of a significant tax effect. For example, MacKie-Mason (1990) states that earlier studies suffered from a failure to fully consider the impact of a firm’s tax shields (tax deductions or investment tax credits) on its effective marginal tax rate. He notes that if a firm has no taxable income, any additional tax shields it receives will have no impact on its marginal tax rate. The marginal rate will be affected only if tax shields lower taxable income to zero. By taking this point into account and investigating incremental financing decisions using discretechoice models instead of debt/equity ratios, MacKie-Mason finds evidence that firms with high tax-loss carryforwards are less likely to rely on debt. This is certainly consistent with both the theoretical models of MM and the debtconstraint model, which predict that as the corporate tax rate decreases, debt should shrink. This more recent finding of a significant tax effect forces the reduced-form research to be more careful in defining its hypotheses. How do taxes influence investment? They could ( 1 – tp ) = 1 + d¢(D ). ( 1 – tc ) ■ 9 See Trezevant (1994), which discusses the contemporary debate surrounding ERTA. The author finds a significant substitution effect in taxes. ■ 8 This follows directly from the relation 23 affect it directly through a change in the posttax price of real investment, or indirectly through a change in corporate financial structure, as demonstrated in the previous section. The indirect influence is the one of interest to corporate finance. Separation of the indirect financial effect from the direct real-price effect requires a clarity that makes reduced-form estimation look more like structural estimation. This clarity is seen in more recent research that concentrates on the impact of taxes on corporate financial structure. Givoly et al. (1992) and Cummins, Hassett, and Hubbard (1994) find evidence of a relationship between debt and taxes. The Givoly study empirically examines the response of business firms to the Tax Reform Act of 1986 (TRA). The authors find evidence of a substitution effect between debt and nondebt tax shields, as hypothesized by DeAngelo and Masulis (1980). In addition, both corporate and personal tax rates appear to affect leverage decisions. Givoly et al. provide a good example of how forming an implicit structure about the effect of taxes provides a precise hypothesis for testing with a reduced-form estimation strategy. Consider how they use their structure and their knowledge of TRA specifics to develop simple statistical hypotheses. For example, they use only 1987 data to describe TRA’s effect, because they assert that the Act was surrounded by uncertainty until its actual passage by the Senate. Their test year was 1987, and their control years were 1984 and 1985, before any tax reform legislation was introduced. Although firms might have behaved in anticipation of a new tax structure, it is unlikely that this Lucas effect would be of overriding importance in the statistical results. Givoly et al. test their hypotheses involving tax code changes by estimating standard crosssection OLS regressions of the change in leverage on the firm’s effective tax rate, nondebt tax shields, dividend yield, Tobin’s q, firm size, business risk, and changes in depreciation and investment tax credits.10 The authors are able to reach definitive conclusions about the effect of the TRA using cross-section analysis because they state their hypotheses carefully. For example, the TRA greatly reduced the statutory corporate tax rates, so that firms faced more similar ■ 10 In Givoly et al., Tobin’s q proxies for bankruptcy costs and the collateral value of the firm when bankrupt. Business risk is proxied by the coefficient of variation in operating income (minus depreciation) over the firm’s last 10 years. rates. Thus, their Hypothesis 1 states that in response to the decline in the statutory corporate tax rate, firms with a high marginal effective rate will decrease their leverage more than will firms with a low marginal effective rate. In other words, the decline in the effective corporate tax rate will have a greater impact on firms with high marginal tax burdens. Low effective tax rates imply a low tax advantage of debt and result in a smaller decrease in leverage stemming from a cut in the statutory tax rate. Hence, the relationship between the effective tax rate and changes in leverage should be inverse. Notice how the hypothesis embodies a solid underlying structure of how the firm reacts to a tax change. This structure gives the hypothesis a clarity and specificity that provide the necessary power for a reduced-form test. The hypothesis also illustrates a clear understanding of the workings of the tax code. Although tax laws specify the marginal statutory corporate tax rate, corporate decisions are based on the marginal effective tax rate, which is the present value of future tax payments resulting from an additional dollar of taxable income. The statutory rate is the same for all firms, but the effective rate differs from firm to firm above a certain dollar amount. Tax shields such as investment tax credits, tax-loss carryforwards, depreciation allowances, and interest expenses lower the effective tax rate. The Givoly results support all of their hypotheses: Each of the relevant coefficients is statistically significant and has the proper sign for 1987, the first year the TRA was in effect. The hypotheses appear to be of moderate economic significance. For example, a firm with an effective tax rate 10 percent above that of another firm would have lowered its debt/equity ratio 1.15 percent more in response to the TRA. Givoly and others provide only part of the answer regarding the effect of financial structure on real input decisions. By showing that taxes influence financial structure, the studies have shown that capital markets are imperfect, thus providing an important clue as to why the MM propositions are not supported by the data. However, they have not clearly shown how these financial decisions impact real input decisions. Furthermore, the magnitude of the effect is far from certain. Thus, the latest findings point out one link of the chain of indirect tax effects through capital structure by showing how taxes influence corporate finance. However, the complete change still suffers from lack of information on the magnitude of the effect of corporate financial structure on real investment decisions. 24 Although earlier empirical work found no significant tax effects, more recent studies have addressed the inherent empirical problems and have produced evidence that supports the importance of taxes on financing decisions. Hence, one link between policy and real investment behavior has been established, but only after very clear statements about the firm’s underlying structural behavior were used to define relevant variables and identifying assumptions. These are formulations that require a careful, informed analysis. Even so, only a fully structural model can provide policymakers with an accurate measure of how changes in tax policy influence real investment. Without parameter estimates from such a model, the short- and long-term effects of tax policy changes on real investment remain uncertain. V. Conclusion Simply estimating a structural model without first determining and reporting general directions in the data is a recipe for disaster. However, the reduced-form strategy, when used without accompanying structural estimates, is distinguished by what it has not done. We do not have precise estimates of the magnitude of the effects. Because the estimating equations are formulated without an explicit structure, the resulting parameters are subject to fewer “reality checks” to determine whether they make economic sense. In addition, fewer comparisons can be made to related empirical literatures to determine the appropriateness of the estimating equations’ specifications. Is this or that estimated parameter comparable to a risk-aversion parameter estimated in the portfolio-balance literature? We cannot tell from a reduced-form estimate because the reduced form resists a structural interpretation that will allow comparison. What have we learned from examining 35 years of research? In each case—the direct test of the MM hypothesis through cross-section regressions, the test of the timing of investment through Granger causality, and the test of whether taxes should matter to corporate financial structure—the findings exhibit the same pattern. Early research often failed to reveal statistical significance in the relationship between corporate real investment and the explanatory variable, be it financial structure or taxes. This seemed to provide prima facie support for the empirical relevance of MM’s assertions. Our present knowledge of corporate financial structure through reduced-form estimation is typical of what a more careful reduced-form strategy can do. The weight of the current evidence seems to reject the MM neutrality hypothesis. Financial structure does matter to a firm’s investment decisions, and taxes do influence these decisions through their effect on financial structure. These are important statements to bear in mind both when deciding on policy and when formulating new theory with which to guide policy. Our cautionary tale does not say that the reduced form is an unwise estimation strategy. Rather, it notes what conditions are necessary if a reduced form is to yield accurate results. In all cases, an underlying structure of behavior (even when not used explicitly in a structural estimation model) guided the research through the crucial steps of data definition and formulating the correct econometric test. It is also important to note that a reduced-form analysis is a critical step in any empirical study of a policy question. References Cummins, J.G., K.A. Hassett, and R.G. Hubbard. “A Reconsideration of Investment Behavior Using Tax Reforms as Natural Experiments,” Brookings Papers on Economic Activity, vol. 2 (1994), pp. 1–59. DeAngelo, H., and R. Masulis. “Optimal Capital Structure under Corporate and Personal Taxation,” Journal of Financial Economics, vol. 8, no. 1 (March 1980), pp. 3–29. Dhrymes, P., and M. Kurz. “Investment, Dividend, and External Finance Behavior of Firms,” in Determinants of Investment Behavior. New York: National Bureau of Economic Research, 1967. Fischer, E., R. Heinkel, and J. Zechner. “Dynamic Capital Structure Choice: Theory and Tests,” Journal of Finance, vol. 44, no. 1 (March 1989), pp. 19–40. Gilchrist, S. and C.P. Himmelberg. “Evidence on the Role of Cash Flow for Investment,” New York University, Stein School of Business, Working Paper, 1995. Givoly, D., C. Hayn, A. Ofer, and O. Sarig. “Taxes and Capital Structure: Evidence from Firms’ Response to the Tax Reform Act of 1986,” Review of Financial Studies, vol. 5, no. 2 (1992), pp. 331–55. 25 Granger, C. “Investigating Causal Relations by Econometric Models and Cross-Spectral Methods,” Econometrica, vol. 37, no. 3 ( July 1969), pp. 424–38. Long, M.S., and I.B. Malitz. “Investment Patterns and Financial Leverage,” in B.M. Friedman, ed., Corporate Capital Structures in the United States. Chicago: University of Chicago Press, 1985, pp. 325–48. MacKie-Mason, J. “Do Taxes Affect Corporate Financing Decisions?” Journal of Finance, vol. 45, no. 5 (December 1990), pp. 1471–93. McCabe, G.M. “The Empirical Relationship between Investment and Financing: A New Look,” Journal of Financial and Quantitative Analysis, vol. 14, no. 1 (March 1979), pp. 119–35. McDonald, J.G., Jacquillat, B., and M. Nussenbaum. “Dividend, Investment, and Financing Decisions: Empirical Evidence on French Firms,” Journal of Financial and Quantitative Analysis, vol. 10, no. 5 (December 1975), pp. 741–55. Modigliani, F., and M. Miller. “The Cost of Capital, Corporate Finance, and the Theory of Investment,” American Economic Review, vol. 48, no. 3 (June 1958), pp. 261–97. _________ and ________. “Corporate Income Taxes and the Cost of Capital: A Correction,” American Economic Review, vol. 53, no. 3 ( June 1963), pp. 433–43. Mougoue, M., and T. K. Mukherjee. “An Investigation into the Causality among Firms’ Dividend, Investment, and Financing Decisions,” Journal of Financial Research, vol. 17, no. 4 (Winter 1994), pp. 517–30. Peterson, P. P., and G.A. Benesh. “A Reexamination of the Empirical Relationship between Investment and Financing Decisions,” Journal of Financial and Quantitative Analysis, vol. 18, no. 4 (December 1983), pp. 439–53. Smirlock, M., and W. Marshall. “An Examination of the Empirical Relationship between the Dividend and Investment Decisions: A Note,” Journal of Finance, vol. 38, no. 5 (December 1983), pp. 1659–67. Titman, S., and R. Wessels. “The Determinants of Capital Structure Choice,” Journal of Finance, vol. 43, no. 1 (March 1988), pp. 1–19. Trezevant, R. “How Did Firms Adjust Their Tax-Deductible Activities in Response to the Economic Recovery Tax Act of 1981?” National Tax Journal, vol. 47, no. 2 ( June 1994), pp. 253–71. 26 Predicting Real Growth Using the Yield Curve by Joseph G. Haubrich and Ann M. Dombrosky Introduction The yield curve, which plots the yield of Treasury bonds against their maturity, is one of the most closely watched financial indicators.1 Many market observers carefully track the yield curve’s shape, which is typically upward sloping and somewhat convex. At times, however, it becomes flat or slopes downward (“inverts,” in Wall Street parlance), configurations that many business economists, financial analysts, and other practitioners regard as harbingers of recession (see figure 1). A recent article in Fortune labeled the yield curve “a near-perfect tool for economic forecasting” (see Clark [1996]). In fact, forecasting with the yield curve does have a number of advantages. Financial market participants truly value accurate forecasts, since they can mean the difference between a large profit and a large loss. Financial data are also available more frequently than other statistics (on a minute-to-minute basis if one has a computer terminal), and such a simple test as an inversion does not require a sophisticated analysis. In this Review, we examine the yield curve’s ability to predict recessions and, more generally, future economic activity. After comparing the Joseph G. Haubrich is a consultant and economist and Ann M. Dombrosky is a financial reports analyst at the Federal Reserve Bank of Cleveland. curve’s forecasts with the historical record, we judge its accuracy against other predictions, including naive forecasts, traditional leading indicators, and sophisticated professional projections. This article builds on a wide range of previous research, but, taking an eclectic approach, differs from the earlier work in a variety of ways. These differences show up mainly in the way we judge forecast performance. Like the important early work of Harvey (1989, 1991, 1993) and Hu (1993), we use outof-sample forecasts and compare yield curve forecasts with other predictions (including professional forecasts), but we extend our data set to the mid-1990s. In addition, we consider how adding the yield curve improves (or reduces) the accuracy of other forecasts. In this, we follow Estrella and Hardouvelis (1991), who do not, however, use out-of-sample forecasts. Finally, building on the recent work of Estrella and Mishkin (1995, 1996), we consider how well the yield curve predicts the severity of recessions, not just their probability, and compare the forecasts with a wider range of alternatives. ■ 1 Yield curve reports appear in the “Credit Markets” section of The Wall Street Journal and the “Business Day” section of The New York Times. 27 F I G U R E 1 Yield Curvesa provides some insight into market sentiment. Of course, it’s always a good idea to check whether the expensive and complicated forecasts actually do perform better. After first reviewing some basics about the yield curve and the reasons it might predict future growth, we look at the actual relationship and compare predictions from the yield curve to those generated by naive statistical models, traditional indicators, professional forecasters, and an econometric model. I. Interest Rates and Real Economic Activity While our main goal is a rather atheoretical a. Three-month and six-month instruments are quoted from the secondary market on a yield basis; all other instruments are constant-maturity series. SOURCE: Board of Governors of the Federal Reserve System. The most distinguishing feature of this paper, however, is that it documents the decline in the yield curve’s predictive ability over the past decade (1985–95) and discusses possible reasons for this phenomenon. By some measures, the yield curve should be an even better predictor now than it has been in the past. Widespread use of the yield curve makes assessing its accuracy a worthwhile exercise for economists. But policymakers, too, need an accurate and timely predictor of future economic growth. The ready availability of term-structure data (as opposed to, say, quarterly GDP numbers) ensures a timely prediction, but accuracy is another question. Central bankers have an added incentive to understand the yield curve, since the federal funds rate and the discount rate are themselves interest rates. Uncovering the “stylized facts” about the curve can help the Federal Reserve to understand the market in which it operates. With sophisticated macroeconometric models and highly paid professional forecasters, is there any place for a simple indicator like the yield curve? Aside from the knowledge gained about the curve itself, there are several reasons to answer that question affirmatively. Simple predictions may serve as a check on more complex models, perhaps highlighting when assumptions or relationships need rethinking. Agreement between predictions increases confidence in the results, while disagreement signals the need for a second look. A simple, popular indicator also assessment of the yield curve’s predictive power, forecasts based on the yield curve are on a sounder economic footing than those based on hemlines or Superbowl victories. The best way to see this is to start with a simple theory of the term structure, called the expectations hypothesis. Under this theory, long-term interest rates are the average of expected future short-term rates. If today’s one-year rate is 5 percent and next year’s one-year rate is expected to be 7 percent, the two-year rate should be 6 percent ([7 + 5] –.. 2 = 6). More generally, the expectations hypothesis equates the yield (at time t) on an n-period bond, Ynt , and a sequence of oneperiod bonds:2 Ynt = Et (Y1, t Y1, t + 1Y1, t + 2 ...Y1, t + n – 1). If low interest rates are associated with recessions, then an inverted term structure—implying that upcoming rates will be lower—predicts a recession.3 One possible reason for expecting low interest rates in recessions might be termed the policy anticipations hypothesis. If policymakers act to reduce short-term interest rates in recessions, market participants who expect a recession would also expect low rates. The yield curve picks up the financial market’s estimate of future policy.4 ■ 2 See chapter 7 of Mishkin (1989) for a fuller treatment of this issue. Mishkin also points out the main flaw in the expectations hypothesis: The term structure normally slopes up, but interest rates do not trend up over time. Campbell (1995) offers a useful discussion of related points. ■ 3 For a classic documentation of this pattern, see Kessel (1965). ■ 4 Rudebusch (1995) takes this approach. 28 Another possibility is that current monetary policy may shift both the yield curve and future output. For example, tight monetary policy might raise short-term interest rates, flattening the yield curve and leading to slower future growth. Conversely, easy policy could reduce short-term interest rates, steepen the yield curve, and stimulate future growth. The yield curve predicts future output because each of these shifts follows from the same underlying cause: monetary policy. Taking this logic one step further, monetary policy may react to output, so that the yield curve picks up a complex intermingling of policy actions, reactions, and real effects. In these explanations, the yield curve reflects future output indirectly, by predicting future interest rates or future monetary policy. It may also reflect future output directly, because the 10-year interest rate may depend on the market’s guess of output in 10 years. The expectations hypothesis certainly marks the beginning of wisdom about the yield curve, but only the beginning. The 30-year bond may have a high interest rate not because people expect interest rates to rise, but because such a bond must offer a high return to get people to hold it in the first place. (This is commonly called the risk premium, though for some theories that may be a misnomer.) Investors may dislike wide swings in prices as market expectations about the distant future change over time. Conversely, there may be reasons why some people would rather hold a 30-year bond than a one-year bond. For example, they may be saving for retirement and prefer the certain payoff on the longer-term note (this is sometimes called the preferred habitat hypothesis). The risk premium provides another reason why the yield curve may be a useful predictor: The premium itself holds information. As a simple example, consider that recessions may make people uncertain about future income and employment, or even about future interest rates. The risk premium on a longer-term bond reflects this. In conjunction with changes working through the expectations hypothesis, the yield curve may take some very strange twists indeed, becoming inverted, humped, or even u-shaped.5 These explanations provide an additional motivation for investigating yield curve predictions. They also hint at the many important issues that transcend the yield curve’s predictive power. It matters, for instance, if the curve reacts to future policy, to movements in output, or to some combination of the two. But these considerations fall by the wayside if the yield curve is not an accurate predictor of future economic activity. In this article, we concentrate on that more basic issue, leaving determination of the underlying causes for another day. II. Data and Computation There are many ways of using the yield curve to predict future real activity. One common method uses inversions (when short rates exceed long rates) as recession indicators. Is it possible, however, to predict the magnitude as well as the direction of future growth? Does a large inversion predict a severe recession? Does a steep yield curve predict a strong recovery? 6 Operationally, this means relating a particular measure of yield curve “steepness” to future real growth. In taking this route, we follow and build on the related work of Estrella and Hardouvelis (1991). Obtaining predictions from the yield curve requires much preliminary work. Three principles guided us through the many decisions that were required: Keep the process simple, preserve comparability with previous work, and avoid data snooping. Thus, we avoided both complicated nonlinear specifications and a detailed search for the “best” predictor. To begin with, there is no unambiguous measure of yield curve steepness. A yield curve may be flat at the short end and steep at the long end. The standard solution uses a spread, or the difference between two rates (in effect, a simple linear approximation of the nonlinear yield curve).7 This means choosing a particular spread, in itself no trivial matter. Among the 10 most commonly watched interest rates (the federal funds rate and the three-month, six-month, and one-, two-, three-, five-, seven-, 10-, and 30year Treasury rates), 45 possible spreads exist.8 An additional problem is that there are several types of yield curves or term structures. In fact, it sometimes helps to draw a distinction between the yield curve and the term structure. The yield curve is the relation between the ■ 5 Stambaugh (1988) makes this point. For a less technical description, see Haubrich (1991). ■ 6 Other approaches also exist. For example, Harvey (1988) examines whether the term structure predicts changes in consumption. ■ 7 Frankel and Lown (1994) is one of the few papers that considers nonlinear measures of steepness. ■ 8 If there are n rates, there are n/2 (n –1) spreads. This is the classic formula for combinations. See Niven (1965), chapter 2. 29 yield on Treasury securities and their maturity. The term structure is a particular yield curve— that for zero-coupon Treasury securities. The term structure is theoretically more interesting. It answers the question, “How much would I pay for one dollar delivered 10 years from today?” The problem is that a zero-coupon Treasury security rarely matures in exactly 10 years. What we actually observe in the market are prices (and thus yields) on existing Treasury securities. These may not mature in precisely 10 years (or whatever maturity you choose), and they often have coupon payments. That is, a 10-year Treasury note pays interest semiannually at a specified coupon rate, so its yield is not the yield called for in the term structure. Finding the desired interest rate almost always involves estimation of some kind. Calculating the theoretically pure term structure is often quite difficult, as it must be estimated from coupon bonds of the wrong maturity, all subject to taxation. (Using zero-coupon bonds may help, but this approach introduces problems of its own, as the market is thinner and the tax treatment of coupons and principal differs.) This means that the pure term structure is not available in real time, when the Federal Reserve must attempt to discern the course of the economy. To avoid stale data, we must turn to the more “rough-and-ready” yield curve. Even here, the problem of matching maturities arises. Fortunately, the Treasury Department publishes a “constant-maturity” series, where market data are used to estimate today’s yield on a 10-year Treasury note, even though no such note exists.9 For our study, we use data from the Federal Reserve’s weekly H.15 statistical release (“Selected Interest Rates”), which compiles interest rates from various sources. For the spread, we chose the 10-year CMT rate minus the secondary-market three-month Treasury bill rate. In addition to allowing a comparison with the work of Estrella and Hardouvelis (1991), Harvey (1989, 1993), and Estrella and Mishkin (1995, 1996), choosing only one spread enables us to minimize the problems of data snooping (Lo and MacKinlay [1990]) and the associated spuriously good results.10 That is, trying every single spread would produce something that looked like a good predictor, but it very likely would be a statistical fluke akin to Superbowl victories and hemlines. We then convert the bill rate, which is published on a discount rate basis, to a coupon-equivalent yield so that it is on the same basis as the 10-year rate.11 Also following Estrella and Hardouvelis, we use quarterly averages for the spread. This smoothes the anomalous rates that appear at the turn of each month.12 A priori there is no presumption that GDP should correlate better with a particular date’s spread than with the quarterly average.13 As our measure of real growth, we use the four-quarter percent change in real (fixedweight) GDP. GDP is, of course, the standard measure of aggregate economic activity, and the four-quarter forecast horizon answers the “what-happens-next-year” type of question without embroiling us in data snooping issues regarding the optimal horizon choice. Our sample period runs from 1961:IQ through 1995:IIIQ. This covers various inflationary experiences, episodes of monetary policy tightening and easing, and several business cycles and recessions. Included are five recessions, inflation rates from 1 percent to more than 13 percent, and a federal funds rate ranging from under 3 percent to over 19 percent. Our basic model, then, is designed to predict real GDP growth four quarters into the future based on the current yield spread. Operationally, we accomplish this by running a series of regressions (detailed below) using real GDP growth and the interest rate spread lagged four quarters (for example, the interest rate spread used for 1961:IQ is actually from 1960:IQ). The next step involves comparing the yield curve forecasts with a sequence of increasingly sophisticated predictions using other techniques. We start with a naive (but surprisingly effective) technique which assumes that GDP growth over the next four quarters will be the same as it was over the last four. (That is, the growth rate is a random walk.) We then regress real GDP growth against the index of leading economic indicators (lagged four quarters). This enables us to make a comparison with another simple and popular forecasting technique. ■ 9 See Smithson (1995) for a good description of constant-maturity Treasuries (CMTs). ■ 10 Work by Knez, Litterman, and Scheinkman (1994) suggests that using more or different rates would not capture much additional information. ■ 11 The three-month CMT rate was not published before May 1995. To keep the data consistent, we use the secondary-market threemonth Treasury bill rate throughout. For such a short rate, the differences are minimal. ■ 12 Park and Reinganum (1986) document this calendar effect. ■ 13 We also reworked the results using data for the last week of each quarter for the 1963–95 period. The findings were comparable, although the predictive power of the spread decreased somewhat. 30 B O X 1 Forecasting Equations Yield spread: in-sample RGDPt + 4 – RGDPt Yield spread: out-of-sample RGDPt + 4 – RGDPt Naive RGDPt + 4 – RGDPt Leading indicators RGDPt + 4 – RGDPt Lagged GDP RGDPt + 4 – RGDPt RGDPt RGDPt RGDPt Lagged GDP plus yield spread RGDPt RGDPt RGDPt + 4 – RGDPt RGDPt = a + b spreadt = a + b spreadt = RGDPt – RGDPt – 4 RGDPt = a + b indext = a+b = a+b We next look at two additional forecasts generated by statistical procedures. We regress real GDP growth against its own lag (again, four quarters) and against its own lag and the 10-year, three-month spread. The final, and most sophisticated, alternative forecasts we consider come from the Blue Chip organization and DRI/McGraw–Hill (hereafter referred to simply as DRI). We first compare the results of our model with forecasts from Blue Chip Economic Indicators, beginning with the July 1984 issue.14 We use the one-year-ahead Blue Chip consensus forecasts for real GDP (or real GNP when GDP forecasts are unavailable), labeled “percent change from same quarter in prior year.” These forecasts are taken from the Blue Chip newsletters corresponding to the first month of each quarter. We next compare our results with predictions from DRI, reported in various issues of its Review of the U.S. Economy. DRI generates these forecasts from an econometric model. Although we tried to collect forecasts for the same period as our Blue Chip forecasts (that is, from issues corresponding to the first month of each quarter), our DRI data set is missing two points: 1985:IIIQ and 1987:IIQ. We use forecasts for real GDP (or real GNP when GDP forecasts are unavailable) one year ahead. Box 1 summarizes the regressions used to forecast future real GDP growth. RGDPt – RGDPt – 4 RGDPt RGDPt – RGDPt – 4 RGDPt + g spreadt III. Forecast Results Does the yield curve accurately predict future GDP? First, look directly at the data. Figure 2 shows the growth of real GDP and the lagged spread between the 10-year and three-month Treasury yields. A decline in the growth of real GDP is usually preceded by a decrease in the yield spread, and a narrowing yield spread often signals a decrease in real GDP growth. A negative yield spread (inverted yield curve) usually precedes recessions, but not always. For example, the yield spread turned negative in the third and fourth quarters of 1966, but no recession occurred for the next three years. (The recession that began in late 1969 was preceded by two quarters of a negative yield spread.) The latest recession, which occurred in 1990–91, was preceded by a yield curve more accurately described as flat than inverted. Figure 3 plots the same data in a different form. It shows a scatterplot, with each point representing a particular combination of real ■ 14 Blue Chip Economic Indicators is a monthly collection of economic forecasts by a panel of economists from some of the top firms in the United States (the so-called Blue Chip companies). The Blue Chip consensus forecast for real GDP is the average of about 50 individual forecasts. See Zarnowitz and Lambros (1987) for evidence that the consensus forecast predicts much better than individual forecasts, and Lamont (1994) for a possible explanation. 31 F I G U R E 2 Real GDP Growth and Lagged Yield Spread generate the GDP predictions, we ran an insample regression, using the entire sample to generate each predicted data point. This is the sort of comparison Estrella and Hardouvelis (1991) make, and our results, presented below, confirm their assessment that the 10-year, threemonth spread has significant predictive power for real GDP growth:15 Real GDP growth = 1.8399 + 0.9791spread (3.89) (4.50) R 2 = 0.291, D – W = 0.352. a. Four-quarter percent change. b. Lagged four quarters. NOTE: Shaded areas indicate recessions. SOURCES: Board of Governors of the Federal Reserve System; and U.S. Department of Commerce, Bureau of Economic Analysis. F I G U R E 3 Scatterplot of Real GDP Growth and Lagged Yield Spread a. Four-quarter percent change. b. Lagged four quarters. SOURCES: Board of Governors of the Federal Reserve System; and U.S. Department of Commerce, Bureau of Economic Analysis. GDP growth and the lagged yield spread. Even a casual look at the results reveals that the relationship between the two variables is usually positive; that is, positive real GDP growth is associated with a positive lagged yield spread, and vice versa. Plotting the data gives a strong, albeit qualitative, impression that the yield spread predicts future real activity. We desire a more quantitative prediction, one that says more than “The yield curve is steep; looks like good times.” To The yield spread emerges as statistically and economically significant, translating almost one for one into expected future growth. Thus, a spread of 100 basis points (1 percent) implies future growth of 2.8 percent (we derive this as 1.8 + 0.98 x 1). The R 2 indicates that much variation remains to be explained. Figure 4 plots our in-sample real GDP predictions versus actual real GDP growth. The in-sample results are somewhat misleading, as the coefficients depend on information not available early in the sample. Figure 5 plots another series of predicted real GDP growth, this time generated from an out-of-sample regression. Each data point in this chart is based on a regression using only the data (yield spreads) before the predicted data point. That is, the predicted GDP growth rate for, say, 1980:IQ is based on the data sample from 1961:IQ through 1979:IVQ. Hence, this regression generates a true forecast because it uses available data to predict future (out-of-sample) real GDP growth.16 The predicted GDP series from the in-sample and out-of-sample regressions are broadly similar and generally follow the actual GDP data. The root mean square error (RMSE) of the predictions is 2.04 for the in-sample and 2.10 for the out-of-sample forecasts. It is not surprising that the in-sample regression performs slightly better. If we calculate the RMSE over the last 10 years of our data set (1985:IIIQ to 1995:IIIQ), the in-sample regression (RMSE 1.07) again does better than the out-of-sample regression (RMSE 2.09). ■ 15 The t statistics are Newey–West corrected with five lags. This offsets the bias created by overlapping prediction intervals (a serious problem in this case, as indicated by the Durbin–Watson statistic). ■ 16 We also corrected the regression for a more subtle problem. Because first-quarter GDP numbers become available only after the first quarter ends, we should not use those numbers in a regression. (That is, as of the first quarter, we still don’t know the four-quarter growth rate over last year’s first quarter.) 32 F I G U R E 4 Real GDP Predictions: In-Sample NOTE: Shaded areas indicate recessions. SOURCES: U.S. Department of Commerce, Bureau of Economic Analysis; and authors’ calculations. F I G U R E 5 Real GDP Predictions: Out-of-Sample NOTE: Shaded areas indicate recessions. SOURCES: U.S. Department of Commerce, Bureau of Economic Analysis; and authors’ calculations. T A B L E 1 RMSE of Forecasts 1965:IVQ– 1995:IIIQ 1985:IIIQ– 1995:IIIQ 2.04 2.10 3.15 2.37 2.49 2.19 n.a. n.a. 1.07 2.09 1.66 1.36 1.46 2.06 0.63 1.13 Yield spread: in-sample Yield spread: out-of-sample Naive Leading indicators Lagged GDP Lagged GDP plus yield spread DRIa Blue Chip a. DRI forecasts are missing two quarters. SOURCE: Authors’ calculations. We use this RMSE criterion to compare the yield spread forecasts with those derived from other techniques. The results are reported in table 1. For the entire sample, the yield curve emerges as the most accurate predictor of real economic growth. Furthermore, adding the yield spread to a lagged GDP regression improves the forecast, while adding lagged GDP to the yield spread worsens the forecast. For in-sample regressions, adding variables never hurts, but it quite commonly reduces the performance of the out-of-sample regressions. Curiously, the 1985–95 subsample completely reverses the results. The yield spread becomes the least accurate forecast, and adding it to lagged GDP actually worsens the fit. The leading indicators emerge as the best of the “low-cost” forecasts, and the two professional services do markedly better than the rest. In part, this may reflect the simple specifications used in the regression forecasts: With more lags, a simple regression forecast is often better than a sophisticated model (see Chatfield [1984], chapter 5). But the change is even more significant than that. Using unpublished data, Harvey (1989) finds that over the 1976–85 period, the yield spread performs as well as or better than seven professional forecasting services (including DRI but not Blue Chip). The dramatic drop in forecasting ability may result from several factors. It certainly reflects a changing relationship between the yield curve and the economy. The coefficients in the termspread regression demonstrate this. At the beginning of the sample, using only 20 data points (five years of quarterly data), the coefficient on the term spread is –0.14 (statistically insignificant). Midway through the sample, after 70 data points, the coefficient is 1.48, and for the whole sample, 0.98. While one advantage of an out-of-sample procedure is that it allows the coefficients to change, the influence of the first 20 years may force the “wrong” coefficient on the last 10. It is also quite reasonable that the relationship between the yield curve and the real economy might have changed over 30 years. Advances in technology, new production processes, changes in market organization or in the way the market reacts to new information, or even shifts in Federal Reserve policy (the famous “Lucas critique”) might have altered the relationship between the yield curve and real activity. Evidence suggests that both the timing and the size of the relationship between the yield curve and real activity has in fact changed. If 33 T A B L E 2 Correlations between Lagged Yield Spread and Real GDP Growth Lag 0 1 2 3 4 5 6 7 8 1960:IQ– 1995:IIIQ 1960:IQ– 1985:IIQ 1985:IIIQ– 1995:IIIQ 0.041 0.206 0.366 0.481 0.540 0.505 0.423 0.331 0.218 0.075 0.252 0.439 0.585 0.666 0.624 0.523 0.398 0.251 0.143 0.358 0.538 0.641 0.684 0.708 0.736 0.721 0.660 SOURCE: Authors’ calculations. F I G U R E 6 Real GDP Predictions: Blue Chip NOTE: Shaded area indicates recession. SOURCES: U.S. Department of Commerce, Bureau of Economic Analysis; and Blue Chip Economic Indicators, various issues. F I G U R E we look at the correlations between the yield spread and real GDP growth at different lags (table 2), we see that the recent period has higher correlations between lags of yield spreads and real GDP growth. We also see that the largest correlation for the early (and total) period, 0.666, occurs at a lag of four quarters, exactly the lag used in our regressions. In the recent period, however, despite a higher correlation at four lags, the highest correlation (0.736) is reached at six quarters. The correlations drop off more slowly in the latter period as well. This accounts for our somewhat paradoxical conclusion: Despite better correlations between the yield curve and real GDP—with an in-sample RMSE that beats even the Blue Chip forecast—regressions using past data are less reliable predictors. It is somewhat instructive to make a more detailed comparison with the sophisticated forecasts. The Blue Chip predictions can be considered out-of-sample because each one is based on data available at the time of the prediction. Figure 6 plots the Blue Chip consensus forecasts against actual real GDP growth. The Blue Chip forecasts appear much smoother than GDP, as they consistently underpredict real GDP when economic growth is high and overpredict real GDP when economic growth is negative. The DRI forecasts are plotted in figure 7. These appear broadly similar to the Blue Chip forecasts, although the DRI series is more volatile. Like the Blue Chip forecasts, the DRI forecasts generally underpredict GDP when economic growth is high and overpredict GDP when economic growth is low. Our out-ofsample forecasts based on the yield spread overpredict GDP growth for all but one quarter during the 1985:IIIQ–1995:IIIQ period (corresponding to the Blue Chip and DRI data sets). 7 Real GDP Predictions: DRI IV. Conclusion Does the yield curve accurately predict real a. 1985:IIIQ and 1987:IIQ data are missing. NOTE: Shaded area indicates recession. SOURCES: U.S. Department of Commerce, Bureau of Economic Analysis; and DRI/McGraw–Hill. economic growth? Answering this seemingly simple question requires a surprising amount of preliminary work. Much of this paper is devoted to refining the initial question to confront the realities of the financial marketplace. Fortunately, the answer is less complex, if somewhat nuanced. The 10-year, three-month spread has substantial predictive power, and in this we confirm a variety of earlier studies. Over the past 30 years, it provides one of the best (in our sample, the best) forecasts of real growth four quarters into the future. Over the past decade, it has been less successful: Indeed, 34 the yield curve was the worst forecast we examined. This shift seemingly results from a change in the relationship between the yield curve and real economic activity—one that has become closer, but nonetheless has made regressions based on past data less useful. An interesting topic for future research would be to examine whether simple fixes, such as a rolling regression model or more lags, could improve the recent performance of the yield curve. Certainly the simple yield curve growth forecast should not serve as a replacement for the consensus predictions of the Blue Chip panel or the DRI econometric model. It does, however, provide enough information to serve as a useful check on the more sophisticated forecasts and to encourage future research into the reasons behind the yield curve’s worsening performance. References Campbell, J.Y. “Some Lessons from the Yield Curve,” Journal of Economic Perspectives, vol. 9, no. 3 (Summer 1995), pp. 129–52. Chatfield, C. The Analysis of Time Series: An Introduction, 3d. ed. New York: Chapman and Hall, 1984. Clark, K. “A Near-Perfect Tool for Economic Forecasting,” Fortune, July 22, 1996, pp. 24–26. Estrella, A., and G.A. Hardouvelis. “The Term Structure as a Predictor of Real Economic Activity,” Journal of Finance, vol. 46, no. 2 ( June 1991), pp. 555–76. Estrella, A., and F. S. Mishkin. “Predicting U.S. Recessions: Financial Variables as Leading Indicators,” National Bureau of Economic Research, Working Paper No. 5379, 1995. ___________ , and ___________ . “The Yield Curve as a Predictor of U.S. Recessions,” Federal Reserve Bank of New York, Current Issues in Economics and Finance, vol. 2, no. 7, June 1996. Frankel, J.A., and C.S. Lown. “An Indicator of Future Inflation Extracted from the Steepness of the Interest Rate Yield Curve along Its Entire Length,” Quarterly Journal of Economics, vol. 109, no. 2 (May 1994), pp. 517–30. Harvey, C.R. “The Real Term Structure and Consumption Growth,” Journal of Financial Economics, vol. 22, no. 2 (December 1988), pp. 305–33. _______________. “Forecasts of Economic Growth from the Bond and Stock Markets,” Financial Analysts Journal, September/ October 1989, pp. 38–45. _______________. “The Term Structure and World Economic Growth,” Journal of Fixed Income, June 1991, pp. 7–19. _______________. “Term Structure Forecasts Economic Growth,” Financial Analysts Journal, vol. 49, no. 3 (May/June 1993), pp. 6–8. Haubrich, J.G. “Wholesale Money Market,” in The New Palgrave Dictionary of Money and Finance. New York: Stockton Press, 1992, pp. 798–800. Hu, Z. “The Yield Curve and Real Activity,” International Monetary Fund Staff Papers, vol. 40, no. 4 (December 1993), pp. 781–806. Kessel, R.A. “ The Cyclical Behavior of the Term Structure of Interest Rates,” National Bureau of Economic Research, Occasional Paper No. 91, 1965. Knez, P. J., R. Litterman, and J. Scheinkman. “Explorations into Factors Explaining Money Market Returns,” Journal of Finance, vol. 49, no. 5 (December 1994), pp. 1861–82. Lamont, O. “Macroeconomic Forecasts and Microeconomic Forecasters,” Princeton University, unpublished manuscript, 1994. Lo, A.W., and A.C. MacKinlay. “Data-Snooping Biases in Tests of Financial Asset Pricing Models,” Review of Financial Studies, vol. 3, no. 3 (1990), pp. 431–67. Mishkin, F. S. The Economics of Money, Banking, and Financial Markets, 2d. ed. Glenview, Ill.: Scott, Foresman, and Company, 1989. Niven, I. Mathematics of Choice: How to Count without Counting. Washington, D.C.: The Mathematical Association of America, 1965. 35 Park, S.Y., and M.R. Reinganum. “The Puzzling Behavior of Treasury Bills that Mature at the Turn of Calendar Months,” Journal of Financial Economics, vol. 16, no. 2 (June 1986), pp. 267–83. Rudebusch, G.D. “Federal Reserve Interest Rate Targeting, Rational Expectations, and the Term Structure,” Journal of Monetary Economics, vol. 35, no. 2 (April 1995), pp. 245–74. Smithson, C. “ABC of CMT,” Risk, vol. 8, no. 9 (September 1995), pp. 30–31. Stambaugh, R.F. “The Information in Forward Rates: Implications for Models of the Term Structure,” Journal of Financial Economics, vol. 21, no. 1 (May 1988), pp. 41–70. Zarnowitz, V., and L.A. Lambros. “Consensus and Uncertainty in Economic Prediction,” Journal of Political Economy, vol. 95, no. 3 ( June 1987), pp. 591–621.