The full text on this page is automatically extracted from the file linked above and may contain errors and inconsistencies.
Economic Quarterly—Volume 94, Number 3—Summer 2008—Pages 197–218 New Keynesian Economics: A Monetary Perspective Stephen D. Williamson S ince John Maynard Keynes wrote the General Theory of Employment, Interest, and Money in 1936, Keynesian economics has been highly influential among academics and policymakers. Keynes has certainly had his detractors, though, with the most influential being Milton Friedman, Robert Lucas, and Edward C. Prescott. Monetarist thought, the desire for stronger theoretical foundations in macroeconomics, and real business cycle theory have at times been at odds with Keynesian economics. However, Keynesianism has remained a strong force, in part because its practitioners periodically adapt by absorbing the views of its detractors into the latest “synthesis.” John Hicks’s IS-LM interpretation of Keynes (Hicks 1937) and the popularization of this approach, particularly in Samuelson’s textbook (Samuelson 1997), gave birth to the “neoclassical synthesis.” Later, the menu cost models developed in the 1980s were a response to a drive for a more serious theory of sticky prices (Mankiw 1985, Caplin and Spulber 1987). More recently, New Keynesian economists have attempted to absorb real business cycle analysis and other ideas from post-1972 macroeconomics into a “new neoclassical synthesis” (Goodfriend and King 1997). The important New Keynesian ideas, as summarized, for example in Clarida, Gal´, and Gertler (1999) and Woodford (2003), are the following: ı 1. The key friction that gives rise to short-run nonneutralities of money and the primary concern of monetary policy is sticky prices. Because some prices are not fully flexible, inflation or deflation induces relative price distortions and welfare losses. Williamson is affiliated with Washington University in St. Louis, the Federal Reserve Bank of Richmond, and the Federal Reserve Bank of St. Louis. He thanks Huberto Ennis, Robert Hetzel, John Weinberg, and Alex Wolman for helpful comments and suggestions. Any opinions expressed in this article are the author’s and do not necessarily reflect those of the Federal Reserve Bank of Richmond, the Federal Reserve Bank of St. Louis, or the Federal Reserve System. E-mail: swilliam@artsci.wustl.edu. 198 Federal Reserve Bank of Richmond Economic Quarterly 2. Modern monetary economics is not part of the New Keynesian synthesis. New Keynesians typically regard the frictions that we encounter in deep (e.g., Lagos and Wright 2005) and not-so-deep (e.g., Lucas and Stokey 1987) monetary economics as being second-order importance. These frictions are absence-of-double-coincidence problems and information frictions that give rise to a fundamental role for monetary exchange, and typically lead to intertemporal distortions that can be corrected by monetary policy (for example, a ubiquitous result in monetary economics is Friedman’s zero-nominal-interest-rate rule for correcting intertemporal monetary distortions). The Friedman rule is certainly not ubiquitous in New Keynesian economics. 3. The central bank is viewed as being able to set a short-term nominal interest rate, and the monetary policy problem is presented as the choice over alternative rules for how this nominal interest rate should be set in response to endogenous and exogenous variables. 4. There is a short-run Phillips curve tradeoff. A monetary policy that produces an increase in unanticipated inflation will tend to increase real aggregate output. The goal of this paper is to construct a simple sticky-price New Keynesian model and then use it to understand and evaluate the ideas above. In this model there are some important departures from the typical New Keynesian models studied by Clarida, Gal´, and Gertler; Woodford; and others. However, these ı departures will highlight where the central ideas and results in New Keynesian analysis are coming from. For monetary economists, key aspects of New Keynesian economics can be puzzling. For example in Woodford (2003), the apparently preferred framework for analysis is a “cashless model” in which no outside money is held in equilibrium. Prices are denominated in terms of some object called money, and these prices are assumed to be sticky. The interest rate on a nominal bond can be determined in the cashless model, and the central bank is assumed capable of setting this nominal interest rate. Then, the monetary policy problem is formulated as the choice over rules for setting this nominal interest rate. This approach can be contrasted with the common practice in monetary economics, where we start with a framework in which money overcomes some friction, serves as a medium of exchange, and is held in equilibrium in spite of being dominated in rate of return by other assets. Then, studying the effects of monetary policy amounts to examining the consequences of changing the stock of outside money through various means: open market operations, central bank lending, or outright “helicopter drops.” It is usually possible to consider monetary policy rules that dictate the contingent behavior of a nominal interest rate, but in most monetary models we can see what underlying actions the central bank must take concerning monetary quantities to support such a S. D. Williamson: New Keynesian Economics 199 policy. What is going on here? Is the New Keynesian approach inconsistent with the principles of monetary economics? Is it misleading? The first task in this article is to construct a cashless model with sticky prices. This model departs from the usual New Keynesian construct in that there are competitive markets rather than Dixit-Stiglitz (1977) monopolistic competition. This departure helps to make the model simple, and yields information on the importance of the noncompetitive behavior of firms for New Keynesian economics. In general, given the emphasis on the sticky-price friction in New Keynesian economics, we would hope that it is something inherent in the functioning of a sticky-price economy, rather than simply strategic behavior, that is at the heart of the New Keynesian mechanism. Our cashless model is consistent with most of the predictions of standard Keynesian models, new and old. The sticky-price friction leads to a relative price distortion in that with inflation (deflation), too large (too small) a quantity of sticky-price goods is produced and consumed relative to flexible-price goods. Optimally, the inflation rate is zero, which eliminates the relative price distortion. One aspect in which this model differs from standard New Keynesian models is that it does not exhibit Phillips curve correlations. If the substitution effect dominates in the labor supply response to a wage increase (which we consider the standard case), then output is decreasing (increasing) in the inflation rate when the inflation rate is positive (negative). This is because the distortion caused by sticky prices rises with the deviation from a constant price level, and the representative consumer supplies less labor in response to a larger sticky-price distortion. Thus, under these circumstances output is maximized when the inflation rate is zero and there is no output/inflation tradeoff. In the case where the income effect dominates in the labor supply response to a change in wages, output increases (decreases) for positive (negative) inflation rates. Here there is a Phillips curve tradeoff if the inflation rate is positive, but a zero inflation rate is optimal. In most New Keynesian models, Phillips curve correlations are generated because of the strategic forward-looking behavior of price-setting firms. A firm, given the opportunity to set the price in units of money for its product, knows it will not have this opportunity again until some time in the future. Roughly, what matters to the firm is its expectation of the path for the price level during the period of time until its next price-setting opportunity. If the inflation rate is unusually high in the future, then the firm’s relative price will be unexpectedly low, and then, by assumption, it will be satisfying higherthan-expected demand for its product. Given that all firms behave in the same way, unanticipated inflation will tend to be associated with high real output. One message from our model is that Phillips curve behavior can disappear in the absence of strategic behavior by firms, even with sticky prices. We live in a world where outside money is held by consumers, firms, and financial institutions in the form of currency and reserve balances, and this 200 Federal Reserve Bank of Richmond Economic Quarterly outside money is supplied by the central bank and used in various transactions at the retail level and among financial institutions. In using a cashless model to analyze monetary policy, we should feel confident that we are not being led astray by a quest for simplicity. To evaluate what we might lose in focusing on a cashless model, we develop a monetary model that is a straightforward cashin-advance extension of the cashless model. Then, in the spirit of Woodford (2003), we explore how the behavior of this model compares to that of the cashless model and study the “cashless limit” of the more elaborate model. As it turns out, it requires very special assumptions for the limiting economy to behave in the same way as the monetary economy, and, in any case, quantity theory principles hold in equilibrium in the monetary economy. It is useful to know how monetary quantities should be manipulated to produce particular time paths of nominal interest rates, prices, and quantities in the economy, as the key instruments that a central bank has available to it are the quantities on its balance sheet. Thus, it would seem preferable to analyze monetary economies rather than cashless economies. This article is organized as follows. In Section 1 we construct the basic cashless model, work through examples, and uncover the general properties of this model. Section 2 contains a monetary model, extending the cashless model as a cash-in-advance construct with money and credit. Section 3 is a detailed discussion of the importance of the results, and Section 4 is a conclusion. 1. CASHLESS MODEL The goal of this section of the paper is to construct a simple sticky-price model that will capture the key aspects of New Keynesian economics, while also taking a somewhat different approach to price determination, in order to simplify and illuminate the important principles at work. The model we construct shares features with typical New Keynesian “cashless” models (see Woodford 2003), which are the following: 1. Money is not useful for overcoming frictions, it does not enter a cashin-advance constraint or a utility function, nor does it economize on transactions costs. 2. Money is a numeraire in which all prices are denominated. 3. The money prices of goods are sticky in that, during any period, some goods prices are predetermined and do not respond to current aggregate shocks. This model captures the essential friction that New Keynesians argue should be the focus of monetary policy—a sticky-price friction. New Keynesians argue that the other frictions that we typically encounter in monetary models—absence-of-double-coincidence problems and intertemporal S. D. Williamson: New Keynesian Economics 201 price distortions, for example—are of second order for the problem at hand. Further, New Keynesians feel that it is important to model the monetary policy problem in terms of the choice of nominal interest rate rules and that this framework is a very convenient vehicle in that respect. The model we will work with here has an infinite-lived representative consumer who maximizes ∞ β t u(ct1 ) + u(ct2 ) − v(nt ) , E0 (1) t=0 denotes consumption of the i th good, i = 1, 2, and nt is labor where supply. Assume that u(·) is strictly increasing, strictly concave, twice continuously differentiable, and has the property u (0) = ∞. As well, v(·) is strictly increasing, strictly convex, and twice differentiable with v (0) = 0 and v (h) = ∞ for some h > 0. We have assumed a separable period utility function for convenience, and a two-good model is sufficient to exposit the ideas of interest. Goods are perishable. There are linear technologies for producing goods from labor input, i.e., cti yti = γ t ni , t (2) where yti is output of good i, γ t is aggregate productivity, and ni is the quantity t of labor input applied to production of good i. Assume that γ t follows an exogenous stochastic process. This model is very simple. There is no investment or capital, and an optimal allocation is a sequence {n1 , n2 , ct1 , ct2 }∞ satisfying n1 = n2 = nt ˜ t ˜ t ˜ ˜ t=0 ˜t ˜t ˜ and ct1 = ct2 = ct , with ct = γ t nt , where nt solves ˜ ˜ ˜ ˜ ˜ ˜ v (2nt ) ˜ = γ t. u (γ t nt ) ˜ (3) Therefore, at the optimum, consumption of the two goods should be equal in each period with the same quantity of labor allocated to production of each good, given symmetry. As well, from (3) the ratio of the marginal disutility of labor to the marginal utility of consumption should be equal to aggregate productivity for each good in each period. There is another object, which we will call money, that plays only the role of numeraire. We will assume a form of price stickiness reminiscent of what obtains in the staggered wage-setting models of Fischer (1977) and Taylor (1979). That is, assume that prices are sticky, in the sense that, if the price of a good is flexible in period t, then it remains fixed at its period t value through period t + 1, and is subsequently flexible in period t + 2, etc. In any given period, one good is flexible and the other is sticky. Now, since in (1) and (2) the two goods are treated symmetrically in preferences and technology, we can let good 1 be the flexible-price good in each period. Then, let Pt denote the price in units of money of the flexible-price good in period t, and then Pt−1 202 Federal Reserve Bank of Richmond Economic Quarterly is the price of the sticky-price good. As well, let Wt denote the nominal wage rate. The Keynesian modeler must always deal with the problem of how firms and consumers behave in the face of price and/or wage stickiness, as well as how quantities are determined. In typical textbook sticky-wage approaches, the nominal wage is exogenous, and the quantity of labor traded is determined by what is optimal for the representative firm. Standard textbook sticky-price models have a single good sold at an exogenous nominal price, and the quantity of output is demand-determined.1 In New Keynesian economics, the emphasis is on price stickiness (as opposed to wage stickiness), the distribution of prices across goods, and relative price distortions. Clearly, if there is a homogeneous good, constant returns to scale, competitive equilibrium, and perfect information, we cannot have anything other than a degenerate distribution of prices in equilibrium, where all firms producing positive output charge the same price. The New Keynesian approach at least requires heterogeneous goods, and the standard model in the literature is currently one with monopolistically competitive firms. For example, a typical approach is to assume monopolistic competition with Calvo (1983) price setting (or “time-dependent pricing”). In such a model, each firm randomly obtains the opportunity to change its nominal price each period, so that with a continuum of firms, some constant fraction of firms changes their prices optimally, while the remaining fraction is constrained to setting prices at the previous period’s values. Alternatively, it could be assumed that each of the monopolistically competitive firms must set their prices one period in advance, before observing aggregate (or possibly idiosyncratic) shocks. Woodford (2003), for example, takes both approaches. Neither Calvo pricing nor price setting one period in advance in monopolistic competition models is without problems. In a Calvo pricing model, each monopolistically competitive firm is constrained to producing one product. There may be states of the world where some firms will earn negative profits, but they are somehow required to produce strictly positive output anyway, and firms cannot reallocate productive factors to the production of flexible-price goods that will yield higher profits. With prices set one period in advance, firms are constrained to produce, even in the face of negative ex post profits. Here, we have assumed differentiated products, but we will maintain competitive pricing. This model is certainly not typical, so it requires some explanation. First, the consumer’s budget constraint will imply that all wage income will be spent on the two consumption goods, so Pt γ t n1 + Pt−1 γ t n2 − Wt (n1 + n2 ) = 0. t t t t (4) 1 See Williamson (2008) for examples of standard Keynesian sticky-wage and sticky-price models. S. D. Williamson: New Keynesian Economics 203 However, this would then seem to imply that, unless Pt = Pt−1 , the production and sale of one good will earn strictly positive profits and production and sale of the other good will yield strictly negative profits. Thus, it seems that it cannot be profit-maximizing for both goods to be produced in equilibrium. However, suppose that we take for granted, as is typical in much of the New Keynesian literature, that there will be some firms that are constrained to producing the sticky-price good, and that the firms who produce this good will satisfy whatever demand arises at the price Pt−1 . How then should we determine which firms produce which good? For this purpose, assume that there is a lottery, which works as follows. Before a firm produces, it enters a lottery where the outcome of the lottery determines whether the firm produces the flexible-price good or the sticky-price good. If it is determined through the lottery that a particular firm produces the flexible-price good, then that firm receives a subsidy of st1 in nominal terms, per unit of output. If it is determined that a particular firm produces the fixed-price good, that firm receives st2 per unit of output. The agent that offers the lottery will set st1 and st2 so that any firm is indifferent between producing the fixed-price and flexible-price goods, i.e., Pt − st1 = Pt−1 − st2 . (5) Further, the agent offering the lottery breaks even, so that st1 n1 + st2 n2 = 0, t t (6) so solving for the subsidy rates, we obtain st1 = n2 (Pt − Pt−1 ) t and (n1 + n2 ) t t st2 = n1 (Pt−1 − Pt ) t . (n1 + n2 ) t t Given these subsidy rates, the agent offering the lottery breaks even and each firm is willing to enter the lottery as profits per unit produced are zero in equilibrium, whether the firm ultimately produces the flexible-price or stickyprice good. Though this cross-subsidization setup may seem unrealistic, it does not seem less palatable than what occurs in typical Keynesian stickyprice models. In fact, the randomness in determining which firms produce which good is reminiscent of the randomness in Calvo (1983) pricing, but our approach is much more tractable. Now, let π t denote the relative price of flexible-price and sticky-price goods, which is also the gross rate of increase in the price of the flexibleprice good. Under some special circumstances, π t will also be the measured inflation rate, but in general that is not the case. However, π t is the relative price that captures the extent of the effects of the sticky-price friction. Letting 204 Federal Reserve Bank of Richmond Economic Quarterly wt denote the relative price of labor and flexible-price goods, we can rewrite equation (4) as γ t n2 t − wt (n1 + n2 ) = 0. (7) t t πt Optimization by the consumer is summarized by the following two marginal conditions: γ t n1 + t u (ct1 ) − π t u (ct2 ) = 0 and (8) wt u (ct1 ) − v (n1 + n2 ) = 0. t t In equilibrium all output must be consumed, so cti = γ t ni t (9) (10) for i = 1, 2. Further, letting qt denote the price at time t of a claim to one unit of money in period t + 1, we can determine qt in the usual fashion by qt = βEt 1 u (ct+1 ) . π t+1 u (ct1 ) (11) An equilibrium is defined to be a stochastic process {π t , wt , qt , n1 , n2 , ct1 , t t ct2 }∞ given an exogenous stochastic process for γ t , that satisfies (7)–(11) and t=0 qt ≤ 1, so that the nominal interest rate is nonnegative in each state of the world. There is clearly indeterminacy here, as there appears to be nothing that will pin down prices. In equilibrium there is an object called money that is in zero supply, and which, for some unspecified reason, serves as a unit of account in which prices are denominated. The path that π t follows in equilibrium clearly matters, because prices are sticky, but the possibilities for equilibrium paths for π t are limitless. Examples One equilibrium is π t = 1 for all t, which from (7)–(11) gives the optimal allocation with n1 = n2 = nt and ct1 = ct2 = ct , where ct = γ t nt , and where ˜t ˜t ˜ ˜ ˜ ˜ ˜ ˜ nt solves (3). Solving for qt from equation (11), we get ˜ ˜ u (γ t+1 nt+1 ) qt = βEt , (12) u (γ t nt ) ˜ and so long as the variability in productivity is not too large, we will have qt ≤ 1 for all t, so that the nominal interest rate is always nonnegative, and this is indeed an equilibrium. Alternatively (for example, see Woodford 2003), we could argue that the 1 central bank can set the nominal interest rate it = qt − 1. In this instance, if the central bank sets the nominal interest rate from equation (12) according to it = βEt ˜ u (γ t+1 nt+1 ) u (γ t nt ) ˜ −1 − 1, (13) S. D. Williamson: New Keynesian Economics 205 an equilibrium with π t = 1 for all t can be achieved, which is optimal. There is nothing in the model that tells us why the central bank can control it , and why it cannot control π t , for example. In New Keynesian models, the justification for treating the market nominal interest rate as a direct instrument of the central bank comes from outside the model, along the lines of “this is what most central banks do.” In any case, an optimal policy implies that, since π t = 1, the price level is constant and the inflation rate is zero. This optimal policy could then be characterized as an inflation rate peg, or as a policy that requires, from (13), that the nominal interest rate target for the central bank fluctuate with the aggregate technology shock. More simply, from (13), the nominal interest rate at the optimum should equal the “Wicksellian natural rate of interest” (see Woodford 2003). There are, of course, many suboptimal equilibria in this model. For example, consider the special case where u(c) = ln c and v(n) = δn, for δ > 0. Though v(·) does not satisfy some of our initial restrictions, this example proves particularly convenient. We will first construct an equilibrium with a constant inflation rate, which has the property that π t = α, where α is a positive constant (recall that π t is not the gross inflation rate, but if π t is constant then the inflation rate is constant). From (7)–(11), the equilibrium solution we obtain is 2γ t , (14) wt = 1+α qt = βγ t Et α 1 γ t+1 , (15) n1 = t 1 , δ(1 + α) (16) n2 = t α , δ(1 + α) (17) γt , and δ(1 + α) (18) αγ t . δ(1 + α) (19) ct1 = ct1 = In this equilibrium, the rate of inflation is α − 1. Note, from (14)–(19), that higher inflation causes a reallocation of consumption and labor supply from flexible-price to sticky-price goods. From equation (15) for qt ≤ 1, it is sufficient that α not be too small and that γ t not be too variable. This equilibrium is of particular interest because it involves an inflation rate peg. Of course, as we showed previously, α = 1 is optimal. 206 Federal Reserve Bank of Richmond Economic Quarterly Alternatively, consider the same example as above, with log utility from consumption and linear disutility to supplying labor, but now suppose that π t is governed by a rule that responds to productivity shocks, for example, γ π t = t−1 . γt Then, from (7)–(11), we obtain the equilibrium solution wt = 2γ 2 t , γ t−1 + γ t (20) γ t (γ t + γ t+1 ) , γ t+1 (γ t−1 + γ t ) (21) n1 = t 2γ t , δ(γ t−1 + γ t ) (22) n2 = t 2γ t−1 , δ(γ t−1 + γ t ) (23) 2γ 2 t , and δ(γ t−1 + γ t ) (24) 2γ t γ t−1 . δ(γ t−1 + γ t ) (25) qt = βEt ct1 = ct1 = From (21) it is sufficient for the existence of equilibrium that γ t not be too variable. Note in the solution, (20)–(25), that equilibrium quantities and prices all exhibit persistence because of the contingent path that prices follow. Complicated dynamics can be induced through the nominal interest rate rule, of which (25) is an example in this case. Indeed, it is possible (see Woodford 2003), given some nominal interest rate rules, to obtain equilibrium solutions where current endogenous variables depend on anticipated future aggregate shocks. Typical New Keynesian models also obtain equilibrium solutions with such properties through the forward-looking price-setting behavior of monopolistically competitive producers. This latter mechanism is not present in our model. General Properties of the Model To further analyze our model, we find it useful to consider how we would solve for an equilibrium in this model in the absence of sticky prices. The model is purely static, so we can solve period-by-period. An equilibrium for period t consists of relative prices π t and wt , and quantities n1 , n2 , ct1 , and t t S. D. Williamson: New Keynesian Economics 207 ct2 that solve the marginal conditions (8) and (9), the equilibrium conditions (10), and two zero-profit conditions: γ t n1 − wt n1 = 0 and t t (26) γ t n2 t − wt n2 = 0. (27) t πt Of course, the solution we get is the optimum, with wt = γ t , π t = 1, n1 = n2 = nt , and ct1 = ct2 = ct = γ t nt , with nt determined as the solution ˜ ˜ ˜ ˜ t t to (3). Now, how should we think about solving the system (7)–(11) under sticky prices? It seems most useful to think of this system in the traditional Keynesian sense, as a model where one price, π t , is fixed exogenously. Given any exogenous π t = 1, it cannot be the case that all agents optimize and all markets clear in equilibrium. The solution we have chosen here, which is in line with standard New Keynesian economics, is to allow for the fact that (26) and (27) do not both hold. Instead, we allow for cross-subsidization with zero net subsidies across production units and zero profits in equilibrium net of subsidies for production of each good, which gives us equation (7). Given this interpretation of the model, how should we interpret qt , as determined by equation (11)? Since π t is simply the relative price of good 2 in terms of good 1 in period t, qt is the price, in units of good 1 in period t, that a consumer would pay for delivery of one unit of good 2 in period t + 1. Why, then, should we require an arbitrage condition that qt ≤ 1, or why should 1 u (ct+1 ) ≤1 (28) π t+1 u (ct1 ) hold? Such a condition requires the existence of a monetary object. Then we can interpret (11) as determining the price in units of money in period t of a claim to one unit of money delivered in period t + 1, and inequality (28) is required so that zero money balances are held in equilibrium. Thus, in equilibrium the model is purely atemporal. There is no intertemporal trade, nevertheless there exist equilibrium prices for money in terms of goods and for the nominal bond in each period. Thus far, this may be somewhat puzzling for most monetary economists, who are accustomed to thinking about situations in which money is not held in equilibrium as ones in which the value of money in units of goods is zero in each period. This does not happen here, but no fundamental principles of economic analysis appear to have been violated. The key question, then, is what we can learn from this model. First, we will get some idea of the operating characteristics of the model through linear approximation. If we treat π t as exogenous, following our interpretation above, then the exogenous variables are γ t and π t . Substitute using equation (10) in (7)–(9), and then linearize around the solution we get with γ t = γ and ¯ βEt 208 Federal Reserve Bank of Richmond Economic Quarterly ¯ π t = 1, where γ is a positive constant. The equilibrium solution we get in this benchmark case is wt = γ , n1 = n2 = n, and ct1 = ct2 = γ n, where n ¯ t ¯ ¯¯ ¯ t solves ¯ γ u (γ n) − v (2n) = 0. ¯ ¯¯ The solution to the linearized model is (leaving out the solution for the wage, wt ) : ¯ n1 = n + t u (−2v + γ 2 u ) u + γ nu ¯ ¯¯ (π t − 1) − 2 (γ t − γ ), ¯ 2 2γ u (γ u − 2v ) ¯ ¯ γ u − 2v ¯ (29) n2 = n + ¯ t u (2v − γ 2 u ) u + γ nu ¯ ¯¯ (π t − 1) − 2 (γ t − γ ), ¯ 2 2γ u (γ u − 2v ) ¯ ¯ γ u − 2v ¯ (30) ¯¯ ct1 = γ n + u (−2v + γ 2 u ) 2nv + γ u ¯ ¯ ¯ (π t − 1) − 2 (γ t − γ ), and (31) ¯ 2 2u (γ u − 2v ) ¯ γ u − 2v ¯ ¯¯ ct2 = γ n + ¯ ¯ u (2v − γ 2 u ) 2nv + γ u ¯ (π t − 1) − 2 (γ t − γ ), ¯ 2 ¯ 2u (γ u − 2v ) γ u − 2v ¯ (32) and aggregate labor supply and output are given, respectively, by n1 + n2 = 2n − ¯ t t ¯¯ 2(u + γ nu ) (γ t − γ ) and ¯ 2 γ u − 2v ¯ (33) ¯ 2(2nv + γ u ) ¯ (γ t − γ ). ¯ 2 γ u − 2v ¯ (34) ¯¯ ct1 + ct2 = 2γ n − Therefore, from (29)–(34), in the neighborhood of an equilibrium with constant prices, an increase in π t , which corresponds to an increase in the inflation rate, results in a decrease in the production and consumption of the flexibleprice good, an increase in production and consumption of the fixed-price good, and no effect on aggregate labor supply and output. Thus, this model does not produce a Phillips curve correlation, at least locally, and increases in the inflation rate serve only to misallocate production and consumption across goods. As well, from (29)–(34), a positive shock to aggregate productivity has the same effect on production and consumption of both goods. If u + γ nu > 0, ¯¯ (35) then the substitution effect of the productivity increase offsets the income effect on labor supply so that aggregate labor supply increases. In what follows, we assume that the substitution effect dominates, i.e., (35) holds. From (34), aggregate output increases with an increase in productivity, regardless of whether the substitution effect dominates. S. D. Williamson: New Keynesian Economics 209 The absence of a Phillips curve effect here might seem puzzling, as a positive relationship between inflation and output often appears as a cornerstone of new and old Keynesian economics. Thus, we should explore this further to see how π t affects output outside of the neighborhood of our baseline equilibrium. Consider an example that will yield closed-form solutions, in particular 1−α −1 u(c) = c 1−α and v (n) = δn, with α > 0 and δ > 0. Solving (7)–(10), we obtain 1 ⎛ ⎞α 1 α −1 1 1 + πt ⎠ 1 −1 n1 = γ tα δ − α ⎝ , (36) t 1 1 + π tα ⎛ 1 α −1 n2 = γ t t 1 α 2 α −1 π + πt 1 δ− α ⎝ t 1 1 + π tα ⎛ 1 α −1 1 ⎞α ⎠ , (37) 1 ⎞α 1 + πt ⎠ 1 , ct1 = γ t δ − α ⎝ 1 1 + π tα 1 α ⎛ 1 α 2 α −1 π + πt 1 ct2 = γ t δ − α ⎝ t 1 1 + π tα 1 α ⎛ wt = γ t ⎝ 1 α −1 1 + πt 1 α (38) 1 ⎞α ⎠ , and (39) ⎞ ⎠. (40) 1 + πt Then, aggregate labor supply and aggregate output are given by 1 α −1 + n2 t = γt ct1 n1 t + ct2 δ 1 α 1 −α = γt δ 1 −α 1 α −1 1 + πt 1 α −1 1 + πt 1 α 1 α 1 1− α 1 α 1 + πt and (41) . (42) 1 1− α 1 α 1 + πt Now, in the solution (36)–(42), the condition (35) is equivalent to α < 1. Given this, labor supply in both sectors is increasing in productivity, as, of course, is consumption of each good and total output. Our primary interest in this example is what it tells us about the relationship between π t and aggregate output. Note from equation (42) that this relationship is determined by the properties of the function 1 G(π ) = 1 + π α −1 1 α 1 1+πα 1 1− α . 210 Federal Reserve Bank of Richmond Economic Quarterly Differentiating, we obtain 1 1 G (π ) = π α −2 1 + π α −1 1 α −1 1 1+πα 1 −α 1 α 1 − 1 (1 − π ). α Therefore, for the case α < 1, we have G (1) = 0, G (π ) > 0 for π < 1, and G (π ) < 0 for π > 1. Thus, output is maximized for π = 1 and the Phillips curve has a negative slope when inflation is positive and a positive slope when inflation is negative. The key to the Phillips curve relationship is how labor supply responds to the distortion created by inflation or deflation due to the the sticky-price friction. When the substitution effect on labor supply of an increase in productivity dominates the income effect, an increase or decrease in the inflation rate from zero implies that the marginal payoff from supplying labor falls, and the consumer therefore reduces labor supply. The interesting aspect of these results is that they point to a nonrobust link between price stickiness and Phillips curve correlations. In spite of the fact that firms do not set prices strategically in a forward-looking manner, intuition might tell us that there should still be a Phillips curve correlation. That is, with higher inflation, it might seem that the additional quantity of output produced by sticky-price firms should be greater than the reduction in output by flexible-price firms, and aggregate output should increase. However, our analysis shows that this need not be the case and that the key to understanding the mechanism at work is labor supply behavior. In our model, since not all prices are sticky, the key effect of inflation on output comes from the relative price distortion, and labor supply may increase or decrease in response to higher inflation, with a decrease occurring when the elasticity of substitution of labor supply is sufficiently high. As we commented earlier, the assumptions on price stickiness and firm behavior in our model seem no less palatable than what is typically assumed. Thus, the nonrobustness of the Phillips curve we find here deserves attention. 2. A MONETARY MODEL AND THE “CASHLESS LIMIT” In New Keynesian economics (e.g., Woodford 2003), baseline “cashless” models are taken seriously as frameworks for monetary policy analysis. As we have seen, the cashless model focuses attention on the sticky-price friction as the key source of short-run nonneutralities of money. New Keynesian arguments for using a cashless model appear to be as follows: (i) the standard intertemporal monetary distortions—for example, labor supply distortions and the tendency for real cash balances to be suboptimally low when the nominal interest rate is greater than zero—are quantitatively unimportant; (ii) in models where there is some motive for holding money, if we take the limit as the motive for holding money goes to zero, then this limiting economy has essentially the same properties as does the cashless economy. The purpose S. D. Williamson: New Keynesian Economics 211 of this section is to evaluate these arguments in the context of a particular monetary model. For our purposes, a convenient expository vehicle is a cash-in-advance model of money and credit where we can parameterize the friction that makes money useful in transactions. There are other types of approaches we could take here; for example, we could use a monetary search and matching framework along the lines of Lagos and Wright (2005),2 but the model we use here allows us to append monetary exchange to the cashless model in Section 2 with the least fuss. Our framework is much like that in Alvarez, Lucas, and Weber (2001), absent limited participation frictions, but including the labor supply decision and the sticky-price friction we have been maintaining throughout. The structure of preferences, technology, and price determination is identical to that which we assumed in the cashless model. Here, suppose that the representative consumer trades on asset markets at the beginning of each period and then takes the remaining money to the goods market, where goods can be purchased with money and credit. The consumer faces the cash-in-advance constraint Pt ct1 + Pt−1 ct2 + qt bt+1 ≤ θ Wt (n1 + n2 ) + mt + τ t + st lt + bt , t t (43) where bt denotes one-period nominal bonds purchased by the consumer in period t − 1, each of which pays off one unit of money in period t; mt denotes nominal money balances carried over from the previous period; τ t is a nominal lump-sum transfer from the government; and st lt is a within-period money loan from the central bank, where lt is the nominal amount that must be returned to the central bank at the end of the period. Also, θ denotes the fraction of current-period wage income that can be accessed in the form of within-period credit when the consumer trades in the goods market. Note that s1t − 1 is the 1 within-period nominal interest rate on central bank loans, and, as above, qt −1 is the one-period nominal interest rate. Here 0 ≤ θ ≤ 1, and θ is the critical parameter that captures the usefulness of money in transactions. With θ = 0, this is a pure monetary economy, and with θ = 1, money is irrelevant. The consumer must also satisfy his or her budget constraint, given by Pt ct1 +Pt−1 ct2 +mt+1 +qt bt+1 = Wt (n1 +n2 )+mt +τ t +(st − 1) lt +bt . (44) t t Let Mt denote the supply of money and Lt denote the supply of central bank loans. Then the asset market equilibrium conditions are mt = Mt ; bt = 0; lt = Lt , (45) or, money demand equals money supply, the demand for bonds equals the supply, and the demand for central bank loans equals the supply. 2 See, for example, Aruoba and Schorfheide (2008), where a monetary search model with nominal rigidities is constructed for use in quantitative work. 212 Federal Reserve Bank of Richmond Economic Quarterly Eliminating Intertemporal Distortions with Central Bank Lending In general, an equilibrium in this model is difficult to characterize, as price stickiness complicates the dynamics. In part, our goal will be to determine the features of a “cashless limit” in this economy, along the lines of Woodford (2003). To that end, given the New Keynesian view that intertemporal distortions are unimportant, suppose that the regime of central bank lending is set up so that those distortions are eliminated. That is, suppose that the central bank supplies no money, except through central bank loans made at the beginning of the period at a zero nominal interest rate. Let st = 1, and suppose that the central bank accommodates whatever demand for central bank loans arises at a zero nominal interest rate. We then have Mt = τ t = 0 for all t and Lt = (1−θ )Wt (n1 +n2 ). Then, given (43)–(45), t t optimization, and goods market equilibrium, we can define an equilibrium in terms of relative prices and quantities, just as in the cashless economy. This monetary regime is then one where all economic agents have access to a daylight overdraft facility, much like the Federal Reserve System uses each day to accommodate payments among financial institutions. Given a zero nominal interest rate on daylight overdrafts, money will not be held between periods, which we can interpret as a system in which holdings of outside money are zero overnight (interpreting a period as a day). This setup is extreme, as it allows universal access to central bank lending facilities and does not admit anything resembling currency-holding in equilibrium. In equilibrium, (7)–(11) must be satisfied, just as in the cashless economy. The key difference in this monetary economy will be in the determination of π t . Supposing for convenience that (43) always holds with equality in each period, then, given (7), we can determine π t by πt = Lt wt−1 (n1 + n2 ) t−1 t−1 , 1 2 Lt−1 wt (nt + nt ) (46) for t = 1, 2, 3, ..., with π 0 given. An equilibrium is a stochastic process {n1 , n2 , ct1 , ct2 , wt , π t , qt }∞ , with π 0 given, solving (7)–(11) and (46). The t t t=0 solution must satisfy qt ≤ 1 for all t, which assures that an equilibrium exists where (43) holds with equality. In general, it is not straightforward to characterize a solution, but it is clear from (46) that the solution is consistent with the quantity theory of money. That is, Lt is the nominal quantity of money γ t n2 available to spend in period t, and wt (n1 + n2 ) = γ t n1 + π t t is total GDP. t t t Therefore, (46) states that the rate of increase in the price of the flexible-price good is roughly equal to the rate of money growth minus the rate of growth in real GDP. Note that the parameter θ does not appear anywhere in (7)–(11) and (46). That is, we can treat the equilibrium solution as the cashless limit, as we will obtain the same solution for any θ > 0. Note here that the cashless limit of this monetary economy is not the cashless economy, and the quantity S. D. Williamson: New Keynesian Economics 213 of money is important for the solution, along quantity theory lines, in spite of the fact that no money is held between periods. Thus, we have followed the logic of Woodford (2003) here, but we do not get Woodford’s results. What is an optimal monetary policy here, given that the within-period nominal interest rate is zero? The key choice for the central bank is Lt , the nominal loan quantity in each period. If Lt can be set so that π t = 1 for all t, then clearly this would be an optimal policy, since from (7)–(11) we will obtain ni = nt for all t and i = 1, 2. From equation (46), this requires that ˜ t ˜ γ t nt Lt = . Lt−1 γ t−1 nt−1 ˜ (47) This optimal policy then implies a nominal bond price u (γ t+1 nt+1 ) ˜ , (48) u (γ t nt ) ˜ and for this optimal policy to support an equilibrium, we require that qt ≤ 1 for all t, which is satisfied provided the variability in γ t is sufficiently small. Thus, we can define the optimal monetary policy as a rule for monetary growth, which accommodates GDP growth according to (47) or as a nominal interest rate rule governed by (48), i.e., the money growth rule and nominal interest rate rule are flip sides of the same monetary policy. On the one hand, the money growth rule in (47) states that the money supply should always grow at the same rate as optimal GDP. On the other hand, the nominal interest rate rule states that the nominal interest rate should move in response to productivity shocks in such a way that it is equal to the optimal real interest rate. Part of the New Keynesian justification for use of a cashless model (see Woodford 2003) is that if intertemporal frictions are insignificant, even if there is a monetary friction in the model, then the prescription for the optimal nominal interest rate rule is the same as in the cashless economy. This is certainly true here, as the nominal interest rate rule implicit in (48) is the same as (12). For our purposes, though, the monetary economy is more informative about policy, as it says something about the monetary policy regime that is necessary to get this result, and gives us a prescription for how the central bank should manipulate the quantities that it has under its control. qt = βEt 3. DISCUSSION In this section we will discuss our results, organized in terms of monetary policy instruments, Phillips curves, and monetary frictions. Monetary Policy Instruments What can a central bank control? Ultimately, if we ignore the central bank’s regulatory powers, a central bank can control two sets of things. First, it 214 Federal Reserve Bank of Richmond Economic Quarterly can determine the quantities on its balance sheet, including lending to private sector economic agents. Second, it can determine the interest rates on deposits (reserves) held with the central bank and the interest rates on the loans it makes, in particular the interest rates on daylight overdrafts associated with payments made using outside money and the interest rates on overnight central bank lending or lending at longer maturities. The key power a central bank holds is its monopoly on the issue of fiat money. Essentially, a central bank is much like any other bank in that its liabilities serve as a means of payment, and it performs a type of liquidity transformation in intermediating assets that are difficult to use as means of payment. Central bankers, and many economists, hold the view that the quantity of intermediation that the central bank carries out, reflected in the quantity of fiat money outstanding, has consequences for real economic activity in the short run and for prices. Central banks cannot set market interest rates, though they might like to. New Keynesians typically model central bank behavior as the determination of a market nominal interest rate as a function of endogenous and exogenous variables. There are good reasons to think that a central bank operating procedure consisting of periodic revision of an overnight nominal interest rate target or inflation rate targeting is preferable to the money growth targeting that Friedman (1968) had in mind. That is, the predominant shocks that are of concern to the central bank in the very short run, say between Federal Open Market Committee meetings, are financial market shocks that cause fluctuations in the demand for outside money. Given that these shocks are difficult to observe, a sensible procedure may be to smooth the overnight nominal interest rate, which may serve to optimally accommodate financial market shocks. Though it may be possible in the short run for a central bank to use the instruments at its disposal to keep a market nominal interest rate within a tight prespecified corridor, it is inappropriate to use this as a justification for a mode of analysis that eliminates monetary considerations. A model that is used to analyze and evaluate monetary policy should tell us how the economy functions under one central bank operating procedure (e.g., monetary targeting) versus another (e.g., nominal interest rate targeting), how the instruments available to the central bank (i.e., monetary quantities) need to be manipulated to implement a particular policy rule, and how using alternative instruments (e.g., central bank lending versus open market operations) makes a difference. Phillips Curves Some type of Phillips curve relationship, i.e., a positive relationship between the “output gap,” on the one hand, and the rate of inflation or the unanticipated component of inflation, on the other hand, is typically found in New Keynesian macroeconomic models. The Phillips curve was an important example in 1970s policy debates of how policy could go wrong in treating an empirical S. D. Williamson: New Keynesian Economics 215 correlation as structural (Lucas 1972). In New Keynesian economics, the Phillips curve has made a comeback as a structural relationship and plays a central role in reduced-form New Keynesian models (e.g., Clarida, Gal´, and ı Gertler 1999). As we have shown here, in a sticky-price model that seems as reasonable as typical monopolistically competitive New Keynesian setups, there is no tradeoff between output and inflation in the standard case where substitution effects dominate in the response of labor supply to a wage increase. With a zero inflation rate, output is maximized. Even in the case where income effects are large, more output can be obtained if the inflation rate deviates from zero, but this is inefficient. Further, in this case, more output can be obtained not only with inflation, but with deflation. Monetary Frictions We know that what makes a modern central bank unique is the power granted to it as the monopoly issuer of outside money, which takes the form of deposits with the central bank and circulating currency. We also know that central banking is not a necessity. Indeed, there are examples of economies that grew and thrived without a central bank. The United States did not have a central bank until 1914, and the private currency systems in place in Scotland from 1716–1845 and in Canada before 1935 are generally regarded as successes. Before asking how a central bank should behave, we might want to ask what justifies its existence in the first place. From the viewpoint of a monetary economist, a theory of central banking should not only tell us what the role of the central bank is in a modern economy, but also why we should grant the central bank a monopoly in supplying key media of exchange. Such a central banking theory must necessarily come to grips with the principal frictions that make money useful as a medium of exchange and the frictions that may make private provision of some types of media of exchange inefficient. New Keynesians argue that we can do a better job of understanding how monetary policy works and how it should be conducted by ignoring these frictions. By using a very simple cash-in-advance construct, we have shown these arguments require some very special assumptions. For our cashless sticky-price economy to work in the same way as does a comparable monetary economy requires that: (i) a monetary regime be in place that corrects intertemporal inefficiencies; (ii) all economic agents be on the receiving end of the central bank’s actions; and (iii) currency holding be unimportant. While some countries, such as Canada and New Zealand, have moved to monetary systems without reserve requirements and with interest on reserves, thus correcting some distortions, most countries are far from the elimination of intertemporal monetary frictions. Thus, in practice it is likely that intertemporal 216 Federal Reserve Bank of Richmond Economic Quarterly distortions play an important role, and arguably the important role if we are considering the effects of long-run anticipated inflation. Also, the fact that not all economic agents are on the receiving end of monetary policy actions, which gives rise to distributional effects of monetary policy, is regarded as important in the segmented markets literature. Market segmentation (in both goods and financial markets) is perhaps of greater significance than stickyprice frictions in generating short-run nonneutralities of money (see Alvarez and Atkeson 1997; Alvarez, Atkeson, and Kehoe 2003; Williamson [ForthcomingA, ForthcomingB]). Finally, currency is still widely used in the world economy (Alvarez and Lippi 2007). In spite of technological improvements in transactions technologies, currency is a wonderfully simple transactions technology that permits exchange in the many circumstances where anonymous individuals need to transact with each other. 4. CONCLUSION Recent events involving turmoil in credit markets and heretofore unheard-of interventions by the Federal Reserve System make it abundantly clear that the monetary policy problem is far from solved. Further, for the key questions that need to be answered in the midst of this crisis, New Keynesian economics appears to be unhelpful. How is central bank lending different from open market operations in terms of the effects on financial markets and goods markets? To which institutions should a central bank be lending and under what conditions? What regulatory power should the Federal Reserve System exercise over the institutions to which it lends? Should the Fed’s direct intervention be limited to conventional banks, or should this intervention be extended to investment banks and government-sponsored financial institutions? Unfortunately, typical New Keynesian models ignore credit markets, monetary frictions, and banking and are, therefore, of little or no use in addressing these pressing questions. What hope do we have of developing a theory of money and central banking that can satisfy monetary economists and also be of practical use to central bankers? Monetary economics and banking theory have come a long way in the last 30 years or more, and perhaps the economics profession needs to be educated as to why modern monetary and banking theory is useful and can be applied to policy problems. We now understand that recordkeeping and the flow of information over space and time is critical to the role of currency as a medium of exchange (Kocherlakota 1998). We know that decentralized exchange with currency can lead to holdup problems that accentuate the welfare losses from inflation (Lagos and Wright 2005). We understand how banks act to insure private agents against liquidity risk (Diamond and Dybvig 1983) and to economize on monitoring costs (Diamond 1984, Williamson 1986). We know that financial market segmentation and goods S. D. Williamson: New Keynesian Economics 217 market segmentation are important for monetary policy (Alvarez and Atkeson 1997; Alvarez, Atkeson, and Kehoe 2003; Williamson [Forthcoming A; Forthcoming B]). Putting together elements of these ideas in a comprehensive theory of central banking is certainly within our grasp, and I very much look forward to future developments. REFERENCES Alvarez, Fernando, and Andrew Atkeson. 1997. “Money and Exchange Rates in the Grossman-Weiss-Rotemberg Model.” Journal of Monetary Economics 40 (December): 619–40. Alvarez, Fernando, Andrew Atkeson, and Patrick Kehoe. 2002. “Money, Interest Rates, and Exchange Rates with Endogenously Segmented Markets.” Journal of Political Economy 110 (February): 73–112. Alvarez, Fernando, and Francesco Lippi. 2007. “Financial Innovation and the Transactions Demand for Cash.” University of Chicago Working Paper. Alvarez, Fernando, Robert Lucas, and Warren Weber. 2001. “Interest Rates and Inflation.” American Economic Review Papers and Proceedings 91: 219–25. Aruoba, S.B., and Frank Schorfheide. 2008. “Insights from an Estimated Search-Based Monetary Model with Nominal Rigidities.” University of Maryland and University of Pennsylvania Working Paper. Calvo, Guillermo. 1983. “Staggered Prices in a Utility Maximizing Framework.” Journal of Monetary Economics 12: 383–98. Caplin, Andrew, and Daniel Spulber. 1987. “Menu Costs and the Neutrality of Money.” Quarterly Journal of Economics 102: 703–26. Clarida, Richard, Jordi Gal´, and Mark Gertler. 1999. “The Science of ı Monetary Policy: A New Keynesian Perspective.” Journal of Economic Literature 37 (December): 1661–1707. Diamond, Douglas. 1984. “Financial Intermediation and Delegated Monitoring.” Review of Economic Studies 51: 393–414. Diamond, Douglas, and Philip Dybvig. 1983. “Bank Runs, Deposit Insurance, and Liquidity.” Journal of Political Economy 91: 401–19. Dixit, Avinash, and Joseph Stiglitz. 1977. “Monopolistic Competition and Optimum Product Diversity.” American Economic Review 67: 297–308. Fischer, Stanley. 1977. “Long-Term Contracts, Rational Expectations, and the Optimal Money Supply Rule.” Journal of Political Economy 85 (February): 191–205. 218 Federal Reserve Bank of Richmond Economic Quarterly Friedman, Milton. 1968. “The Role of Monetary Policy.” American Economic Review 58: 1–16. Goodfriend, Marvin, and Robert King. 1997. “The New Neoclassical Synthesis and the Role of Monetary Policy.” In NBER Macroeconomics Annual, Vol. 12. Chicago: University of Chicago Press, 231–83. Hicks, John R. 1937. “Mr. Keynes and the ‘Classics:’ A Suggested Interpretation.” Econometrica 5 (April): 147–59. Kocherlakota, Narayana. 1998. “Money is Memory.” Journal of Economic Theory 81 (August): 232–51. Lagos, Ricardo, and Randall Wright. 2005. “A Unified Framework for Monetary Theory and Policy Analysis.” Journal of Political Economy 113 (June): 463–84. Lucas, Robert. 1972. “Expectations and the Neutrality of Money.” Journal of Economic Theory 4: 103–24. Lucas, Robert, and Nancy Stokey. 1987. “Money and Interest in a Cash-in-Advance Economy.” Econometrica 55 (May): 491–513. Mankiw, N. Gregory. 1985. “Small Menu Costs and Large Business Cycles: A Macroeconomic Model of Monopoly.” Quarterly Journal of Economics 100 (May): 529–37. Samuelson, Paul A. 1997. Economics: The Original 1948 Edition. New York: McGraw Hill/Irwin. Taylor, John B. 1979. “Staggered Wage Setting in a Macro Model.” American Economic Review Papers and Proceedings 69 (May): 108–13. Williamson, Stephen. 1986. “Costly Monitoring, Financial Intermediation and Equilibrium Credit Rationing.” Journal of Monetary Economics 18 (September): 159–79. Williamson, Stephen. 2008. Macroeconomics. Boston: Pearson (Addison-Wesley). Williamson, Stephen. Forthcoming A. “Monetary Policy and Distribution.” Journal of Monetary Economics. Williamson, Stephen. Forthcoming B. “Transactions, Credit, and Central Banking in a Model of Segmented Markets.” Review of Economic Dynamics. Woodford, Michael. 2003. Interest and Prices. Princeton, N.J.: Princeton University Press. Economic Quarterly—Volume 94, Number 3—Summer 2008—Pages 219–233 Nominal Frictions, Relative Price Adjustment, and the Limits to Monetary Policy Alexander L. Wolman T here are two broad classes of sticky-price models that have become popular in recent years. In the first class, prices adjust infrequently by assumption (so-called time-dependent models) and in the second class prices adjust infrequently because there is assumed to be a fixed cost of price adjustment (so-called state-dependent models). In both types of models it is common to assume that there are many goods, each produced with identical technologies. Consumers have a preference for variety, but their preferences treat all goods symmetrically. These assumptions mean that it is efficient for all goods to be produced in the same quantities. For that to happen, all goods must sell for the same price at any point in time. Assuming that price adjustment is staggered (as opposed to synchronized), the prices of all goods must be constant over time in order for all goods to be produced in the same quantities. If the aggregate price level were changing over time—even at a constant rate—then with staggered price adjustment prices would necessarily differ across goods. If there are multiple sectors that possess changing relative technologies or that face changing relative demand conditions (because consumers’ preferences are changing across goods), then in general it will no longer be efficient to produce all the goods in the same quantities. Equating marginal rates of transformation to marginal rates of substitution may require relative prices to change over time. These efficient changes in relative prices across sectors require nominal prices to change within sectors. With frictions in nominal The views here are the author’s and should not be attributed to the Federal Reserve Bank of Richmond, the Federal Reserve System, or the Board of Governors of the Federal Reserve System. For helpful comments, the author would like to thank Juan Carlos Hatchondo, Andreas Hornstein, Yash Mehra, and Anne Stilwell. E-mail: alexander.wolman@rich.frb.org. 220 Federal Reserve Bank of Richmond Economic Quarterly price adjustment, nominal price changes bring with them costly misallocations within (and perhaps across) sectors.1 In the circumstances just described, where efficiency across sectors requires nominal price changes within a sector or sectors, a zero inflation rate for the consumption price index may no longer be the optimal prescription for monetary policy in the presence of sticky prices. Which inflation rate results in the smallest distortions from price stickiness depends on the details of the environment: chiefly, the rates of relative price change across sectors and the degree of price stickiness in each sector. To cite an extreme example, suppose there are two sectors with different average rates of productivity growth. Suppose further that the sector with low productivity growth (increasing relative price) has sticky prices, whereas the sector with high productivity growth has flexible prices. Then it would be optimal to have deflation overall. Deflation would allow the desired relative price increase to occur with zero nominal price changes for the sticky-price goods and, thus, with no misallocation from the nominal frictions. The principles for optimal monetary policy I have discussed thus far involve only frictions associated with nominal price adjustment. In reality, monetary policy must balance other frictions as well. As the literature on the Friedman rule emphasizes, the fact that money is nearly costless to produce means that it is socially optimal for individuals to face nearly zero private costs of holding money. This requires a near-zero nominal interest rate, which corresponds to moderate deflation. Other frictions are less well understood, but may be just as important. Many central banks have mandates to achieve price stability, and the fact that my models do not necessarily support this objective does not mean it is misguided; that is, my models may still be lacking. The message of this paper is not that monetary policy should deviate from zero inflation in order to minimize distortions associated with nominal price adjustment. Rather, it is that in the presence of fundamental relative price changes and nominal price adjustment frictions, there is no monetary policy—zero inflation or otherwise—that can render those frictions costless. Section 1 works through the optimality of price stability in a benchmark one-sector model. Section 2 describes a two-sector model where a trend in relative productivities means that all prices cannot be stabilized. In Section 2 I also display U.S. data for broad categories of consumption, which display trends in relative prices. The data show that even on average, it is not possible to stabilize all nominal prices; trends in relative prices mean that if the price of one category of consumption goods is stabilized then prices for the other categories must have a trend. Furthermore, relative price trends in the United 1 These statements are qualitative ones. Unless monetary policy is extremely volatile or generates very large inflation or deflation, most current models attribute relatively small welfare costs to nominal frictions. A. L. Wolman: Relative Prices and Monetary Policy 221 States have been rather large; since 1947, the price of the services component of personal consumption expenditures has risen by a factor of five relative to the price of the durable goods component. Sections 3 and 4 provide a brief review of existing literature on relative price variation and monetary policy. In contrast to the material in Section 2, the existing literature has concentrated on random fluctuations in relative prices around a steady state where relative prices are constant. Section 3 reviews the literature on cyclical variation in relative prices, and Section 4 summarizes related work on wages (a form of relative price) and prices across locations. Section 5 concludes. 1. OPTIMALITY OF PRICE STABILITY IN A ONE-SECTOR MODEL Here I formalize the explanation for how price stability eliminates the distortions associated with price stickiness in a one-sector model. Suppose there are a large number of goods, specifically a continuum of goods indexed by z ∈ [0, 1], and suppose that consumers’ utility from consuming ct (z) of each good is given by ct , where that utility is determined by the following aggregator function: ε/(ε−1) 1 ct = ct (z)(ε−1)/ε dz , (1) 0 where ε is the elasticity of substitution in consumption between different goods. Suppose that each good is produced with a technology that uses only labor input, and that one unit of the consumption good can be produced with 1/at units of labor input: ct (z) = at nt (z) , for z ∈ [0, 1]. (2) Thus, at is a productivity factor common to all goods. I assume that at is exogenous. In particular, monetary policy has no effect on at . Finally, there is a constraint on the total quantity of labor input:2 1 nt (z) ≤ Nt . (3) 0 Without specifying anything about the structure of markets or price-setting behavior, I can discuss efficient production of consumption goods in this model. Efficiency dictates that the marginal rate of substitution in consumption be equated to the marginal rate of transformation in production. That is, the rate at which consumers trade off goods according to their preferences 2 In models with inelastic labor supply, N would be a constant equal to the time endowment. t Otherwise, Nt would be equal to the difference between the time endowment and the endogenous quantity of leisure. 222 Federal Reserve Bank of Richmond Economic Quarterly (represented by [1]) should be equal to the rate at which the technology (represented by [2]) allows goods to be traded off against one another in production. For the aggregator function in (1), consumers’marginal rate of substitution between any two goods, ct (z0 ) and ct (z1 ), is given by ct (z0 ) −1/ε ∂ct /∂ct (z0 ) . (4) = ∂ct /∂ct (z1 ) ct (z1 ) For the simple linear technology in (2), the marginal rate of transformation between any two goods indexed by z0 and z1 is unity: Reducing the labor used in the production of z0 by one unit yields a 1/at unit reduction in ct (z0 ) , and transferring that labor to the production of z1 yields an identical 1/at unit increase in ct (z1 ). Given my assumptions about consumers’ preferences and the technology for producing goods, equating the marginal rate of substitution to the marginal rate of transformation requires that each good, z, be produced in the same quantity. Only then can it be the case that mrs (ct (z0 ) , ct (z1 )) = (ct (z0 ) /ct (z1 ))−1/ε = 1 (5) for all z0 , z1 ∈ [0, 1] . At this point I know that efficiency requires all goods be produced in the same quantity. Under what conditions are the allocations in sticky-price models efficient? A standard assumption in sticky-price models, and an assumption I will make here, is that each individual good is produced by a separate monopolist. Because the Dixit-Stiglitz aggregator function (1) means that each good has many close substitutes, monopoly production of each good leads to an overall market structure known as monopolistic competition. The demand curve faced by the monopoly producer of any good, z, is Pt (z) −ε ct , Pt where Pt is the price index for the consumption basket and is given by ct (z) = 1/(1−ε) 1 Pt = (6) Pt (z)1−ε dz . (7) 0 The demand curve and the price index can be derived from the consumer’s problem of choosing consumption of individual goods in order to minimize the cost of one unit of the consumption basket (see, for example, Wolman 1999). From the demand functions and the efficiency condition, it is clear that efficiency requires all goods to have the same price. If price adjustment is infrequent (i.e., if prices are sticky) and if price adjustment is staggered across firms, then all goods can have the same price only if the aggregate price level is constant. If the price level varied over time, then changes in the price level would occur with only some firms adjusting their price, which would be inconsistent with all firms charging the same price. In somewhat A. L. Wolman: Relative Prices and Monetary Policy 223 simplified form, this is the reasoning behind optimality of price stability in sticky-price models (see, for example, Goodfriend and King 1997, Rotemberg and Woodford 1997, and King and Wolman 1999). 2. TREND VARIATION IN RELATIVE PRICES The model sketched in the previous section is a useful benchmark, but it is obviously unrealistic to suppose that the consumption goods valued by households are all “identical” in the sense of entering preferences symmetrically (1) and being produced with identical technologies (2). Departing from that benchmark, research on monetary policy in multisector sticky-price models has concentrated on the extent to which cyclical fluctuations in the determinants of relative prices interfere with the ability of monetary policy to eliminate sticky-price distortions on a period-by-period basis, and the related question of whether overall price stability remains optimal in such environments. However, more fundamental is the question of whether a trend in relative prices affects the ability of monetary policy to eliminate distortions even in steady state, and the related question of whether price stability is optimal on average, i.e., whether the optimal rate of inflation is zero. I consider these questions in Wolman (2008) and I draw on that analysis in what follows, emphasizing the former question (Can distortions be eliminated in a steady state?). Theory In contrast to the one-sector framework, suppose that consumers have CobbDouglas preferences over two composite goods, and that each of those composites has the characteristics of the single consumption aggregate (ct ) in the previous section. Here, ct will denote the overall consumption basket comprised of the two types of goods, and c1,t and c2,t will denote the sectoral baskets each comprised of a continuum of individual goods. The overall basket is now ν 1−ν ct = c1,t c2,t , (8) and the sectoral baskets are ε/(ε−1) 1 ck,t = ck,t (z)(ε−1)/ε dz , for k = 1, 2. (9) 0 As before, ε is the elasticity of substitution between individual goods within a sector. The elasticity of substitution across sectors is unity, and the sectoral expenditure shares for the two sectors are ν and 1 − ν. The constraint on labor input is 1 0 1 n1,t (z) + 0 n2,t (z) ≤ Nt . (10) 224 Federal Reserve Bank of Richmond Economic Quarterly Technology for producing individual goods is the same as above, ck,t (z) = ak,t nk,t (z) , for k = 1, 2, (11) except that now I allow for different levels of productivity (ak,t ) in the two sectors. Again, productivity is exogenous, or unaffected by monetary policy. Quantities and efficiency I can analyze efficiency just as I did in the one-sector model. However, here there are two dimensions of efficiency to be concerned with: efficiency within sectors and efficiency across sectors. Within either sector, the analysis is identical to that in the one-sector model. Efficiency within a sector requires equal production of each good, ck,t (z0 ) /ck,t (z1 ) −1/ε = 1 for z0 , z1 ∈ [0, 1] , for k = 1, 2, (12) because of symmetry in preferences and identical technologies. Across sectors, the marginal rate of substitution is mrs c1,t (z0 ) , c2,t (z1 ) for z0 , z1 ν = 1−ν ∈ [0, 1] , c2,t c1,t c1,t (z0 ) /c1,t c2,t (z1 ) /c2,t −1 ε , (13) and the marginal rate of transformation is mrt c1,t (z0 ) , c2,t (z1 ) = a2,t a1,t , for z0 , z1 ∈ [0, 1] . (14) Note that in order for efficiency to hold across sectors, it must hold within sectors; if within-sector efficiency does not hold, then from (12) the marginal rate of substitution varies across combinations of z0 , z1 . With the marginal rate of transformation independent of z (from [14]), it is not possible for the marginal rate of substitution to be equated to the marginal rate of transformation for all combinations of z0 , z1 unless there is efficiency within each sector. Efficiency within and across sectors then holds if and only if ck,t (z0 ) = ck,t (z1 ) for z0 , z1 ∈ [0, 1] , for k = 1, 2, (15) and a1,t = a2,t 1−ν ν c1,t c2,t . (16) The former condition states that quantities must be identical for all goods within a sector. The latter condition states that the ratio of sectoral consumptions should be proportional to the ratio of sectoral productivities; thus, if the ratio of sectoral productivities changes over time, then the ratio of sectoral consumptions must change in order to maintain efficiency. A. L. Wolman: Relative Prices and Monetary Policy 225 Prices and efficiency As in the one-sector model, in order to determine the conditions under which efficiency holds I need to specify market structure and pricing behavior. I make analogous assumptions to the one-sector model, namely that individual goods are produced by monopolists, which implies monopolistic competition among producers. The demand curve faced by the monopoly producer of a good z in sector k is Pk,t (z) Pk,t ck,t (z) = −ε ck,t , k = 1, 2, (17) where Pk,t is the price index for the sector k consumption basket, 1/(1−ε) 1 Pk,t = Pk,t (z)1−ε dz . (18) 0 The index of sector k consumption in (17) can be replaced by the appropriate demand function, P1,t Pt c1,t = ν c2,t = (1 − ν) −1 ct , or P2,t Pt (19) −1 ct . (20) These demand functions, as well as the overall price index in the two-sector model (Pt ), are derived from the consumer’s problem of choosing sectoral consumption in order to minimize the cost of one unit of the consumption basket. The price index is given by Pt = P1,t ν ν P2,t 1−ν 1−ν . (21) Note from (19) and (20) that the share of consumption spending (expenditure share) going to sector one (sector two) is constant and equal to ν (equal to 1 − ν). From the demand curves for individual goods (17) and the within-sector efficiency condition (15), it is again clear that efficiency requires all goods within a sector to have the same price: Pk,t (z0 ) = Pk,t (z1 ) for z0 , z1 ∈ [0, 1] , for k = 1, 2. (22) Across sectors, because efficiency requires relative consumptions to move with relative productivities, sectoral relative prices must vary with relative productivities. Combining (16) with (19) and (20) yields P2,t a1,t = . a2,t P1,t (23) 226 Federal Reserve Bank of Richmond Economic Quarterly Working now in terms of prices instead of quantities, conditions (22) and (23) are necessary and sufficient for efficiency. Earlier I stated that productivity in each sector was exogenous. Now I will make the further assumption that there is a trend in the growth rate of sector one’s productivity relative to sector two’s productivity: a1,t = (1 + γ )t , t = 0, 1, 2, ..., γ > 1, (24) a2,t where γ is an exogenous parameter representing relative productivity growth in sector one. Substituting this relationship into the second efficiency condition (23) implies P2,t = (1 + γ )t ; P1,t (25) efficiency requires the price of sector two’s composite good to rise over time relative to the price of sector one’s composite good. The requirement that there be a trend in the relative price P2,t /P1,t could be satisfied with a variety of different combinations of nominal price behavior for P2,t and P1,t , but each of those combinations involves at least one nominal price having a nonzero rate of change. In other words, when there is a trend in relative productivity growth across the two sectors, some nominal prices must change in order for efficiency to hold. But now there is a contradiction, because from the requirement that prices be identical for all goods within a sector (22), I can use the same reasoning as in the one-sector model of Section 1 to conclude that efficiency with price stickiness requires zero price changes within each sector. It is not possible to have both zero price changes within each sector and a nonzero rate of price change in at least one sector. Wolman (2008) shows how one can determine the optimal rate of inflation in a sticky-price model that has the features described here. For my purpose, it is enough to conclude that when there are different trend productivity growth rates across sectors, price stickiness inevitably leads to some real distortions that cannot be undone by monetary policy. Measurement: Price Stickiness and Relative Price Trends It is one thing to make a theoretical argument showing that, given certain assumptions, monetary policy cannot overcome the frictions associated with price stickiness. Here I take the next step and argue that those conditions seem to exist in the United States. The conditions I discuss are, first, that there is some price stickiness (infrequent price adjustment), and second, that there are trends in the relative prices of different categories of consumption goods. Concerning infrequent price adjustment, a vast literature has arisen in recent years reporting on the behavior of the prices of large numbers of individual A. L. Wolman: Relative Prices and Monetary Policy 227 Figure 1 Sectoral Prices Indexes 14 12 Services 10 Overall PCE Price Index 8 6 Nondurables 4 2 Durables 0 1947 1950 1953 1956 1959 1962 1965 1968 1971 1974 1977 1980 1983 1986 1989 1992 1995 1998 2001 2004 2007 Q1 Q1 Q1 Q1 Q1 Q1 Q1 Q1 Q1 Q1 Q1 Q1 Q1 Q1 Q1 Q1 Q1 Q1 Q1 Q1 Q1 goods. The seminal paper in this literature is by Bils and Klenow (2004), who study individual prices that serve as inputs into the United States Bureau of Labor Statistics (BLS) Consumer Price Index computations. Although their headline result is about price adjustment being more frequent than previous studies had estimated, they show that there is substantial heterogeneity in the frequency of price adjustment. The median price duration in their sample is 4.3 months, but one-third of consumption expenditure is on goods and services with price durations greater than 6.8 months, and one-quarter is on goods and services with price durations greater than 9 months. In the model in Section 2, there were two sectors, and thus only one independent sectoral relative price. In contrast, the BLS compiles price indexes for hundreds of categories of consumption goods. Thus, there are hundreds of sectoral relative prices one could study. Here I will report on relative prices from highly aggregated categories of consumption: durable goods, nondurable goods, and services. Price indexes for these categories are reported by the United States Commerce Department’s Bureau of Economic Analysis. Figure 1 plots the price indexes for durables, nondurables, and services from 1947 to 2008. There are clear positive trends in the prices of services and nondurables relative to durables. Since 1947, the price of nondurables relative to durables 228 Federal Reserve Bank of Richmond Economic Quarterly Figure 2 “Zero-Inflation” Price Indexes 1.6 1.4 Services 1.2 1.0 0.8 Nondurables 0.6 0.4 Durables 0.2 0.0 1947 1950 1953 1956 1959 1962 1965 1968 1971 1974 1977 1980 1983 1986 1989 1992 1995 1998 2001 2004 2007 Q1 Q1 Q1 Q1 Q1 Q1 Q1 Q1 Q1 Q1 Q1 Q1 Q1 Q1 Q1 Q1 Q1 Q1 Q1 Q1 Q1 has risen by a factor of more than two, and the price of services relative to durables has risen by a factor of more than five. With these trends in relative prices, it would not have been possible to stabilize the individual prices of each consumption category. Figure 2 displays the counterfactual paths of the price indexes from Figure 1 that are consistent with zero inflation in each quarter, given the historical values for relative prices and expenditure shares; stabilizing the overall price level would have required the nominal price of durables to fall by almost 80 percent and the nominal price of services to rise by almost 40 percent.3 Together, the presence of infrequent individual price adjustment and trends in the relative prices of different consumption categories makes clear that monetary policy cannot eliminate the frictions associated with nominal price 3 Figure 2 is constructed as follows: First, I normalize the levels of the sectoral price indexes at one in the first quarter of 1947. Then, for each subsequent quarter, I divide the gross rate of sectoral price change for each sector by the expenditure share weighted average of the gross rates of sectoral price changes (that is, I divide by the overall inflation rate). The resulting quotient for each sector is the rate of change of the zero-inflation price indexes in Figure 2. By construction, the expenditure share weighted average of the rates of price change for the zero-inflation sectoral price indexes is zero. A. L. Wolman: Relative Prices and Monetary Policy 229 adjustment. In other words, while monetary policy can stabilize an overall price index, it cannot stabilize all individual prices. To the extent that individual price adjustment is costly, either directly (state-dependent pricing) or indirectly (time-dependent pricing), stability of individual prices is required if nominal frictions are to be eliminated. 3. CYCLICAL VARIATION IN RELATIVE PRICES Thus far I have concentrated on the limits to monetary policy when there is a trend in the relative prices of goods with sticky nominal prices. In this setting monetary policy cannot stabilize all nominal prices, even in the absence of shocks. In contrast, most of the growing literature on relative prices and monetary policy has focused on optimal stabilization. That is, the literature has assumed that the optimal average inflation rate (the inflation target) is zero—price stability—and then gone on to study how monetary policy should make inflation behave in response to various shocks. A number of papers in this literature have used multisector models to study cyclical analogues of the kind of issues discussed in the previous section. I review some of them here. The most influential early paper in this line of research is by Aoki (2001). In Aoki’s paper there are sector-specific productivity shocks that make it infeasible to stabilize all nominal prices. However, one of the two sectors in Aoki’s model has flexible prices. Thus, inability to stabilize all prices does not prevent monetary policy from neutralizing nominal frictions. Optimal monetary policy involves stabilizing prices in the sticky-price sector, allowing prices in the other sector to fluctuate with relative productivity. Subsequently, other authors have extended Aoki’s analysis to environments where price stickiness in both (or all) sectors means that monetary policy cannot neutralize the nominal frictions. Huang and Liu (2005) study a model with both intermediate and final goods sectors, with price stickiness in both sectors. As in Aoki’s paper, productivity shocks are sector-specific, but they inevitably lead to distortions because of price stickiness in both sectors. Huang and Liu emphasize that stabilizing consumer prices at the expense of highly volatile producer prices can be quite costly; optimal monetary policy should place nonnegligible weight on producer price inflation. Like Huang and Liu, Erceg and Levin (2006) study a model with two sticky-price sectors. Instead of intermediate and final goods, the sectors produce durable and nondurable final goods.4 As in Huang and Liu’s paper, sector-specific productivity shocks ought to involve relative price changes, and with price stickiness these relative price changes necessarily involve distortions. However, the presence of durable goods gives the model some 4 Erceg and Levin’s paper also includes wage stickiness, which I will discuss in the next section. 230 Federal Reserve Bank of Richmond Economic Quarterly additional interesting properties. A shock to government spending—an aggregate shock—now also should involve relative price changes, because it raises the real interest rate, making durable goods less attractive. With sticky prices in both sectors, stabilizing prices in both sectors is infeasible in response to a government spending shock. Thus, even an aggregate shock can inevitably lead to nominal distortions in a multisector model.5 4. WAGES AND PRICES ACROSS LOCATIONS By now it should be clear that when individual goods prices are sticky, shocks that optimally change the relative price across sectors inevitably restrict the monetary authority’s ability to achieve efficient allocations. Elements of this reasoning also apply to wage stickiness and to prices across regions in a currency union.6 The labor market can be thought of as a sector, and if nominal prices in that sector are sticky (i.e., if nominal wages are sticky) then aggregate shocks that require real wage adjustment will lead to inefficiencies, even under optimal monetary policy. Erceg, Henderson, and Levin (2000) work through the details in a model with sticky wages and prices. Wage stickiness is introduced in a similar manner to price stickiness: Firms must assemble a range of different types of labor inputs, and the supplier of each input has monopoly power and sets her wage only occasionally. In this framework, it is not optimal to completely stabilize prices or wages, but in general higher welfare is achieved by stabilizing wage inflation than by stabilizing price inflation. Wage dispersion has two costs in the model: it makes production less efficient, and it is disliked by households, who would prefer to spread their labor input evenly over all firms. Price dispersion has only the analogue to the first cost (that is, it makes consumption of the aggregate good less efficient), and this helps to explain why wage inflation takes priority over price inflation. Another factor that works toward stabilizing wage inflation is that the productive inefficiency from wage dispersion affects each intermediate good and feeds through into inefficient production of final goods. In contrast, price dispersion leads only to inefficient production of final goods.7 Amano et al. (2007) use a model similar to that of Erceg, Henderson, and Levin to address trend instead of cyclical issues. They assume there is trend productivity growth so that the real wage should rise over time. In a steady state (balanced growth path), a rising real wage means that the nominal wage 5 There are many other recent papers that study multisector models with nominal rigidities. They include Mankiw and Reis (2003, to be discussed in the next section), Carlstrom et al. (2006), Carvalho (2006), and Nakamura and Steinsson (2008). 6 Erceg, Henderson, and Levin (2000) recognized the generality of this point. 7 Similar reasoning may help to explain Huang and Liu’s (2005) quantitative findings regarding producer price inflation, mentioned previously. A. L. Wolman: Relative Prices and Monetary Policy 231 must be rising or nominal prices must be falling. With infrequent adjustment for both wages and prices, any such scenario involves distortions. Amano et al. show that optimal monetary policy involves slight deflation so that nominal wages are rising at a rate lower than real wages—a compromise between constant wages and constant prices. Mankiw and Reis (2003) provide a general framework for thinking about optimal monetary policy in the presence of wage and price stickiness as well as sectoral considerations. They frame the monetary policy problem as choosing the appropriate index of prices and wages to stabilize. Consistent with Erceg, Henderson, and Levin (2000), Mankiw and Reis find that nominal wages carry large weight in the “price index” that should be stabilized. Benigno (2004) studies optimal monetary policy in a two-region currency area. If nominal prices are sticky in both regions and real factors lead to efficient relative price variation across regions, then once again monetary policy cannot eliminate the real effects of price stickiness. The optimal monetary policy problem involves trading off price distortions in the two regions. 5. CONCLUSION If prices or wages are sticky in only one sector of an economy, and if there is no heterogeneity across regions, then monetary policy can undo the effects of price stickiness. However, if there is more than one sector with sticky prices, or if wages and prices are sticky, or if there are heterogeneous regions, then nominal rigidities cause distortions under any monetary policy.8 I described several examples of these distortions, emphasizing an underappreciated one, trending relative prices across sectors. Macroeconomists are acutely aware of the limited ability of monetary policy to counteract real distortions that may be present in the economy (for example, search frictions in labor markets or monopoly power in goods markets). However, we have been perhaps less modest about the ability of monetary policy to counteract nominal distortions—in particular price adjustment frictions. A recent literature on monetary policy in multisector models with price stickiness has served to make us more modest, and this paper aims to draw attention to that literature. 8 Loyo (2002) points out that if different sectors have different currencies, then nominal rigidities can be undone. That possibility is intriguing but not currently relevant. 232 Federal Reserve Bank of Richmond Economic Quarterly REFERENCES Amano, Robert, Kevin Moran, Stephen Murchison, and Andrew Rennison. 2007. “Trend Inflation, Wage and Price Rigidities, and Welfare.” Bank of Canada Working Paper 07-42. Aoki, Kosuke. 2001. “Optimal Monetary Policy Response to Relative Price Changes.” Journal of Monetary Economics 48: 55–80. Benigno, Pierpaolo. 2004. “Optimal Monetary Policy in a Currency Area.” Journal of International Economics 63 (July): 293–320. Bils, Mark, and Peter Klenow. 2004. “Some Evidence on the Importance of Sticky Prices.” Journal of Political Economy 112 (October): 947–85. Carlstrom, Charles T., Timothy S. Fuerst, Fabio Ghironi, and K´ lver o Hern´ ndez. 2006. “Relative Price Dynamics and the Aggregate a Economy.” Manuscript, http://www2.bc.edu/˜ghironi/CaFuGhiroHeRelPrices082206.pdf (August). Carvalho, Carlos. 2006. “Heterogeneity in Price Stickiness and the Real Effects of Monetary Shocks.” Frontiers of Macroeconomics, Vol. 2, Article 1. Erceg, Christopher, and Andrew Levin. 2006. “Optimal Monetary Policy with Durable Consumption Goods.” Journal of Monetary Economics 53: 1341–59. Erceg, Christopher, Dale Henderson, and Andrew Levin. 2000. “Optimal Monetary Policy with Staggered Wage and Price Contracts.” Journal of Monetary Economics 46 (October): 281–313. Goodfriend, Marvin, and Robert King. 1997. “The New Neoclassical Synthesis and the Role of Monetary Policy.” In NBER Macroeconomics Annual, edited by Ben S. Bernanke and Julio Rotemberg. Cambridge, Mass.: MIT Press, 231–82. Huang, Kevin, and Zheng Liu. 2005. “Inflation Targeting: What Inflation Rate to Target?” Journal of Monetary Economics 52 (November): 1435–62. King, Robert G., and Alexander L. Wolman. 1999. “What Should the Monetary Authority do When Prices are Sticky?” In Monetary Policy Rules, edited by John B. Taylor. Chicago: University of Chicago Press, 349–98. Loyo, Eduardo. 2002. “Imaginary Money Against Sticky Relative Prices.” European Economic Review 46 (June): 1073–92. A. L. Wolman: Relative Prices and Monetary Policy 233 Mankiw, N. Gregory, and Ricardo Reis. 2003. “What Measure of Inflation Should a Central Bank Target?” Journal of the European Economic Association 1 (September): 1058–86. Nakamura, Emi, and J´ n Steinsson. 2008. “Monetary Non-Neutrality in a o Multi-Sector Menu Cost Model.” Columbia University Working Paper. Rotemberg, Julio, and Michael Woodford. 1997. “An Optimization-Based Econometric Framework for the Evaluation of Monetary Policy.” In NBER Macroeconomics Annual, edited by Ben S. Bernanke and Julio Rotemberg. Cambridge, Mass.: MIT Press, 297–345. Wolman, Alexander L. 1999. “Sticky Prices, Marginal Cost, and the Behavior of Inflation.” Federal Reserve Bank of Richmond Economic Quarterly 85 (Fall): 29–48. Wolman, Alexander L. 2008. “The Optimal Rate of Inflation with Trending Relative Prices.” Manuscript, Federal Reserve Bank of Richmond. Economic Quarterly—Volume 94, Number 3—Summer 2008—Pages 235–263 Understanding Monetary Policy Implementation Huberto M. Ennis and Todd Keister O ver the last two decades, central banks around the world have adopted a common approach to monetary policy that involves targeting the value of a short-term interest rate. In the United States, for example, the Federal Open Market Committee (FOMC) announces a rate that it wishes to prevail in the federal funds market, where commercial banks lend balances held at the Federal Reserve to each other overnight. Changes in this shortterm interest rate eventually translate into changes in other interest rates in the economy and thereby influence the overall level of prices and of real economic activity. Once a target interest rate is announced, the problem of implementation arises: How can a central bank ensure that the relevant market interest rate stays at or near the chosen target? The Federal Reserve has a variety of tools available to influence the behavior of the interest rate in the federal funds market (called the fed funds rate). In general, the Fed aims to adjust the total supply of reserve balances so that it equals demand at exactly the target rate of interest. This process necessarily involves some estimation, since the Fed does not know the exact demand for reserve balances, nor does it completely control the supply in the market. A critical issue in the implementation process, therefore, is the sensitivity of the market interest rate to unanticipated changes in supply and/or demand. Some of the material in this article resulted from our participation in the Federal Reserve System task force created to study paying interest on reserves. We are very grateful to the other members of this group, who patiently taught us many of the things that we discuss here. We also would like to thank Kevin Bryan, Yash Mehra, Rafael Repullo, John Walter, John Weinberg, and the participants at the 2008 Columbia Business School/New York Fed conference on “The Role of Money Markets” for useful comments on a previous draft. All remaining errors are, of course, our own. The views expressed here do not necessarily represent those of the Federal Reserve Bank of New York, the Federal Reserve Bank of Richmond, or the Federal Reserve System. Ennis is on leave from the Richmond Fed at University Carlos III of Madrid and Keister is at the Federal Reserve Bank of New York. E-mails: hennis@eco.uc3m.es, Todd.Keister@ny.frb.org. 236 Federal Reserve Bank of Richmond Economic Quarterly If small estimation errors lead to large swings in the interest rate, a central bank will find it difficult to effectively implement monetary policy, that is, to consistently hit the target rate. The degree of sensitivity depends on a variety of factors related to the design of the implementation process, such as the time period over which banks are required to hold reserves and the interest rate, if any, that a central bank pays on reserve balances. The ability to hit a target interest rate consistently plays a critical role in a central bank’s communication policy. The overall effectiveness of monetary policy depends, in part, on individuals’ perceptions of the central bank’s actions and objectives. If the market interest rate were to deviate consistently from the central bank’s announced target, individuals might question whether these deviations simply represent glitches in the implementation process or whether they instead represent an unannounced change in the stance of monetary policy. Sustained deviations of the average fed funds rate from the FOMC’s target in August 2007, for example, led some media commentators to claim that the Fed had engaged in a “stealth easing,” taking actions that lowered the market interest rate without announcing a change in the official target.1 In such times, the ability to hit a target interest rate consistently allows the central bank to clearly (and credibly) communicate its policy to market participants. Under most circumstances, the Fed changes the total supply of reserve balances available to commercial banks by exchanging government bonds or other securities for reserves in an open market operation. Occasionally, the Fed also provides reserves directly to certain banks through its discount window. In some situations, the Fed has developed other, ad hoc methods of influencing the supply and distribution of reserves in the market. For example, during the recent period of financial turmoil, the market’s ability to smoothly distribute reserves across banks became partially impaired, which led to significant fluctuations in the average fed funds rate both during the day and across days. In December 2007, partly to address these problems, the Fed introduced the Term Auction Facility (TAF), a bimonthly auction of a fixed quantity of reserve balances to all banks eligible to borrow at the discount window. In principle, the TAF has increased these banks’ ability to access reserves directly and, in this way, has helped ease the pressure on the market to redistribute reserves and avoid abnormal fluctuations in the market rate. Such operations, of course, need to be managed so as to achieve the ultimate goal of implementing the chosen target interest rate. Balancing the demand and supply of reserves is at the very core of this problem. This article presents a simple analytical framework for understanding the process of monetary policy implementation and the factors that influence a 1 See, for example, “A ‘Stealth Easing’ by the Fed?” (Coy 2007). H. M. Ennis and T. Keister: Monetary Policy Implementation 237 central bank’s ability to keep the market interest rate close to a target level. We present this framework graphically, focusing on how various features of the implementation process affect the sensitivity of the market interest rate to unanticipated changes in supply or demand. We discuss the current approach used by the Fed, including the use of reserve maintenance periods to decrease this sensitivity. We also show how this framework can be used to study a wide range of issues related to monetary policy implementation. In 2006, the U.S. Congress enacted legislation that will give the Fed the authority to pay interest on reserve balances beginning in October 2011.2 We use our simple framework to illustrate how the ability to pay interest on reserves can be a useful policy tool for a central bank. In particular, we show how paying interest on reserves can decrease the sensitivity of the market interest rate to estimation errors and, thus, enable a central bank to better achieve its desired interest rate. The model we present uses the basic approach to reserve management introduced by Poole (1968) and subsequently advanced by many others (see, for example, Dotsey 1991; Guthrie and Wright 2000; Bartolini, Bertola, and Prati 2002; and Clouse and Dow 2002). The specific details of our formalization closely follow those in Ennis and Weinberg (2007), after some additional simplifications that allow us to conduct all of our analysis graphically. Ennis and Weinberg (2007) focused on the interplay between daylight credit and the Fed’s overnight treatment of bank reserves. In this article, we take a more comprehensive view of the process of monetary policy implementation and we investigate several important topics, such as the role of reserve maintenance periods, which were left unexplored by Ennis and Weinberg (2007). 1. U.S. MONETARY POLICY IMPLEMENTATION Banks hold reserve balances in accounts at the Federal Reserve in order to satisfy reserve requirements and to be able to make interbank payments. During the day, banks can also access funds by obtaining an overdraft from their reserve accounts at the Fed. The terms by which the Fed provides daylight credit are one of the factors determining the demand for reserves by banks. To adjust their reserve holdings, banks can borrow and lend balances in the fed funds market, which operates weekdays from 9:30 a.m. to 6:30 p.m. A bank wanting to decrease its reserve holdings, for example, can do so in this market by making unsecured, overnight loans to other banks. The fed funds market plays a crucial role in monetary policy implementation because this is where the Federal Reserve intervenes to pursue its policy objectives. The stance of monetary policy is decided by the FOMC, which 2 After this article was written, the effective date for the authority to pay interest on reserves was moved to October 1, 2008, by the Emergency Economic Stabilization Act of 2008. 238 Federal Reserve Bank of Richmond Economic Quarterly selects a target for the overnight interest rate prevailing in this market. The Committee then instructs the Open Market Desk to adjust, via open market operations, the supply of reserve balances so as to steer the market interest rate toward the selected target.3 The Desk conducts open market operations largely by arranging repurchase agreements (repos) with primary securities dealers in a sealed-bid, discriminatory price auction. Repos involve using reserve balances to purchase securities with the explicit agreement that the transaction will be reversed at maturity. Repos usually have overnight maturity, but the Desk also employs other maturities (for example, two-day and two-week repos are commonly used). Open market operations are typically conducted early in the morning when the market for repos is most active. The new reserves created in an open market operation are deposited in the participating securities dealers’ bank accounts and, hence, increase the total supply of reserves in the banking system. In this way, each day the Desk tries to move the supply of reserve balances as close as possible to the level that would leave the market-clearing interest rate equal to the target rate. An essential step in this process is accurately forecasting both aggregate reserve demand and those changes in the existing supply of reserve balances that are due to autonomous factors beyond the Fed’s control, such as payments into and out of the Treasury’s account and changes in the quantity of currency in circulation. Forecasting errors will lead the actual supply of reserve balances to deviate from the intended level and, hence, will cause the market rate to diverge from the target rate, even if reserve demand is perfectly anticipated. Reserve requirements in the United States are calculated as a proportion of the quantity of transaction deposits on a bank’s balance sheet during a twoweek computation period prior to the start of the maintenance period. These requirements can be met through a combination of vault cash and reserve balances held at the Fed. During the two-week reserve maintenance period, a bank’s end-of-day reserve balances must, on average, equal the reserve requirement minus the quantity of vault cash held during the computation period. Reserve requirements make a large portion of the demand for reserve balances fairly predictable, which simplifies monetary policy implementation. Reserve maintenance periods allow banks to spread out their reserve holdings over time without having to scramble for funds to meet a requirement at the end of each day. However, near the end of the maintenance period, this averaging effect tends to lose force. On the last day of the period, a bank has some level of remaining requirement that must be met on that day. This generates a fairly inelastic demand for reserve balances and makes implementing a target interest rate more challenging. For this reason, the Fed allows banks 3 See Hilton and Hrung (2007) for a more detailed overview of the Fed’s monetary policy implementation procedures. H. M. Ennis and T. Keister: Monetary Policy Implementation 239 holding excess or deficient balances at the end of a maintenance period to carry over those balances and use them to satisfy up to 4 percent of their requirement in the following period. If a bank finds itself short of reserves at the end of the maintenance period, even after taking into account the carryover possibilities, it has several options. It can try to find a counterparty late in the day offering an acceptable interest rate. However, this may not be feasible because of an aggregate shortage of reserve balances or because of the existence of trading frictions in this market. A second alternative is to borrow at the discount window of its corresponding Federal Reserve Bank.4 The discount window offers collateralized overnight loans of reserves to banks that have previously pledged appropriate collateral. Discount window loans are typically charged an interest rate that is 100 basis points above the target fed funds rate, although changing the size of this gap is possible and has been used, at times, as a policy instrument. Finally, if the bank does not have the appropriate collateral or chooses not to borrow at the discount window for other reasons, it will be charged a penalty fee proportional to the amount of the shortage. Currently, banks earn no interest on the reserve balances they hold in their accounts at the Federal Reserve.5 This situation may soon change: The Financial Services Regulatory Relief Act of 2006 allows the Fed to begin paying interest on reserve balances in October 2011. The Act also includes provisions that give the Fed more flexibility in determining reserve requirements, including the ability to eliminate the requirements altogether. Thus, this legislation opens the door to potentially substantial changes in the way the Fed implements monetary policy. To evaluate the best approach within the new, broader set of alternatives, it seems useful to develop a simple analytical framework that is able to address many of the relevant aspects of the problem. We introduce and discuss such a framework in the sections that follow. 2. THE DEMAND FOR RESERVES In this section, we present a simple framework that is useful for understanding banks’demand for reserves. In this framework, a bank holds reserves primarily to satisfy reserve requirements, although other factors, such as the desire to make interbank payments, may also play a role. Since banks cannot fully predict the timing of payments, they face uncertainty about the net outflows from their reserve accounts and, therefore, are typically unable to exactly satisfy their reserve requirements. Instead, they must balance the possibility 4 There are 12 regions and corresponding Reserve Banks in the Federal Reserve System. For each commercial bank, the corresponding Reserve Bank is determined by the region where the commercial bank is headquartered. 5 See footnote 2. 240 Federal Reserve Bank of Richmond Economic Quarterly of holding excess reserve balances—and the associated opportunity cost— against the possibility of being penalized for a reserve deficiency. A bank’s demand for reserves results from optimally balancing these two concerns. The Basic Framework We assume banks are risk-neutral and maximize expected profits. At the beginning of the day, banks can borrow and lend reserves in a competitive interbank market. Let R be the quantity of reserves chosen by a bank in the interbank market. The central bank affects the supply of reserves in this market by conducting open market operations. Total reserve supply is equal to the quantity set by the central bank through its operations, adjusted by a potentially random amount to reflect unpredictable changes in autonomous factors. During the day, each bank makes payments to and receives payments from other banks. To keep things as simple as possible, suppose that each bank will make exactly one payment and receive exactly one payment during the “middle” part of the day. Furthermore, suppose that these two payment flows are of exactly the same size, PD > 0, and that this size is nonstochastic. However, the order in which these payments occur during the day is random; some banks will receive the incoming payment before making the outgoing one, while others will make the outgoing payment before receiving the incoming one. At the end of the day, after the interbank market has closed, each bank experiences another payment shock, P , that affects its end-of-day reserve balance. The value of P can be either positive, indicating a net outflow of funds, or negative, indicating a net inflow of funds. We assume that the payment shock, P , is uniformly distributed on the interval −P , P . The value of this shock is not yet known when the interbank market is open; hence, a bank’s demand for reserves in this market is affected by the distribution of the payment shock and not the realization. We assume, as a starting point, that a bank must meet a given reserve requirement, K, at the end of each day.6 If the bank finds itself holding fewer than K reserves at the end of the day, after the payment shock P has been realized, it must borrow funds at a “penalty” rate of interest, rP , to satisfy the requirement. This rate can be thought of as the rate charged by the central bank on discount window loans, adjusted to take into account any “stigma” associated with using this facility. In reality, a bank may pay a deficiency fee instead of borrowing from the discount window or it may borrow funds 6 We discuss more complicated systems of reserve requirements later, including multiple-day maintenance periods. For the logic in the derivations that follow, the particular value of K does not matter. The case of K = 0 corresponds to a system without reserve requirements. H. M. Ennis and T. Keister: Monetary Policy Implementation 241 in the interbank market very late in the day when this market is illiquid. In the model, the rate rP simply represents the cost associated with a late-day reserve deficiency, whatever the source of that cost may be. The specific assumptions we make about the number and size of payments that a bank sends are not important; they only serve to keep the analysis free of unnecessary complications. Two basic features of the model are important. First, the bank cannot perfectly anticipate its end-of-day reserve position. This uncertainty creates a “precautionary” demand for reserves that smoothly responds to changes in the interest rate. Second, a bank makes payments during the day as a part of its normal operations and the pattern of these payments can potentially lead to an overdraft in the bank’s reserve account. We initially assume that the central bank offers daylight credit to banks to cover such overdrafts at no charge. We study the case where daylight overdrafts are costly later in this section. The Benchmark Case We begin by analyzing a simple benchmark case; we show later in this section how the framework can be extended to include a variety of features that are important in reality. In the benchmark case, banks must meet their reserve requirement at the end of each day, and the central bank pays no interest on reserves held by banks overnight. Furthermore, the central bank offers daylight credit free of charge. Figure 1 depicts an individual bank’s demand for reserves in the interbank market under this benchmark scenario. On the horizontal axis we measure the bank’s choice of reserve holdings before the late-day payment is realized. On the vertical axis we measure the market interest rate for overnight loans. To draw the demand curve, we ask: Given a particular value for the interest rate, what quantity of reserves would the bank demand to hold if that rate prevailed in the market? A bank would be unwilling to hold any reserves if the market interest rate were higher than rP . If the market rate were higher than the penalty rate, the bank would choose to meet its requirement entirely by borrowing from the discount window. It would actually like to borrow even more than its requirement and lend the rest out at the higher market rate, but this fact is not important for the analysis. The important point is simply that there will be no demand for (nonborrowed) reserves for any interest rate larger than rP . When the market interest rate exactly equals the penalty rate, rP , a bank would be indifferent between holding any amount of reserves between zero and K − P and, hence, the demand curve is horizontal at rP . As long as the bank’s reserve holdings, R, are smaller than K − P , the bank will need to borrow at the discount window to satisfy its reserve requirement, K, even if the late-day inflow of funds into the bank’s reserve account is the largest 242 Federal Reserve Bank of Richmond Economic Quarterly Figure 1 Benchmark Demand for Reserves r r P Demand for Reserves r T 0 K-P K S T K+P R possible value, P .7 The alternative would be to borrow more reserves in the market to reduce this potential need for discount window lending. Since the market rate is equal to the penalty rate, both strategies deliver the same level of profit and the bank is indifferent between them. For market interest rates below the penalty rate, however, a bank will choose to hold at least K − P reserves. As discussed above, if the bank held fewer than K − P reserves it would be certain to need to borrow from the discount window, which would not be an optimal choice when the market rate is lower than the discount rate. The bank’s demand for reserves in this situation can be described as “precautionary” in the sense that the bank chooses its reserve holdings to balance the possibility of falling short of the requirement against the possibility of ending up with extra reserves in its account at the end of the day. 7 To see this, note that even in the best case scenario the bank will find itself holding R + P reserves after the arrival of the late-day payment flow. When R < K − P , the bank’s end-of-day holdings of reserves is insufficient to satisfy its reserve requirement, K, unless it takes a loan at the discount window. H. M. Ennis and T. Keister: Monetary Policy Implementation 243 If the market interest rate were very low—close to zero—the opportunity cost of holding reserves would be very small. In this case, the bank would hold enough precautionary reserves so that it is virtually certain that unforeseen movements on its balance sheet will not decrease its reserves below the required level. In other words, the bank will hold K + P reserves in this case. If the market interest rate were exactly zero, there would be no opportunity cost of holding reserves. The demand curve is, therefore, flat along the horizontal axis after K + P . In between the two extremes, K − P and K + P , the demand for reserves will vary inversely with the market interest rate measured on the vertical axis; this portion of the demand curve is represented by the downward-sloping line in Figure 1. The curve is downward-sloping for two reasons. First, the market interest rate represents the opportunity cost of holding reserves overnight. When this rate is lower, finding itself with excess balances is less costly for the bank and, hence, the bank is more willing to hold precautionary balances. Second, when the market rate is lower, the relative cost of having to access the discount window is larger, which also tends to increase the bank’s precautionary demand for reserves. The linearity of the downward-sloping part of the demand curve results from the assumption that the late-day payment shock is uniformly distributed. With other probability distributions, the demand curve will be nonlinear, but its basic shape will remain unchanged. In particular, the points where the demand curve intersects the penalty rate, rP , and the horizontal axis will be the same for any distribution with support −P , P .8 The Equilibrium Interest Rate Suppose, for the moment, that there is a single bank in the economy. Then the demand curve in Figure 1 also represents the total demand for reserves. Let S denote the total supply of reserves in the interbank market, as jointly determined by the central bank’s open market operations and autonomous factors. Then the equilibrium interest rate is determined by the height of the demand curve at point S. As shown in the diagram, there is a unique level of reserve supply, ST , that will generate a given target interest rate, rT . Now suppose there are many banks in the economy, but they are all identical in that they have the same level of required reserves, face the same payment shock, etc. When there are many banks, the total demand for reserves can be found by simply “adding up” the individual demand curves. For any interest 8 The support of the probability distribution is the set of values of the payment shock that is assigned positive probability. An explicit formula for the demand curve in the uniform case is derived in Ennis and Weinberg (2007). If the shock instead had an unbounded distribution, such as the normal distribution used by Whitesell (2006) and others, the demand curve would asymptote to the penalty rate and the horizontal axis but never intersect them. 244 Federal Reserve Bank of Richmond Economic Quarterly rate r, total demand is simply the sum of the quantity of reserves demanded by each individual bank. For presentation purposes, it is useful to look at the average demand for reserves, that is, the total demand divided by the number of banks. When all banks are identical, the average demand is exactly equal to the demand of each individual bank. In other words, in the benchmark case where banks are identical, the demand curve in Figure 1 also represents the aggregate demand for reserves, expressed in per-bank terms. The determination of the equilibrium interest rate then proceeds exactly as in the single-bank case. In particular, the market-clearing interest rate will be equal to the target rate, rT , if and only if reserve supply (expressed in per-bank terms) is equal to ST . Note that the central bank has two distinct ways in which it can potentially affect the market interest rate: changing the supply of reserves available in the market and changing (either directly or indirectly) the penalty rate. Suppose, for example, that the central bank wishes to decrease the market interest rate. It could either increase the supply of reserves through open market operations, leading to a movement down the demand curve, or it could decrease the penalty rate, which would rotate the demand curve downward while leaving the supply of reserves unchanged. Both policies would cause the market interest rate to fall. Heterogeneity While the assumption that all banks are identical was useful for simplifying the presentation above, it is clearly a poor representation of reality in most economies. The United States, for example, has thousands of banks and other depository institutions that differ dramatically in size, range of activities, etc. We now show how the analysis above changes when there is heterogeneity among banks and, in particular, how the size distribution of banks might affect the aggregate demand for reserves. Each bank still has a demand curve of the form depicted in Figure 1, but now these curves can be different from each other because banks may have different levels of required reserves, face different distributions of the payment shock, and/or face different penalty rates. These individual demand curves can be aggregated exactly as before: For any interest rate r, the total demand for reserves is simply the sum of the quantity of reserves demanded by each individual bank. The aggregate demand curve, expressed in per-bank terms, will again be similar to that presented in Figure 1, with the exact shape being determined by the properties of the various individual demands. If different banks have different levels of required reserves, for example, the requirement K in the aggregate demand curve will be equal to the average of the individual banks’ requirements. H. M. Ennis and T. Keister: Monetary Policy Implementation 245 Our interest here is in studying how bank heterogeneity affects the properties of this demand curve. We focus on heterogeneity in bank size, which is particularly relevant in the United States, where there are some very large banks and thousands of smaller banks. We ask how large banks may differ from small banks in the context of the simple framework and how the presence of both large and small banks might affect the properties of the aggregate demand curve. To simplify the presentation, we study the three possible dimensions of heterogeneity addressed by the model one at a time. In reality, of course, the three cases are closely intertwined. Size of Requirements Perhaps the most natural way of capturing differences in bank size is by allowing for heterogeneity in reserve requirements. When requirements are calculated as a percentage of the deposit base, larger banks will tend to have a larger level of required reserves in absolute terms. Suppose, then, that banks have different levels of K, but they face the same late-day payment shock and the same penalty rate for a reserve deficiency. How would the size distribution of banks affect the aggregate demand for reserves in this case? To begin, note that in Figure 1 the slope of the demand curve is independent of the size of the bank’s reserve requirement, K. To see why this is the case, consider an increase in the value of K. Since both K − P and K + P become larger numbers, the demand curve in Figure 1 shifts to the right. Notice that these two points shift exactly the same distance, leaving the slope of the downward-sloping segment of the demand curve unchanged. Simple aggregation then shows that the slope of the aggregate demand curve will be independent of the size distribution of banks. In other words, for the case of heterogeneity in K, the sensitivity of reserve demand to changes in the interest rate does not depend at all on whether the economy is comprised of only large banks or, as in the United States, has a few large banks and very many small ones. Adding heterogeneity in reserve requirements does generate an interesting implication for the distribution of excess reserve holdings across banks. If large and small banks face similar (effective) penalty rates and are not too different in their exposure to late-day payment uncertainty, then the framework suggests that all banks should hold similar quantities of precautionary reserves. In other words, for a given level of the interest rate, the difference between the chosen reserve balances, R, and the requirement, K, should be similar for all banks. After the payment shocks are realized, of course, some banks will end up holding excess reserves and others will end up needing to borrow. On average, however, a large bank and a small one should finish the period with comparable levels of excess reserves. If the banking system is composed of a relatively small number of large banks and a much larger number of small 246 Federal Reserve Bank of Richmond Economic Quarterly Figure 2 Heterogeneity Panel A r Small Banks r s P r L P r T Large Banks 0 sL s s K-P K+P R Panel B r Low-Variance Demand r P HighVariance Demand r T 0 K-P H K-P L K K+P L K+P H R banks, then the majority of the excess reserves in the banking system will be held by small banks, simply because there are so many more of them. Even if large banks hold the majority of total reserve balances because of their larger requirements, most of the excess reserve balances will be held by small banks. This implication is broadly in line with the data for the United States. The Penalty Rate Another way in which small banks might differ from large ones is the penalty rate they face if they need to borrow to avoid a reserve deficiency. To be H. M. Ennis and T. Keister: Monetary Policy Implementation 247 eligible to borrow at the discount window, for example, a bank must establish an agreement with its Reserve Bank and post collateral. This fixed cost may lead some smaller banks to forgo accessing the discount window and instead borrow at a very high rate in the market (or pay the reserve deficiency fee) when necessary. Smaller banks may also have fewer established relationships with counterparties in the fed funds market and, as a consequence, may find it more difficult to borrow at a favorable interest rate late in the day (see Ashcraft, McAndrews, and Skeie 2007). S Suppose small banks do face a higher penalty rate, such as the value rP L depicted in Figure 2, Panel A, while larger banks face a lower rate, rP . The figure is drawn as if the two banks have the same level of requirements, but this is done only to make the comparison between the curves clear. The figure shows two immediate implications of this type of heterogeneity. First, at any given interest rate, small banks will hold a higher level of precautionary reserves, that is, they will choose a larger reserve balance relative to their level of required reserves. In the figure, the smaller bank will hold a quantity SS while the larger bank holds only SL , even though—in this example—both face the same requirement and the same uncertainty about their end-of-day balance. As a result, the distribution of excess reserves in the economy will tend to be skewed even more heavily toward small banks than the earlier discussion would suggest. The second implication shown in Figure 2, Panel A is that the demand curve for small banks has a steeper slope. In an economy with a large number of small banks, therefore, the aggregate demand curve will tend to be steeper, meaning that average reserve balances will be less sensitive to changes in the market interest rate. Notice that this result obtains even though there are no costs of reserve management in the model. Support of the Payment Shock A third way in which banks potentially differ from each other is in the distribution of the late-day payment shock they face. Figure 2, Panel B depicts two demand curves, one for a bank facing a higher variance of this distribution and one for a bank facing a lower variance. The figure shows that having more uncertainty about the end-of-day reserve position leads to a flatter demand curve and, hence, a reserve balance that is more responsive to changes in the interest rate. In this case, it is not completely clear which curve corresponds better to large banks and which to small banks. Banks with larger and more complex operations might be expected to face much larger day-to-day variations in their payment flows. However, such banks also tend to have sophisticated reserve management systems in place. As a result, it is not clear whether the end-of-day uncertainty faced by a large bank is higher or lower than that faced 248 Federal Reserve Bank of Richmond Economic Quarterly by a small bank.9 The effect of the size distribution of banks on the shape of the aggregate demand curve is, therefore, ambiguous in this case. Daylight Credit Fees So far, we have proceeded under the assumption that banks are free to hold negative balances in their reserve accounts during the day and that no fees are associated with such daylight overdrafts. Most central banks, however, place some restriction on banks’ access to overdrafts. In many cases, banks must post collateral at the central bank in order to be allowed to overdraft their account. The Federal Reserve currently charges an explicit fee for daylight overdrafts to compensate for credit risk. We now investigate how reserve demand changes in the basic framework when access to daylight credit is costly. Suppose a bank sends its daytime payment, PD , before receiving the incoming payment. If PD is larger than R (the bank’s reserve holdings), the bank’s account will be overdrawn until the offsetting payment arrives. Let re denote the interest rate the central bank charges on daylight credit, δ denote the time period between the two payment flows during the day, and π denote the probability that a bank sends the outgoing payment before receiving the incoming one. Then the bank’s expected cost of daylight credit is π re δ (PD − R). This expression shows that an additional dollar of reserve holdings will decrease the bank’s expected cost of daylight credit by π re δ. In this way, the terms at which the central bank offers daylight credit can influence the bank’s choice of reserve position.10 Figure 3 depicts a bank’s demand for reserves when daylight credit is costly (that is, when re > 0). The case studied in Figure 1 (that is, when re = 0) is included in the figure for reference. It is still true that there will be no demand for reserves if the market rate is above the penalty rate rP . The interest rate measured on the vertical axis is (as in all of our figures) the rate for a 24-hour loan. If the market rate were above the penalty rate, a bank would prefer to lend out all of its reserves at the (high) market rate and satisfy its requirements by borrowing at the penalty rate. By arranging these loans to settle at approximately the same time on both days, this plan would have no 9 One possibility is that large banks face a wider support of the shock because of their larger operations, but face a smaller variance because of economies of scale in reserve management. This distinction cannot be captured in the figures here, which are drawn under the assumption that the distribution of the payment shock is uniform. For other distributions, the variance generally plays a more important role in the analysis than the support. 10 The treatment of overnight reserves can, in turn, influence the level of daylight credit usage. See Ennis and Weinberg (2007) for an investigation of this effect in a closely-related framework. See, also, the discussion in Keister, Martin, and McAndrews (2008). H. M. Ennis and T. Keister: Monetary Policy Implementation 249 Figure 3 Daylight Credit Fees r r P No Fee Positive Fee rT re 0 K-P S S K+P P R effect on the bank’s daylight credit usage and, hence, would generate a pure profit. It is also still true that whenever the market rate is below the penalty rate, the bank will choose to hold at least K − P reserves, since otherwise it would be certain to need a discount window loan to meet its requirement. As the figure shows, the downward-sloping part of the demand curve is flatter when daylight credit is costly. For any market interest rate below the discount rate, the bank will choose to hold a higher quantity of reserves because these reserves now have the added benefit of reducing daylight credit fees. Rather than decreasing all the way to the horizontal axis as in Figure 1, the demand curve now becomes flat at the bank’s expected marginal cost of intraday funds, π re δ. As long as R is smaller than PD , the bank would not be willing to lend out funds at an interest rate below π re δ, because the expected increase in daylight credit fees would be more than the interest earned on the loan. For values of R larger than PD , the bank is holding sufficient reserves to cover all of its intraday payments and the demand curve drops to the horizontal axis.11 11 The analysis here assumes a particular form of daylight credit usage; if an overdraft occurs, the size of the overdraft is constant over time. Alternative assumptions about the process of daytime payments would lead to minor changes in the figure, but the qualitative properties would be largely 250 Federal Reserve Bank of Richmond Economic Quarterly As the figure shows, when daylight credit is costly, the level of reserves required to implement a given target rate is higher (S2 rather than S1 in the diagram). In other words, costly daylight credit tends to increase banks’ reserve holdings. The demand curve is also flatter, meaning that reserve holdings are more sensitive to changes in the interest rate. 3. INTEREST RATE VOLATILITY One of the key determinants of a central bank’s ability to consistently achieve its target interest rate is the slope of the aggregate demand curve for reserves. In this section, we describe the relationship between this slope and the volatility of the market interest rate in the basic framework. The next two sections then discuss policy tools that can be used to limit this volatility. While the central bank can use open market operations to affect the supply of reserves available in the market, it typically cannot completely control this supply. Payments into and out of the Treasury account, as well as changes in the amount of cash in circulation, also affect the total supply of reserves. The central bank can anticipate much of the change in such autonomous factors, but there will often be significant unanticipated changes that cause the total supply of reserves to be different from what the central bank intended. As is clear from Figure 1, if the supply of reserves ends up being different from the intended amount, ST , the market interest rate will deviate from the target rate, rT . Figure 4 illustrates the fact that a flatter demand curve for reserves is associated with less volatility in the market interest rate, given a particular level of uncertainty associated with autonomous factors. Suppose this uncertainty implies that, after a given open market operation, the total supply of reserves will be equal to either S or S in the figure. With the steeper (thick) demand curve, this uncertainty about the supply of reserves leads to a relatively wide range of uncertainty about the market rate. With the flatter (thin) demand curve, in contrast, the variation in the market rate is smaller. For this reason, the slope of the demand curve, and those policies that affect the slope, are important determinants of the observed degree of volatility of the market interest rate around the target. As discussed in the previous section, a variety of factors affect the slope of the aggregate demand for reserves. Figure 4 can be viewed, for example, as comparing a situation where all banks face relatively little late-day uncertainty with one where all banks face more uncertainty; the latter case corresponds unaffected. The analysis also takes the size and timing of payments as given. Several papers have studied the interesting question of how banks respond to incentives in choosing the timing of their outgoing payments and, hence, their daylight credit usage. See, for example, McAndrews and Rajan (2000) and Bech and Garratt (2003). H. M. Ennis and T. Keister: Monetary Policy Implementation 251 Figure 4 Interest Rate Volatility r rP rH High Volatility Low Volatility rL rL 0 S S R to the thin line in the figure. However, it should be clear that the reasoning presented above does not depend on this particular interpretation. The exact same results about interest rate volatility would obtain if the demand curves had different slopes because banks face different penalty rates in the two scenarios or because of some other factor(s). What the figure shows is that there is a direct relationship between the slope of the demand curve and the amount of interest rate volatility caused by forecast errors or other unanticipated changes in the supply of reserves. Central banks generally aim to limit the volatility of the interest rate around their target level to the extent possible. For this reason, a variety of policy arrangements have been designed in an attempt to decrease the slope of the demand curve, at least in the region that is considered “relevant.” In the remainder of the article, we show how some of these tools can be analyzed in the context of our simple framework. In Section 4 we discuss reserve maintenance periods, while in Section 5 we discuss approaches that become feasible when the central bank pays interest on reserves. 4. RESERVE MAINTENANCE PERIODS Perhaps the most significant arrangement designed to flatten the demand curve for reserves is the introduction of reserve maintenance periods. In a system 252 Federal Reserve Bank of Richmond Economic Quarterly with a reserve maintenance period, banks are not required to hold a particular quantity of reserves each day. Rather, each bank is required to hold a certain average level of reserves over the maintenance period. In the United States, the length of the maintenance period is currently two weeks. The presence of a reserve maintenance period gives banks some flexibility in determining when they hold reserves to meet their requirement. In general, banks will try to hold more reserves on days in which they expect the market interest rate to be lower and fewer reserves on days when they expect the rate to be higher. This flexibility implies that a bank’s reserve holdings will tend to be more responsive to changes in the interest rate on any given day. In other words, having a reserve maintenance period tends to make the demand curve flatter, at least on days prior to the last day of the maintenance period. We illustrate this effect by studying a two-day maintenance period in the context of the simple framework. We then briefly explain how the same logic applies to longer periods. A Two-Day Maintenance Period Let K denote the average daily requirement so that the total requirement for the two-day maintenance period is 2K. The derivation of the demand curve for reserves on the second (and final) day of the maintenance period follows exactly the same logic as in our benchmark case. The only difference with Figure 1 is that the reserve requirement will be given by the amount of reserves that the bank has left to hold in order to satisfy the requirement for the period. In other words, the reserve requirement on the second day is equal to 2K minus the quantity of reserves the bank held at the end of the first day. On the first day of the maintenance period, a bank’s demand for reserves depends crucially on its belief about what the market interest rate will be on the second day. Suppose the bank expects the market interest rate on the second day to equal the target rate, rT . Figure 5 depicts the demand for reserves on the first day under this assumption.12 As in the basic case presented in Figure 1, there would be no demand for reserves if the market interest rate were greater than rP . Suppose instead that the market interest rate on the first day is close to, but smaller than, the penalty rate, rP . Then the bank will want to satisfy as much of its reserve requirement as possible on the second day, when it expects the rate to be substantially lower. However, if the bank’s reserve balance after the late-day payment shock is negative, it will be forced to borrow funds at the penalty rate to avoid incurring an overnight overdraft. As long as the market rate is below the penalty rate, the bank will choose a reserve position of at least −P . Note that this reserve position represents the amount of reserves held by 12 For simplicity, Figure 5 is drawn with no discounting on the part of the bank. The effect of discounting is very small and inessential for understanding the basic logic described here. H. M. Ennis and T. Keister: Monetary Policy Implementation 253 Figure 5 A Two-Day Maintenance Period r rP rT P 0 P 2K-P 2K+P R the bank before the late-day payment shock is realized. Even if this position is negative, as would be the case when the market rate is close to rP in Figure 5, it is still possible that the bank will receive a late-day inflow of reserves such that it does not need to borrow funds at the penalty rate to avoid an overnight overdraft. However, if the bank were to choose a position smaller than −P , it would be certain to need to borrow at the penalty rate, which cannot be an optimal choice as long as the market rate is lower. For interest rates below rP , but still larger than the target rate, the bank will choose to hold some “precautionary” reserves to decrease the probability that it will need to borrow at the penalty rate. This precautionary motive generates the first downward-sloping part of the demand curve in the figure. As long as the day-one interest rate is above the target rate, however, the bank will not hold more than P in reserves on the first day. By holding P , the bank is assured that it will have a positive reserve balance after the late-day payment shock. If the bank were holding more than P on the first day, it could lend those reserves out at the (relatively high) market rate and meet its requirement by borrowing reserves on the second day in the event that the interest rate is expected to be at the (lower) target rate, yielding a positive profit. Hence, the first downward-sloping part of the demand curve must end at P . Now suppose the first-day interest rate is exactly equal to the target rate, rT . In this case, the bank expects the rate to be the same on both days and is, 254 Federal Reserve Bank of Richmond Economic Quarterly therefore, indifferent between holding reserves on either day for the purpose of meeting reserve requirements. In choosing its first-day reserve position, the bank will consider the following issues. It will choose to hold at least enough reserves to ensure that it will not need to borrow at the penalty rate at the end of the first day. In other words, reserve holdings will be at least as large as the largest possible payment P . The bank is willing to hold more reserves than P for the purpose of satisfying some of its requirement. However, it wants to avoid the possibility of over-satisfying the requirement on the first day (that is, becoming “locked-in”), since it must hold a non-negative quantity of reserves on the second day to avoid an overnight overdraft. This implies that the bank will not be willing to hold more than the total requirement, 2K, minus the largest possible payment inflow, P , on the first day. The demand curve is flat between these two points (that is, P and 2K −P ), indicating that the bank is indifferent between the various levels of reserves in this interval. Finally, suppose the market interest rate on the first day is smaller than the target rate. Then the bank wants to satisfy most of the requirement on the first day, since it expects the market rate to be higher on the second day. In this case, the bank will hold at least 2K − P reserves on the first day. If it held any less than this amount, it would be certain to have some requirement remaining on the second day, which would not be an optimal choice given that the rate will be higher on the second day. As the interest rate moves farther below the target rate, the bank will hold more reserves for the usual precautionary reasons. In this case, the bank is balancing the possibility of being locked-in after the first day against the possibility of needing to meet some of its requirement on the more-expensive second day. The larger the difference between the rates on the two days, the larger the quantity the bank will choose to hold on the first day. This trade-off generates the second downward-sloping part of the demand curve. The intermediate flat portion of the demand curve in Figure 5 can help to reduce the volatility of the interest rate on days prior to the settlement day. As long as movements in autonomous factors are small enough such that the supply of reserves stays in this portion of the demand curve, interest rate fluctuations will be minimal. For a central bank that is interested in minimizing volatility around its target rate, this represents a substantial improvement over the situation depicted in Figure 1.13 13 It should be noted that Figure 5 is drawn under the assumption that the reserve requirement is relatively large. Specifically, K > P is assumed to hold, so that the total reserve requirement for the period, 2K, is larger than the width of the support of the late-day payment shock, 2P . If this inequality were reserved, the flat portion of the demand curve would not exist. In general, reserve maintenance periods are most useful as a policy tool when the underlying reserve requirements are sufficiently large relative to the end-of-day balance uncertainty. H. M. Ennis and T. Keister: Monetary Policy Implementation 255 There are, however, some issues that make implementing the target rate through reserve maintenance periods more difficult than a simple interpretation of Figure 5 might suggest. First, the position of the flat portion of the demand curve at the exact level of the target rate depends on the central bank’s ability to hit the target rate (on average) on settlement day. If banks expected the settlement-day interest rate to be lower than the current target, for example, the flat portion of the first-day demand curve would also lie below the target. This issue is particularly problematic when market participants expect the central bank’s target rate to change during the course of a reserve maintenance period. A second difficulty is that the flat portion of the demand curve disappears on the settlement day and the curve reverts to that in Figure 1.14 This feature of the model indicates why market interest rates are likely to be more volatile on settlement days. Multiple-Day Maintenance Periods Maintenance periods with three or more days can be easily analyzed in a similar way. Consider, for example, the case of a three-day maintenance period with an average daily requirement equal to K. As before, suppose that the central bank is expected to hit the target rate on the subsequent days of the maintenance period and consider the demand for reserves on the first day. This demand will be flat between the points P and 3K − P . That is, the demand curve will be similar to that plotted in Figure 5, but the flat portion will be wider. To determine the shape of the demand curve for reserves on the second day we need to know how many reserves the bank held on the first day of the maintenance period. Suppose the bank held R1 reserves with R1 < 3K. Then on the second day of the maintenance period, the demand curve for reserves would be flat between the points P and 3K − R1 − P . Hence, we see that as the bank approaches the final day of the maintenance period, the flat portion of its demand curve is likely to become smaller, potentially opening the door to increases in interest rate volatility. For the interested reader, Bartolini, Bertola, and Prati (2002) provide a more thorough analysis of the implications of multiple-day maintenance periods on the behavior of the overnight market interest rate using a model similar to, but more general than, ours. 14 In practice, central banks often use carryover provisions in an attempt to generate a small flat region in the demand curve on a settlement day. Another alternative would be to stagger the reserve maintenance periods for different groups of banks. This idea goes back to as early as the 1960s (see, for example, the discussion between Sternlight 1964 and Cox and Leach 1964 in the Journal of Finance). One common argument against staggering the periods is that it could make the task of predicting reserve demand more difficult. Whether the benefits of reducing settlement day variability outweigh the potential costs of staggering is difficult to determine. 256 5. Federal Reserve Bank of Richmond Economic Quarterly PAYING INTEREST ON RESERVES We now introduce the possibility that the central bank pays interest on the reserve balances held overnight by banks in their accounts at the central bank. As discussed in Section 1, most central banks currently pay interest on reserves in some form, and Congress has authorized the Federal Reserve to begin doing so in October 2011. The ability to pay interest on reserves gives a central bank an additional policy tool that can be used to help minimize the volatility of the market interest rate and steer this rate to the target level. This tool can be especially useful during periods of financial distress. For example, during the recent financial turmoil, the fed funds rate has experienced increased volatility during the day and has, in many cases, collapsed to values near zero late in the day. As we will see below, the ability to pay interest on reserves allows the central bank to effectively put a floor on the values of the interest rate that can be observed in the market. Such a floor reduces volatility and potentially increases the ability of the central bank to achieve its target rate. In this section, we describe two approaches to monetary policy implementation that rely on paying interest on reserves: an interest rate corridor and a system with clearing bands. We explain the basic components of each approach and how each tends to flatten the demand curve for reserves. Interest Rate Corridors One simple policy a central bank could follow would be to pay a fixed interest rate, rD , on all reserve balances that a bank holds in its account at the central bank.15 This policy places a floor on the market interest rate: No bank would be willing to lend reserves at an interest rate lower than rD , since they could instead earn rD by simply holding the reserves on deposit at the central bank. Together, the penalty rate, rP , and the deposit rate, rD , form a “corridor” in which the market interest rate will remain.16 Figure 6 depicts the demand for reserves under a corridor system. As in the earlier figures, there is no demand for reserves if the market interest rate is higher than the penalty rate, rP . For values of the market interest rate below rP , a bank will choose to hold at least K − P reserves for exactly the same 15 In practice, reserve balances held to meet requirements are often compensated at a different rate than those that are held in excess of a bank’s requirement. For the daily process of targeting the overnight market interest rate, the rate paid on excess reserves is what matters; this is the rate we denote rD in our analysis. 16 A central bank may prefer to use a lending facility that is distinct from its discount window to form the upper bound of the corridor. Banks may be reluctant to borrow from the discount window, which serves as a lender of last resort, because they fear that others would interpret this borrowing as a sign of poor financial health. The terms associated with the lending facility could be designed to minimize this type of stigma effect and, thus, create a more reliable upper bound on the market interest rate. H. M. Ennis and T. Keister: Monetary Policy Implementation 257 Figure 6 A Conventional Corridor r r P Demand for Reserves r T rD 0 K-P K S T K+P R reason as in Figure 1: if it held a lower level of reserves, it would be certain to need to borrow at the penalty rate, rP . Also as before, the demand for reserves is downward-sloping in this region. The big change from Figure 1 is that the demand curve now becomes flat at the deposit rate. If the market rate were lower than the deposit rate, a bank’s demand for reserves would be essentially infinite, as it would try to borrow at the market rate and earn a profit by simply holding the reserves overnight. The figure shows that, regardless of the level of reserve supply, S, the market interest rate will always stay in the corridor formed by the rates rP and rD . The width of the corridor, rP − rD , is then a policy choice. Choosing a relatively narrow corridor will clearly limit the range and volatility of the market interest rate. Note that narrowing the corridor also implies that the downward-sloping part of the demand curve becomes flatter (to see this, notice that the boundary points K − P and K + P do not depend on rP or rD ). Hence, the size of the interest rate movement associated with a given shock to an autonomous factor is smaller, even when the shock is small enough to keep the rate within the corridor. An interesting case to consider is one in which the lending and deposit rates are set the same distance on either side of the target rate (x basis points above and below the target, respectively). This system is called a symmetric 258 Federal Reserve Bank of Richmond Economic Quarterly corridor. A change in policy stance that involves increasing the target rate, then, effectively amounts to changing the levels of the lending and deposit rates, which shifts the demand curve along with them. The supply of reserves needed to maintain a higher target rate, for example, may not be lower. In fact— perhaps surprisingly—in the simple model studied here, the target level of the supply of reserves would not change at all when the policy rate changes. If the demand curve in Figure 6 is too steep to allow the central bank to effectively achieve its goal of keeping the market rate close to the target, a corridor system could be combined with a reserve maintenance period of the type described in Section 4. The presence of a reserve maintenance period would generate a flat region in the demand curve as in Figure 5. The features of the corridor would make the two downward-sloping parts of the demand curve in Figure 5 less steep, which would limit the interest rate volatility associated with events where reserve supply exits the flat region of the demand curve, as well as on the last day of the maintenance period when the flat region is not present. Another way to limit interest rate volatility is for the central bank to set the deposit rate equal to the target rate and then provide enough reserves to make the supply, ST , intersect the demand curve well into the flat portion of the demand curve at rate rD . This “floor system” has been recently advocated as a way to simplify monetary policy implementation (see, for example, Woodford 2000, Goodfriend 2002, and Lacker 2006). Note that such a system does not rely on a reserve maintenance period to generate the flat region of the demand curve, nor does it rely on reserve requirements to induce banks to hold reserves. To the extent that reserve requirements, and the associated reporting procedures, place significant administrative burdens on both banks and the central bank, setting the floor of the corridor at the target rate and simplifying, or even eliminating, reserve requirements could potentially be an attractive system for monetary policy implementation. It should be noted, however, that the market interest rate will always remain some distance above the floor in such a system, since lenders in the market must be compensated for transactions costs and for assuming some counterparty credit risk. In other words, in a floor system the central bank is able to fully control the risk-free interest rate, but not necessarily the market rate. In normal times, the gap between the market rate and the rate paid on reserves would likely be stable and small. In periods of financial distress, however, elevated credit risk premia may drive the average market interest rate significantly above the interest rate paid on reserves. Our simple model abstracts from these important considerations.17 17 The central bank could also set an upper limit for the quantity of reserves on which it would pay the target rate of interest to a bank; reserves above this limit would earn a lower rate (possibly zero). Whitesell (2006) proposed that banks be allowed to choose their own upper H. M. Ennis and T. Keister: Monetary Policy Implementation 259 Clearing Bands Another approach to generating a flat region in the demand curve for reserves is the use of daily clearing bands. This approach does not rely on a reserve maintenance period. Instead, the central bank pays interest on a bank’s reserve holdings at the target rate, rT , as long as those holdings fall within a pre-specified band. Let K and K denote the lower and upper bounds of this band, respectively. If the bank’s reserve balance falls below K, it must borrow at the penalty rate, rP , to bring its balance up to at least K. If, on the other hand, the bank’s reserve balance is higher than K, it will earn the target rate, rT , on all balances up to K but will earn a lower rate, rE , beyond that bound. The demand curve for reserves under such a system is depicted in Figure 7. The figure is drawn under the assumption that the clearing band is fairly wide relative to the support of the late-day payment shock. In particular, we assume that K + P < K − P . Let us call the interval K + P , K − P the “intermediate region” for reserves. By choosing any level of reserves in this intermediate region, a bank can ensure that its end-of-day reserve balance will fall within the clearing band. The bank would then be sure that it will earn the target rate of interest on all of the reserves it ends up holding overnight. When the market interest rate is equal to the target rate, rT , a bank is indifferent between choosing any level of reserves in the intermediate region. For example, if the bank borrows in the market to slightly increase its reserve holdings, the cost it would pay in the market for those reserves would be exactly offset by the extra interest it would earn from the central bank. Similarly, lending out reserves to slightly decrease the bank’s holdings would also leave profit unchanged. This reasoning shows that the demand curve for reserves will be flat in the intermediate region between K + P and K − P . As long as the central bank is able to keep the supply of reserves within this region, the market interest rate will equal the target rate, rT , regardless of the exact level of reserve supply. Outside the intermediate region, the logic behind the shape of the demand curve is very similar to that explained in our benchmark case. When the market interest rate is higher than rT , a bank can earn more by lending reserves in the market than by holding them on deposit at the central bank. It would, therefore, prefer not to hold more than the minimum level of reserves needed to avoid being penalized, K. Of course, the bank would be willing to hold some precautionary reserves to guard against the possibility that the lateday payment shock will drive their reserve balance below K. The quantity of precautionary reserves it would choose to hold is, as before, an inverse function of the market interest rate; this reasoning generates the first downward-sloping part of the demand curve in Figure 7. limits by paying a facility fee per unit of capacity. Such an approach leads to a demand curve for reserves that is flat at the target rate over a wide region. 260 Federal Reserve Bank of Richmond Economic Quarterly Figure 7 A Clearing Band r rP rT r E 0 K-P K+P K-P K+P R When the market rate is below rT , on the other hand, the bank would like to take full advantage of its ability to earn the target interest rate by holding reserves at the central bank. It would, however, take into consideration the possibility that a late-day inflow of funds will leave it with a final balance higher than K, in which case it would earn the lower interest rate, rE , on the excess funds. The resulting decision process generates a downward-sloping region of the demand curve between the rates rT and rE . As in Figure 6, the demand curve never falls below the interest rate paid on excess reserves (now labeled rE ); thus, this rate creates a floor for the market interest rate. The demand curve in Figure 7 has the same basic shape as the one generated by a reserve maintenance period, which was depicted in Figure 5. It is important to keep in mind, however, that the forces generating the flat portion of the demand curve in the intermediate region are fundamentally different in the two cases. The reserve maintenance period approach relies on intertemporal arbitrage: banks will want to hold more reserves on days when the market interest rate is low and fewer reserves when the market rate is high. This activity will tend to equate the current market interest rate to the expected future rate (as long as the supply of reserves is in the intermediate region). The clearing band system relies instead on intraday arbitrage to generate the flat portion of the demand curve: banks will want to hold more reserves when H. M. Ennis and T. Keister: Monetary Policy Implementation 261 the market interest rate is low, for example, simply to earn the higher interest rate paid by the central bank. The intertemporal aspect of reserve maintenance periods has two clear drawbacks. First, if—for whatever reason—the expected future rate differs from the target rate, rT , it becomes difficult for the central bank to achieve the target rate in the current period. Second, large shocks to the supply of reserves on one day can have spillover effects on subsequent days in the maintenance period. If, for example, the supply of reserves is unusually high one day, banks will satisfy an unusually large portion of their reserve requirements and, as a result, the flat portion of the demand curve will be smaller on all subsequent days, increasing the potential for rate volatility on those days. The clearing band approach, in contrast, generates a flat portion in the demand curve that always lies at the current target interest rate, even if market participants expect the target rate to change in the near future. Moreover, the width of the flat portion is “reset” every day; it does not depend on past events. These features are important potential advantages of the clearing band approach. We should again point out, however, that our simple model has abstracted from transaction costs and credit risk. As with the floor system discussed above, these considerations could result in the average market interest rate being higher than the rate rT , as the latter represents a risk-free rate. 6. CONCLUSION A recent change in legislation that allows the Federal Reserve to pay interest on reserves has renewed interest in the debate over the most effective way to implement monetary policy. In this article, we have provided a basic framework that can be useful for analyzing the main properties of the various alternatives. While we have conducted all our analysis graphically, our simplifying assumptions permit a fairly precise description of the alternatives and their effectiveness at implementing a target interest rate. Many extensions of our basic framework are possible and we have analyzed several of them in this article. However, some important issues remain unexplored. For example, we only briefly mentioned the difficulties that fluctuations in aggregate credit risk can introduce in the implementation process. Also, as the debate continues, new questions will arise. We believe that the framework introduced in this article can be a useful first step in the search for much-needed answers. 262 Federal Reserve Bank of Richmond Economic Quarterly REFERENCES Ashcraft, Adam, James McAndrews, and David Skeie. 2007. “Precautionary Reserves and the Interbank Market.” Mimeo, Federal Reserve Bank of New York (July). Bartolini, Leonardo, Giuseppe Bertola, and Alessandro Prati. 2002. “Day-To-Day Monetary Policy and the Volatility of the Federal Funds Interest Rate.” Journal of Money, Credit, and Banking 34 (February): 137–59. Bech, M. L., and Rod Garratt. 2003. “The Intraday Liquidity Management Game.” Journal of Economic Theory 109 (April): 198–219. Clouse, James A., and James P. Dow, Jr. 2002. “A Computational Model of Banks’ Optimal Reserve Management Policy.” Journal of Economic Dynamics and Control 26 (September): 1787–814. Cox, Albert H., Jr., and Ralph F. Leach. 1964. “Open Market Operations and Reserve Settlement Periods: A Proposed Experiment.” Journal of Finance 19 (September): 534–9. Coy, Peter. 2007. “A ‘Stealth Easing’ by the Fed?” BusinessWeek. http://www.businessweek.com/investor/content/aug2007/ pi20070817 445336.htm [17 August]. Dotsey, Michael. 1991. “Monetary Policy and Operating Procedures in New Zealand.” Federal Reserve Bank of Richmond Economic Review (September/October): 13–9. Ennis, Huberto M., and John A. Weinberg. 2007. “Interest on Reserves and Daylight Credit.” Federal Reserve Bank of Richmond Economic Quarterly 93 (Spring): 111–42. Goodfriend, Marvin. 2002. “Interest on Reserves and Monetary Policy.” Federal Reserve Bank of New York Economic Policy Review 8 (May): 77–84. Guthrie, Graeme, and Julian Wright. 2000. “Open Mouth Operations.” Journal of Monetary Economics 46 (October): 489–516. Hilton, Spence, and Warren B. Hrung. 2007. “Reserve Levels and Intraday Federal Funds Rate Behavior.” Federal Reserve Bank of New York Staff Report 284 (May). Keister, Todd, Antoine Martin, and James McAndrews. 2008. “Divorcing Money from Monetary Policy.” Federal Reserve Bank of New York Economic Policy Review 14 (September): 41–56. H. M. Ennis and T. Keister: Monetary Policy Implementation 263 Lacker, Jeffrey M. 2006. “Central Bank Credit in the Theory of Money and Payments.” Speech. http://www.richmondfed.org/news and speeches/ presidents speeches/index.cfm/2006/id=88/pdf=true. McAndrews, James, and Samira Rajan. 2000. “The Timing and Funding of Fedwire Funds Transfers.” Federal Reserve Bank of New York Economic Policy Review 6 (July): 17–32. Poole, William. 1968. “Commercial Bank Reserve Management in a Stochastic Model: Implications for Monetary Policy.” Journal of Finance 23 (December): 769–91. Sternlight, Peter D. 1964. “Reserve Settlement Periods of Member Banks: Comment.” Journal of Finance 19 (March): 94–8. Whitesell, William. 2006. “Interest Rate Corridors and Reserves.” Journal of Monetary Economics 53 (September): 1177–95. Woodford, Michael. 2000. “Monetary Policy in a World Without Money.” International Finance 3 (July): 229–60. Economic Quarterly—Volume 94, Number 3—Summer 2008—Pages 265–300 CEO Compensation: Trends, Market Changes, and Regulation Arantxa Jarque C ompensation figures for the top managers of large firms are on the news frequently. Newspapers report the salaries, the bonuses, and the profits from selling stock options of the highest paid executives, often under headlines suggesting excessive levels of pay or a very weak relation of pay to the performance of the firms.1 Especially after the fraud scandals at Enron and other important corporations around the world, executive performance and pay have been carefully scrutinized by the public. Academic economists, however, have long ago recognized the importance of understanding the issues involved in determining executive pay and have been studying them for decades. In short, the main economic problem behind executive compensation design is that firm owners need to align the incentives of the executives to their own interests—typically to maximize firm value. To achieve this alignment, the compensation of the manager is usually made contingent on the performance of the firm. While in the largest firms executive compensation typically represents a small fraction of the total firm value, the decisions that a top executive makes can be potentially important for firm performance. Therefore, the way in which the dependence of compensation on firm performance is structured can have a significant impact on the added value that the executive brings to the firm. There is by now a large body of academic studies that document practices in executive pay and study the The author would like to thank Brian Gaines, Borys Grochulski, Leornardo Martinez, and John Weinberg for helpful comments. Jarque is an assistant professor at the Universidad Carlos III de Madrid. This article was written while the author was a visiting economist at the Federal Reserve Bank of Richmond. The views expressed in this article do not necessarily represent the views of the Federal Reserve Bank of Richmond or the Federal Reserve System. E-mail: ajarque@eco.uc3m.es. 1 Some recent examples, following the release of an Associated Press study for 2007, include Beck and Fordahl (2008a, 2008b) and Associated Press (2008). 266 Federal Reserve Bank of Richmond Economic Quarterly optimal design of compensation contracts. Some of the most important and recent findings in the literature are summarized in this article. In Section 1, I present the various instruments commonly used to compensate executives and the main empirical regularities about executive pay. Over the last two decades, the average pay of a chief executive officer (CEO) working in one of the 500 largest firms in the United States has increased six-fold. This increase has occurred simultaneously with a change in the composition of pay of these CEOs, moving away from salary and increasingly toward performance-based compensation in the form of stock grants and stock option grants. This shift has resulted in a clear increase in the sensitivity of the total pay of CEOs to the performance of their firms. Although the increase in the level and the sensitivity are most likely related, I separate the discussion into two sections for a more detailed exposition of the evidence and a discussion of possible explanations. In Section 2, I summarize how the level of pay has evolved since 1936. Then I briefly review some of the explanations in the academic literature for the sharp increase in the last two decades. I focus on a recent strand of the literature that has built on ideas from the classic papers of Lucas (1978) and Rosen (1981, 1982) that model firms’ competition for the scarce talent of managers. These studies argue that the sharp increase in firm size and value of the last decades could be important to explain the six-fold increase in the level of CEO pay over the same period. In Section 3, I focus on how the pay of CEOs depends on their performance. I introduce several measures of sensitivity of the CEO’s pay to the results of the firm that are widely used in the literature. The empirical studies provide a wide range of estimates for sensitivity, but there is consensus about two facts: an increase in sensitivity over the last two decades and a negative relation of sensitivity with firm size (as measured by market capitalization value). I discuss some of the recent explanations for these regularities. Many of the explanations are based on studying the interaction between the manager and the owners of the firm as a moral hazard problem. In the model, shareholders minimize the cost of bringing the (risk averse) CEO to exert an unobservable effort that improves the results of the firm. This analysis is typically performed in partial equilibrium and aims to describe the form of optimal compensation contracts, i.e., the optimal sensitivity of pay to firm performance. Some recent studies suggest that a seemingly low sensitivity of pay to performance, as well as the negative relation between sensitivity of pay and firm size, could be features of dynamic optimal contracts that are used to solve the moral hazard problem. Also in this moral hazard framework, several recent articles demonstrate the efficiency of some seemingly unintuitive pay practices that are often discussed by the media, such as bonuses in bad earning years or repricing of out-of-the-money options. The level of market competition across firms, as well as the relative demand of general A. Jarque: CEO Compensation 267 versus firm-specific skills, has also been shown to be empirically significant in explaining pay trends. In Section 4, I describe the main regulatory changes affecting executive compensation in the last 15 years: changes in personal, corporate, and capital gains taxes; new limits on the deductibility of CEO pay expenses that favor performance-based compensation; an increase in the disclosure requirements about CEO pay; a standardization of the expensing of option grants; and several initiatives fostering shareholders’ activism and independence of the board of directors. Amidst the headlines on excessive pay, the popular press has been debating these changes in regulation, their potential role in the recent rise in pay, and the need for new government intervention.2 To shed some light on the role of regulation, I review the findings of academic studies that have rigorously tried to quantify the effect of several specific measures on the level and sensitivity of compensation. As it turns out, these studies find little evidence of a sizable effect on pay practices coming from tax advantages or salary caps. Following the main regulatory changes, some studies find evidence of a small shift of compensation from salaries and bonuses toward performance-based compensation (stocks and options), which translates into a slight increase in the sensitivity of pay to performance. Better corporate governance and the increase in the proportion of institutional shareholders appear to be associated with higher company returns and higher incentives for CEOs. This suggests that regulation efforts to improve corporate governance and transparency have been moving in the right direction, although it is difficult to evaluate the relative importance of regulation versus the marketinduced changes in governance practices. Designing executive compensation packages is, no doubt, complicated. Judging the appropriateness of those packages is, consequently, a very difficult task in which models and sophisticated econometric tools are a necessity. I now proceed to review the most recent attempts at this task and their conclusions. 1. UNDERSTANDING AND MEASURING CEO COMPENSATION Today companies pay their top executives through some or all of the following instruments: a salary, a bonus program, stock grants (usually with restrictions 2 On one hand, in the most recent special report on CEO compensation published by the Wall Street Journal (2008), one of the articles is dedicated to the unintended consequences of regulation changes. On the other hand, the “say on pay” proposals, which seek to force boards of directors to have a (nonbinding) vote on the compensation plan that they design for a CEO, have been receiving particular attention in the press—see The Economist (2008b, “Pay Attention”) on the recent trend of pay in Europe and the regulation responses of some European governments, and The Economist (2008a, “Fair or Foul”) on the “say on pay” proposals both in Europe and the United States, which cites both U.S. presidential candidates as being in favor of forcing the shareholders’ vote. 268 Federal Reserve Bank of Richmond Economic Quarterly on the ability to sell them), grants of options on the stock of the firm, and long-term incentive plans that specify retirement and severance payments, as well as pension plans and deferred benefits. The most accepted explanation for these variable compensation instruments is the existence of a moral hazard problem: The separation of ownership and control of the firm implies the need to provide incentives to the CEO that align his interests with those of the firm owners. In the presence of moral hazard, the optimal contract prescribes that the pay of the executive should vary with the results of the firm. However, in spite of the need for incentives, limited funds on the part of CEOs or risk aversion considerations imply that exposing the CEO to the same risk as shareholders is typically either an unfeasible or inefficient arrangement. The optimal contract should balance incentives and insurance. Therefore, part of the compensation should be variable and could be provided through stock or stock options, while part of the compensation, such as the annual salary, should not be subject to risk, providing some insurance to the CEO against bad firm performance due to factors that he cannot control. Data on the level of annual compensation of workers classified as CEOs are available from the Bureau of Labor Statistics (BLS). These are wages representative of the whole economy. However, details are not available on the specific forms of the contract (i.e., the compensation instruments that were used in providing that compensation). Data from the BLS show that the average CEO earns an annual wage of $151,370 (May 2007)—less than the average doctor (internist) and about $25,000 more per year than a lawyer. The annual wage of the average CEO today is about 3.5 times that of the average worker in the economy, and the evolution of this comparison over the last seven years has followed a similar pattern as that of other white-collar professions. However, the distribution of wages of CEOs is extremely skewed. Figure 1 shows the evolution of the total level of pay for the CEOs of the 500 largest firms in the economy, according to the figures in Forbes, a magazine that publishes surveys of top CEO pay annually. The average annual wage of this sample of executives in 2007 was about $15 million, approximately 300 times that of the average worker in the U.S. economy. For the top 500 firms, information about CEO pay is readily available. Since these companies are public, they file their proxy statements with the Securities and Exchange Commission (SEC), making their reports on their top executives’compensation public. Most academic studies restrict themselves to studying the compensation of the CEOs in this subsample of firms. Therefore, I too concentrate on this subsample in the rest of the article. From this point on, the acronym “CEO” refers to the chief executive officer of one of the top 500 firms in the United States. Figure 1 also shows the evolution of the three main components of CEO pay from 1989 to 2007. Stock gains refers to the value realized during the given fiscal year by exercising vested options granted in previous years. The A. Jarque: CEO Compensation 269 Figure 1 Compensation Data Based on Forbes Magazine’s Annual Survey (2007 Constant Dollars) 16 Salary and Bonus Stock Gains 14 4.4 Other Compensation 12 $ Millions 4.1 3.2 3.2 1.3 1.9 10 8 2.2 1.3 6 1.3 1.1 0.6 0.7 4 0.2 0.2 0.5 0.4 0.7 0.6 2.2 7.5 6.3 7.4 1.3 1.7 6 5.9 6.3 6.3 5 3.2 3 3.1 1.2 1.8 1.7 0.6 3.3 3.3 3.5 3.6 3.3 2.6 2.6 2.9 2.2 2.1 2.1 2.6 1.6 1.8 1.9 1.5 1.5 1.5 1.4 1.5 2 0.6 0.6 0.6 0.8 0 1989 1991 1993 1995 1997 1999 2001 2003 2005 2007 gain is the difference between the stock price on the date of exercise and the exercise price of the option. Other compensation includes long-term incentive payouts and the value realized from vesting of restricted stock and performance shares, and other perks such as premiums for supplemental life insurance, annual medical examinations, tax preparation and financial counseling fees, club memberships, security services, and the use of corporate aircraft. The main trend that we observe in this figure is that, although annual salaries have been increasing, the proportion of total pay that they represent has decreased in the last 20 years, while compensation through options has become the most important component, increasing from a low of 35 percent to 77 percent. The total levels of pay have risen six-fold.3 The shift of compensation practices toward options and restricted stocks has one immediate implication: the total compensation of the CEO is now 3 The data is from Forbes magazine’s annual survey. Prior to 1992, the data on options is a record of the gains from exercise. After 1992, Forbes records the value of the option grants at the date of granting, using the valuation method that the firm uses in its proxy statement. This method is mostly the Black and Scholes formula, for which the company reports the parameters as established in the SEC disclosure rules of 1993. See Section 4 of this article on regulation. Also, note that good data on pensions is not available and most studies leave it out of their analysis. 270 Federal Reserve Bank of Richmond Economic Quarterly Figure 2 Annual Compensation (Forbes Magazine) and Dow Jones Industrial Average During the Previous Year 16 Total Compensation Dow Jones Industrial Average (Monthly) 14,000 14 12,000 12 $ Millions 10,000 10 8,000 8 6,000 6 4,000 4 2,000 2 2007 2005 2006 2004 2003 2001 2002 2000 1999 1998 1996 1997 1995 1993 1994 1991 1990 1989 1988 1992 0 0 more closely tied to the performance of the firm than it was in the 1980s. In Figure 2, I plot the Forbes measure of total compensation from Figure 1, together with the Dow Jones Industrial Average during the previous year, as an approximation for the average performance of the firms managed by the CEOs included in the compensation surveys. Apparently, on average, the compensation of top CEOs is closely correlated with the performance of the top firms that they manage. A more detailed analysis is needed to determine this relation at the individual firm level. In the next two sections, I review the empirical literature that has performed a detailed analysis, documenting the trends in the level and sensitivity of pay. In a widely cited paper, Jensen and Murphy (1990a) studied changes in compensation facts from 1974 to 1986. Following this paper, many studies on pay have been published. Rosen (1992) provided a first survey of the empirical findings on CEO compensation and Murphy (1999) is a very complete survey of the literature in the 1990s. The general picture, according to the most recent studies, is that the rates of growth in the level of CEO pay have increased dramatically since the 1980s, but pay has become, over the same period, more closely linked to firm performance. I present the evidence in two separate sections, one focusing on the level of pay and one on the sensitivity of pay—although, as will become clear A. Jarque: CEO Compensation 271 Table 1 Median CEO Pay Period 1936–1939 1940–1945 1946–1949 1950–1959 1960–1969 1970–1979 1980–1989 1990–1999 2000–2005 Median CEO Pay (Million $) 1.11 1.07 0.90 0.97 0.99 1.17 1.81 4.09 9.20 Avg. Annual Growth Rate −0.01 −0.04 0.01 0.00 0.02 0.04 0.08 0.14 Source: Frydman and Saks (2007). CEO pay is in constant year 2000 dollars. from the discussion, the two features are related through risk and incentive considerations. 2. THE LEVEL OF PAY The bulk of the literature on CEO compensation (see Jensen, Murphy, and Wruck 2004; Frydman and Saks 2007; Gabaix and Landier 2008) has documented that, after moderate rates of growth in the 1970s, CEO pay rose much faster in the most recent decades. Total compensation reached a median value of $9.20 million (in 2000 constant dollars) in 2005. Frydman and Saks (2007) present previously unavailable historical evidence starting in the 1930s. They find that compensation decreased sharply after World War II and it had a modest rate of growth of about 0.8 percent for the following 30 years. Hence, the recent growth in pay (see Table 1 and Figure 1), with annual growth rates above 10 percent in the period 1998–2007, represents a remarkable change from previous trends. In the first subsection, I review some estimations of the relevance and appropriateness of the levels of CEO pay that we observe today. I follow with a subsection on the proposed explanations for the time trends of the level of pay in the last century, which are divided into two groups: theories based on competitive markets and optimal contracts and theories based on inefficient practices. Are Today’s Levels of Pay Justified? According to some recent studies, the cost of CEO compensation for firms may not be so economically significant. However, the economic consequences of not providing the right incentives appear to be potentially large. Gabaix and 272 Federal Reserve Bank of Richmond Economic Quarterly Landier (2008), for example, report an average value of CEO pay over firm earnings of 0.5 percent during the period 1992–2003. Margiotta and Miller (2000), relying on a moral hazard model and historical data on compensation contracts of 34 firms from 1948 to 1977, find that companies benefit highly from providing the right level of incentives (ranging from $83 million to $263 million in 1977 dollars, depending on the industry). The observed contracts in their sample, however, imply a modest cost of providing incentives to induce high effort (about $0.5 million in 1967 dollars).4 More recently, Bennedsen, P´ rez-Gonz´ lez, and Wolfenzon (2007) use data from Danish firms from 1992 e a to 2003 to estimate the effect of the CEO’s death on firm performance. They find that the industry-adjusted operating returns on assets decrease by 1.7 percent following the absence of the CEO. They also report a decrease of 0.7 percent after a close family member of the CEO passes away. Both measures indicate the importance of CEO input into the performance of the firm, either through effort, talent, or both. Using a competitive model, Tervi¨ (2008) o estimates the talent distribution for CEOs and finds that if the managers of the 1,000 largest U.S. companies in 2004 had been substituted by a CEO of the same talent as that of the 1,000th company, the combined value of those companies would have been between $21.3 billion and $25 billion lower. His calculations imply that the contribution of the talented CEOs (net of pay) was between $16.9 billion and $20.6 billion (approximately 15 percent of the total market capitalization of the largest 1,000 firms), while the combined payments to the 1,000 CEOs that year were $7.1 billion. Explanations for the Increase in the Level of Pay The widely documented increase in the level and sensitivity of pay in the past two decades has motivated research efforts to construct robust explanations for these trends. Gabaix and Landier (2008) argue in a recent influential article that the six-fold increase in pay since the 1980s can be explained by the six-fold increase in the market value of U.S. firms in the same period. Building on work by Lucas (1978) and Rosen (1981, 1982), Gabaix and Landier’s model presents a competitive economy in which firms compete for a manager’s talent.5 In their characterization of the equilibrium, they show that the wage of the CEO increases both with the size of the firm and the average size of the firm in the economy. Using extreme value theory, they calibrate the distribution for a manager’s talent and combine this with a calibration of the size distribution of firms. Their results are consistent with the findings of Tervi¨ (2008), who o 4 Margiotta and Miller (2000) use exponential utility specifications in their empirical estimations, for which incentive considerations are independent of the level of wealth. 5 See Prescott (2003) for a related model. A. Jarque: CEO Compensation 273 independently used a competitive model to estimate the distribution of CEO talent that does not rely on extreme value theory. Although the model in Gabaix and Landier (2008) can explain trends in CEO compensation since the 1980s, recent evidence in Frydman and Saks (2007) challenges their hypothesis for earlier periods. Frydman and Saks provide a comprehensive empirical account of the evolution of CEO pay in the last century. Their database includes previously unavailable data for the period 1936–1992, hand-collected from the 10-K filings of firms with the SEC. They study an unbalanced panel of 101 firms, selected for being the largest according to market value at three different dates. They find, consistent with previous work, that the increase in CEO pay was moderate in the 1970s and picked up in the last three decades. However, their new historical evidence starting in the 1930s shows that prior to the mid-1970s, while the level of pay was positively correlated with the relative size ranking of the firm in the economy, the correlation between changes in the market value of firms and the changes in the level of compensation was very weak. In particular, their data shows that CEO pay was stagnant from the 1940s until the 1970s, a period during which there was considerable growth of firms—particularly during the 1950s and 1960s. Some authors in the literature have departed from the usual modeling approach by proposing that the observed contracts are in fact suboptimal. The following are three representative proposals of this sort. Misperception of the Cost of Options Murphy (1999) and Jensen, Murphy, and Wruck (2004) have argued that the significant increase in the use of option grants in recent years can be explained in part by the fact that compensation boards do not correctly valuate the cost of granting options. They argue that the tax and accounting advantages of options make it an attractive method of compensation.6 They claim that, in fact, options are an expensive way of compensating the CEO: risk aversion, combined with the impossibility of hedging, as well as the high percentages of human and monetary wealth that CEOs invest in their own firm, and the positive probability of being unable to exercise the options if they are fired before the options become vested, all imply that managers ought to demand a high risk premium for the options. Thus, while firms are experiencing a cost that is well approximated by the Black-Scholes value of the grants disclosed in the proxy statements and listed in the executive compensation databases, CEOs highly discount their value. Hall and Murphy (2002) explicitly model these issues and provide calculations on how the CEO’s risk aversion influences the 6 Recent changes in regulation extend the expensing requirements to grants that have an exercise price equal to or above the market price at the time of granting. 274 Federal Reserve Bank of Richmond Economic Quarterly valuation of the stock grants. They find that the increase of “risk-adjusted” pay from 1992 to 1999 has not been nearly as dramatic as that of total pay figures taken at face value. Ratchet Effect Murphy (1999) provides a description of how firms decide on the annual option grant to be awarded to executives. He points to the fact that firms typically use a measure of the average compensation given to executives in a peer group for reference. His data also indicates that 40 percent of firms have compensation plans that specify the number of options to be granted, fixed for several years, as opposed to the value of the options. In times of a growing stock market, such as the late 1990s, these two compensation practices together would result in a “rachet effect”—an escalation in total pay based on mutual benchmarking, as opposed to rewarding exceptional performance of the individual CEOs. Entrenchment Bebchuck and several coauthors (see, for example, Bebchuck and Fried [2003, 2004] and references therein) have argued that there is ample evidence suggesting that executives control the boards of directors and effectively determine their own pay. Kuhnen and Zwiebel (2008) provide a formal model of this hypothesis, which presents the problem of the manager maximizing his own pay subject to an “acceptable” outrage level of the shareholders. The mainstream literature accepts that some incentive problems remain unsolved given the current compensation and corporate practices in effect (see Holmstr¨ m and o Kaplan 2003 or Jensen, Murphy, and Wruck 2004). Yet, there is no consensus on whether the entrenchment model is a better description of real life than the classic moral hazard model, which has shareholders maximizing their value subject to the incentive constraints of CEOs (see Grossman and Hart 1983). It is also worth noticing that the entrenchment explanation per se does not help in understanding the upward trend in the level of pay observed in the last 15 years, since corporate governance practices have been improving during this period (see Holmstr¨ m and Kaplan 2003). o 3. THE SENSITIVITY OF PAY TO PERFORMANCE The literature on CEO compensation uses the term “sensitivity of pay to performance” to refer to the changes in the total pay of a CEO that are implied by a given measure of the performance of the firm. The most common measure of firm performance used in the empirical estimations of the sensitivity of pay is the change in the value to shareholders, i.e., stock price change. Theoretical studies based on moral hazard models have shown that the design of the optimal compensation scheme is a complicated task. The way in which the level A. Jarque: CEO Compensation 275 of risk aversion of undiversified CEOs decreases with their wealth (which may originate both from their firm pay and from outside sources), for example, has not been well established empirically (Haubrich 1994). Also, finding empirical evidence on the parameters of the model is difficult: How does the hidden action of the CEO affect the results of the firm and how costly is this action to the CEO? These considerations make it difficult to assess quantitatively the optimal sensitivity of direct pay. Moreover, the sensitivity itself is difficult to estimate empirically, since the pay of the CEO changes with firm performance both through direct and indirect channels: The competition for talented CEOs in the market, for example, implies that career concerns and the risk of being fired impose some incentives on the executives that are not captured by their compensation contracts. In the first subsection, I review the academic work that attempts to quantify the incentives provided through direct pay. In the second subsection, I review the work that quantifies incentives through indirect pay, consisting mainly of studies that document historical trends of the probability of employment termination for CEOs. Although there is no agreement in the literature about the right measure of sensitivity, two stylized facts are widely accepted: total (direct plus indirect) sensitivity of pay has been increasing in the last two decades and sensitivity is negatively correlated with firm size. In the third subsection, I review some of the proposed explanations for these stylized facts, as well as several models that justify the diversity of instruments in real life compensation packages. Sensitivity of Direct Pay In an influential paper, Jensen and Murphy (1990a) point out the seemingly low sensitivity of executive pay to performance: an increase in total wealth of $3.25 for each $1,000 of increase in the firm value. This measure of sensitivity is, in the data, highly negatively correlated with the size of the firm. As an alternative that allows for variation of the Jensen and Murphy (1990a) measure across firm size, other studies have estimated the elasticity of pay to firm performance. For both measures, a sharp increase in the sensitivity of pay over the last two decades has been widely documented: estimations suggest an increase in sensitivity of more than five times over that period. I now describe in detail each of the two measures and the main findings, including a discussion of the regularities with respect to firm size. The Jensen-Murphy Statistic Jensen and Murphy (1990a) use data on compensation of approximately 2,000 CEOs of large corporations collected by Forbes magazine from 1974 to 1986. In their paper, they focus on a particular measure of pay sensitivity to 276 Federal Reserve Bank of Richmond Economic Quarterly performance: They estimate how a $1,000 change in the value of a firm translates into dollar changes in the wealth of its CEO. This measure of sensitivity is often reported in the literature as the Jensen-Murphy statistic. In their paper, they run regressions of the form w = α + β ( V ) + θ (CON T ROLS) , (1) where V includes lagged as well as contemporaneous measures of changes in the dollar value of the firm. They report that the median CEO in their sample experiences an increase in his total wealth of $3.25 for each $1,000 of increase in the value of the firm he manages. Of this change in wealth, 2 cents correspond to year-to-year changes in salary and bonus, while the rest is because of changes in the value of the stock and option holdings of the CEO, the effect of performance on future wages and bonus payments, and the wealth losses associated with variations in the probability of dismissal. The authors qualify this sensitivity of pay as “low” and inconsistent with the predictions of any agency model. Murphy (1999) updates these estimates using Execucomp data up to 1996 and reports an increase in the sensitivity of pay to $6 per $1,000 by that year. Garen (1994) and Haubrich (1994) point out that the seemingly low estimates of Jensen and Murphy are consistent with optimal incentives in the presence of moral hazard if the right parameters of risk aversion are chosen for the specific functional forms. Garen (1994) presents a static moral hazard model with closed forms solutions and, as an alternative test for the theory, derives comparative static predictions, which do not rely on the particular values of parameters and are more robust to functional form specifications. He concludes that most implications of moral hazard are consistent with the data. In the same spirit, Wang (1997) provides a computed solution to a dynamic agency (repeated moral hazard) model that, under a reasonable range of parameters, is able to replicate the low sensitivity of pay results in Jensen and Murphy (1990a). The model shows that deferred compensation is sometimes a more efficient way of providing incentives than current pay because of incentive-smoothing over time. Changes in Sensitivities with Firm Size The Jensen-Murphy statistic has been documented to be very different across different firm size subsamples: Jensen and Murphy (1990a) divide their sample according to market value and find a total pay sensitivity value of $1.85 for the subsample of larger firms versus $8.05 for that of smaller firms.7 Updates of this measure provided in Murphy (1999) suggest that the increase in median pay sensitivity from 1992 to 1996 was a lot higher for large firms than 7 See, also, Jensen and Murphy (1990b). A. Jarque: CEO Compensation 277 for smaller ones (64 percent [from $2.65 to $4.36] for the largest half of the S&P 500, 5 percent [from $7.33 to $7.69] for the smallest half of the S&P 500, 28 percent [from $12.04 to $15.38] for the S&P Mid-Cap corporations, and 24 percent [from $22.84 to $28.23] for the S&P Small-Cap corporations). Garen (1994) also finds a negative relationship between sensitivity of pay and firm size. Schaefer (1998) finds that the Jensen and Murphy (1990a) measure is approximately inversely proportional to the square root of firm size. Recent estimations of this measure use quantile regressions to prevent estimations from being driven by big firms in the samples.8 The literature agrees that the Jensen-Murphy measure is adequate for firms in which the marginal effect of effort is independent of firm size. However, it has been repeatedly pointed out that it does not correctly capture incentives when the marginal product increases with firm size. Baker and Hall (2004) propose a model that relates compensation to the marginal product of the CEO. They confirm the previous estimates of decreasing sensitivity of pay to firm size and they identify higher marginal products of CEOs working in bigger firms as the main cause.9 Their model implies that, although the Jensen and Murphy (1990a) measure of sensitivity falls sharply with firm size, the total incentives of CEOs remain constant or fall only slightly since the effect of CEOs’ actions increases with firm size. A similar assumption is used in the competitive market model of Edmans, Gabaix, and Landier (forthcoming), in which the effort cost for the CEO is bounded above, but his marginal impact depends on the size of the firm he manages. Elasticity of Pay to Firm Value An alternative measure of sensitivity of pay is the elasticity of compensation to performance: the percent increase in executive wealth for a 1 percent improvement in the performance of the firm. Note that assuming a constant elasticity across firms implies variation in the Jensen-Murphy statistic across firm sizes. A convenient way of estimating the elasticity of pay to firm performance is to regress the logarithm of earnings on the return of the firm’s stock: ln wt = a + b × rt . (2) Then b = d(ln w)/dr measures the semi-elasticity of earnings to firm value and the elasticity is recovered for each particular return value as approximately b = r b. Initial estimates of the semi-elasticity of cash compensation (salary and bonus) to stock returns, surveyed in Rosen (1992), had reported 8 See, for example, Murphy (1999) and Frydman and Saks (2007). 9 This is consistent with the hypothesis that more talented CEOs work for bigger firms, as mentioned in Rosen (1992). See, also, the competitive market for talent models of Gabaix and Landier (2008) and Tervi¨ (2008) reviewed later in this article. o 278 Federal Reserve Bank of Richmond Economic Quarterly a median value for b of about 0.10—an increase in a firm’s return of 10 percent ( r = 0.1) implies a 1 percent increase in pay.10 Using average values for w and r, Rosen (1992) reports that these semi-elasticity values imply a Jensen-Murphy statistic for cash compensation (salary plus bonus) of 10 cents, compared with their finding of 2 cents.11 He concludes that lower sensitivities of pay at bigger firms probably influence the Jensen-Murphy estimation, especially since they do not log the compensation figures. Hall and Liebman (1998) provide estimates of the elasticity of compensation using detailed data on CEO pay from 1980 to 1994. To correctly estimate the elasticity and several other measures of sensitivity of pay (including the Jensen-Murphy statistic), they collect data on stock and options grants from the proxy statements of firms. They construct the portfolio of stock and options for each CEO in their databases in 1994: This way they can include in their measure of total pay the variation in the wealth of the CEO because of the changes in the price of the firm’s stock.12 Also, they can evaluate the potential changes in wealth for different realizations of the stock price (the ex ante sensitivity of pay). Since changes in the wealth of the CEO are sometimes negative because of the decrease in the value of their stock holdings, Hall and Liebman cannot directly run the regression in (2) to calculate the semi-elasticity. Instead, with information both on current total compensation as a function of the firm’s return in 1994, w (rt ) , and on the distribution of annual stock returns, they predict changes in compensation for a given change in firm return, a hypothetical w (rt+1 ). With this predicted total compensation measure, they calculate the semi-elasticity directly from the formula w (rt+1 ) − w (rt ) = b × (rt+1 − rt ) . w (rt ) In their data for 1994, Hall and Liebman evaluate the effects of a typical change in firm return: from a median performance (r = 5.9 percent) to a 70th percentile performance (r = 20.5 percent), which implies a 14.6 increase in r.13 For this typical change, they calculate the semi-elasticity for each CEO 10 Rosen (1992) points out the discrepancy of the findings in the literature with those of Jensen and Murphy. The sensitivity of salary and bonus in Jensen and Murphy (1990a) is 1.35 cents for $1,000. This number implies an elasticity at the mean firm size of 0.0325; equivalently, an elasticity of 0.1 translates into a change of 10 cents per $1,000 increase in firm value. Rosen attributes these differences to functional forms and to the sensitivity measure, which is dominated by the large firms’ observations, with significantly lower sensitivities. 11 Rosen (1992) uses an average value for w/V of 10−3 . Following his calculations, the w Jensen-Murphy statistic from equation (1) is β ≈ b V . 12 Jensen and Murphy (1990a) also include in their total compensation measures the changes in the wealth of the CEO attributable to his holdings of stocks and options. Their sample, however, ends in 1983, before the significant increase in option grants that is captured in the Hall and Liebman (1998) sample up to 1994. Moreover, the sample in Jensen and Murphy (1990a) is from Forbes 800 and includes the value of exercised options as opposed to the value of options at the time of granting. After 1992, Execucomp started collecting this information. 13 The median standard deviation of r is about 32 percent in their data. A. Jarque: CEO Compensation 279 in their data set: b= w(0.205)−w(0.059) w(0.059) . 0.146 The implied mean elasticity is 4.9 and the median is 3.9. Hall and Liebman also compute the Jensen-Murphy median sensitivity based on variation in stock and options only: they find a value of $6, compared to $3.25 reported in Jensen and Murphy (1990a) for the period 1974–1986.14 They confirm the finding in Jensen and Murphy (1990a) that incentives are provided mainly through the granting of stocks and stock options. The sensitivity of salary and bonuses to firm performance is very low, while the changes in the wealth of the CEO that originate from his stock and option holdings are very big. They find a large increase in the use of option grants in the period covered by their sample, which translates into a significant increase in the sensitivity of various pay to performance measures: Between 1980 and 1994, the median elasticity went from 1.2 to 3.9, while the median wealth change per $1,000 (the Jensen-Murphy statistic) went from $2.5 to $5.3. Indirect Sensitivity: Provision of Incentives through Career Concerns or Threat of Dismissal Even if the total pay of a CEO was independent from the performance of his firm, the manager still would have some incentives to exert effort if the threat of dismissal was high enough or career concerns were present. Career concerns is the term used to summarize the fact that workers expect to receive offers for better paying jobs after an above-average performance. Several studies have tried to quantify the importance of these two implicit incentive channels. Jensen and Murphy (1990a) recognize the existence of indirect provision of incentives that is implicit in the threat of dismissal of the CEO following poor performance. To account for those indirect incentives in their sensitivity measure, Jensen and Murphy provide an approximation of the wealth variations for the CEO that would follow if he were fired from his job. To estimate this wealth loss, they first estimate the probability of CEO turnover as a function of firm performance (net-of-market return in the current and previous year). They find that a CEO in a firm with average market returns in the past two years has a .111 dismissal probability, while a performance 50 percent below market returns for two years increases the dismissal probability to .175. Since they do not have information on whether the separation was voluntary 14 Hall and Liebman also report the mean sensitivity of pay in their sample, equal to $25 without evaluating potential changes in wealth due to the threat of dismissal. They claim that the large differences between the mean and the median values are due to the high skewness of stock and option holdings in the sample of CEOs. Jensen and Murphy (1990a) do not report mean values. 280 Federal Reserve Bank of Richmond Economic Quarterly or because of retirement, these estimated values represent a rough measure of the real probability of dismissal. They also find evidence of greater effects of performance on turnover for younger CEOs and for smaller firms. They use these estimated probabilities and a simplifying assumption that a fired executive will earn a salary of $1 million elsewhere until retirement to calculate the dismissal-related wealth loss of CEOs of various ages and as a function of their net-of-market return. For example, a 62-year-old CEO would suffer a loss of 26.4 cents for every $1,000 lost by shareholders if his firm performed 75 percent below the market, as opposed to performing the same as the market.15 The highest sensitivity they find is for younger CEOs, who would experience a loss of 89 cents per $1,000. Subsequent studies have found an increase in job turnover probabilities in recent years, as well as evidence of the importance of relative performance as a determinant of firing decisions. Kaplan and Minton (2006) extend the analysis for the period 1992–2005, and they find an increase in the probability of turnover with respect to previous periods. They report a decrease of the average tenure of a CEO from over ten years, as reported in Jensen and Murphy (1990a), to less than seven years for the period 1992–2003 (which corresponds to a probability of turnover of 0.149). The average for 1998–2003 is significantly lower, at just over six years (a probability of 0.165), and this subperiod shows higher sensitivity to current measures of performance than in previous periods. Kaplan and Minton include in their analysis three measures of performance: firm performance relative to industry, industry performance relative to the overall market, and the performance of the overall stock market. They find that turnover initiated by boards is sensitive not only to firm performance measures but also to bad industry or economy-wide performance. They interpret this fact as indicative of a change in corporate governance since the 1970s and 1980s, when bad industry or economy-wide performance influenced CEO turnover in the form of hostile takeovers. In a related study, Jenter and Kanaan (2006), using a new sample including both voluntary and involuntary turnovers during 1993 to 2001, confirm the results in Kaplan and Minton (2006). They find that both poor industry-wide performance and poor market-wide performance significantly increase the probability of a CEO being fired. A decline in the industry performance from the 75th to the 25th percentile, for example, increases the probability of a forced CEO turnover by approximately 50 percent, controlling for firm-specific performance. The firm-specific performance measures are also weighted by boards more heavily when the overall market and industry performance is worse. 15 1986 constant dollars. A. Jarque: CEO Compensation 281 Gibbons and Murphy (1992) derive a model with explicit contracts and career concerns that builds on Holmstr¨ m (1999). Career concerns represent o the expectation over future wages based on the market beliefs about the quality of the CEO. They present empirical evidence that explicit incentives increase as the CEO gets closer to retirement age: for one, two, or three years left in office, respectively, they find elasticities of pay to performance of .178, .203, and .183, compared with elasticities of .119 and .116 when they have five or six years left. They interpret this evidence as support for their theoretical findings that career concerns complement the explicit incentives provided by the ties of current wealth to performance in their current employment. Overall, the available measures of indirect sensitivity of pay to performance appear to have been increasing in the last three decades, in accord with the increase in the direct sensitivity of pay. Explanations of Sensitivity Facts Two main empirical regularities emerge from the studies reviewed previously. First, the sensitivity of pay is smaller for CEOs of larger firms. Second, the sensitivity of pay across all firm sizes has increased in the last two decades. In the following subsections, I review the main theoretical explanations proposed for these two facts. Finally, I include a subsection that presents justifications for a set of characteristics of real life compensation contracts. These characteristics may, to the uninformed eye, seem unappealing for the provision of incentives. The theoretical models show, however, that they may just be the optimal instruments to implement sensitivity of pay to performance. Fact 1: The Sensitivity of Pay Decreases with Firm Size Baker and Hall (2004) show that, in the presence of moral hazard, the sensitivity of pay optimally decreases with the size of the firm. They rely on a partial equilibrium model with multitasking and heterogeneity in the marginal productivity of the CEO as a function of the firm size. In recent work, Edmans, Gabaix, and Landier (forthcoming) extend Gabaix and Landier (2008) and present an agency model embedded in a competitive market model for talent. They propose a measure of sensitivity of pay that their theory predicts as independent of firm size: the dollar change in wealth for a percentage change in firm value, scaled by annual pay. Fact 2: There has been an Increase in Sensitivity of Pay in the Last Two Decades The explanations for the increase in sensitivity of pay to performance fall mostly into two categories: those that maintain that the increase has been driven by changes in firm characteristics and increased market competition 282 Federal Reserve Bank of Richmond Economic Quarterly and those that maintain that the increase has been driven by an improvement in corporate governance practices. A good example of the first category of explanations is the work of Cu˜ at n and Guadalupe who, in a series of three papers (2004, 2005, 2006), evaluate the changes in CEO compensation following several changes in market conditions:16 Cu˜ at and Guadalupe (2004) study two episodes of deregulation of n the U.S. banking system and find that the increase in competition is correlated with an increase in sensitivity of pay; Cu˜ at and Guadalupe (2005) find that in n the period following the sudden appreciation of the pound in 1996, which implied different levels of competition for tradables and nontradables, we observe an increase in sensitivity of pay for executives in the tradables sector; Cu˜ at n and Guadalupe (2006) find, for a sample of U.S. executives, that industries more exposed to globalization (higher foreign competition as instrumented by tariffs and exchange rates) exhibit higher sensitivity of pay, higher dispersion of wages across different levels of executives, and higher demand for talent at the top, which drives up the salary of the CEO. This evidence supports the conclusions of some theoretical papers that analyze the effects of increased competition for talented CEOs. Frydman (2005), for example, presents a dynamic model of executive compensation and promotions in which the superior information of firms about their incumbent workers implies that an increase in the relative importance of general skills versus firm-specific skills triggers an increase in manager turnover, increased differences in pay among executives of different ranks of the same firm, and higher levels of pay. Frydman contrasts these implications with historical data on compensation, careers, and education of top U.S. executives from 1936 to 2003. She concludes that facts are consistent with the predictions of her model under the hypothesis of an increase in the relative demand of general skills after the 1970s. The work of Holmstr¨ m and Kaplan (2003) falls under the second o (possibly complementary) category of explanations. They argue that the changes in corporate governance of U.S. firms in the last two decades have had a net positive impact, translating into higher sensitivity of pay to performance. Although they recognize that the increase in options compensation has probably led to new incentive problems that may be behind the corporate scandals of recent years, they argue that the evidence is still in favor of an improvement in the management of firms. They identify three main positive signs: the good comparative performance of the U.S. stock market with respect to the rest of the world, the undisputable increase in the link between CEO compensation and firm performance, and the use of compensation schemes by buyout investors and venture capital investors that resemble those of the average public firm. They review the trends in corporate governance since the 16 See, also, references in Gabaix and Landier (2008). A. Jarque: CEO Compensation 283 1980s and they identify, as a sign of success, the increasing convergence of international governance practices to those of the United States, which were drastically different from those of, for example, Germany and Japan during the 1980s. They also point to an increase in shareholder activism, which may be due to an increased institutional shareholders’ stake in public companies: their overall share in the total stock market went from 30 percent in 1980 to more than 50 percent in 1994. Gompers and Metrick (2001) report evidence supporting the hypothesis in Holmstr¨ m and Kaplan (2003) that institutional o investors improve corporate governance. They find that over the period 1980– 1996, firms with higher shares of institutional shareholders had significantly higher stock returns. Gompers, Ishii, and Metrick (2003) construct an index to proxy for the level of good governance of firms based on data on completion of a number of governance provisions aimed at protecting shareholder’s rights. They collect data on four dates between 1990 and 1998 and include a very high proportion of the largest U.S. firms according to market capitalization. They find strong correlation between good governance and good firm performance in terms of Tobin’s Q (market value over book value), profits, and sales growth. Holmstr¨ m and Kaplan (2003) also identify signs of improved board o of directors’ independence: an increase in CEO turnover, a decrease in the size of boards, and an increase in the proportion of the compensation of board members that is in the form of stock and options of the firm.17 An earlier theoretical paper, Hermalin and Weisback (1998), provides support for the importance of board independence on sensitivity of pay and, in general, CEO performance. This paper models explicitly the determination of independence levels of the board of directors. The CEO uses his perceived higher ability with respect to any alternative hire to bargain with the board and influence its composition. Their model has implications consistent with empirical facts for the 1980s and 1990s, such as CEO turnover being negatively related to prior performance, with greater effect for firms with more independent boards and with accounting measures being better predictors of turnover than stock price performance. Also, the independence of the board is likely to increase after poor firm performance and decrease over a CEO’s career. Diversity of Compensation Practices In spite of the existence of some stylized facts on sensitivity of pay, real life compensation packages are highly complex and diverse. It is easy to find examples of features of compensation instruments and practices that may seem counterintuitive to the uneducated eye, such as the lack of relative 17 Compensation boards are independent committees that design the compensation of the top managers of the firm. See Section 4 on regulation for details on the legal requirements for defining this board and the board of directors as “independent.” 284 Federal Reserve Bank of Richmond Economic Quarterly performance evaluation, or the repricing of options gone out-of-the-money. However, changing environments and dynamic considerations sometimes imply that such features are part of an efficient provision of incentives. What follows is a brief (nonexhaustive) list of recent theoretical articles that provide justification for some controversial compensation practices. Commitment. Clementi, Cooley, and Wang (2006) show in a dynamic model of CEO compensation that implementing optimal incentives for executives in the presence of moral hazard and limited commitment through the issue of securities dominates deferred cash compensation arrangements; granting securities improves commitment and helps retain CEOs. Dynamic Incentives. Wang (1997), as discussed previously, provides a model of repeated moral hazard that can explain very small (or even negative) sensitivities of CEO pay based on the optimal smoothing of incentives over time. Learning. Celentani and Loveira (2006) show how learning about common shocks in the economy can explain the apparent absence of relative performance evaluation documented in the literature.18 In the presence of moral hazard, if the productivity of the CEO’s effort depends on aggregate conditions, the optimal compensation contract is not necessarily decreasing in poor relative performance. Short-term Stock Price Performance. Current compensation plans are often criticized for providing incentives for CEOs to engage in actions that increase stock prices in the short term as opposed to creating long-term value for the shareholders.19 Bolton, Scheinkman, and Xiong (2006) present a model that explicitly accounts for liquidity needs as a reason to participate in the stock market. They show that shareholders may in some circumstances benefit from artificial increases in today’s stock prices, which influence the belief of speculators and allow the firm to sell its stock at higher prices in the future. This implies that compensation packages are sometimes optimally designed to make CEOs take actions that make short-term profits higher. Targeting Efforts of the CEO. Kadan and Swinkels (2008) present a model in which the choice between restricted stock grants or option grants is linked to the potential effect of a manager’s actions on firm survival. In financially distressed firms or startups where this effect is higher, stock is more efficient than options. The choice of compensation instruments helps the firm tailor the sensitivity of pay to the firm’s idiosyncratic position. They find empirical evidence of stock being more widely used than options in firms with a higher probability of bankruptcy. Repricing. The public perception of the practice of decreasing the exercise price of options that are out-of-the-money is that it “undoes” incentives 18 See, for example, Gibbons and Murphy (1990). 19 See, for example, Jensen, Murphy, and Wruck (2004). A. Jarque: CEO Compensation 285 and is thus interpreted as a sign of management entrenchment. This practice of repricing is fairly uncommon: Brenner, Sundaram, andYermack (2000) report that 1.3 percent of the executives in their sample (top five officers in a sample of 1,500 firms between 1992 and 1995) had options repriced in a given year. However, it has received academic attention, perhaps motivated by its reputation as a bad compensation practice. Chance, Kumar, and Todd (2000) identify size as the main predictor for firm reprices, with smaller firms repricing more often. Brenner, Sundaram, and Yermack (2000) find that higher volatility also significantly raises the probability of repricing. Carter and Lynch (2001) find that young high technology firms and those whose outstanding options are more out-of-the-money are more likely to reprice. Chen (2004) finds that firms that restrict repricing have a higher probability of losing their CEO after a decline in their stock price and that they typically grant new options in those circumstances, possibly in an effort to retain the CEO. Acharya, John, and Sundaram (2000) present a theoretical model that implies that the practice of repricing can be optimal in a wide range of circumstances. The intuition is that there are benefits to adjusting incentives to information that becomes available after the initial granting and before the expiration of the options; in many cases, these benefits offset any loss in ex ante incentive provisions because of the lack of credibility from possible repricings. 4. REGULATION The current levels of CEO compensation are viewed by many as excessive and unjustified. It is not uncommon to see demands in the popular press for government regulation of CEO pay.20 However, the effects of regulation in a complicated matter such as the design of compensation packages for top executives are not always clear. An example of this is the change in regulation that was introduced in 1993 that limits tax deductions for any compensation above $1 million. This limit was introduced partly as a response to the popular rejection of the big increases in CEO pay that took place in the early 1990s. The law, however, established an exception to the limit: “performance-based” compensation (such as, for example, bonuses and options granted at the money).21 Empirical studies cited below have found that firms have shifted compensation away from salary and toward stocks and options in response to the limit. A popular view today is that the inclusion 20 As a recent example, the Emergency Economic Stabilization Act of 2008 passed by Congress on October 3, 2008, explicitly includes some limits on CEO compensation for firms in which the Treasury acquires assets or is given a meaningful equity or debt position. The compensation practices of such firms “must observe standards limiting incentives, allowing clawback and prohibiting golden parachutes.” Restrictions also apply to firms that sell more than $300 million in assets to the Treasury through auction. 21 For references, see United States Senate (2006). 286 Federal Reserve Bank of Richmond Economic Quarterly of stocks and options in the compensation packages, together with the stock market boom, is partly responsible for the recent increase in CEO pay. Did the $1 million limit contribute sizably to the increase in pay by encouraging the use of performance-based compensation? In this section, I review the academic studies that have studied this question, as well as the effect of other changes in the tax treatment of CEO compensation. Regulation has also been introduced in recent years to improve corporate governance in the United States.22 Mainly, the measures have targeted improved coordination and power of shareholders, as well as the transparency of compensation practices. As a recent example, in the aftermath of corporate accounting scandals at Enron, WorldCom, and other important U.S. firms, the Sarbanes-Oxley Act of 2002 (SOX) increased the legal responsibilities of CEOs and of compensation committees. There are academic studies on the effect of regulation on corporate governance and I summarize their main conclusions in the second part of this section. Tax Advantages Tax exemptions for deferred compensation have been in place since the early 1980s. These tax exemptions apply to the capital gains tax owed by a CEO, which is levied for capital gains that are part of his compensation. We may wonder whether the savings on capital gains tax for stock option grants could have played a role in explaining the spectacular increase in the use of options in compensation packages. Also, some of the regulatory efforts that have stated specific limits for tax deductions have triggered criticisms since firms that were paying less to their executives may now raise their salaries to the limit, which is rendered by the regulation as acceptable. For example, with the passing of the Deficit Reduction Act of 1984, the U.S. government imposed a tax on excessive (three times the executive’s recent average remuneration) “golden parachutes,” i.e., payments to incumbent managers who were being fired in relation to a change in control. Jensen, Murphy, and Wruck (2004) argue that these types of agreements were fairly uncommon before the law was passed and that the regulation had the effect of endorsing the practice, which became a standard in the following years. A similar argument has been made for the Omnibus Budget Reconciliation Act resolution 162(m) of 1992, which imposed a $1 million cap on the amount of the CEO’s nonperformance-based compensation that qualifies for a tax deduction. In this section, I review a few of the academic studies that have tried to quantify the effect of tax advantages on CEO compensation. The main conclusion is that there exists a tax advantage of deferred compensation, although 22 See Weinberg (2003) for a discussion of the financial reporting issues in corporate governance and of the Sarbanes-Oxley Act of 2002. A. Jarque: CEO Compensation 287 a modest one. Also, studies of the effect of resolution 162(m) have found evidence of a certain increase in salaries of companies that were paying less than $1 million before the new law was enacted, as well as some shifting of compensation toward performance-based instruments and away from salaries. There is evidence that this shift translated into a slight increase in total compensation. The studies, however, do not find evidence that the majority of the spectacular rises in the level of compensation and the increases in the use of options can be attributed to tax reasons. Quantifying the Effect of the Special Tax Treatment for Compensation in the Form of Options Currently, pay in the form of options has a tax advantage for CEOs when compared to payments in salary: There are no capital gains tax charges on the increase in value of a stock from the grant date to the exercise date. If, after exercise, the executive holds the stock and there is further increase in its value, capital gains tax is owed only on the appreciation from the date of exercise to the date of sale. Motivated by different tax treatments of various instruments of compensation, Miller and Scholes (2002) explicitly compare the after-tax value (together for the firm and the CEO) of compensation in wages and several deferred compensation instruments (deferred salary schemes, stock purchase plans, deferred stock plans, nonqualified option plans, and insurance plans). Miller and Scholes document that by 1980 there had been already a rapid increase in the usage of deferred compensation versus bonus and salary. The use of these instruments is usually interpreted as a sign of incentive provision: With the granting of stock options, the compensation board attempts to align the incentives of the CEO with those of the shareholders, which helps increase the long-term value of the company. However, they find that none of the instruments they analyze have clear tax disadvantages with respect to wage payments: they are all either beneficial or tax-neutral. They argue that the tax savings makes it difficult to identify incentive provision as the only reason for deferred compensation. The advantage of options versus pay is also analyzed in Hall and Liebman (2000) with data for 1980–1994.23 A simplified version of their analysis is presented here. We must compare two alternative ways of compensating the CEO, both with net present value of P for the firm. On one hand, the firm could grant options to the CEO for a present value of P , which would vest in N years. The return of this compensation to the CEO, assuming an annual return of r for his firm, is P [1 + r (1 − Tc )]N 1 − Tp , 23 For details on current tax treatment of CEO compensation, see United States Senate (2006). 288 Federal Reserve Bank of Richmond Economic Quarterly Figure 3 Evolution of Tax Advantage Compared with the Proportion of Pay Given in the Form of Stock and Options 20 10 0 0 2006 1 2004 2 2002 30 2000 3 1998 40 1996 4 1994 50 1992 5 1990 60 1988 6 1986 70 1984 7 1980 80 1982 8 Percent 90 Percent 9 Tax Advantage (A) (Percent of Total Pay) Left Axis Percent of Stock and Options (Percent of Total Pay [Source: Forbes]) Right Axis Source: IRS and Forbes magazine. where Tp , Tc , and Tcg are, respectively, the personal, corporate, and capital gains tax rates. On the other hand, the firm could schedule a cash payment of P today. Assuming the CEO invests the whole after-income tax amount in a nondividend-paying security with the same return, r, for the same number of years, N, the return of the cash payment to the CEO is P 1 − Tp [1 + r (1 − Tc )]N − Tcg P 1 − Tp (1 + r (1 − Tc ))N − 1 . Therefore, the options advantage, A, depends on the three tax rates: A Tp , Tc , Tcg = Tcg P 1 − Tp (1 + r (1 − Tc ))N − 1 . Hall and Liebman (2000) report the value of A for the period 1980– 1998 by substituting the corresponding rates, keeping everything else equal. They find that the advantage of options versus cash compensation reached a maximum of $7 per $100 of compensation P in 1988–1990 and was down to $4 per $100 in 1998. Figure 3 plots the value of the advantage reported in the paper, plus the value for the period 1980–2007 using recent tax rates. The Bush tax breaks have decreased the advantage of options in recent years; A. Jarque: CEO Compensation 289 after they expire, and with no further changes, the advantage should go back to the $4 calculated for 1998. In the same figure, on the right axis, I have plotted the ratio of option grants over total compensation from 1992 to 2007 from the Forbes data previously shown. Comparing the two series, we can see that, although tax advantages were decreasing in the later period, options were being used relatively more intensively than salaries. Hall and Liebman (2000) complement their analysis with regressions of the use of options on the several tax rates and their calculation of the tax advantage and find no evidence of a sizable effect. Quantifying the Effect of Caps on Deductibility of CEO Salaries When firms calculate their corporate tax liability, under the Internal Revenue Code, they are allowed to subtract from their revenues the compensation of their workers. Prior to 1993, this deduction included all compensation given to the top executives of the firm.24 In 1993, new regulation took effect that limited the deductions of the compensation of each of the five highest ranked executives of a firm to $1 million (section 162(m) of the Internal Revenue Code). The regulation specifies an exception to the $1 million limit: Compensation that is performance-based can be deducted regardless of being worth more than $1 million. The requirements to qualify as performance-based for each form of compensation can be summarized as follows: • Salaries are, by nature, nonperformance-based. Any company paying its CEO a salary higher than $1 million would face higher tax costs after the change in regulation. • A bonus program qualifies if it specifies payments contingent on objective performance measures and is approved in advance by the vote of the shareholders.25 • Stock options are considered entirely performance-based compensation. In order to qualify for tax deduction, firms have to detail in an option grant plan the maximum number of options that can be awarded 24 The accounting of compensation expenses is briefly discussed later in relation to regulation of disclosure of information to shareholders and the valuation of options. 25 Murphy (1999) documents that companies do include subjective measures of the performance of the CEO in their bonus plans, usually labeled “individual performance.” In the typical bonus plan in his data source, this subjective component rarely exceeds 25 percent of the executive’s bonus. 290 Federal Reserve Bank of Richmond Economic Quarterly to a given executive in the specified time frame. The plan needs to be approved by the shareholders.26 • Stock grants and other types of deferred compensation (pensions, severance payments, etc.) are only considered performance-based if they are granted or their selling restrictions vest based on performance targets. I now review the three main academic studies of the effect of this change in regulation. A simple strategy of testing for the effect of section 162(m) on pay practices is to regress changes in compensation on changes in firm value: ln (wit ) = β 0 rit + δ t + ε it , where δ t captures the annual growth rates.27 Lower values of δ after the reform would imply the regulation succeeded in slowing the growth of executive compensation. However, the trends in CEO compensation may be responding to changes in the environment other than regulation. To identify the effect of the reform, the empirical studies have exploited the cross-sectional variation in the population of firms. By separating the firms in which compensation practices were not affected by the cap on deductions, they can use those as a control group and isolate the effect of the regulation: ln (wit ) = β 0 rit + α t AF F ECT ED + δ t + ε it , where the difference in α t before and after the change in regulation captures the effect of deduction limits on the growth of pay in the affected firms.28 Deciding which firms should be classified as affected is not such a simple task, however. Hall and Liebman (2000), for this purpose, use data on the compensation in 1992, the only observation available in the Execucomp data set before the change in the law. They construct a variable equal to 1 if the salary observation in 1992 is above $1 million, and equal to $1 million divided by the salary for firms below that level. They find that a CEO with a $1 million salary in 1992 saw, after the new regulation, his salary grow at an annual rate of approximately 0.6 percent less than an executive earning $500,000. Perry and Zenner (2001) propose three alternatives for identifying the set of firms affected by the regulation. One is the same indicator constructed in 26 These plans are usually specified for several years. If, after approval, the compensation board finds it necessary to exceed the total number of options specified in the plan for a given CEO, the excess amount of options does not qualify for a deduction. 27 In the empirical studies, several measures of performance (sometimes logged, sometimes in differences) and lagged performance are included as explanatory variables, as well as other controls like CEO tenure and other relevant available information. In this section, I use a simplified statement of the regression equations meant only for illustration. 28 It has also been argued that setting the deduction cap at $1 million encouraged firms that were far below this level of compensation to increase their manager’s pay, since $1 million became “acceptable” (see Wall Street Journal 2008). If this effect really existed, it would bias the estimates in the above regression. A. Jarque: CEO Compensation 291 Hall and Liebman (2000), a second one is an indicator that takes a value of 1 if the salary in 1992 is above $900,000, and a third is an indicator that takes a value of 1 if the executive had a combined salary and bonus payment above $1 million in any year before the change in regulation. They find similar results in the three specifications. Consistent with Hall and Liebman (2000), Perry and Zenner estimate lower growth of salaries for affected firms and evidence of an increase in the sensitivity of bonus and total compensation to stock performance after 1993. Rose and Wolfram (2002) point out that using 1992 levels of pay to classify firms creates statistical correlation between compensation and the variable that indicates whether firms are affected by the regulation that does not correspond to an effect of the change in the law: Firms that had higher compensation levels in 1992 had different characteristics than low compensation firms and these characteristics likely influence their pay growth rate. They propose an alternative estimation using differences in differences that circumvents this problem.29 To correct for potential bias, Rose and Wolfram construct their indicator for affected firms using predicted instead of observed compensation for years after 1993: They use data in 1992 to construct what the level of compensation of each firm in the sample would have been had there not been a change in regulation. They construct the indicator based on combined salary and bonus payments, since they find that using only salary or broader measures of compensation constitutes a less precise predictor. Based on their classification methodology, the sample mean growth rates of compensation after 1993 were higher for affected firms than for those that were unaffected. An important feature in Rose and Wolfram’s analysis is that they include data on the decisions of qualification of salary, bonus, and stock plans (although this is limited to a small sample of firms, based on a consultant firm survey). Stock option plan qualification is the most common choice; in their sample, this implies a tax savings of about 25 percent of ex ante total compensation. About two-thirds of the sample of firms had qualified their stock plans by 1997. Rose and Wolfram find that affected firms were more likely than unaffected firms to qualify their long-term incentive plans (more than twice as likely) and bonus plans (three times more likely), but both groups qualify their option plans in the same proportion. They find evidence of salary compression at the $1 million level after 1993 and a flattening of the distribution of cash payments (salary plus bonus) for firms that choose to qualify bonus plans. However, they do not find strong statistical evidence for an overall compression effect for all firms. Hence, they find no evidence of a perverse consequence of the law suggested in the financial press, whereby the argument was that nonaffected firms raised their salaries to $1 million because the limit was in fact made into 29 See, also, Rose and Wolfram (2000) for a detailed explanation. 292 Federal Reserve Bank of Richmond Economic Quarterly an “acceptable” standard by the new law. Interestingly, Rose and Wolfram find that firms that choose to qualify their option plan have higher growth of salary, bonus, and total compensation after 1993 than both unaffected firms and affected firms that do not qualify. Their estimates are, in general, very noisy, but their evidence is consistent with the hypothesis that qualification and limiting the compensation growth are alternative means of responding to political pressures on executive compensation. Regulating Corporate Governance Several regulations passed in the last 15 years have focused on improving corporate governance in U.S. firms. I now summarize the most important initiatives. In 1992 the SEC increased the disclosure requirements in proxy statements, asking firms to include detailed information about the compensation of the CEO, chief financial officer, and the three other highest-paid executives. Along with those details, the statements had to include an explanation of the compensation policies of the firm, as well as performance measures of the firm in the past five years. In particular, firms were required to report the value of option grants given to the executives. Options could be valued at the time of grant using several alternatives. If Black-Scholes valuation was used, companies were not required to specify the value of the parameters used in the formula, such as the risk-free interest rate, expected volatility, and dividend yield, or any adjustment to the valuation made to take into account the nontransferability of the options and the risk of forfeiture. Alternatively, firms could choose any other accepted pricing methodology, even simply the value of the options under the assumption that stock prices would grow at 5 percent or 10 percent annually during the term of the option (the “potential” or “intrinsic” value of options). In November 1993, the SEC amended its rules and required the disclosure of the parameters used for the valuation and details of any discount.30 The reforms of 1992 also expanded the set of allowable topics for shareholder proposals to include executive compensation and decreased the minimum share or capital stake necessary to initiate a proxy vote. Following the accounting fraud scandals at Enron, Tyco, WorldCom, and a number of other firms, the government passed the Sarbanes-Oxley Act in July 2002.31 The Act had several consequences for corporate governance practices, mainly the requirement of independent accounting auditing, mandatory executive certification of financial reports (accompanied by an increase in penalties for corporate fraud), forfeiture of certain bonuses to executives 30 Murphy (1996) finds evidence that, in the year in which they had a choice, managers chose the valuation method that minimized the value of their pay. 31 In November 2001, after weeks of SEC investigations, Enron filed a restatement of its financial results that revealed that the company had underreported its debt exposure. A. Jarque: CEO Compensation 293 after a financial restatement resulting from malpractice, or the prohibition of personal loans extended by the firm to the executives. In November 2003, the SEC approved proposals by the NYSE and NASDAQ to guarantee the independence of directors, compensation committees, and auditors. In 2004, the Financial Accounting Standards Board (FASB) modified their recommendations for the valuation of stock grants: Reporting the fair value of options became the norm, eliminating the previous alternative of reporting their intrinsic value, and expensing requirements were extended to options granted with an exercise price equal to or higher than the market price at the time of granting (FAS 123[R]). In 2006, the SEC made these recommendations compulsory. At the same time, it increased the disclosure requirements of compensation of the executives (salary and bonus payments for the past three years, a detailed separate account of stock and option grants, as well as holdings that originated in previous year’s grants, including their vesting status, and retirement and post-employment compensation details). The SEC also demanded an explanation of compensation objectives and implementation in “plain English.” It required a classification of the members of the board of directors and of the compensation committees as independent or not independent, with an explicit statement of the arguments used for this classification. Finally, it asked for the disclosure of compensation of directors following rules similar to those for the top executives. Recently, the proposal of making mandatory a “say on pay” nonbinding vote of shareholders on the compensation of the CEO has received attention both from the media and politicians, and has been voluntarily adopted by a few U.S. corporations.32 This proposal is in line with the above described efforts of making compensation practices as transparent as possible to shareholders. Quantifying Effects of Regulation on Corporate Governance Most of the above described changes in regulation could directly or indirectly influence compensation contracts, as well as firm performance. A few studies have tried to quantify these potential effects. Johnson, Nelson, and Shackell (2001) document that the 1992 SEC reforms related to shareholder participation translated into an increase in the number of shareholder-initiated proposals on CEO pay. However, they do not conclude that the probability of a proposal is higher for firms with poor incentive alignment.33 Instead, this probability of proposal is higher for firms with higher levels (or lower sensitivity) of CEO compensation. They attribute this correlation to the higher “exposure” of the compensation practices of firms 32 The “say on pay” practice is already in place in the United Kingdom, the Netherlands, and Australia. See, for example, the article “Fair or Foul” (The Economist 2008a). 33 Poor incentive alignment is approximated by the proportion of compensation that is not predictable by observable variables, following Core, Holthausen, and Larcker (1999). 294 Federal Reserve Bank of Richmond Economic Quarterly with higher salaries. They find, however, a positive effect on corporate governance when they look at the probability of acceptance of the proposals: it is higher in firms with poor incentive alignment and higher for institutional investor proposals. Chhaochharia and Grinstein (2007) provide some evidence on the effects of SOX and of the changes in the stock exchange independence requirements, as evaluated by investors in the stock market. They analyze the effects of the announcements of those regulatory changes on the stock prices of firms with different compliance levels. They construct portfolios of firms based on their degrees of compliance and find that less compliant firms (for example, firms that restated their financial statements or that do not have independent boards) earn abnormal returns of about 6 to 20 percent around the announcement of the new rules. They also report evidence that the market believes that small firms find it more difficult to comply with requirements for internal control and independence of directors: Their portfolios have small negative abnormal returns after the announcements, in contrast with positive abnormal returns of larger noncompliant firms. The imposition of the $1 million limit with the passing of section 162(m) discussed earlier arguably could have consequences for corporate governance—on top of any direct effect on executive compensation. The evidence discussed in the previous subsection points to an increase in sensitivity of pay, which may have translated to better alignment of the incentives of the CEO with those of shareholders. This hypothesis is consistent with the evidence in Maisondieu-Laforge, Kim, and Kim (2007): They find that stock returns and operating returns improve for firms affected by the cap. However, one may argue that compensation committees may have been compelled to forego some of the subjective components in the granting of bonuses and options in order to qualify compensation plans for tax exemptions. These committees may have been compelled to qualify their compensation plans solely for tax savings purposes, or to comply with what may have been viewed as the “correct” practice after section 162(m) was approved. Hayes and Schaefer (2000) find that the part of CEO compensation that is unexplained by observable performance measures is a good predictor for future performance. This suggests that discretion of the compensation board may in fact be rewarding good executive performance, and the requirements for qualification of bonus schemes may be distorting good compensation practices. 5. CONCLUSION Real life compensation contracts are complicated. Even in the absence of illegal actions on the part of CEOs, it is easy to find examples in which the incentives of managers are not perfectly aligned with those of the shareholders. The recent cases of fraudulent behavior, such as the accounting scandals A. Jarque: CEO Compensation 295 at Enron, the backdating of options, or insider trading, obviously demand government intervention, as does any other breach of the law. However, a close look at the problem of CEO pay design demonstrates that providing incentives is a complicated matter and that many seemingly unintuitive features of compensation packages may, in fact, be helpful in solving incentive problems. There is ample evidence that, despite rising levels of CEO pay in the last two decades, improved corporate governance practices have translated into stronger ties between pay and performance over the same period. Recent regulation of corporate governance seems to have been aiding market-driven changes in internal control. However, it is also easy to find indications that some of the compensation regulation may impose unnecessary burdens and distortions on firms. REFERENCES Acharya, Viral V., Kose John, and Rangarajan K. Sundaram. 2000. “Contract Renegotiation and the Optimality of Resetting Executive Stock Options.” Journal of Financial Economics 57: 65–101. Associated Press. 2008. “CEO Pay Keeps Rising Even as Economy Slows.” Richmond Times Dispatch. http://www.inrich.com/cva/ric/search.apx.content-articles-RTD-2008-06-16-0104.html [16 June]. Baker, George P., and Brian J. Hall. 2004. “CEO Incentives and Firm Size.” Journal of Labor Economics 22 (October): 767–98. Bebchuk, Lucian, and Jesse Fried. 2003. “Executive Compensation as an Agency Problem.” Journal of Economic Perspectives 17 (January): 71–92. Bebchuk, Lucian, and Jesse Fried. 2004. Pay without Performance: The Unfulfilled Promise of Executive Compensation. Cambridge, Mass.: Harvard University Press. Beck, Rachel, and Matthew Fordahl. 2008a. “CEO Pay Chugs Up in ’07 Despite Economy.” http://www.hermes-press.com/CEO pay1.htm [16 June]. Beck, Rachel, and Matthew Fordahl. 2008b. “CEO Pay Climbs Despite Companies’ Struggles.” USA Today. http://www.usatoday.com/money/ companies/management/2008-06-15-ceo-pay N.htm [20 June]. 296 Federal Reserve Bank of Richmond Economic Quarterly Bennedsen, Morten, Francisco P´ rez-Gonz´ lez, and Daniel Wolfenzon. e a 2007. “Do CEOs Matter?” Copenhagen Business School Working Paper. Brenner, Menachem, Rangarajan Sundaram, and David Yermack. 2000. “Altering the Terms of Executive Stock Options.” Journal of Financial Economics 57 (July): 103–28 Bolton, Patrick, Jos´ Scheinkman, and Wei Xiong. 2006. “Executive e Compensation and Short-termist Behavior in Speculative Markets.” Review of Economic Studies 73: 577–610. Carter, Mary Ellen, and Luann J. Lynch. 2001. “An Examination of Executive Stock Option Repricing.” Journal of Financial Economics 61 (August): 207–25. Chance, Don, Raman Kumar, and Rebecca Todd. 2000. “The ‘Repricing’ of Executive Stock Options.” Journal of Financial Economics 57: 129–54. Chhaochharia, Vidhi, and Yaniv Grinstein. 2007. “Corporate Governance and Firm Value: The Impact of the 2002 Governance Rules.” Journal of Finance 62: 1789–825. Celentani, Marco, and Rosa Loveira. 2006. “A Simple Explanation of the Relative Performance Evaluation Puzzle.” Review of Economic Dynamics 9: 525–40. Clementi, Gian Luca, Thomas F. Cooley, and Cheng Wang. 2006. “Stock Grants as a Commitment Device.” Journal of Economic Dynamics and Control 30 (November): 2191–216. Core, John E., Robert W. Holthausen, and David F. Larcker. 1999. “Corporate Governance, Chief Executive Officer Compensation, and Firm Performance.” Journal of Financial Economics 51: 371–406. Cu˜ at, Vicente, and Maria Guadalupe. 2004. “Executive Compensation and n Competition in the Banking and Financial Sectors.” Working Paper (April). Cu˜ at, Vicente, and Maria Guadalupe. 2005. “How Does Product Market n Competition Shape Incentive Contracts?” Journal of the European Economic Association 3: 1058–82. Cu˜ at, Vicente, and Maria Guadalupe. 2006. “Globalization and the n Provision of Incentives Inside the Firm.” Columbia University Working Paper (November). Edmans, Alex, Xavier Gabaix, and Augustin Landier. Forthcoming. “A Multiplicative Model of Optimal CEO Incentives in Market Equilibrium.” Review of Financial Studies. A. Jarque: CEO Compensation 297 Frydman, Carola. 2005. “Rising Through the Ranks: The Evolution of the Market for Corporate Executives, 1936–2003.” Working Paper (November). Frydman, Carola, and Raven Saks. 2007. “Historical Trends in Executive Compensation, 1936–2003.” Harvard University Working Paper (November). Gabaix, Xavier, and Augustin Landier. 2008. “Why Has CEO Pay Increased So Much?” The Quarterly Journal of Economics 123: 49–100. Garen, John. 1994. “Executive Compensation and Principal-Agent Theory.” Journal of Political Economy 102 (December): 1175–99. Gibbons, Robert, and Kevin J. Murphy. 1990. “Relative Performance Evaluation for Chief Executive Officers.” Industrial and Labor Relations Review 43: 30S–51S. Gibbons, Robert, and Kevin J. Murphy. 1992. “Optimal Incentive Contracts in the Presence of Career Concerns: Theory and Evidence.” Journal of Political Economy 100 (June): 468–505. Gompers, Paul A., and Andrew Metrick. 2001. “Institutional Investors and Equity Prices.” Quarterly Journal of Economics CXIV: 229–60. Gompers, Paul A., Joy L. Ishii, and Andrew Metrick. 2003. “Corporate Governance and Equity Prices.” Quarterly Journal of Economics 118 (February): 107–55. Grossman, S., and O. Hart. 1983. “An Analysis of the Principal-Agent Problem.” Econometrica 51: 7–45. Hall, Brian, and Jeffrey Liebman. 1998. “Are CEOs Really Paid Like Bureaucrats?” Quarterly Journal of Economics 113 (August): 653–91. Hall, Brian, and Jeffrey Liebman. 2000. “The Taxation of Executive Compensation.” Working Paper 7596. Cambridge, Mass.: National Bureau of Economic Research. (March). Hall, Brian J., and Kevin J. Murphy. 2000. “Optimal Exercise Prices for Executive Stock Options.” American Economics Review Papers and Proceedings 90: 209–14. Haubrich, Joseph. 1994. “Risk Aversion, Performance Pay, and the Principal-Agent Problem.” Journal of Political Economy 102 (April): 258–76. Hayes, Rachel, and Scott Schaefer. 2000. “Implicit Contracts and the Explanatory Power of Top Executive Compensation for Future Performance.” RAND Journal of Economics 31: 273–93. 298 Federal Reserve Bank of Richmond Economic Quarterly Hermalin, Benjamin E., and Michael S. Weisbach. 1998. “Endogenously Chosen Boards of Directors and Their Monitoring of the CEO.” American Economic Review 88 (April): 96–118. Holmstr¨ m, Bengt. 1999. “Managerial Incentive Problems: A Dynamic o Perspective.” Review of Economic Studies 66 (January): 169–82. Holmstr¨ m, Bengt, and Steven Kaplan. 2003. “The State of U.S. Corporate o Governance: What’s Right and What’s Wrong?” Journal of Applied Corporate Finance 15: 8–20. Jensen, Michael, and Kevin J. Murphy. 1990a. “Performance Pay and Top-Management Incentives.” Journal of Political Economy 98 (April): 225–64. Jensen, Michael C., and Kevin J. Murphy. 1990b. “CEO Incentives: It’s Not How Much You Pay, But How.” Harvard Business Review 3 (May-June). Jensen, Michael, Kevin J. Murphy, and Eric Wruck. 2004. “Remuneration: Where We’ve Been, How We Got to Here, What are the Problems, and How to Fix Them.” Harvard NOM Working Paper 04-28. (July). Jenter, Dirk, and Fadi Kanaan. 2006. “CEO Turnover and Relative Performance Evaluation.” Working Paper 12068. Cambridge, Mass.: National Bureau of Economic Research. (March). Johnson, Marilyn, Karen Nelson, and Margaret Shackell. 2001. “An Empirical Analysis of the SEC’s 1992 Proxy Reforms on Executive Compensation.” Stanford University Graduate School of Business Research Paper 1679. Kadan, Ohad, and Jeroen Swinkels. 2008. “Stocks or Options? Moral Hazard, Firm Viability, and the Design of Compensation.” Review of Financial Studies 21 (January): 451–82. Kaplan, Steven N., and Bernadette A. Minton. 2006. “How has CEO Turnover Changed? Increasingly Performance Sensitive Boards and Increasingly Uneasy CEOs.” Working Paper 12465. Cambridge, Mass.: National Bureau of Economic Research. Kole, Stacey R. 1997. “The Complexity of Compensation Contracts.” Journal of Financial Economics 43 (January): 79–104. Kuhnen, Camelia, and Jeffrey Zwiebel. 2008. “Executive Pay, Hidden Compensation and Managerial Entrenchment.” Working Paper (July). Lucas, Robert E., Jr. 1978. “On the Size Distribution of Business Firms.” Bell Journal of Economics 9: 508–23. Margiotta, Mary M., and Robert A. Miller. 2000. “Managerial Compensation and the Cost of Moral Hazard.” International Economic Review 41 (August): 669–719. A. Jarque: CEO Compensation 299 Maisondieu-Laforge, Olivier, Yong H. Kim, and Young Sang Kim. 2007. “Financial Contracting and Operating Performance: The Case for OBRA and Efficient Contracting.” Corporate Ownership & Control 4: 217–27. Miller, Merton H., and Myron S. Scholes. 2002. “Executive Compensation, Taxes, and Incentives.” In Selected Works of Merton H. Miller: A Celebration of Markets, edited by Bruce D. Grundy. Chicago: The University of Chicago Press, 254–78. Murphy, Kevin J. 1996. “Reporting Choice and the 1992 Proxy Disclosure Rules.” Journal of Accounting, Auditing, and Finance 11: 497–515. Murphy, Kevin J. 1999. “Executive Compensation.” In Handbook of Labor Economics, edited by Orley Ashenfelter and David Card. New York: Elsevier Science North Holland, 2485–563. Perry, Tod, and Marc Zenner. 2001. “Pay for Performance? Government Regulation and the Structure of Compensation Contracts.” Journal of Financial Economics 62 (December): 453–88. Prescott, Edward S. 2003. “Firms, Assignments, and Earnings.” Federal Reserve Bank of Richmond Economic Quarterly 89 (Fall): 69–81. Rose, Nancy, and Catherine Wolfram. 2000. “Has the ‘Million-Dollar Cap’ Affected CEO Pay?” American Economic Review Papers and Proceedings 90 (May): 197–202. Rose, Nancy, and Catherine Wolfram. 2002. “Regulating Executive Pay: Using the Tax Code to Influence Chief Executive Officer Compensation.” Journal of Labor Economics 20 (April): S138–S175. Rosen, Sherwin. 1981. “The Economics of Superstars.” American Economic Review 71: 845–58. Rosen, Sherwin. 1982. “Authority, Control and the Distribution of Earnings.” Bell Journal of Economics 13: 311–23. Rosen, Sherwin. 1992. “Contracts and the Market for Executives.” In Contract Economics, edited by Lars Werin and Hans Wijkander. Cambridge, Mass.: Blackwell, 181–211. Schaefer, Scott. 1998. “The Dependence of Pay-Performance Sensitivity on the Size of the Firm.” The Review of Economics and Statistics 80 (August): 436–43. Tervi¨ , Marko. 2008. “The Difference that CEOs Make: An Assignment o Model Approach.” American Economic Review 98 (June): 642–68. The Economist. 2008a. “Fair or Foul.” http://www.economist.com/business/ displaystory.cfm?story id=11543754 [12 June]. The Economist. 2008b. “Pay Attention.” http://www.economist.com/ business/displaystory.cfm?story id=11543665 [12 June]. 300 Federal Reserve Bank of Richmond Economic Quarterly United States Senate, Joint Committee on Taxation. 2006. “Present Law and Background Relating to Executive Compensation.” Staff report JCX-39-06 (5 September). Wall Street Journal. 2008. “Rising Pay and Unintended Consequences.” http://online.wsj.com/article/SB120793959097608465.html? mod=2 1565 leftbox [14 April]. Wang, Cheng. 1997. “Incentives, CEO Compensation, and Shareholder Wealth in a Dynamic Agency Model.” Journal of Economic Theory 76 (September): 72–105. Weinberg, John A. 2003. “Accounting for Corporate Behavior.” Federal Reserve Bank of Richmond Economic Quarterly 89 (Summer): 1–20.