View original document

The full text on this page is automatically extracted from the file linked above and may contain errors and inconsistencies.

ni J

Federal Reserve Bank
of Chicago

fll

g H
SEP

t!

81999

RB RESEARCH I IRRAR\
Third Quarter 1999

Economic..

perspectives

2

State budgets and the business cycle: Implications
for the federal balanced budget amendment debate

18

Birth, growth, and life or death of newly chartered banks

36

New facts in finance

59

Portfolio advice for a multifactor world

79

Audio tapes for 1999 Bank Structure Conference

Economic .

perspectives

President
Michael H. Moskow
Senior Vice President and Director of Research
William C. Hunter
Research Department
Financial Studies
Douglas Evanoff, Vice President

Macroeconomic Policy
Charles Evans, Vice President
Microeconomic Policy
Daniel Sullivan, Vice President

Regional Programs
William A. Testa, Vice President
Economics Editor
David Marshall
Editor
HelenO’D. Koshy

Production
Rita Molloy, Kathryn Moran, Yvonne Peeples,
Roger Thryselius, Nancy Wellman

Economic Perspectives is published by the Research

Department of the Federal Reserve Bank of Chicago. The
views expressed are the authors’ and do not necessarily
reflect the views of the Federal Reserve Bank of Chicago
or the Federal Reserve System.

Single-copy subscriptions are available free of charge. Please
send requests for single- and multiple-copy subscriptions, back
issues, and address changes to the Public Information Center,
Federal Reserve Bank of Chicago, P.O. Box 834, Chicago,
Illinois 60690-0834, telephone 312-322-511 1
or fax 312-322-5515.
Economic Perspectives and other Bank
publications are available on the World Wide Web
at http://www.frbchi.org.

Articles may be reprinted provided the source is credited
and the Public Information Center is sent a copy of the
published material. Citations should include the following
information: author, year, title of article, Federal Reserve
Bank of Chicago, Economic Perspectives, quarter, and
page numbers.

ISSN 0164-0682

Contents

Third Quarter 1999, Volume XXIII, Issue 3

2

State budgets and the business cycle: Implications for the
federal balanced budget amendment debate
Leslie McGranahan

Balanced budget amendment proponents often use the experience of the states with
balanced budget restrictions as an argument in favor of a federal balanced budget
amendment. However, the state experience is not directly relevant to the federal
government. State restrictions are more lenient than those considered at the federal
level, and many of the techniques used by the states to balance their budgets over the
business cycle are not available to the federal government.

18

Birth, growth, and life or death of newly chartered banks
Robert DeYoung

Thousands of new commercial banks have been chartered in the U.S. over the past two
decades. This article documents how the financial characteristics of new banks evolve
over time, develops a simple theory of why and when new banks fail, and tests the
theory using a variety of methods.

36

New facts in finance
John H. Cochrane
In the last 15 years, the cherished “random walk” view that stock returns are unpre­
dictable, the “CAPM” view that the market is the only benchmark and market
exposure the only source of returns, and the “expectations hypothesis” relating
interest rates of various maturities and countries have all been abandoned. This article
surveys this revolution in finance, explaining and integrating the new view of the facts.

59

Portfolio advice for a multifactor world
John H. Cochrane

How does traditional portfolio theory adapt to the new facts? The old “two-fund”
theorem becomes a “many-fund” theorem; some investors can improve returns by
investing in portfolio strategies that let them take on nonmarket sources of risk; and
other investors can shed nonmarket risks in the same way. Investors can, if willing to
take on the risks, improve returns by some modest market timing. However, the
average investor must always hold the market, so only investors who are different
from average can benefit from holding new and unusual portfolios.

79

Audio tapes for 1999 Bank Structure Conference

State budgets and the business cycle: Implications
for the federal balanced budget amendment debate
Leslie McGranahan

Introduction and summary
A proposal to amend the U.S. Constitution to require
that the federal budget be balanced has been a part
of the national debate for over 25 years. Following its
inclusion as one of the central planks of the Republican Contract with America in 1994, the balanced budget
amendment became a prominent item on the congressional agenda. The amendment easily passed the
House by a vote of 300 to 132 in January 1995, but
failed to achieve the two-thirds majority required in
the Senate to send it back to the states. Since the
proposal’s most recent failure in the Senate, by one
vote on March 4, 1997, it has been a less important
agenda item because of the strength of the economy
and the surplus in the federal budget. However, the
issue is by no means dead. In January 1999, the
amendment was again proposed in the House with
the cosponsorship of 117 representatives.
Balanced budget amendment supporters frequently
cite the experience of the states, most of which have
statutory or constitutional balanced budget restrictions.1 In this article, I question how the state experience
with balanced budget restrictions can inform the federal debate on a balanced budget amendment. First,
I address how the longstanding state restrictions
compare with those contemplated at the federal level.
I then investigate how state government revenues,
expenditures, debt issuance, and asset holdings have
responded to changes in the states’ economic conditions, as measured by the unemployment rate, during
the last two decades. I use regression analysis to ask
how, controlling for a time trend and state fixed effects,
state finances have reacted to fiscal year state unemployment rates from 1977 to 1997. I further question
whether similar responses on the part of the federal
government would be either feasible or prudent.
In my investigation of how state finances respond
to business cycle conditions, I discover that states
use four main mechanisms to maintain budget balances

2

during downturns: they issue more short- and longterm debt; they rely more heavily on the federal
government for funds while giving less to local governments; they increase tax rates; and they lower
capital spending. This is not a feasible policy combination for the federal government for a number of
reasons. Most importantly, the provisions of the balanced budget amendment would not allow the federal
government to issue any new debt without a legislative supermajority. In this way, the federal balanced
budget amendment differs significantly from the restrictions in place in the states. While the states use the issuance of debt as an important safety valve, this option
would not be available to the federal government.
Of course, the opportunity to receive more from
a higher level of government would also not be available to the federal government. However, the federal
government could follow the states’ lead by transferring less money to the states during difficult times.
This would reverse the current relationship between
federal government intergovernmental spending and
the business cycle and would make it more difficult
for the state governments to balance their budgets.
Importantly, this suggests that one of the reasons
that the states are able to balance their budgets is that
the federal government does not.
The federal government could follow the states
by increasing tax rates during economic downturns.
This would be an unpopular policy for two main reasons. First, tax increases are always unpopular and
difficult to pass. Second, unlike the state governments,
the federal government is responsible for the condition

Leslie McGranahan is an economist in the Economic
Research Department of the Federal Reserve Bank of
Chicago. The author would like to thank Loula Sassaris
for research assistance and colleagues at the Federal
Reserve Bank of Chicago for help and comments.

Economic Perspectives

of the macroeconomy. Tax increases during recessions would further depress disposable incomes and
consumption and could prolong downturns.
The other state behavior open to the federal
government would be to decrease capital spending
during economic downturns. States get a lot of leverage out of their ability to cut capital spending during
difficult times; my results show that this is among the
most pronounced state responses to a deteriorating
economic situation. The federal government may be
unwilling to follow the states’ lead by cutting capital
spending during recessions because the bulk of federal capital spending, over 80 percent, is in the area
of national defense (U.S. Government, Office of Management and Budget, 1999). By contrast, the majority
of state capital spending is on highways (57 percent)
and institutions of higher education (14 percent)
(U.S. Department of Commerce, Bureau of the Census,
1977–87 and 1988–97). Whether it is prudent for the
federal government to structure defense capital
spending to maintain budget balance during downturns is an open question.
Because of the differences in the proposed federal balanced budget amendment and the measures in
place in the states and the different responsibilities
of the federal versus state governments, none of the
four methods used by state governments during economic downturns is an obvious choice for the federal
government. In summary, my results suggest that the
ability of the states to function under their current
balanced budget restrictions should not be used to
argue in favor of the balanced budget amendment
most recently proposed in Congress. However, this
does not necessarily imply that other reasons advanced in favor of a balanced budget amendment are
invalid or that the amendment should not be justified
on other grounds.
Comparing state and federal balanced
budget requirements
The provisions of the proposed federal balanced
budget amendment are quite basic. The amendment
as voted on in 1997 simply states that “[t]otal outlays
for any fiscal year shall not exceed total receipts for
that fiscal year, unless three-fifths of the whole number of each House of Congress shall provide by law
for a specific excess of outlays over receipts by a rollcall vote.” Additional provisions require a three-fifths
majority to increase the debt limit or to increase revenues (U.S. Senate, 1997).
The amendment does not provide for separate
funds to finance capital projects and, therefore, in the
absence of a super-majority vote, does not allow the

Federal Reserve Bank of Chicago

government to issue any new debt. In addition, the
amendment does not provide for a reserve fund that
can be used to carry over surpluses from one year to
the next. Instead, surpluses that were neither spent
nor returned to citizens would be used to reduce the
existing debt. This arises from the provisions that
outlays must be financed by total receipts from the
same fiscal year—it does not allow for the use of
receipts from previous years. Both of these features
would be in contrast to the provisions in the states.
In short, the amendment simply requires that the budget be balanced every year.
State balanced budget restrictions are far more
complex than the federal proposal. There is no prototypical requirement at state level; each state has a
unique set of provisions. However, the following state
provisions are comparable to the federal proposal: balanced budget requirements, restrictions on deficit carryovers, and restrictions on long-term debt issuance.
Before addressing how these three types of
restrictions interact to affect state behavior, it is important to briefly explain the role of capital budgeting
in the states. Most states have capital budgets that
are separate from their operating budgets.2 The construction of new facilities and the repair, maintenance,
remodeling, and rehabilitation of existing facilities are
funded separately.3 One important feature that distinguishes state balanced budget requirements from
those at the federal level is that most of these requirements only mandate that the operating budget be balanced. In cases where the capital budget also needs
to be balanced, proceeds from the issuance of debt
are counted as revenues. Therefore, the balanced
budget restrictions do not stop states from issuing
debt. This contrasts with the federal proposal, which
explicitly excludes receipts derived from borrowing
from government revenues. The ability of states to
borrow for capital projects reconciles the common
perception that states have balanced budgets with
a thriving and substantial municipal bond market.
Submitting, passing, or signing a balanced budget
When commentators write that most states have
balanced budget restrictions, they are usually referring
to constitutional or statutory provisions that require
that the governor must submit, the legislature must
pass, or the governor must sign a balanced budget.
These provisions do not explicitly require that the
year-end budget end up balanced, but rather that the
budget as proposed, passed, or signed be balanced
in expectation. For example, the Illinois constitution
requires that the governor submit and the legislature
pass a balanced budget. The document states, “[t]he
Governor shall prepare and submit to the General

3

Assembly, at a time prescribed by law, a State budget
for the ensuing fiscal year. ... Proposed expenditures
shall not exceed funds estimated to be available for
the fiscal year as shown in the budget.” It further
states that “[a]ppropriations for a fiscal year shall not
exceed funds estimated by the General Assembly to
be available during that year” (italics added) (State of
Illinois, 1970, Article 8, Section 2). Note that in both
cases expenditures cannot exceed estimated revenues.
Deficit carryover provisions
In the event that circumstances change during
the year and a budget that was expected or estimated
to be balanced was not, state provisions either allow
or do not allow deficits to be carried over from one
fiscal year to the next. If the state does not allow
deficits to be carried over, the state must either cut
spending or increase revenues to eradicate the deficit
by fiscal year-end. Such deficit carryover provisions
represent the teeth of the balanced budget requirements, because they prohibit the state from issuing
debt to finance a shortfall. The National Conference
of State Legislatures reports that 13 states have no
restriction on carrying over a deficit and a total of 21
may carry over a deficit if necessary (Snell, 1999).
Illinois is one of the states allowed to carry over deficits. The Illinois constitution states that “[s]tate debt
may be incurred by law in an amount not exceeding
15 percent of the State’s appropriations for that fiscal
year to meet deficits caused by emergencies or failures
of revenue” (State of Illinois, 1970, Article 9, Section 9).4
Note that all states do allow surpluses to be carried over
from one year to the next and 45 states have special
“rainy day funds” for surplus carryovers (Eckl, 1998).
State long-term debt provisions
The final parts of states’ budget restrictions are
provisions limiting their ability to issue long-term
debt. Nearly all long-term debt is used to finance specific capital projects in conjunction with the state’s
capital budget. While federal Treasury bonds, notes,
and bills are very general in nature, most state government debt is very specific and is issued to benefit
particular capital projects. State debt can be backed
by either the full faith and credit or the taxing power
of the government, and can be redeemed from general
revenues or be nonguaranteed and be backed by
specific income streams.
Most states have a restriction limiting the issuance
of long-term debt. Some state constitutions require
that debt cannot be issued until it receives majority
support in a statewide referendum; in some states
debt can only be issued up to a prespecified limit;
and other states allow no debt to be issued at all.5

4

However, state courts have interpreted these constitutional requirements as only applying to debt backed
by the full faith and credit of the government. As a
result, states can issue nonguaranteed debt limited
only by the constraints of the capital market. In fact,
despite restrictions on long-term debt that in some
cases seem quite severe, in every year since 1977
every state has issued some long-term debt.
In sum, the restrictions on the states are far
more lenient than that contemplated for the federal
government. In particular, all states can and do issue
long-term debt and many states can issue debt to
finance deficits.
Nonetheless, the states’ experience with budget
restrictions is frequently used to support balanced
budget restrictions at the federal level. For example,
Michigan’s Governor John Engler in his State of the
State Address in 1997 said, “I support the balanced
budget amendment and so do Michigan voters. When
Congress takes up this historical amendment next
month, I urge them to pass it and submit it to the states.
I invite this legislature to join the debate, call upon
your colleagues in Congress to act and help the
federal budget look more like Michigan’s budget—
balanced” (Engler, 1997). Similarly, in his 1997 State of
the State address Oklahoma Governor Frank Keating
stated that “We Oklahomans know the wisdom of a
constitutional mandate for fiscal common sense. Let’s
send some Oklahoma values to Washington by being
the first to ratify this vital amendment” (Keating, 1997).
While the current state restrictions and the contemplated federal restrictions are quite different, the
general perception that states are more fiscally responsible is warranted. States do a better job on two
dimensions. First, they have a lower level of overall
debt relative to their financial obligations. Between
1977 and 1997, net interest payments on the federal
debt averaged 12.7 percent of outlays and 15.0 percent of receipts (U.S. Government, Office of Management and Budget, 1999), while state interest payments
averaged 3.7 percent of expenditures and 3.4 percent
of revenues (U.S. Department of Commerce, Bureau
of the Census, 1977–87 and 1988–97).6 Similarly, gross
federal debt outstanding averaged 2.3 times outlays
and 2.7 times receipts, while state gross debt outstanding averaged 0.5 times revenues and 0.6 times
outlays over the same period. Second, the states do
a better job of smoothing over the business cycle.
A 1 percentage point increase in the state unemployment rate increases the average state’s budget deficit
(expenditures – revenues) by $23 per capita or about
9 percent (relative to the mean), while a 1 percentage

Economic Perspectives

point increase in the national unemployment rate
increases the federal government deficit by $134 or
about 16 percent.
Next, I investigate how state budget items respond
to business cycle conditions. If a federal balanced
budget amendment were to pass, the federal government would need to find ways to either raise additional
funds or cut expenditures to compensate for the
decline in tax revenues that accompanies downturns.
The assumption that the federal government could
mimic the cyclical behavior of the states is implicit in
the argument that state experience is a valid example
for the federal government. I ask what the states do
and whether the state experience could or should be
mimicked by the federal government.
Data and methodology
To look at how state finances change over the
business cycle, I need data on both business cycle
conditions within a state and on various attributes
of state government finances.
Measuring the business cycle
To measure business conditions in the state,
I use the average monthly state unemployment rate
during the fiscal year for which the state finance data
are collected. For the most part, the analysis focuses
on state fiscal years (FY) 1977–97. Most states’ fiscal
year begins on July 1 and ends on June 30. 7 Since
January 1978, the Bureau of Labor Statistics has calculated a monthly unemployment rate for every state
(expect California, first calculated in 1980). Since FY
1979, I calculate the fiscal year unemployment rate as
the average monthly unemployment rate during the
fiscal year. Prior to FY 1979, I calculate the fiscal year
unemployment rate as a weighted average of the
unemployment rates in the state in the two calendar
years that comprise the fiscal years. The weights
depend on the fraction of months for which the fiscal
and calendar years overlap.
While the national business cycle is usually discussed in terms of changes in gross domestic product
(GDP), the unemployment rate is a better measure of
economic conditions in the state than gross state
product (GSP). There are problems concerning the
accuracy of GSP numbers. GSP is gross output minus
the value of intermediate inputs. Evaluating the worth
of intermediate inputs for the same company across
different states is surely a daunting task. While such
transfer pricing issues also arise for GDP, linkages
across nations are both weaker and more carefully
monitored than those across states. The final advantage of the unemployment rate is that during most of

Federal Reserve Bank of Chicago

the period of study, it was measured monthly. This
allows me to calculate a measure that corresponds in
timing to the state financial year. By contrast, GSP is
measured only yearly and is therefore more difficult
to match accurately with the financial data. However,
if I were to use the percentage change in GSP per capita
as the measure of state fiscal condition instead of the
fiscal year unemployment, I would arrive at a set of
results broadly similar to those discussed below.8
Fiscal data
The data I use to measure state financial variables
come from the annual survey of state government
finances conducted by the U.S. Census Bureau (U.S.
Department of Commerce, Bureau of the Census,
1977–87 and 1988–97). The survey measures approximately 450 different aspects of state revenues, expenditures, debts, and assets. I use the survey data from
1977–97; the 1998 data have not yet been released
and the data prior to 1977 are not available in electronic
form. Importantly, this is not accounting data drawn
from state budgets, but is statistical in nature. Budgetary data would not be as comparable across states
or over time. The variables measured over this period
have been relatively consistent. One important exception is that major changes in measurement of debt
occurred in 1988. (Throughout, dollar numbers are in
GDP-deflated 1997 dollars.)
Methodology
In analyzing state fiscal behavior, I look at how a
change in the fiscal year unemployment rate changes
a variable measuring a fiscal outcome. I measure all
fiscal outcomes in per capita terms to make the numbers comparable across states. Throughout, the unit
of analysis is an individual state and states are not
weighted in terms of population. I look at how a 1
percentage point change in the fiscal year unemployment rate (say, a jump from 4.2 percent to 5.2 percent)
affects the per capita measure of a fiscal variable.
Throughout the remainder of the analysis, I omit the
state of Alaska. Alaska’s fiscal behavior differs drastically from that of the other 49 states, mainly due to
the revenues Alaska receives from oil production.
I also include a series of state fixed effects. This
allows the average value of a variable to differ across
states. This is especially important when looking at
state expenditure patterns because the role of the
local governments in service provision differs quite
dramatically across states. Importantly, I do not include any measures of the nature or severity of state
balanced budget requirements. One might want to
include these interacted with the unemployment rate

5

to investigate whether fiscal variables in states with
stricter requirements are more responsive to changes
in the unemployment than states with more lax requirements; however, I do not do so here. I believe that
the issuance of debt by all states implies that their
provisions are more similar than different.9 I am more
interested in how all states behave because states as
a whole are perceived as being more fiscally responsible than the federal government. I also include both
a linear and a quadratic time trend to account for the
fact that there was an upward secular trend in state
spending during this entire period.
The regression estimated for each fiscal
variable is:
fiscal variablest
= a + b œ unemployment ratest
populationst
+ c œ year -1977 t

1
6
+ d œ 1year -19776

2

t

+ f œ state dummiess + e.
In the tables, I only present the coefficient on
the unemployment rate, β. This coefficient can be
interpreted as the effect of a 1 percentage point
increase in the unemployment rate on the per capita
amount of the fiscal variable. Note that the typical
peak to trough difference in the unemployment rate
is greater than 1 percent. For example, the average
fiscal year state unemployment rate rose from 6.0 percent in 1981 to 9.8 percent in 1983. In the milder 1991
recession, the average fiscal year state unemployment rate increased from 5.2 percent in 1990 to 6.7
percent in 1992; it retreated to 5.0 percent in FY 1997.
In some places I compare the behavior of the states

to the behavior of the federal government. To do so, I
use federal data from the Budget of the United States
(U.S. Government, Office of Management and Budget, 1999). This is accounting data,
unlike the state data. In the case of the federal data, I
estimate the same regression presented above, excluding the series of state dummies.
I break the analysis into four parts—first, I look
at the gap between state expenditures and revenues
(the deficit or surplus); second, at state revenues;
third, at expenditures; and finally, at state indebtedness and asset accumulation. In each section, I look
separately at finances inside and outside the insurance
trust funds run by the states. The states administer a
number of different insurance trust systems, including
employee retirement systems, unemployment compensation, workers’ compensation, and other smaller funds
(including disability and sickness policies). The budget items outside the insurance trust system are
considered “general” budget items.10
Responsiveness of the surplus
to the business cycle
Between 1977 and 1997, average state general
fund revenues exceeded average state general fund
expenditures by almost $64 per capita while average
state insurance trust fund revenues exceeded average
state insurance trust fund expenditures by nearly
$189 per capita (see table 1, column 1). While these
calculations imply that states operate with a general
fund surplus on average, this is somewhat misleading
because state expenditure data exclude state payments
into their insurance trust systems. State contributions
to their insurance trust systems average just over
$70 per capita yearly. These contributions are almost

TABLE 1

Per capita budget deficit or surplus, 1977–97
(dollars)
Average
per capita value

Effect of 1 percentage
point increase in
unemployment rate

Total surplus (revenues – expenditures)
General fund surplus
Insurance trust surplus

252.00
63.50
188.50

–23.03 (6.943)
–10.85 (4.642)
–12.18 (5.822)

General fund surplus net of interest payments

163.87

–10.92 (4.526)

Budget category

Notes: Absolute t-statistics are in parentheses. The final column of each row represents the
coefficient on the unemployment rate in a separate regression. Other variables included in all
regressions are a linear and quadratic time trend, a constant, and a series of state fixed effects.
Source: Author’s calculations from U.S. Department of Commerce, Bureau of the Census, 1977–87 and 1988–99,
State Government Finances.

6

Economic Perspectives

exclusively payments by states into their employee retirement systems. If these intragovernmental transfers
were included as general fund expenditures and insurance trust revenues, the average general surplus would
become slightly negative and the average insurance
trust surplus would increase.
When I run the regression specified above to
look at how state surpluses are affected by changes
in the unemployment rate, I find that a 1 percentage
point increase in the unemployment rate decreases
state surpluses by $23.03 per capita, 11 as shown in
the last column of table 1. This combines a $10.85
($2.34)—number in parentheses indicates the standard
error—per capita drop in the general fund surplus with
a $12.18 ($2.09) per capita drop in the insurance trust
surplus. This result suggests that state budgets as a
whole do respond to the business cycle. Below, I investigate the sources of this business cycle variation by
exploring revenues and expenditures separately.

FIGURE 1

Expenditures and revenues, 1977–97
dollars per capita

Total revenue

Total expenditures

Note: Dollars are measured in 1997 dollars.
Source: Author’s calculations from U.S. Department of
Commerce, Bureau of the Census, 1977–87, Survey of State
Government Finances, Washington, DC: Government Printing
Office, and U.S. Department of Commerce, Bureau of the
Census, 1988–97, Survey of State Government Finances,
available on the Internet at www.census.gov/govs/state/
www.census.gov/govs/state/.

Responsiveness of revenues to
business cycle
Between 1977 and 1997 average state yearly revenues per capita were $2,893. This breaks down into
$2,448 raised by the general fund and $445 raised by
the insurance trust systems. Total revenues per capita were growing rather steadily over the period, from
$2,220 in 1977 to $3,908 in 1997 (see figure 1). These
revenues come from five distinct sources: taxes, intergovernmental transfers from both the federal government and local governments, government charges for
service provision, 12 funds from miscellaneous other
sources including lotteries and property sales, and
contributions to the trust systems run by the state.
Table 2 presents both totals and the breakdown of
average yearly per capita revenues during this period
and the responsiveness of budget items to the unemployment rate. Figure 2 depicts the percentage contribution to total revenues from each of these sources.
The table and figure demonstrate that the great majority of state government funds come from taxes, intergovernmental transfers from the federal government,
and insurance trust contributions.
Overall per capita revenues are somewhat responsive to changes in the fiscal condition in the state as
measured by the state fiscal year unemployment rate.
In particular, as presented in table 2, I find that a 1
percentage point increase in the state unemployment
rate decreases total state revenues by $13.80 ($4.16)
per capita. This combines a $20.08 ($3.47) decrease in
general revenues with a $6.28 ($2.12) increase in the
revenues of the insurance trust funds. The changes
mask considerable variation within the various categories in the budget.

Federal Reserve Bank of Chicago

Taxes
Not surprisingly, taxes are among the most fiscally
sensitive of state revenue sources. Although the lion’s
share of such revenues comes from sales and income
taxes, state governments also assess license taxes
and taxes on miscellaneous items such as stock
transfers. Table 2 shows the breakdown in per capita
tax revenues into these three categories and their responsiveness to a 1 percentage point change in the
FIGURE 2

State revenue sources, 1977–97

Miscellaneous
general
7%

Insurance
trust
15%
Taxes
46%

Charges
9%

Intergovernmental
from local
governments 1%

Intergovernmental
from federal
government
22%

Source: Author’s calculations from U.S. Department of Commerce,
Bureau of the Census, 1977–87, Survey of State Government
Finances, Washington, DC: Government Printing Office, and U.S.
Department of Commerce, Bureau of the Census, 1988–97,
Survey of State Government Finances, available on the Internet at
www.census.gov/govs/state/www.census.gov/govs/state/.

7

TABLE 2

Average yearly revenue per capita, 1977–97
(dollars)
Budget category

Average
per capita value

Effect of 1 percentage
point increase in
unemployment rate

Total revenues

2,892.54

–13.80 (3.313)

General fund revenues
Tax revenues
Sales taxes
Income taxes
Other taxes
Intergovernmental revenues
Federal intergovernmental revenues
Public welfare
Education
Other
State intergovernmental revenues
Charges
Miscellaneous general revenues

2,447.97
1,318.45
662.80
460.83
194.82
649.95
625.70
276.85
111.96
236.88
24.26
261.09
218.47

–20.08
–21.04
–10.90
–10.50
0.36
3.24
2.74
4.51
–1.27
–0.50
0.50
–2.01
–0.26

Insurance trust
Contributions
Investment revenue
Federal unemployment
insurance advances

(5.788)
(8.728)
(7.524)
(7.331)
(0.270)
(2.040)
(1.850)
(4.047)
(4.707)
(0.513)
(1.409)
(2.898)
(0.188)

444.57
223.68
208.22

6.28 (2.958)
0.55 (0.669)
–2.73 (1.390)

12.67

8.46(12.479)

Notes and source: See table 1.

unemployment rate. 13 Some tax revenues are more
sensitive to the business cycle than others. As table
2 indicates, sales and income tax receipts are far more
sensitive to the business cycle than other taxes.
While I find that income and sales taxes are
equally sensitive to the business cycle, I would
expect income taxes to be far more sensitive. This
expectation arises from the fact that while income is
highly sensitive to the unemployment rate, individuals
dip into savings in order to smooth consumption during downturns. As a result of this smoothing, total
sales, and hence sales tax receipts, are not thought to
be as sensitive as income taxes to the business cycle.
The lower than expected income tax numbers can
be explained by the fact that these numbers represent
the change in actual tax collections and do not account
for the fact that states often make statutory changes
in their tax structures to counteract the effects of the
business cycle. In particular, states tend to raise tax
rates during times of economic difficulty and lower
taxes in times of economic strength. In the absence
of such statutory changes, the cyclicality of state
revenues would be more pronounced.14 One potential

8

explanation for the income tax number not being larger
than the sales tax number is that income tax levels are
more often statutorily adjusted than sales tax levels
in response to economic conditions. This conjecture
certainly holds true of the current economic expansion.
In their yearly reports on State Tax Actions from 1995
to 1998, the National Conference of State Legislatures
(NCSL) reported that income tax reductions and, in
particular, reductions in the personal income tax
“dominated state tax reduction efforts” (NCSL, 1995);
were “the primary focus of state tax cuts” (NCSL, 1996);
“dominated legislative tax actions” (NCSL, 1997); and
were “the main focus of cuts” (NCSL, 1998). In contrast, in most years excise and sales tax changes were
relatively small. In total, the tax reductions put into
effect between 1995 and 1998 reduced state taxes by
a staggering $16.8 billion dollars.
Even though states counteract some of the effects of the business cycle on tax receipts by changing
tax rates, states are still faced with declining resources
during times of economic difficulty. The tax rate
changes do not totally counteract the fiscal effects
of recession.

Economic Perspectives

Intergovernmental revenues
Intergovernmental transfers are the second major
source of state revenue. While states receive payments
from both the federal and local governments, the
amount from the federal government far exceeds the
amount from the local governments (see table 2). As
shown in table 2, intergovernmental revenues are relatively unresponsive to business cycle conditions.
Looking at the breakdown into local and federal intergovernmental revenues yields a similar picture—
in both categories per capita revenues increase
slightly when the unemployment rate increases.
To look at the relationship between federal intergovernmental revenues and the business cycle a bit
more closely, I break revenues into three categories—
education, public welfare and other. Public welfare
consists of grants for income support and medical
assistance programs. I expect intergovernmental
spending on public welfare revenues to be more cyclically sensitive than spending in the other categories.
The results in table 2 support this picture. Intergovernmental revenues for public welfare increase when
the economy worsens, while spending in the other two
categories declines. Importantly, the welfare reform
legislation passed in 1996 will reduce the cyclicality
of public welfare grants because it replaced an openended matching grant with a fixed block grant.15
Charges
Charges include government fees for service
provision and revenues from the sale of products in
connection with general government activities. For
example, the air transportation measure of charges
includes landing fees at airports and rents for concession stands. I also include the revenues of public
utilities and liquor stores in this category. As is shown
in table 2, revenues from charges only decline slightly
during a downturn.
Miscellaneous revenue sources
Miscellaneous revenue sources consist of monies coming into the state that cannot be easily categorized elsewhere. These include proceeds from special
assessments and property sales and monies from
interest earnings, rents, royalties, fines, forfeits, and
state lotteries. The analysis of miscellaneous revenues
differs from that of other revenue sources because a
major code change in FY 1988 makes a couple of the
subcategories noncomparable before and after this
date. Since 1988, a 1 percentage point change in the
unemployment rate has increased miscellaneous revenues by $4.82 ($2.06), while prior to 1988, a 1 percentage point change in the unemployment rate decreased
revenues by $3.39 ($1.88). (I present the regressions

Federal Reserve Bank of Chicago

for the entire period in table 2 so that the subcategories can add up to the total). The more recent experience suggests that state governments can expect
revenues to go up slightly in the future when the
economy worsens.
One argument regarding how the federal government might adjust its budgeting in order to achieve
budget balance in times of economic stress is that it
might engage in “increased sales of public lands”
(Eisner et al., 1997). I explore whether the state governments engage in the analogous activity by increasing
property sales during times of high unemployment.
Because the category “property sales” did not experience a definitional change in 1988, I look at behavior
over the entire sample period.16 I find no evidence of
increased property sales during times of economic
stress. While this does not mean that the federal government, with its far more extensive land holdings,
would not engage in this behavior, it does suggest
that states do not sell property to compensate for
budget shortfalls.17
Insurance trust revenues
Revenues for insurance trust programs come from
three different sources (aside from within the state itself): contributions from employees, contributions
from other governments (both local and federal), and
interest revenues.18 As shown in table 2, overall insurance trust revenues are countercyclical, increasing
$6.28 ($2.13) when the unemployment rate increases
by 1 percentage point.
Not surprisingly, most of the variation within this
category over the business cycle occurs in unemployment compensation. In particular, federal advance
contributions, the amounts credited to the states
when contributions and interest cannot pay unemployment benefits due, increase by $8.46 ($0.68) per
capita when the unemployment rate increases by 1
percentage point. By contrast, contributions and
investment revenues are much less sensitive to the
state of the economy.
Revenue results and implications
During times of economic difficulty, state revenues
drop by about $23 per capita. This drop is mostly
driven by declining tax revenues and in particular by
declining income and sales tax receipts. There are
three principal reasons that this decline is not more
pronounced. First, state income tax rates are often
increased when times are bad. Although this does
not emerge directly from this analysis, the recent
declines in state tax rates highlight this phenomenon.
Second, the states get more money from the federal
government during downturns, particularly in terms

9

of intergovernmental funds for public welfare and
federal advances from the unemployment insurance
system. Third, state governments rely on a number
of income sources that are fairly acyclical. Only 44
percent of state revenues come from taxes and only
15 percent come from the highly sensitive income tax.
By contrast, 53 percent of federal government revenues came from taxes in 1991 and 47 percent came from
income taxes (U.S. Department of Commerce, Bureau
of the Census, 1994).
While state revenues decline in recessions, federal government revenues have historically declined
even more. Between 1977 and 1997, a 1 percentage
point increase in the national unemployment rate
reduced federal government revenues per capita by
$115.75 ($30.00), 2.5 percent of the mean federal revenue level of $4,674.06; by contrast the drop in state
revenues is about 0.8 percent of mean revenues ($23.03
of $2,892.54).
The methods that states use to mitigate this decline, heavier reliance on the federal government, tax
increases, and use of less cyclically sensitive revenue
sources, would not be as readily available to the federal government. Heavier reliance on a higher level of
government is obviously not an option for the federal
government. Tax increases during downturns are a possibility but would aggravate recessions by decreasing
disposable income and consumption during recessions. States are able to increase tax rates because
they are not responsible for the condition of the macroeconomy. Eventually the federal government may
want to seek out less cyclically sensitive revenue
sources. One such possibility would be a national
sales tax that could be less sensitive than the income
tax to downturns.
Because of the super-majority requirement for
revenue increases enshrined in most balanced budget
proposals, it is unlikely that much of the adjustment
in recessions would occur via revenues. Indeed, this
is exactly the point for some proponents of the measure—they seek an amendment that would force
Congress to cut spending during difficult times. Next,
I investigate what happens to state expenditures during recessions.
Responsiveness of expenditures to
business cycle
State government expenditure is divided into
five different categories—current spending, capital
spending, intergovernmental expenditures, interest
on the debt, and insurance trust expenditures. The
breakdown of expenditures is presented in the first
column of table 3 and in figure 3. Like revenues, state

10

per capita expenditures have been steadily increasing
since 1977 (see figure 1).
Overall expenditures are somewhat sensitive to
business cycle conditions, although less so than
revenues. The first row of table 3 shows that a 1 percentage point increase in the unemployment rate
increases overall expenditures by $9.23 ($4.14) per
capita. This is the combination of a $9.23 ($3.75) decline in general fund expenditures with an $18.46
($0.95) increase in insurance trust expenditures. Falling general fund expenditures are more than offset
by rising insurance trust spending.
Current expenditure
Current expenditure represents the biggest portion of state government expenditure at just over half
of the entire category. Current operations include
spending on a vast array of goods and services including transportation, hospitals, state educational institutions, and public welfare.19 As shown in table 3, current
expenditure is rather flat over the business cycle,
increasing by an insignificant amount when the unemployment rate rises.
Breaking current operations expenditures down
by the function they support, I find that during downturns public welfare spending increases, while spending on education (mostly higher education) and other
services falls. The increase in public welfare is not
surprising given that during downturns a greater
fraction of the population relies on the government
for support.
Capital expenditure
Capital expenditure is much more sensitive to the
business cycle than current expenditure. Table 3 shows
that capital expenditure per capita drops by $6.94
($1.23) when the unemployment rate increases by 1
percentage point. This drop is evenly split between
a decline in spending on construction and a decline
in other capital outlay (mostly comprising land and
equipment).20
Given that the benefits of capital projects are less
immediately apparent, spending on capital projects
may be politically easier to cut. In addition, states
have more discretion over capital spending because
it is less likely than current spending to arise from entitlement programs. Capital spending is also naturally
less persistent. Although a state cannot easily close a
university to bring about budget balance, it can slow
down major capital projects or wait to begin new ones.
The role of this reduction in capital spending is
interesting in light of the fact that state capital budgets are outside the operating budgets directly affected by balanced budget restrictions. It suggests that

Economic Perspectives

TABLE 3

Average yearly expenditure per capita, 1977–97
(dollars)
Average
per capita value

Effect of 1 percentage
point increase in
unemployment rate

Total expenditures

2,640.54

9.23 (2.232)

General fund expenditures
Current operations
Education
Public welfare
Other current operations
Capital expenditure
Construction
Other capital outlay
Intergovernmental expenditures
To school districts
To other local
To federal government
To education
To public welfare
To other
Interest payments on the debt

2,384.47
1,349.71
358.64
404.66
586.41
239.85
192.98
46.88
694.54
386.76
302.19
5.59
464.57
43.42
186.55
100.37

Budget category

Insurance trust expenditure
Unemployment benefits
Other trust payments

256.07
99.23
156.84

–9.23
2.75
–2.43
7.09
–1.91
–6.94
–3.57
–3.37
–4.97
–4.22
–0.77
0.02
–3.74
2.00
–3.23
–0.07

(2.462)
(1.102)
(4.190)
(4.776)
(1.288)
(5.640)
(3.398)
(7.996)
(3.180)
(3.158)
(0.648)
(0.294)
(3.077)
(4.254)
(4.248)
(0.118)

18.46 (19.500)
17.71 (28.736)
0.75 (1.039)

Notes and source: See table 1.

FIGURE 3

State expenditure areas
Interest payments
on the debt 4%

Insurance
trust
10%

Intergovernmental
to other local
governments
11%

Intergovernmental
to school districts
15%

Current
operations
51%

Capital
9%

Source: Author’s calculations from U.S. Department of
Commerce, Bureau of the Census, 1977–87, Survey of State
Government Finances, Washington, DC: Government Printing
Office, and U.S. Department of Commerce, Bureau of the
Census, 1988–97, Survey of State Government Finances ,
available on the Internet at www.census.gov/govs/state/
www.census.gov/govs/state/.

Federal Reserve Bank of Chicago

states reduce pressure on their operating budget
by reducing capital spending. When I compare debt
issuance to capital spending, I find that if all revenues
from debt issuance were spent on capital projects,
only 60 percent of the money for capital projects
would be financed by debt.21 This indicates that
states finance a large portion of capital expenditure
out of current revenues.
Intergovernmental expenditure
States transfer money to local governments and
to the federal government. The great majority of these
funds go to school districts and to general-purpose
local governments, such as county, municipal, and
township governments. Only a small sum is transferred to the federal government. As shown in table 3,
overall intergovernmental expenditures fall when the
economy worsens.
I break up intergovernmental expenditures in two
different ways. First, I divide them by recipient government: school districts, other local governments,
and federal government. Second, I divide them by

11

function: education, public welfare, and other. While
transfers to the federal government and to local governments are relatively flat over the business cycle,
transfers to school districts drop off significantly
when the economy worsens. The functional breakdown yields the same picture, with declines in education spending being the main explanation for the
overall reduction in intergovernmental revenues. By
contrast, as with federal intergovernmental revenues
and current operations, public welfare intergovernmental spending increases during downturns as
states transfer more money to localities to support
swelling public assistance rolls.
Interest expenditures
States pay interest on general debt and interest
on the debts of public utilities, with the general debt
accounting for the bulk of interest paid. As shown in
table 3, interest expenditures are largely acyclical.
Although state debt may increase during difficult
economic times, as explained further below, the stock
of debt and, hence, interest payments are quite flat
over time.
Insurance trust
Insurance trust expenditures are benefit payments
to recipients under the state’s employee retirement,
workers compensation, unemployment insurance,
and other trust funds. In total, as shown in table 3,
insurance trust expenditures are highly procyclical,
increasing by $18.46 ($0.95) or about 7 percent of the
mean when the unemployment rate increases by 1
percentage point.
Given that unemployment benefits are one source
of insurance trust expenditures, the size of this increase
is not surprising. During times of high unemployment,
unemployment benefit benefits greatly increase. In
fact, all of the increase in insurance trust spending
that occurs when unemployment is high can be attributed to increases in spending for unemployment benefits.
Expenditure results and implications
During times of economic difficulty, states are
able to decrease their general fund expenditures by
$9 per capita in spite of increasing pressure on public
welfare spending. States do this in three ways: They
decrease higher education current expenditure; they
drastically reduce capital expenditure; and they cut
the funds going to school districts.
The implications of this for the federal government are mixed. There is no reason to believe that the
federal government would not be able to cut current
expenditure in some areas in response to recessions.
While the size of federal government entitlement

12

programs limits government flexibility, the federal
government has some areas of responsibility that are
akin to state governments’ higher education responsibilities. The most obvious area is that of education,
training, employment, and social services, but cuts in
other areas would also be possible.
The states’ ability to decrease capital spending
is important in helping them to achieve budget balance.
In fact, the drop in state capital spending almost totally
offsets the increase in current public welfare expenditure brought about by a 1 percentage point increase
in the unemployment rate. However, whether the federal budget would or should follow the states’ lead in
this arena is a difficult question. Some of the same
factors causing the states to decrease capital spending during recessions may also affect the federal government. In particular, because capital spending has
current costs and longer term benefits, cuts in capital
spending may be politically easier to swallow than
cuts in federal spending on education or job training.
In addition, the absence of a federal capital budget
may make federal capital spending even more responsive to economic conditions. It is possible that states
do not reduce their capital spending further because
they can issue debt for capital projects. Therefore,
their ability to alleviate general budget pressures is
limited by the portion of capital spending that is being
financed by current revenues.
However, there is one important reason that federal capital spending may not be as susceptible to the
business cycle as state capital spending. While the
majority of state capital spending is for highways and
higher education, projects that may be easy to delay,
the great majority of federal capital expenditure goes
to finance defense. Between 1977 and 1997, 82 percent
of the money spent on direct federal capital expenditure was used for defense.22 In no year did defense
spending drop below 70 percent of total direct capital
expenditure. It seems unlikely that federal defense
spending would or should be a function of business
cycle conditions. A brief glance at the numbers demonstrates that, historically, defense capital spending
has been more a function of the political climate and
whether the nation is at war than of the unemployment
situation. For example, from 1943–46, at the height of
U.S. involvement in World War II, defense capital
goods represented about 99 percent of federal capital
expenditure on average. The federal government
could cut capital spending in other areas, but nondefense capital spending is a very small part of the budget—averaging only 1.6 percent between 1977 and
1997 (total capital spending averaged 9 percent of the
federal budget over the same period).

Economic Perspectives

In addition to reducing current spending for education and capital expenditure, state governments
reduce overall intergovernmental grants, especially
those to school districts. In general the states take
advantage of their unique position in the intergovernmental structure by procuring additional grants
from the federal government while sending less money to the local governments. The federal government
could follow the states lead here by reducing intergovernmental expenditures to the states during times
of economic stress. While this may improve the federal government’s budgeting position, it would make
it more difficult for the states to balance their budgets.
Part of the reason state governments are able to come
close to balancing their budgets is that the federal
government does not achieve a balanced budget.
The federal government could not avail of the
overall expenditure strategy relied on in the states
because of its unique responsibility to provide for
national defense. By contrast, the federal government
may be able to follow the states’ lead in cutting current
expenditure and in cutting grants to lower levels of
government. The wording of the federal balanced
budget amendment implies that the government would
need to cut spending to compensate for the entire
drop in revenues. However, state governments have
an important safety valve in their ability to issue debt
to fund capital projects. Next, I investigate the extent
to which they take advantage of this safety valve.
What happens to debt and assets?
The combination of the revenue and expenditure
pictures for both the general and insurance trust funds
is not very consistent with the common notion of
budget balance. During difficult times, general fund
revenues fall more than expenditures, and trust fund
expenditures increase more than revenues. This implies that states must either deplete assets or issue
debt when the economy deteriorates. In other words,
their net asset position must worsen. Below, I look
at what happens to state debt issuance and state
reserve funds, both inside and outside the insurance
trust system.
Short-term debt
Short-term debt is issued to account for unexpected shortfalls. This category includes debt payable
within one year of issuance or debt backed by taxes
to be collected in the same year. It includes items
such as tax anticipation notes and short-term warrants
and obligations, but excludes accounts payable and
similar less formal non-interest-bearing obligations.
States that are not allowed to carry over deficits still

Federal Reserve Bank of Chicago

sometimes have short-term debt in the form of tax
obligation notes and similar liabilities.
The Census Bureau only collects two short-term
debt items (in stark contrast to the approximately 50
different measures of long-term debt)—the amounts
outstanding at the beginning and the end of the fiscal
year. I use the amount outstanding at the end of the
year; given that most short-term debt has a maturity
of under one year, this is a reasonably good proxy for
issuance. 23 As table 4 shows, short-term debt is fairly
responsive to the business cycle, increasing by about
$2.41 ($0.56) for a 1 percentage point increase in the
unemployment rate. However, this only goes part of
the way in explaining how states finance the growing
gap between revenues and expenditures during
downturns. States also rely on additional long-term
debt issuance.
Long-term debt and government assets
Because long-term debt and asset data before
and after 1988 are not comparable (due to a classification change in 1988), I use data from 1989 onwards.
State government long-term debt and asset data are
far more complicated than other financial data for
three main reasons. First, over 40 percent of state
government debt is “public debt for private purposes.” This debt is issued using the tax-exempt status
of state governments to finance expenditures by private firms. I analyze this debt separately from government purpose debt.24 Second, not all debt issuance
funds contemporaneous expenditures. Some debt is
issued to refund previously issued debt. Because a
lot of state debt is callable (that is, it can be redeemed
prior to maturity for a prespecified premium), when
interest rates are falling, states can realize savings if
they call debt and refund it at a lower interest rate.
Because I am interested in debt issuance that contributes to the state’s concurrent fiscal situation, I would
ideally like to look only at new government purpose
debt issued, that is, net of refunding. Unfortunately,
I cannot do this because debt issued for refunding
cannot be separated into public and private purpose
debt. Third, an analysis of debt cannot be separated
from an analysis of government assets because two
of the three state government asset measures are
directly related to debt. The sinking fund contains
money explicitly saved for debt redemption, while
the bond fund contains the proceeds of bond issuance prior to disbursement. Only the “other funds”
category contains assets not explicitly linked to debt.
Because these assets are all stocks rather than flows,
I look at the change in value per capita from one
year to the next as the appropriate measure of government assets.25

13

TABLE 4

Debts and assets, 1989–97
(dollars)

Budget category
Short-term debt, 1977–97
Outstanding at end
Long-term debt, 1989–97
Issuance
Governmental purposes
Private purposes
Refunding
Redemption/retirement
Governmental purposes
Private purposes
Retired by refunding
Government assets, 1989–97
Change in sinking fund
Change in bond fund
Change in other funds
Insurance trust assets, 1978–97
Change in employee retirement
Change in unemployment insurance
Change in worker’s compensation
Change in other trust assets

Average
per capita value

Effect of 1 percentage
point increase in
unemployment rate

14.94

2.41 (4.299)

155.86
145.37
37.50

16.06 (2.604)
10.90 (1.755)
10.89 (2.680)

104.77
109.64
36.11

18.99 (4.228)
8.26 (1.815)
9.93 (2.619)

–8.10
0.71
35.83

–10.97 (2.553)
–7.51 (2.432)
3.61 (0.489)

183.01
7.43
8.35
0.11

2.84 (0.466)
–8.54 (7.730)
–0.33 (0.268)
0.04 (0.524)

Notes and source: See table 1.

Table 4 shows the relationship between the state
unemployment rate and the state long-term debt issuance, redemption, and asset measures. The first thing
to notice is that all measures of debt issuance increase
significantly during downturns. Because nearly all
long-term debt is used to finance capital projects and
because capital spending drops off quite significantly during downturns, the increase in debt issuance
suggests that state governments finance a higher percent of their capital spending with debt during recessions. This implies that states use debt issuance as
an important safety valve during recessions. The decrease in the state bond fund, also shown in table 4,
supports this finding. Although states spend less on
capital projects, they both draw down unspent monies
from previous bond issuance and issue more bonds.
As with issuance, all three measures of debt
redemption also increase during downturns (also in
table 4). This result is more intuitive than it may appear
when combined with the information on the change
in the value of the sinking fund.26 States redeem more
debt during downturns, but it appears that this extra
money is coming from a combination of debt refunding

14

(which increases by $9.93 per capita) and a drop in
the value of the sinking fund (which decreases by
$10.97 per capita) rather than from current revenue
sources. The transfers from the sinking fund probably
occur because of cyclical changes in financial market
conditions. In particular, states have an incentive to
pay off debt using sinking fund assets when they are
paying more interest on existing debt than they are
receiving from fund assets. In short, during good
times, states accumulate assets in their sinking funds
that are then spent to call bonds when the economy
worsens and interest rates fall. Finally, there is no
evidence of changes in the assets of non-bondrelated funds.
Assets of the trust funds
One of the most frequently articulated worries
about a balanced budget requirement is that it would
lead to the depleting of social security reserves in
a downturn. Do state government deplete the assets
of state managed trust funds in downturns? I look
at the change in the assets of all four types of government trust funds—employee retirement, workers

Economic Perspectives

compensation, unemployment insurance, and others.
The employee retirement trust fund is the only one
that is directly comparable to social security. The other funds, particularly the unemployment insurance
trust fund, are supposed to fall during recessions.
Table 4 shows that there is little evidence of systematic raiding of the trust funds. While state unemployment insurance trust funds decline dramatically
during downturns, there is no evidence that the assets
of other trust funds fall.
Debt and assets results
I find that states issue more short-term and longterm debt during recessions. As mentioned above, the
federal balanced budget amendment does not allow
any new debt issuance short of a super-majority
vote. Therefore, this avenue would not be open to
the federal government. Instead, the federal government would be compelled to find areas in which to
cut spending in order to confront revenue declines.
Conclusion
Both state and national balanced budget supporters frequently cite the experience of the states to demonstrate the feasibility of a federal balanced budget
amendment. State governors and U.S. presidents alike
have claimed that the state experience is a relevant
example to the federal government. In this analysis of
the way that state budgets respond to the business
cycle, I find few examples of methods for budget balance used by the states that are directly relevant to
the federal government. This is the case for four principal reasons.
First, state balanced budget requirements differ
in one major way from the amendment currently contemplated at the federal level. State governments can
and do issue both short-term and long-term debt to

finance shortfalls and capital projects, respectively.
The states are able to issue long-term debt because
state capital projects are outside the restrictions
imposed by the balanced budget amendments.
Second, despite the fact that state capital budgets
are separate, states cut capital spending quite drastically during downturns in order to relax budgetary
pressures. The current costs and delayed benefits of
capital spending make it politically easier to cut. The
federal government may not find capital spending so
easy a target because most federal capital spending
is for defense.
Third, states take advantage of their unique position in the federal system to cut funds going to local
governments while drawing on increased funds from
the federal government. The federal government can
not draw down more money from a higher level of
government, but could potentially decrease the money
it sends to the states.
Finally, states increase tax rates during downturns and decrease them during booms. The states
are able to engage in this behavior because, unlike
the federal government, they are not perceived as
being responsible for the macroeconomy.
Overall, the state experience with budget balance
and business cycles is not a very relevant model for
the federal government. State governors are not responsible for the macroeconomy or for national defense
and, in general, confront a more relaxed budget restriction than that proposed for the federal government.
Policymakers need to consider carefully how budget
balance at the federal level could be achieved during
an economic downturn under a balanced budget
restriction—for example, which taxes could be increased, which programs could be cut, or which
assets could be sold.

NOTES
Briffault (1996) provides an interesting set of quotations
suggesting that the state experience is relevant to the federal
government.
1

For a further discussion of limits on long-term debt, see
McGranahan (1999b).
5

These comparisons actually underestimate the difference
between the states and the federal government because, while
the federal numbers are net of trust fund interest revenues, the
state numbers are gross. I do not net out state interest revenues
because the definition of interest revenues changed in 1988 to
include revenues from public debt for private purposes. Therefore, it is impossible to calculate a net number for the states
that is consistent over time. The gross numbers for the federal
government would be 18 percent for expenditures and 21 percent for revenues.
6

The National Association of State Budget Officers (1997)
states that 40 of 48 states that responded to a survey report
that their capital planning occurs in a capital budget.
2

The exact definition of what capital spending consists of differs
by state. This is the most common definition.
3

Forty-eight states have either a constitutional or statutory
balanced budget requirement. One state that does not is not
permitted to carry over deficits. These combine to generate
the frequently cited figure that 49 states have balanced budget
restrictions. The exception is Vermont.
4

Federal Reserve Bank of Chicago

The year refers to the calendar year in which the fiscal year ends,
so fiscal 1999 ended in most states on June 30, 1999. Some states
7

15

have different fiscal years. I take these differences into
account when calculating the fiscal year unemployment rate.
One disadvantage of using the unemployment rate is that it
is often viewed as a lagging indicator of economic activity.

Contributions by the state to its own insurance trust systems
are considered within government transfers and do not enter
the revenue tabulations.
18

8

For a discussion of the effects of different balanced budget
restrictions in the states, see Poterba (1994).
9

This division is analogous to the separation between onbudget and off-budget in the federal context, because the federal
budget excludes most social security funds.

I include spending on assistance and subsidies in the current
expenditure category. It is only a small portion of total current
expenditure. In published Census tables, assistance and subsidies
(which include scholarships, veterans benefits, and some welfare
payments) are usually presented separately.
19

10

With a standard error of $3.32; note that table 1 shows
t-statistics rather than standard errors.
11

I include receipts of utilities and liquor stores run by the state
in charges. In Census Bureau statistics, these are treated separately. They are generally very small and do not warrant separate treatment.
12

Sales taxes refer to all sales and gross receipt taxes, including
general sales, gas, and tobacco taxes. Income taxes refer to
both individual and corporate income tax collections.
13

For a discussion of tax revenue changes taking statutory
changes into account, see Dye and McGuire (1998).
14

15

For further discussion of this issue, see McGranahan (1999a).

The classification manual defines property sales as “amounts
received from sale of real property, buildings, improvements
to them, land easements, rights-of-way, and other capital assets
(buses, automobiles, etc.), including proceeds from sale of
operating and nonoperating property of utilities. Includes
sale of property to other governments.”
16

Interestingly, the historical relationship between federal
property sales and the unemployment rate has been negative,
indicating that the federal government sells less when the
economy is bad.

There were some minor changes in coding of some of the
capital outlay variables in 1988. Looking only at data from
after this change yields very similar conclusions—capital
expenditure falls off, mostly driven by changes in spending
on equipment and existing land and structures.
20

The 60 percent number represents the average of debt issuance
divided by capital spending from 1988–97. Debt issuance
excludes debt for private purposes but is not net of refunding.
21

Direct capital expenditure excludes grants. Many grants are
to state governments for highways and other programs.
22

The U.S. Department of Commerce, Bureau of the Census
(1995) reports that “obligations having no fixed maturity date
(even where outstanding for more than one year if payable
from a tax levied for collection in the same year it was issued)”
are included in short-term debt.
23

The major reclassification in 1988 pertains to changes in
the categorization of public debt for private purposes. Prior
to 1988 it is not possible to fully separate it from other debts.
The spending supported by public debt for private purposes
does not show up in the states’ expenditure measures.
24

This is not the per capita change, but the change per capita
where the population is the population in the second year.
25

17

I subtract the value of public debt for private purposes outstanding from the sinking fund numbers to account for the fact
that the value of collateral pledged for private purpose debt is
included in the sinking fund numbers.
26

REFERENCES

Bond Market Association, The, 1999, “Daily report of
municipal bond transactions” available on the Internet
at www.investinginbonds.com/, accessed April 20.
Branstad, Terry E., 1997, “Balancing the budget:
What Washington can learn from the states,” Heritage
Lecture, No. 586, available on the Internet at www.
heritage.org/library/categories/budgettax/hl586.html,
May 13.
Briffault, Richard, 1996, Balancing Acts: The Reality
Behind State Balanced Budget Requirements, New
York: The Twentieth Century Fund Report.
Dye, Richard F., and Therese J. McGuire, 1998,
“Block grants and the sensitivity of state revenues to
recession,” in National Tax Association Proceedings
of the Annual Conference on Taxation, Washington,
DC, pp. 15–23.

16

Eckl, Corina, 1998, “States broaden the scope of
rainy day funds,” Washington, DC: National Association of State Legislatures, available on the Internet
at www.ncsl.org/programs/fiscal/rdf97.htm.
Eisner, Robert, with Robert M. Solow, and James Tobin, 1997, “Petition in opposition to Balanced Budget
Amendment,” in “Economists Oppose the Balanced
Budget Amendment,” Max Sawicky (lead contributor),
Internet discussion board of Communications for a
Sustainable Future, available at http://csf.colorado.
edu/mail/pkt/jan97/0338.html, January 16.
Engler, John, 1997, “Text of the State of the State
address,” The Detroit News, available on the
Internet at http://detnews.com/1997/metro/9701/29/
01290057.htm, January 29.

Economic Perspectives

Illinois, State of, 1970, Constitution of the State of
Illinois, adopted at special election on December 15,
available on the Internet at www.legis.state.il.us/
commission/lrb/conmain.htm.
Keating, Frank, 1997, “The State of the State,” available on the Internet at www.oklaosf.state.ok.us/
osfdocs/sos97.html, February 3.
McGranahan, Leslie, 1999a, “Welfare reform and state
budgets,” Chicago Fed Letter, Federal Reserve Bank
of Chicago, No. 137, January.
__________, 1999b, “Voter preferences for capital
and debt spending: Evidence from state debt referenda,” Proceedings of the Ninety First Annual Conference on Taxation, Washington, DC: National Tax
Association, forthcoming
National Association of State Budget Officers, 1997,
“Capital budgeting in the states: Paths to success,”
available on the Internet at www.nasbo.org/pubs/
capbud97/capbud97.htm/.
National Conference of State Legislatures (NCSL),
1998, State Tax Actions 1998, Washington, DC:
National Conference of State Legislatures, available
on the Internet at www.ncsl.org/programs/fiscal/
sta97sum.htm, accessed April 1, 1999.
__________, 1997a, Capital Budgeting in the
States, Washington, DC: National Association of
State Budget Officers, September.
__________, 1997b, State Tax Actions 1997, Washington, DC: National Conference of State Legislatures,
available on the Internet at www.ncsl.org/programs/
fiscal/sta97sum.htm, accessed April 1, 1999.

Poterba, J., 1994, “State responses to fiscal crises: The
effects of budgetary institutions and politics,” Journal
of Political Economy, Vol. 102, No. 4, pp. 799–821.
Snell, Ronald K., 1999, “State balanced budget requirements: Provisions and practice,” National Conference of State Legislatures Fiscal Letter, available
on the Internet at www.ncsl.org/programs/fiscal/
0796fl.htm.
U.S. Department of Commerce, Bureau of the Census, 1999, Governments Finance and Employment
Classification Manual, Washington, DC, available
on the Internet at www.census.gov/govs/www/
class.html/.
__________, 1998, 1997 State Government Finance
Tables by State, available on the Internet at
www.census.gov/govs/www/stsum97.html.
__________, 1994, Government Finances: Summary
of Federal Government Finances—1991 to 1994,
available on the Internet tat www.census.gov/govs/
fedfin/federal.txt.
__________, 1990, State Government Finances in
1989, Washington, DC: U.S. Government Printing
Office, Series GF-89-3.
__________, 1988–97, State Government Finances,
available on the Internet at www.census.gov/govs/
state/.
__________, 1977–87, State Government Finances,
Washington, DC: U.S. Government Printing Office,
electronic data provided by the Census Bureau Governments Division, Series GF-Year-3.

__________, 1996, State Tax Actions 1996, Washington, DC: National Conference of State Legislatures,
available on the Internet at www.ncsl.org/programs/
fiscal/STAEX.HTM, accessed April 1, 1999.

U.S. Government, Office of Management and Budget,
1999, “Budget of the United States government, fiscal
year 2000, historical tables,” Washington, DC: U.S.
Government Printing Office, available on the Internet
at www.access.gpo.gov/usbudget/fy2000/.

__________, 1995, State Tax Actions 1995, Washington, DC: National Conference of State Legislatures,
available on the Internet at www.ncsl.org/programs/fiscal/STA95P1.HTM, accessed April 1, 1999.

U.S. Senate, 1997, “S.J. Res. 1, proposing an amendment to the Constitution of the United States to
require a balanced budget,” available on the Internet
at http://thomas.loc.gov/, accessed April 19, 1999.

Federal Reserve Bank of Chicago

17

Birth, growth, and life or death of
newly chartered banks
Robert DeYoung

Introduction and summary
Thousands of new commercial banks have been chartered in the U.S. over the past two decades. As the
U.S. banking industry continues to consolidate, these
de novo banks are potentially important for preserving competition and providing credit in local markets.
However, like other new business ventures, newly
chartered banks initially struggle to earn profits, and
this financial fragility makes them especially prone to
failure. In this article, I document the financial evolution of the typical de novo bank and develop and test
a simple theory of why and when new banks fail.
Recent decades have seen an upsurge in the number of mergers and failures among new banks. Figure 1,
panel A shows the annual change in the number of
commercial bank charters in the U.S. since 1966. Prior
to 1980, the reduction in bank charters due to mergers
and failures was relatively stable at about 100 charters
per year, or about 1 percent of the industry total (figure 1, panel B). The pace accelerated greatly after 1980,
and since 1986 about 600 charters, or 5 percent to 6
percent of the industry total, have disappeared each
year due to mergers and failures.
To a large extent, this tremendous consolidation
can be explained by the repeal of federal and state
laws that restricted branch banking and interstate
banking. As these restrictions gradually were relaxed,
banking companies expanded their geographic reach
by acquiring thousands of other banks, and reduced
their overhead expenses by converting thousands of
affiliate banks into branch offices. This geographic
expansion, combined with newly deregulated deposit
rates, increased competition between commercial banks
just when new information technology was allowing
mutual funds, insurance companies, and the commercial paper market to compete for banks’ traditional loan
and deposit businesses. Under these new competitive
conditions, many commercial banks became more vulnerable to economic downturns, and thousands of

18

banks failed during the 1980s and early 1990s. Over the
past two decades, the combined effect of these mergers and failures has reduced the number of commercial
banks in the U.S. by nearly 40 percent.
This consolidation has been partially offset by
a recurring wave of new bank charters. As shown in
figure 1, panel A, over 3,000 de novo commercial banks
have been chartered by state and federal banking authorities since 1980. It is generally believed that these
newly chartered banks can help restore competition
in local markets that have experienced a large amount
of consolidation. It is also commonly believed that
these newly chartered banks can help replace credit
relationships for small businesses whose banks failed
or were acquired or reorganized. However, before a
newly chartered bank can provide strong competition
for established banks and before it can be a dependable
source of credit for small businesses, it must survive
long enough to become financially viable.
I begin by examining the conditions under which
investors are likely to start up new banks, including
the influence of business cycles, merger activity in
local banking markets, and the policies of federal and
state chartering authorities. Next, I track the evolution
of profits, growth rates, capital levels, asset quality,
overhead costs, and funding mix at more than 1,500
commercial banks chartered between 1980 and 1994.
These data suggest that newly chartered banks pass
through a period of financial fragility during which
they are more vulnerable to failure than established
Robert DeYoung is a senior economist and economic
advisor at the Federal Reserve Bank of Chicago. The
author thanks Iftekhar Hasan and Curt Hunter for their
advice; Richard Cahill, Philip Jackson, and participants at a Federal Reserve Bank of Chicago seminar for
helpful comments; Eli Brewer and David Marshall for
reading an earlier draft of this article; and especially,
Nancy Andrews for outstanding data support.

Economic Perspectives

FIGURE 1

Entry and exit of commercial banks
A. Change in number of U.S. commercial bank charters
number

New charters
Failed banks
Merged banks

B. Change in number of U.S. commercial bank charters, as a percent of total charters
percent

New charters
Failed banks
Merged banks

Source: Federal Deposit Insurance Corporation, 1966–98, “Change in number of insured commercial banks,”
available on the Internet at www2.fdic.gov/hsob/.

banks. Specifically, new bank capital ratios quickly
decline to established bank levels, but new bank
profits improve more slowly over time before attaining
established bank levels.
Based on these empirical observations, I develop
a simple life-cycle theory of de novo bank failure, in
which the probability of failure at first rises, and then
declines with the age of the new bank. I use hazard
function analysis to test this simple theory for 303
new commercial banks chartered in 1985, just as the
wave of bank failures shown in figure 1 was picking
up steam. The tests offer support for the simple theory.
On average, the results suggest that newly chartered
banks are less likely to fail than established banks
during the first few years of their lives; however, new
banks quickly become substantially more likely to fail
than established banks; and, over an extended period
of time, new bank failure rates gradually converge to
the failure rates of established banks.
What are the implications of these results for
bank supervision and bank competition policy? The
results suggest that the policies in place during the
1980s successfully insulated new banks from economic

Federal Reserve Bank of Chicago

disruptions early in their lives, but were less successful in preventing new banks from failing after the initial years. Clearly, de novo bank failure rates could be
reduced by requiring investors to supply higher
amounts of start-up capital or by requiring banks to
maintain extranormal capital-to-asset ratios in the early
years—indeed, the latter policy option was adopted
by federal bank supervisors during the 1990s. However, failure-proofing de novo banks is not an optimal
policy. The social costs of small bank failure are relatively low, and setting higher capital requirements
would at some point discourage investment in new
banks and thereby limit the competitive benefits of
de novo entry.
Birth of new banks
As illustrated in figure 1, panel A, the number of
new banks started up each year has ebbed and flowed
over the past three decades. There are a number of
explanations for these patterns. Like all new business
ventures, new banks are more likely to form when
business conditions are good. For example, new bank
charters bulged to well over 250 per year during the

19

general economic expansion of the mid-1980s. This
high rate of bank start-ups also coincided with the
relaxation of unit banking laws in a number of states,
laws that had prevented banking companies from
operating affiliates in multiple locations. The steady
decline in new charters during the late 1980s and early
1990s, which bottomed out at about 50 new banks per
year, also had multiple causes. Difficult times in regional
banking markets made new bank start-ups unprofitable in many regions (bank failures reached their peak
in 1988), and a national recession in the early 1990s
reinforced this trend. New bank charters have been
on the increase since then, reaching over 100 per year
in 1997 and 1998, in large part due to the extended
economic expansion of the 1990s.
Conditions in local banking markets also influence
bank start-ups. Moore and Skelton (1998) find that
there are more de novo banks 1) in markets that are
experiencing healthy economic growth, 2) in highly
concentrated banking markets in which competition
among existing banks is weak, and 3) in markets where
small banks are under-represented and, hence, small
businesses are not being adequately served. These
results imply that new banks will be more likely to start
up in local markets where mergers have reduced the
number of competing banks, and where the resulting
market power has reduced the level of banking services. In such markets, new banks should receive a
profitable welcome from customers unhappy with
paying high prices for financial services or from businesses whose credit relationships were disrupted
when their bank was acquired or failed. Researchers
only recently began investigating these phenomena,
so there is not yet a consensus on the results. In a
study of de novo bank entry in all U.S. markets between 1980 and 1998, Berger, Bonime, Goldberg, and
White (1999) find that the probability of de novo entry
is higher in local markets that have experienced mergers
or acquisitions during the previous three years, particularly mergers and acquisitions involving large banking organizations. In contrast, Seelig and Critchfield
(1999) find that local market entry by acquisition deters
entry by de novo banks and thrifts.1 Their results
are based on a study of de novo banks and thrifts
between 1995 and 1997, a time when banking conditions were exceptional and restrictions on geographic
mobility were virtually nonexistent.
Differences in the policies of the legal authorities
that grant commercial bank charters can also affect the
rate and location of new bank start-ups. A de novo
national bank receives its charter from the Office of
the Comptroller of the Currency (OCC), while a de novo
state bank receives its charter from the banking

20

commission of the home state. The OCC has historically been more liberal in granting charters than most
state authorities. Its policy has been that market forces,
not the chartering authority, should determine which
local markets need and can support new commercial
banks. In contrast, many state chartering authorities
have historically applied convenience and needs tests
when considering applications for new bank charters,
denying applications if they judge that the convenience and needs of the banking public are already
adequately served. Although this federal–state difference in chartering philosophy has diminished over
time, DeYoung and Hasan (1998) find that national
banks were chartered with greater frequency than
state banks during the 1980s and early 1990s, and that
the financial performance of de novo national banks
initially lagged that of de novo state chartered banks.2
This suggests that national banks chartered during
the 1980s were likely to have had a higher probability
of failure than newly chartered state banks operating
under similar economic and market conditions.
A concern shared by all chartering authorities is
that newly chartered banks start out with enough
equity capital to survive through the several years
of negative earnings and rapid asset growth that is
typical of de novo banks. The dollar amount of startup financial capital required for approval might be $3
million, $10 million, or even as much as $20 million, depending on the proposed location and business plan
of the prospective bank. Larger amounts of start-up
capital are generally required for urban banks, for
banks locating in vibrant economic markets, and for
banks with business strategies that feature fast growth
(for example, a new Internet bank).
Once a new bank opens its doors for business,
regulatory scrutiny shifts from the applications staff
to the examination staff. Bank supervisors pay closer
attention to newly chartered banks than to similarly
situated established banks, although the difference in
treatment varies depending on the new bank’s primary regulator. Federal Reserve supervisors will conduct
full scope examinations for safety and soundness at
a newly chartered bank at six-month intervals (established banks are examined every 12 to 18 months) and
will continue to schedule exams at this frequency
until the bank receives a strong composite CAMEL
rating (that is, a rating of 1 or 2) in two consecutive
exams. The Federal Deposit Insurance Corporation
requires that all newly chartered state and national
banks maintain an 8 percent tier 1 equity capital-to-riskbased assets ratio for their first three years of operation, while the Federal Reserve requires new state
chartered Fed member banks to hold this ratio above

Economic Perspectives

9 percent for three years. These temporary extranormal capital requirements for new banks (the tier 1
requirement for established banks to be considered
adequately capitalized is only 4 percent) are a relatively recent supervisory response to de novo failure
experience of the 1980s. Bank supervisors also prohibit
new banks from paying out dividends for several
years and, in some cases, require new banks to maintain minimum levels of loan loss reserves.
Evolution of new banks
Relatively few research studies have examined
how banks grow and evolve in the years immediately
after they receive their charters.3 Brislin and Santomero
(1991) show that the financial statements of a new
bank can fluctuate rapidly and dramatically during its
first year. A handful of studies have examined how the
profitability of de novo banks grows over time (for example, Hunter and Srinivasan, 1990, and DeYoung
and Hasan, 1998). Another strand of research documents how small business lending becomes less important to de novo banks as they mature (for example,
DeYoung, Goldberg, and White, 1999). In this section,
I analyze how a broad group of de novo bank characteristics not typically considered in the literature evolve
over time, including de novo bank profits, growth rates,
capital ratios, sources of income, financing mix, overhead ratios, and loan quality.
Each of the eight panels in figure 2 examines a
different financial ratio and compares its average value for a sample of de novo commercial banks to its average value for a sample of established commercial
banks. The de novo bank sample includes 4,305 observations of commercial banks that were chartered between 1980 and 1994, were between one and 14 years
old when they were observed, and were located in
urban banking markets. The established bank sample
includes 4,305 observations of commercial banks that
were at least 14 years old when they were observed,
operated in urban banking markets, and were similar
to the de novo banks in terms of asset size. These
two samples of banks were originally constructed by
DeYoung and Hasan (1998). Box 1 contains additional
details about the two bank samples.
To construct each of the graphs in figure 2, I divided the de novo banks into 14 separate age groups
(one-year old banks, two-year old banks, etc.). I then
calculated the median average for the financial ratio
in question—say, return on assets (ROA)—for each
age group. Plotting these 14 average values in chronological order creates a time path showing how ROA
evolves as the typical de novo bank matures. Finally,
I superimposed the value of ROA at the 25th, 50th,

Federal Reserve Bank of Chicago

and 75th percentiles of the established bank sample
as horizontal lines over the de novo bank time path.
These horizontal lines serve as maturity benchmarks
against which to compare the progress of de novo
banks over time. The rate at which the de novo time
path converges with the maturity benchmarks indicates the speed at which the de novo banks mature.
While each of the individual graphs in figure 2
has a straightforward interpretation when considered
in isolation, these eight panels reveal a richer story
when they are interpreted in conjunction with each
other. For example, by itself the return on assets (ROA)
graph (panel A) merely confirms the results of existing
studies of de novo bank profitability, that is, that the
typical new bank loses money until it is about 18
months old and continues to underperform the average established bank for about a decade. But when
the ROA graph is considered together with the assetgrowth (panel B) and equity-to-asset (panel C) graphs,
a simple theory of de novo failure begins to emerge.
De novo banks average an extraordinary 20 percent
annual rate of growth during the first three years of
their lives. While this fast growth rate is increasing
the amount of assets against which new banks need
to hold equity capital, the losses suffered during the
first and second years of these banks’ lives are depleting their equity capital. Despite initially high capital
levels, the equity-to-asset ratio of the typical new
bank declines very quickly, entering the established
bank range after just three years. Thus, panels A, B,
and C suggest the probability of failure should increase as new banks pass their third year of life—
their capital has declined to established bank levels
by year three, but their asset growth and profitability
do not converge with those of established banks
until at least year ten.
The remaining five panels in figure 2 are consistent with the simple theory of de novo bank failure
suggested by the ROA, asset growth, and equity-toasset panels. For example, newly chartered banks initially have almost no nonperforming loans (panel D).
This is because these banks’ loan portfolios are composed disproportionately of unseasoned loans made
recently to borrowers who demonstrated strong financial fundamentals. However, as time passes some
of these new borrowers will naturally run into trouble,
and the quality of de novo banks’ loan portfolios will
naturally decline. This happens quite quickly for the
typical de novo bank, as its level of nonperforming
loans rises slightly above the median level for established banks after three years—just as de novo
banks are depleting their excess capital cushions and
well before new bank profitability rates have matured.

21

FIGURE 2

Financial ratio time paths for de novo banks
A. Return on assets

B. Annual asset-growth rate

percent

percent

years

years

C. Equity-to-asset ratio

D. Ratio of nonperforming banks

percent

percent

years

years

E. Ratio of interest-bearing assets

F. Fee income ratio

percent

percent

years

years

G. Ratio of large deposits

H. Accounting efficiency ratio

percent

percent

years

years

Notes: The data are described in box 1. The three colored horizontal bands are maturity benchmarks that indicate the twenty-fifth, fiftieth, and seventyfifth percentiles of the distribution of the ratio in question for the established bank sample. The black line plots the median value of the ratio in
question for the banks of various ages in the de novo bank sample. Return on assets is net income divided by total assets. Annual asset-growth rate
is the percent increase in total assets over the previous year-end total. Equity-to-asset ratio is the book value of equity divided by total assets. Ratio
of nonperforming loans is loans past due 90 or more days plus nonaccruing loans divided by total loans. Ratio of interest-bearing assets is total
performing loans plus total securities divided by total assets. Fee income ratio is noninterest income divided by net interest income plus noninterest
income. Ratio of large deposits is deposits in accounts greater than $100,000 divided by total deposits. Accounting efficiency ratio is noninterest
expense divided by net interest income plus noninterest income.
Source: National Information Center, 1988, 1990, 1992, and 1994, “Report of income and condition,” selected banks, December 31.

22

Economic Perspectives

The slow rate at which de novo bank profitability
improves appears to be attributable more to cost factors
than to revenue factors. Although the percentage of
de novo bank assets invested in interest-bearing
assets, such as loans and securities, starts out relatively
low and increases only slowly over time (panel E), the
typical de novo bank outperforms one-quarter of the
established banks in this area after only three years.
(De novo bank ROA does not reach the 25th percentile benchmark until year six.) Even more impressive is

the speed at which new banks develop the ability to
generate fee income (panel F). The typical de novo
bank outstrips the average established bank in feebased revenues after only three years, and outperforms three-quarters of the older banks in this area
after about nine years. By virtue of their newness, de
novo banks may be less constrained by the inertia of
existing customer relationships and existing employee habits and, therefore, may be better able to impose
fees on retail customers or to enter into less traditional

BOX 1

Financial ratio time path data
Both the de novo bank sample and the established
bank sample were taken from a primary data set
used originally in a study by DeYoung and Hasan
(1998). For the current study, I added variables
from the “Reports of income and condition” (call
reports). The primary dataset is an unbalanced
panel consisting of 16,282 observations of 5,435
small, urban commercial banks at year-end 1988,
1990, 1992, and 1994. Not all of the banks are present
in each of the four years because some banks failed,
were acquired, or received their charters during
the sample period. There are 2,611 banks present
in all four years, 977 banks in three of the four years,
1,005 banks in two years only, and 842 banks in just
one year.
Banks had to meet a number of conditions to
be included in the primary dataset. First, banks
had to have less than $500 million of assets (in 1994
dollars). By definition, newly chartered banks are
small, and established banks that are large will not
serve as good benchmarks against which to judge
the progress of young banks. Large banks have
access to production methods, risk strategies,
distribution channels, and managerial talent not
available to small banks. Second, all banks had to
be headquartered in metropolitan statistical areas
(MSAs). Demand for banking products, as well
as competitive rivalry among banks, can be quite
different in rural and urban markets, and may cause
young banks to develop differently in these two
environments. Third, banks had to be at least 12
months old at the time of observation. For example,
a bank that was chartered during 1993, but was
observed at year-end 1994, is referred to as a oneyear old bank. Brislin and Santomero (1991) find
that financial statements are quite volatile during
the first year of a bank’s operations, which makes
performance difficult to measure. Fourth, all banks
had to make loans and take deposits, eliminating
special purpose banks such as credit card banks.
Fifth, banks that were 14 years old or less (that is,

Federal Reserve Bank of Chicago

banks that would be in the de novo sample) were
excluded if they held more than $50 million in assets at the end of their first year. This filter prevents established banks that received new charters
as part of regulatory reorganizations and established thrift institutions converting to bank charters from being identified as de novo banks.
The resulting de novo sample comprises
4,305 observations of 1,579 different banks 14
years old or younger. Roughly 47 percent of these
de novo banks hold federal charters and roughly
21 percent are affiliates in multibank holding companies. The established bank sample was constructed by choosing 4,305 observations of 1,514
different banks, each more than 14 years old, from
the primary dataset. Roughly 25 percent of these
established banks hold federal charters and
roughly 27 percent are affiliates in multibank
holding companies. The established banks were
chosen to have roughly the same asset-size distribution as the de novo bank sample, as follows:
Banks more than 14 years old were grouped into
ten asset categories ($0–$50 million, $50–$100
million, ..., $450–$500 million). Established banks
were drawn at random from each of these size categories, depending on the number of de novo banks
of each asset size. The assets of the resulting established bank sample average $55.97 million with a
standard deviation of $49.64 million, compared
with the de novo bank sample average of $54.39
million and standard deviation of $48.70 million.
Obviously, there is no bright line that separates de novo banks from established banks. I chose
the 14-year old threshold for two reasons. First,
it is the maximum age at which previous studies
refer to commercial banks as de novo (see Huyser,
1986, and DeYoung and Hasan, 1998). Second,
choosing a relatively large number for this threshold ensures that the maturity benchmarks in figure
2 contain only banks that are fully mature.

23

fee-generating lines of business. In addition, de novo
banks tend to start up in markets where business
conditions are strong, and selling fee-based financial
services may be easier in these markets.
In contrast to their reasonably strong ability to
generate revenue, newly chartered banks have a difficult time controlling expenses. De novo banks initially
use large deposits twice as intensively as do established banks, and this disparity only slowly disappears (panel G). This suggests that de novo banks
tend to finance their fast asset growth by purchasing
funds rather than by growing their core deposit base.
All else being equal, this is an expensive and potentially risky financing strategy, because large depositors
are more sensitive to changes in interest rates than are
retail depositors and require higher rates to leave their
funds in the bank. The accounting efficiency ratio
graph (panel H) indicates that newly chartered banks
also have relatively high levels of overhead expenses
(for example, branch locations, labor expenses, and
computer equipment) and that these fixed factors of
production are not used at near full capacity for a
number of years. Excess overhead capacity not only
depresses bank profitability but, by increasing operating leverage, it makes bank profits more sensitive
to fluctuations in bank revenues.
Note that each of the panels in figure 2 exhibits
what is known as survivor bias, because some de novo
banks fail before they are 14 years old. For example,
average de novo ROA equals approximately 0.4 percent
for the three-year old banks, which is about twice as
large as the average ROA of 0.2 percent for the twoyear old banks. For the most part, this substantial improvement can be attributed to better performance as
young banks grow older and larger. But some amount
of this improvement occurs because some of the most
unprofitable de novo banks failed between years two
and three and dropped out of the sample. Although
this second explanation is responsible for only a small
amount of the large increase in ROA (as we shall see,
very few de novo banks fail after only two years of
operation), it is a good illustration of how survivor
bias can affect our results. Thus, the most exact way
to interpret the ROA time path is as follows: If a newly chartered bank survives to be three years old, one
would expect its ROA to be about 0.4 percent. I revisit
the issue of survivor bias when I estimate time to failure models in a later section (see Estimating hazard
functions section, starting on page 26).
Hypothetical hazard rates

likely to fail than established banks. Despite the losses
typically incurred during their first year of operation,
de novo banks initially have very high cushions of
equity capital and very low levels of nonperforming
loans. But the time paths in figure 2 also imply that
de novo banks become dramatically more likely to fail
as time passes, and quickly may become more likely
to fail than established banks. As de novo banks age,
their initially high capital cushions and low nonperforming loan ratios move rapidly toward established
bank levels—much more rapidly than their profitability reaches established bank levels.
The combined effect of these financial ratio time
paths on the timing and probability of de novo bank
failure is suggested by the hypothetical hazard functions in figure 3. A hazard function tracks changes over
time in the hazard rate, which is simply the probability
that a bank will fail at a particular time, given that it
has survived through all of the previous periods
leading up to that time.4 The horizontal line at P* represents the hypothetical hazard rate for established
banks, and the curved line plots the hypothetical hazard rate for newly chartered banks. Although this figure is highly stylized, the relative shapes of the two
functions are consistent with the combined financial
ratio time paths shown in figure 2.
The constant, non-zero hazard rate depicted in
figure 3 for established banks is an obvious simplification. Historically, established banks are more prone
to failure during recessionary periods, and almost
completely unlikely to fail during expansionary periods.
This simplification focuses attention on the issue of primary interest here, the failure rate of newly chartered
banks relative to the failure rate of established banks.
The hypothetical hazard rate for newly chartered
banks starts out at zero in figure 3, which makes sense
because these banks are so heavily capitalized at the
outset. But, as we saw in figure 2, de novo bank capital
FIGURE 3

Hypothetical probabilities of failure
New bank

P*

Established bank

0

time

The time paths in figure 2 imply that de novo banks
will at first be very unlikely to fail, perhaps even less

24

Economic Perspectives

percent with a standard deviation of 1 percent, then
its Z-score would equal 6.00. In this case, the bank’s
ROA would have to decline by 6 standard deviations
below its average (to –5 percent) for its losses to exhaust its capital cushion. Thus, the higher a bank’s
Z-score, the lower its probability of failure. Z will increase (that is, the probability of failure will decrease)
with higher levels of average ROA; Z will increase with
higher levels of equity to assets; and Z will increase
with lower variability in ROA.5
Table 1 displays Z-scores for the established
bank sample, for the de novo bank sample, and for
several subsamples of de novo banks. All of these
calculations employ the data used to construct the
graphs in figure 2. For each sample or subsample of
banks, Z is calculated using the median average of
ROA, the mean average of equity/assets, and the crosssectional standard deviation of ROA. (I use the median
ROA because the mean ROA is skewed downward
by banks that incurred large losses.) Because these
Z-scores are averages, they represent the likelihood
of failure for the typical bank in each sample.
In general, the calculations shown in table 1 suggest that becoming insolvent is a relatively unlikely
event for the typical bank in these samples. For example, the lowest Z-score (highest probability of insolvency) is 3.01, or about 3 standard deviations, for the
average three- to five-year old bank. Assuming that Z
is normally distributed, this implies only a 13 in 1,000
(0.13 percent) chance of becoming insolvent. Given
the large number of bank failures during the sample
period (see figure 1, panel A), the level of the failure

ratios decline to established bank averages after
about three years, while de novo bank profits, asset
quality, and growth rates do not reach (or return to)
established bank levels for around ten years. When
these time paths are considered simultaneously, they
imply the hypothetical patterns displayed in figure 3.
The hypothetical hazard rate for new banks increases
at first (for example, between ages one and three) as
new banks become increasingly vulnerable to economic fluctuations; it exceeds the established bank hazard
rate for a time (for example, after year three); and it
eventually declines to converge with established
bank levels (for example, around year ten). Regardless
of the exact shape and timing of the de novo bank
hazard function, it must eventually converge with the
established bank hazard function, because by definition new banks that survive eventually turn into established banks.
A rough way to check the relative accuracy of
the hypothetical hazard functions drawn in figure 3 is
to calculate Z-score probabilities of failure for de
novo banks and established banks. The Z-scores are
constructed as follows:

Z =

ROA + equity / assets
.
standard deviation of ROA

The Z-score indicates the number of standard deviations that ROA would have to fall below its average
value in order to wipe out 100 percent of the bank’s
equity capital. For example, if a bank has 5 percent
equity capital and, on average, it earns ROA of 1

TABLE 1

Average Z-scores
Components of average Z-score
Number
of banks

Median ROA

De novo banks
1 to 14 years old
Less than 3 years old
3 to 5 years old
6 to 10 years old
11 to 14 years old

4,305
667
1,424
1,570
644

.0057
.0006
.0050
.0074
.0089

Established banks
More than 14 years old

4,305

.0097

Cross-sectional
standard
deviation of ROA

Average
Z-score

.0957
.1231
.0814
.0762
.0795

.0230
.0206
.0287
.0195
.0155

3.97
6.00
3.01
4.28
5.70

.0867

.0129

7.48

Mean capitalto-assets ratio

Notes: Z-scores were calculated using formula described in the text and data described in box 1.
The selected commercial banks are also described in box 1.
Source: Author’s calculations based on year-end data from the National Information Center,
1988, 1990, 1992, and 1994, “Report of income and condition,” selected banks.

Federal Reserve Bank of Chicago

25

probabilities implied by these Z-scores is probably
too low.6 However, these Z-scores are still useful,
because they summarize the information in figure 2
into a single number that ranks the probability of failure across banks of different ages.
Overall, the analysis suggests that newly chartered banks are more likely to fail than established
banks: The average de novo Z-score of 3.97 is considerably smaller than the average established bank
Z-score of 7.48. On average, de novo banks and established banks have nearly identical capital-to-asset
ratios, so any difference in their implied failure rates
must be due to the level and variability of ROA. Indeed,
the median de novo bank ROA is only about half as
large as the median established bank ROA (.0057 versus .0097), and ROA is nearly twice as variable across
the de novo banks than across the established banks
(.0230 versus .0129).
Analyzing the Z-scores across de novo banks
of different ages provides some support for the shape
of the de novo bank hazard function in figure 3. The
implied probability of failure is relatively low for banks
less than three years old (Z = 6.00); is substantially
higher for three- to five-year old banks (Z = 3.01); and
then gradually declines toward established bank levels
for banks that survive beyond five years (Z = 4.28)
and beyond ten years (Z = 5.70). Looking at the components of these average Z-scores reveals why the
probability of failure changes as new banks mature.
The youngest group of de novo banks are the least
likely to fail because their earnings are relatively stable
(although they average near zero) and their capital
cushions are large. The three- to five-year old de novo
banks are more likely to fail because, although they
have higher average earnings, their capital cushions
have been depleted and their earnings are highly variable. Once banks are five to ten years old, increasing
earnings, increasing capital, and declining earnings
volatility all contribute to a reduced probability of
bank failure.
Estimating hazard functions
Next, I test whether the hypothetical hazard functions in figure 3 accurately depict the relative rates at
which newly chartered banks and established banks
fail. The Z-score analysis discussed above provides
some support for these hypothetical hazard functions,
but that evidence is crude at best and suffers from
survivor bias in the data. In this section, I employ
more sophisticated techniques to estimate hazard
functions for both newly chartered and established
banks. These techniques explicitly account for survivor

26

bias caused by failures and acquisitions during the
sample period. In addition, these techniques generate
continuous (or nearly continuous) hazard functions
that can be plotted against time, making them easy to
compare with the shape of the hypothetical hazard
functions in figure 3. Finally, one of these techniques
tests whether differences in de novo and established
bank failure rates are caused by differences in these
banks’ locational, regulatory, or organizational characteristics.
Data on bank failures
Table 2 displays some summary statistics for a
bank failure dataset Federal Reserve Bank of Chicago
staff created for the purpose of this study. This dataset
contains 56 quarters of information on 2,653 banks
from 1985 through 1998, and is constructed from the
“Reports of income and condition” (call reports) and
from the failures, transformations, and attributes tables
in the National Information Center database. The
dataset includes 303 newly chartered commercial
banks that opened their doors during 1985 and 2,350
established commercial banks that had been in operation for at least 25 years in 1985. The established
banks each had less than $25 million in assets (1985
dollars); had equity capital equal to at least 5 percent
of their assets; and were located in states in which
at least four de novo banks started up in 1985. The
dataset tracks each of these 2,653 banks across time
and records the quarters in which banks left the
dataset because they either failed or were acquired
by another bank.
These data cover a period during which there
were economic disruptions of sufficient magnitude to
cause a statistically meaningful number of bank failures.
Commercial bank failures were extremely rare in the
U.S. during the 1950s, 1960s, and 1970s, due to generally good economic times, regulatory limits on the risks
that banks could take, and legal entry barriers that
protected banks from competition. But the combination
of banking deregulation and volatile interest rates
during the 1970s and 1980s exposed banks to greater
risks and more competition. As seen in figure 1, panel
A, bank failures accelerated from near zero in 1980 to
over 100 failures per year from the mid-1980s through
the early 1990s. The catalyst for these bank failures
was a series of substantial economic disruptions,
including a general recession in the early 1990s and
a number of regional recessions in the mid- to late
1980s, the most disruptive of which was due to land
price deflations in Texas and other oil-producing states.
For the purposes of this article, I consider a bank
to have failed when at least one of the following

Economic Perspectives

TABLE 2

Descriptive statistics for hazard function data sets

characteristics affect the probability of
bank failure.

Nonparametric hazard functions
I use the bank failure data, summarized in table 2, to estimate separate hazNumber of banks
303
2,350
ard functions for newly chartered banks
and established banks, and then compare
<1
> 25
Age of banks in 1985 (years)
these estimated hazard functions with
43.23
84.17
Federal charters (%)
the hypothetical hazard functions in
78.55
18.81
Urban locations (%)
figure 3. I employ two different hazard
32.34
11.62
Multibank holding company (%)
function techniques to produce these
Southwest states (%),
estimates—a nonparametric, or actuarial,
(Texas, Louisiana, Oklahoma)
32.67
10.68
approach, and a parametric, or duration
Mean equity/assets
0.368
0.098
model, approach.
Median assets
An actuarial hazard function is sim(current dollars in thousands)
6,204
14,415
ply a series of actuarial hazard rates
Outcome
strung together in chronological order.
(number and % of sample)
Calculating the actuarial hazard rates is
Failed before 1999
50 (16.5)
185 (7.9)
straightforward and intuitive. For example,
Acquired before 1999
144 (47.5)
302 (12.9)
to calculate the 1990 hazard rate for a set
Survived to 1999
109 (36.0)
1,863 (79.2)
of banks that were chartered in 1985, one
Notes: The de novo banks began operations during 1985. The established
simply divides the number of these banks
banks were operating in the same states as the de novo banks and were
at least 25 years old in 1985. For further details of the data sources and
that failed during 1990 by the number of
data selection process, see “Estimating hazard functions” section of the
the banks that still existed at the begintext. These data are used to estimate the hazard functions shown in
figures 4 and 5.
ning of 1990. Thus, the hazard rate tells
Sources: Federal Deposit Insurance Corporation, 1985–98, “Report of
us the probability of failure in 1990 conincome and condition,” Washington, DC, and National Information Center,
1988, 1990, 1992, and 1994, “Report of income and condition.”
ditional on having survived for five years.
The following, more exact, formula can
be used to calculate the actuarial hazard
rate
for
any
time
period, T:
occurs: 1) the bank is declared insolvent by its regulator; 2) the bank receives regulatory assistance (for
example, a capital injection) without which it would
no. of bank failures during T
hazard(T) =
become insolvent; or 3) the bank is acquired soon
no. of banks surviving at start of T
after its net worth has declined to less than 1 percent
of assets. In terms of raw percentages, 16.5 percent
f(T)
£
,
of the de novo banks failed before the end of the 14T -1
1
n(t = 0) - Í f(t)+ m(t) - m(T)
year sample period. While this is over twice the 7.9
2
t=0
percent failure rate for the established banks in the
sample, it is well below the reported failure rates for
where n(t = 0) is the number of banks present at the
new (nonbank) business ventures. (See box 2 for a
beginning of the analysis; f(t) represents the number
short discussion of new bank failures versus new
of these banks that failed during time period t; m(t)
business failures.)
represents the number of these banks that were acBoth the sample de novo banks and the sample
quired in mergers during time period t, and T indicates
established banks were more likely to be acquired
the current time period. Note the subtle adjustment to
than to fail during the sample period. The new banks
the denominator in the second line of this formula:
were more likely to hold state charters; to be located
The denominator is reduced by one-half the number
in urban areas; to be located in the Southwest (primaof banks that were acquired during the current time
rily Texas, but also Louisiana and Oklahoma); and to
period. These banks clearly did not survive until the
be affiliates in multibank holding companies. Some of
end of time period T, and subtracting some portion of
the hazard functions I estimate below include tests of
these banks from the denominator acknowledges the
whether these locational, organizational, and regulatory
possibility that they might have failed during time T
De novo
banks

Established
banks

1

Federal Reserve Bank of Chicago

6

27

BOX 2

New bank failures and new business failures
During the 14 years covered by the bank failure
dataset (see table 2), 16.5 percent of the newly
chartered banks failed, compared with only 7.9
percent of the established banks of comparable
size and location. To put these new bank failure
rates into perspective, note that a 16.5 percent failure rate over 14 years is substantially lower than
the failure rates typically reported for business
start-ups in general. Raw data reported by the
U.S. Small Business Administration (1992) suggest
that at least 60 percent of new business ventures
with less than 500 employees that started in 1977–78
failed within six years. Kirchhoff (1994, pp. 153–169)
argues convincingly that these raw data overstate
the new business failure rate because, among other things, the data in many instances define firms
that changed owners or voluntarily shut down as
having failed. After adjusting for these and other

had they not been acquired. Although weighting
these banks by one-half is a crude and ad hoc adjustment, it is important to make some kind of adjustment
because, as shown in table 2, acquired banks greatly
outnumbered failed banks between 1985 and 1998.
I use the above formula to calculate 14 separate
hazard rates (one rate for each of the 14 years from
1985 through 1998) for the 303 newly chartered banks.
I repeat this exercise for the 2,350 established banks.
Plotting the resulting hazard rates in chronological
order generates two nonparametric hazard functions,
which are displayed in figure 4.
In general, the nonparametric hazard functions in
figure 4 resemble the hypothetical hazard functions
posited in figure 3. The hazard rate for newly chartered banks is initially zero, and it remains below the
established bank hazard rate for several years. As
discussed above, this is most likely because the typical new bank holds a healthy equity cushion at the
outset. After year three, the new bank hazard rate
exceeds the established bank hazard rate, and it remains substantially higher than the established bank
hazard rate until year eight. The hazard rate for newly
chartered banks peaks in years five, six, and seven at
about 1.2 percent—that is, if a newly chartered bank
reaches the beginning of any of these years without
failing or being acquired, it has about a 1.2 percent
chance of failing before the year is out. At this point,
the typical new bank’s capital ratio has declined to
established bank levels, but its profitability has not

28

factors, Kirchhoff concludes that, in a best case
scenario, 18 percent of new business ventures fail
within eight years of start-up—about the same
rate of failure as the de novo banks but in half the
number of years. Furthermore, the 16.5 percent failure rate for new banks occurred during the worst
period of bank failures since the Great Depression.
It should not be surprising that new banks
have a better rate of survival than other new businesses. Both federal and state bank regulators
deny charters to applicants with questionable financial credentials, restrict business activities, require
high amounts of capital, apply regular scrutiny via
on-site exams, and have the power to revoke bank
charters. Banking start-ups face more severe entry
barriers and ongoing scrutiny than new businesses in most other industries, and this selection bias
naturally leads to a higher survival for new banks.

yet attained the level or degree of stability found at
established banks. After year eight, the new bank
hazard rate approaches the established bank hazard
rate from above, suggesting that the maturation of
new banks is well under way at this point.
The results displayed in figure 4 are consistent
with the simple life-cycle theory of de novo bank failure. The nonparametric techniques used to generate
figure 4 paint a good general picture of the rate at
which new banks fail relative to established banks.
But these nonparametric techniques do not control
for the survivor bias in the data and, as a result, they
understate the hazard rate at any given point in time.
FIGURE 4

Nonparametric hazard rates
percent

De novo
banks

Established
banks

years after 1985

Economic Perspectives

Furthermore, these techniques are not useful for testing how much, if any, of the difference between the new
bank and established bank failure rates is caused by
the economic, regulatory, and organizational conditions under which newly chartered banks operate.
In the final step in this analysis, I use econometric
duration analysis to estimate hazard functions. These
parametric methods account for survivor bias and
control for environmental conditions that can affect
the probability of failure.
Parametric hazard functions
Duration analysis is a statistical regression approach. The dependent variable in these regressions
is t, the length of time that passes between a new
bank’s start-up date and its subsequent failure. For
established banks, t is the length of time between its
first observation in the dataset (in this case, the first
quarter of 1985) and its subsequent failure. The period
measured by t is often referred to as a bank’s duration.
Because the banks in this dataset are observed quarterly, duration will range from t = 1 for banks that fail
during the quarter in which they begin operations, to
t = 56 for banks that fail in the fourth quarter of 1998.
The simplest duration approach includes no
explanatory variables. The analyst starts by selecting
a probability distribution formula that has a shape
that is roughly similar to the actual distribution of the
duration variable t, and uses maximum likelihood techniques to estimate parameter values that shape that
probability distribution formula more exactly to the
actual duration data. Here, I use a log-logistic distribution formula, because this is capable of producing
hazard functions that have shapes similar to the hazard
functions in figures 3 and 4. (Details of these duration
model procedures can be found in the appendix to this
article or in Greene, 1997). Once the parameters of the
distribution formula have been estimated, they can be
used to construct hazard functions as follows:

hazard(T) =

f(T)
,
1 – F(T)

where f(T) is the probability that a bank fails at time T
(that is, the log-logistic probability density) and F(T)
is the probability that a bank fails before time T (that
is, the log-logistic cumulative probability distribution).
The denominator, 1 – F(T), is the log-logistic survival
function, which is the probability that a bank neither
fails nor is acquired before time T. This parametric
hazard function has the same general interpretation
as the nonparametric hazard function calculated in
the previous section—they are both estimates of the

Federal Reserve Bank of Chicago

probability that a bank will fail at time T given that it
has survived until time T. One difference is that the
hazard function generated by this parametric approach
will be a smooth and continuous function of time similar to the hypothetical hazard functions in figure 3,
as opposed to the segmented nonparametric hazard
function in figure 4.
The duration models I estimate here control for
survivor problems in the data. Recall that many of the
sample banks either survived beyond the end of the
sample period or were acquired during the sample period. These banks are known as censored observations.
We cannot assign a duration value t to these banks
because we cannot observe their ultimate fate (failure
or survival). Furthermore, history suggests that very
few of these banks will eventually fail, so including
them in hazard rate calculations creates a downward
bias by inflating the survival function 1 – F(t). Duration models can adjust for this problem by estimating
the probability that censored banks will eventually
fail, and then weighting the censored observations by
this probability before estimating the parameters of the
hazard function. (See the appendix for more details.)
The sample banks differ in terms of their geographic location, their organizational form, and their
primary regulator. These characteristics could make
a bank more or less likely to fail, or given that a bank
does fail, these characteristics could influence how
quickly it fails. For example, banks located in depressed
economic regions will be more likely to fail and, absent
regulatory intervention, will fail more quickly than
banks located in economically healthy markets. Duration models can include a vector of independent variables, typically known as covariates, measuring the
characteristics that vary across banks but remain
constant for each bank over the sample period. I use
a split population approach which estimates two regression coefficients for each of the covariates in the
duration model. The first coefficient measures the
covariate’s impact on the probability that a bank will
survive—a negative coefficient indicates that the
covariate is associated with a lower probability of
survival (higher probability of failure). The second
coefficient measures the covariate’s impact on a
bank’s duration—given that a bank will eventually
fail, a negative coefficient indicates that the covariate
is associated with a shorter duration (a faster failure).
The duration models I estimate include four covariates, each of which is expressed as a (0, 1) dummy
variable. OCC = 1 if the bank holds a federal charter
(as opposed to a state charter). The OCC has traditionally practiced a more lenient chartering policy than
most state chartering authorities, relying on market

29

forces rather than administrative rules to determine
the number of banks a market could support.7 A negative coefficient on OCC would suggest that this policy caused new national banks to fail more often and/
or more quickly, on average, than new state-chartered
banks. INDEPENDENT = 1 if the bank is either a freestanding business or a one-bank holding company
(as opposed to being an affiliate of a multibank holding company) throughout the sample period. A negative coefficient on INDEPENDENT would suggest
that banks not having access to the financial strength
and managerial expertise of a multibank holding company tend to fail more often and/or more quickly.
MSA = 1 if the bank is located in an urban area. Banks
in urban areas face greater competition than rural
banks, but also may have greater opportunities for
diversification. A negative coefficient on MSA would
suggest that, on balance, these conditions cause
banks in urban areas to fail more often and/or more
quickly than rural banks. SW = 1 if the bank is located
in the southwestern states of Louisiana, Texas, or
Oklahoma, which experienced large numbers of bank
failures during the mid- to late 1980s due to disruptions
in energy-related industries. One would expect the
coefficients on SW to be negative, reflecting lower
survival probabilities and shorter duration times for
banks in this region.
I add these four covariates to the duration model
merely to illustrate how conditions and events external to the bank can affect its probability of failure and
its time to failure. These four variables are not meant
to be an exhaustive list of such conditions. Similarly,
the duration model I estimate here is by no means definitive of the duration model techniques available to
researchers. Other duration approaches do exist, including those that allow for time-varying covariates
(for example, changes in economic, regulatory, or
competitive conditions during each bank’s duration).
However, the multiple approaches I employ (including
the Z-score and actuarial hazard function analysis
conducted above) serve the purpose of this study,
which is to test the simple life-cycle theory of de novo
bank failure summarized in figure 3.
Table 3 displays the results of the duration models
estimated separately for newly chartered banks and
established banks. The estimated probability that the
average bank will eventually fail is 19.65 percent for
de novo banks and 8.93 percent for established banks.
Note that these estimated failure probabilities are somewhat higher than the raw failure percentages shown
at the bottom of table 2. In each case, the estimated
probability is higher than the raw percentage because
of the possibility that some of the censored observations will eventually fail.

30

Although established banks are less likely to fail,
those that do fail have relatively short durations. Of
the established banks that are expected to eventually
fail, half of them will fail within an estimated 9.8 quarters (about 2.5 years) after the beginning of the sample
period. Consistent with the life-cycle theory, newly
chartered banks fail more slowly than established
banks. It takes an estimated 21.1 quarters (about 5.25
years) for half of the de novo banks that are expected
to fail to do so.
These differences in average duration can be
seen clearly in figure 5, which charts the estimated
hazard rates from the de novo and established bank
duration models. Each of these functions is plotted
based on the estimated coefficients shown in table 3
and the average values of the covariates for each
sample. In general, these two estimated hazard functions resemble the shapes displayed above in figures
3 and 4. Thus, after controlling for censored data and
a variety of environmental conditions, the failure patterns of newly chartered banks still differ substantially from the failure patterns of established banks.
The estimated probability of failure for established
banks starts out above zero; peaks at about 8 percent
for banks that survive for two years; and then slowly
declines as the bank failure wave dissipates (see figure 1). In contrast, the estimated probability of failure
for de novo banks starts out at zero and remains lower
than the established bank hazard rate for three years;
increases rapidly and peaks at nearly 14 percent for
banks the survive for seven years; and then declines
relatively quickly and begins to approach the established bank hazard rate. Note that both of these hazard
functions peak much higher on the vertical scale than
did the actuarial hazard functions plotted in figure 4.
Thus, by not controlling for censored observations
and the overall low probability of eventual failure, the
actuarial model substantially understated the hazard
rates. Also, note that the hazard rates in figure 5 are
in decline but are still positive at year 14, which reflects
the non-zero probability of failure for the censored
observations.
As expected, being located in one of the southwestern states reduces the probability of survival (or
increases the probability of failure) for both de novo
and established banks. Failing de novo banks also
failed more quickly in this region, but failing established
banks had longer than average durations. The latter
result may indicate that regulators allowed troubled
banks with longstanding business relationships (and,
hence, more franchise value) more time to recover
before stepping in to resolve them.8 Being located in
a metropolitan statistical area reduced the probability

Economic Perspectives

TABLE 3

Selected results from parametric duration models
De novo banks
Number of banks in sample

Established banks

303

2,350

Average predicted failure probability

19.65%

8.93%

Predicted time for 50 percent of banks to fail

21.1 quarters

9.8 quarters

1.2930**
(0.5144)

1.2475***
(0.2145)

0.0197
(0.2242)

–0.1296
(0.1117)

Probability of survival parameter estimates
Constant
OCC (= 1 if national bank)
SW (= 1 if in Louisiana, Oklahoma, or Texas)

–0.3928*
(0.2219)

–0.7419***
(0.0992)

MSA (= 1 if urban bank)

–0.5266*
(0.3205)

0.2530**
(0.1163)

–0.5265**
(0.2386)

#

INDEPENDENT (= 1 if independent or sole bank in
a one-bank holding company)
Survival time parameter estimates
Constant
OCC

3.3045***
(0.6124)

2.3369***
(0.4489)

0.0075
(0.1376)

0.1354
(0.2339)

SW

–0.3243**
(0.1386)

0.4753**
(0.1971)

MSA

–0.2204
(0.5318)

0.5391**
(0.2699)

INDEPENDENT

–0.1195
(0.1547)

#

*, **, and *** indicate significance at the 10 percent, 5 percent, and 1 percent levels, respectively.
Notes: Both models are estimated using the data sets described in table 2. Standard errors are in
parentheses. # indicates that it was necessary to exclude the variable INDEPENDENT to make the
established bank model converge. Further details on these models can be found in the appendix to this article.

FIGURE 5

Parametric hazard functions
percent

De novo
banks

Established
banks

years after 1985

Federal Reserve Bank of Chicago

of survival for de novo banks, but increased both the
probability of survival for established banks and the
survival time for established banks likely to eventually fail. Recall that intense competitive rivalry can cause
banks to fail in urban markets, and that the lack of diversification opportunities can cause banks to fail in
rural markets. The results suggest that these two
phenomena affect de novo banks and established
banks differently—on balance, de novo banks may
be more sensitive to competition than to diversification risk, while small established banks may be more
affected by a lack of diversification than by competitive rivalry. Being an independent bank or banking
organization also reduces the probability of survival
for de novo banks, which suggests that having access
to the resources of multibank holding companies helps

31

new banks survive. (I excluded this covariate from the
established bank model because its presence prevented the model from converging.) The identity of
a bank’s primary regulator (OCC or state) is not a significant determinant of the probability of survival or
the survival time for either set of banks.
Conclusion
Like all new business ventures, banks start with
a business plan but no guarantee of success. So,
despite the regulatory safeguards of on-site examinations, capital requirements, and other risk controls, we
should not be surprised to find that new banks are
more likely to fail than established banks. This article
offers a simple framework that explains not only why
but also when new banks are likely to fail.
My results suggest that the primary determinant
of new bank failure is how new the bank is. Ironically,
de novo banks are relatively unlikely to fail during
their first few years of operation when they are earning negative profits. They are relatively more likely to
fail during the years of positive profits that follow.
Brand new, but unprofitable, banks are typically protected from failure by large initial capital cushions.
However, equity cushions at de novo banks typically
decline to established bank levels several years before
their earnings become stable enough to justify these
relatively low levels of capital.
What are the implications of this result for capital
regulation at newly chartered banks? If ensuring a
high rate of survival for de novo banks is a regulatory
objective, then this result offers support for requiring
high levels of start-up capital for new banks, and for
holding young banks to higher capital requirements.
Higher levels of required capital will make newly chartered banks less vulnerable to failure. Under such
policies, de novo entrants might be a more credible
long-run deterrent to market power in consolidating
local markets. Indeed, in the wake of the wave of de
novo bank failures during the 1980s, federal bank
supervisory agencies did impose higher capital requirements on newly chartered banks.

32

On the other hand, promoting the safety and
soundness of the banking system does not require
that regulators prevent all bank failures, much less
all failures of new banks. At some point, attempting
to improve the survival rate of de novo banks by increasing the amount of capital necessary for investors
to secure the charter will act as an entry barrier. Similarly, increasing the required capital ratios for young
banks with charters already in hand will, at some
point, depress investors’ expected rates of return and
discourage investment in new banks. Higher capital
requirements for young banks could also slow the rate
at which they can grow their balance sheets, hampering the beneficial impact of new banks in markets
where existing banks (perhaps with market power)
are not adequately serving the banking public.
What are the implications of this study for the
bank chartering decision? During the period covered
by this study, some state chartering authorities would
approve or deny a charter application only after considering whether a local market “needed” an additional
bank, based on the number of banks already serving
the market and the expected rate of local economic
growth. These restrictive chartering policies sought
to reduce bank failure rates, and the financial disruptions that accompany them, by limiting competition
in local banking markets. In contrast, the federal chartering authority practiced a liberal entry policy that
explicitly ignored these “convenience and needs”
issues, stressing instead the potential procompetitive
benefits of de novo entry. My results indicate that
the de novo national banks chartered in 1985 were no
more likely to fail, or to fail quickly, than the de novo
state banks chartered in that same year. This suggests
that the benefits of a liberal chartering policy can be
achieved without substantial increases in de novo
bank failure rates. Additional research might confirm
whether these findings, which are based on data from
just 303 new banks chartered in a single year, also
hold for banks chartered in other years and/or under
different economic and regulatory circumstances.

Economic Perspectives

APPENDIX

Split population duration models
The parametric hazard functions described in the text
begin with the assumption that a population of N
banks will fail over time period (0,t) according to
some probability distribution:

1 6 I f (t )dt ,
t

F t =

0

where t represents time and f(t) is the probability density function associated with F(t). The hazard function
h(t) can then be written as a function of F(t) and f(t)
as follows:

h( T ) =

f (T )
f (T )
=
,
1 - F (T )
S (T )

where S(t) = 1 – F(t) is the survival function and
0<T<t. Thus, h(T) gives the probability (that is, the
hazard rate) that a bank will fail at T conditional on
surviving until T.
The general shape of the estimated hazard function will depend on the underlying probability distribution chosen to fit the data. I use the log-logistic
distribution because it is capable of producing the
hazard function shapes hypothesized in figure 3. The
log-logistic distribution imposes the following functional forms on the hazard and survival functions:
h( t ) =

lp(lp) p -1
1 + (lt ) p

S (t ) =

1
,
1 + (lt ) p

where the parameters p and λ give the hazard function
its exact shape. The parameter p captures duration
dependence or whether the hazard rate increases or
decreases across time. The parameter λ captures the
portion of the hazard rate that is time-invariant. This
parameter, which can take on different values for different banks, is expressed as follows:

li = e-bŠXi ,
where the bank index i ranges from 1 to N, and Xi is a
vector of bank-specific covariates that do not vary
over time. Table 3 reports the estimated values of β
under the heading “survival time parameter estimates.”
I use these β estimates to evaluate the above expression at the means of the covariates, which results in

Federal Reserve Bank of Chicago

λ = 0.0474 for the average de novo bank and 0.1019
for the average established bank. The parameter p is
a constant that does not vary across banks; it equals
5.0175 for the de novo bank model and 1.6333 for the
established bank model.
All of the parameters of the duration models in
this study were estimated using maximum likelihood
techniques. The standard estimation procedure for
duration models starts with the following likelihood
function:
N

L = ½ f (t i | p, b)
i =1

Qi

S (ti | p, b)

1- Qi

,

where Qi = 1 if bank i failed during the sample period
(an uncensored observation) and Qi = 0 if bank i
survived or was acquired during the sample period
(a censored observation). Substituting h(t)/S(t) for
f(t) and performing a log transformation produces the
log-likelihood function to be maximized:
UC

N

i =1

i =1

lnL = Í lnh(t i | p, b) + Í lnS (ti | p, b) ,
where i≤UC are the uncensored observations. Once I
have estimated the parameters p and β, I can calculate
the median time to failure by setting S(t) = 0.50 and
solving for t.
This standard approach is based on the assumption that all of the censored banks will eventually fail
(or would have eventually failed had they not been
acquired). This assumption is inappropriate for the
data used here, however, because over 80 percent of
the de novo banks were censored observations, and
over 90 percent of the established banks were censored observations. Given the nature of the data, I
use a more general framework that avoids making this
assumption. In the split population duration model,
an additional estimable parameter δ, the probability
that a bank eventually fails, enters the likelihood
function as follows:
N

L = ½ df (t i | p, b)
i =1

Qi

(1 - d) + dS (ti | p, b)

1- Qi

.

Both Cole and Gunther (1995) and Hunter,
Verbrugge, and Whidbee (1996) estimate split population models of financial institution failure. Note that
this formulation collapses to the standard framework
when δ = 1. But when δ<1, the functions S(t) and f(t)
become conditional on the bank eventually failing.
Thus, the estimated hazard function h(t) = f(t)/S(t)

33

will not be unduly influenced by censored observations of banks that have little chance of ever failing.
The parameter δ can vary across banks as a function
of a bank’s covariate values:

di =

1
.
1 + e a ŠX i

Table 3 reports the estimated values of α under
the heading is “probability of survival parameter estimates.” I use these α estimates to evaluate the above

expression at the means of the covariates, which results
in δ = 0.1965 for the average de novo bank and 0.0893
for the average established bank.
The hazard functions plotted in figure 5 show
the probability that the average bank will fail at time t,
given that the bank has not yet failed but will eventually fail. Thus, the shapes of the plotted hazard
functions are based on the estimated values of λ and
p, but not on the estimated values of δ.

NOTES
Note that such ambiguity is largely absent from studies that
examine the determinants of local market entry by already
established banks. For a recent example, see Amel and Liang
(1997). In general, these studies tend to find that established
banks are more likely to enter highly profitable local banking
markets, but less likely to enter highly concentrated local
banking markets. Of course, the causes and consequences of a
new bank start-up may be quite different from the causes and
consequences of market entry by an already established bank.
1

Seelig and Critchfield (1999) find that, on average, the state
chartering authorities remain relatively more likely than the
OCC to consider the ability of local banking markets to support
an additional bank when evaluating a charter application. The
authors show that income per capita per branch in the local
banking market was a substantially stronger predictor of de
novo state bank entry than of de novo national bank entry
between 1995 and 1997.
2

There are a number of possible reasons for this. In general,
Z-score analysis performs best when used to represent the
likelihood that an individual firm will become insolvent and,
as such, Z-scores are typically constructed based on the known
distribution (the mean and standard deviation) of ROA for an
individual firm. In contrast, these average Z-scores are constructed for groups of banks, and rely on the cross-sectional
distribution (the median and standard deviation) of ROA for
each group of banks. As mentioned in the text, the distribution
of ROA is not normally distributed, but rather is relatively
skewed. Hence, it would be inappropriate to use the absolute
levels of these average Z-scores to draw statistical inferences
about the probability of bank failure.
6

See Hunter and Srinivasan (1990) and DeYoung and Hasan
(1998) for discussions that compare historical federal and
state chartering policies.
7

There are many potential reasons for the high bank failure
rates in Texas, and for the relatively shorter durations for de
novo banks in Texas, during the 1980s and 1990s. These reasons
include unexpected economic shocks, unit banking restrictions
that limited geographic diversification, a relatively undiversified regional economy, and regulatory failure.
8

See DeYoung and Hasan (1998) for a more complete review of
this literature.
3

Examples of studies that have used hazard rates to analyze
financial institution failure include Whalen (1991); Wheelock
and Wilson (1995); Cole and Gunther (1995); Helwege (1996);
and Hunter, Verbrugge, and Whidbee (1996).
4

The Z-score is a measure of the probability that a firm’s losses
(negative profits) will exceed its equity capital. See Brewer
(1989) for a discussion of the Z-score and its use in banking
research.
5

REFERENCES

Amel, Dean F., and J. Nellie Liang, 1997, “Determinants of entry and profits in local banking markets,”
Review of Industrial Organization, Vol. 12, No. 1,
pp. 59–78.

Brewer III, Elijah, 1989, “Relationships between
bank holding company risk and nonbank activity,”
Journal of Economics and Business, Vol. 41, November, pp. 337–353.

Berger, Allen N., Seth D. Bonime, Lawrence G.
Goldberg, and Lawrence J. White, 1999, “The dynamics of market entry: The effects of mergers and
acquisitions on de novo entry and customer service
in banking,” Proceedings from a Conference on
Bank Structure and Regulation, Federal Reserve
Bank of Chicago, forthcoming.

Brislin, Patricia, and Anthony Santomero, 1991,
“De novo banking in the Third District,” Business
Review, Federal Reserve Bank of Philadelphia, January,
pp. 3–12.

34

Economic Perspectives

Cole, Rebel A., and Jeffery W. Gunther, 1995, “Separating the likelihood and timing of bank failure,” Journal
of Banking and Finance, Vol. 19, No. 6, September,
pp. 1073–1089.
DeYoung, Robert, Lawrence G. Goldberg, and
Lawrence J. White, 1999, “Youth, adolescence, and
maturity of banks: Credit availability to small business in an era of banking consolidation,” Journal
of Banking and Finance, Vol. 23, No. 2-4, February,
pp. 463–492.
DeYoung, Robert, and Iftekhar Hasan, 1998, “The
performance of de novo commercial banks: A profit
efficiency approach,” Journal of Banking and
Finance, Vol. 22, May, pp. 565–587.
Greene, William H., 1997, Econometric Analysis,
Upper Saddle River, NJ: Prentice Hall.
Helwege, Jean, 1996, “Determinants of savings and
loan failure,” Journal of Financial Services Research,
Vol. 10, No. 4, pp. 373–392.
Hunter, William C. and Aruna Srinivasan, 1990,
“Determinants of de novo bank performance,”
Economic Review, Federal Reserve Bank of Atlanta,
March, pp. 14–25.
Hunter, William C., James A. Verbrugge, and David
A. Whidbee, 1996, “Risk taking and failure in de novo
savings and loans in the 1980s,” Journal of Financial
Services Research, Vol. 10, No. 3, pp. 235–272.

Federal Reserve Bank of Chicago

Huyser, Daniel, 1986, “De novo bank performance in
the Seventh District states,” Banking Studies, Federal
Reserve Bank of Kansas City, pp. 13–22.
Kirchhoff, Bruce A., 1994, Entrepreneurship and
Dynamic Capitalism, Westport, CT and London:
Greenwood, Praeger.
Moore, Robert R., and Edward C. Skelton, 1998,
“New banks: Why enter when others exit?,” Financial Industry Issues, Federal Reserve Bank of Dallas,
First Quarter, pp. 1–7.
Seelig, Stephen, and Timothy Critchfield, 1999,
“Determinants of de novo entry in banking,” Federal
Deposit Insurance Corporation, working paper, No.
99-1, January.
U.S. Small Business Administration, 1992, The State
of Small Business, Washington, DC: U.S. Government
Printing Office.
Whalen, Gary, 1991, “A proportional hazards model
of bank failure: An examination of its usefulness as
an early warning tool,” Economic Review, Federal
Reserve Bank of Cleveland, Vol. 27, No. 1, First Quarter,
pp. 20–31.
Wheelock, David C., and Paul W. Wilson, 1995,
“Explaining bank failures: Deposit insurance, regulation, and efficiency,” Review of Economics and Statistics, November, pp. 689–700.

35

New facts in finance

John H. Cochrane

Introduction and summary
The last 15 years have seen a revolution in the way
financial economists understand the investment world.
We once thought that stock and bond returns were
essentially unpredictable. Now we recognize that
stock and bond returns have a substantial predictable
component at long horizons. We once thought that
the capital asset pricing model (CAPM) provided a
good description of why average returns on some
stocks, portfolios, funds, or strategies were higher than
others. Now we recognize that the average returns of
many investment opportunities cannot be explained
by the CAPM, and “multifactor models” are used in
its place. We once thought that long-term interest
rates reflected expectations of future short-term rates
and that interest rate differentials across countries
reflected expectations of exchange rate depreciation.
Now, we see time-varying risk premiums in bond and
foreign exchange markets as well as in stock markets.
We once thought that mutual fund average returns
were well explained by the CAPM. Now, we see that
funds can earn average returns not explained by the
CAPM, that is, unrelated to market risks, by following
a variety of investment “styles.”
In this article, I survey these new facts, and I show
how they are variations on a common theme. Each
case uses price variables to infer market expectations
of future returns; each case notices that an offsetting
adjustment (to dividends, interest rates, or exchange
rates) seems to be absent or sluggish. Each case suggests that financial markets offer rewards in the form
of average returns for holding risks related to recessions and financial distress, in addition to the risks
represented by overall market movements. In a companion article in this issue, “Portfolio advice for a multifactor world,” I survey and interpret recent advances
in portfolio theory that address the question, What
should an investor do about all these new facts?
First, a slightly more detailed overview of the
facts then and now. Until the mid-1980s, financial

36

economists’ view of the investment world was based
on three bedrocks:
1. The CAPM is a good measure of risk and thus
a good explanation of the fact that some assets (stocks,
portfolios, strategies, or mutual funds) earn higher
average returns than others. The CAPM states that
assets can only earn a high average return if they
have a high “beta,” which measures the tendency
of the individual asset to move up or down with the
market as a whole. Beta drives average returns because
beta measures how much adding a bit of the asset to
a diversified portfolio increases the volatility of the
portfolio. Investors care about portfolio returns, not
about the behavior of specific assets.
2. Returns are unpredictable, like a coin flip. This
is the random walk theory of stock prices. Though
there are bull and bear markets; long sequences of
good and bad past returns; the expected future return
is always about the same. Technical analysis that
tries to divine future returns from patterns of past
returns and prices is nearly useless. Any apparent
predictability is either a statistical artifact which will
quickly vanish out of sample or cannot be exploited
after transaction costs.
Bond returns are not predictable. This is the
expectations model of the term structure. If long-term
bond yields are higher than short-term yields—if the
yield curve is upward sloping—this does not mean
that you expect a higher return by holding long-term
bonds rather than short-term bonds. Rather, it means
John H. Cochrane is the Sigmund E. Edelstone
Professor of Finance in the Graduate School of
Business at the University of Chicago, a consultant to
the Federal Reserve Bank of Chicago, and a research
associate at the National Bureau of Economic Research
(NBER). The author thanks Andrea Eisfeldt for research
assistance and David Marshall, John Campbell, and
Robert Shiller for comments. The author’s research is
supported by the Graduate School of Business and by
a grant from the National Science Foundation,
administered by the NBER.

Economic Perspectives

that short-term interest rates are expected to rise in
the future. Over one year, the rise in interest rates will
limit the capital gain on long-term bonds, so they earn
the same as the short-term bonds over the year. Over
many years, the rise in short rates improves the rate
of return from rolling over short-term bonds to equal
that of holding the long-term bond. Thus, you expect
to earn about the same amount on short-term or longterm bonds at any horizon.
Foreign exchange bets are not predictable. If a
country has higher interest rates than are available in
the U.S. for bonds of a similar risk class, its exchange
rate is expected to depreciate. Then, after you convert your investment back to dollars, you expect to
make the same amount of money holding foreign or
domestic bonds.
In addition, stock market volatility does not
change much through time. Not only are returns close
to unpredictable, they are nearly identically distributed
as well. Each day, the stock market return is like the
result of flipping the same coin, over and over again.
3. Professional managers do not reliably outperform simple indexes and passive portfolios once one
corrects for risk (beta). While some do better than the
market in any given year, some do worse, and the
outcomes look very much like luck. Funds that do well
in one year are not more likely to do better than average the next year. The average actively managed fund
performs about 1 percent worse than the market index.
The more actively a fund trades, the lower the returns
to investors.
Together, these views reflect a guiding principle
that asset markets are, to a good approximation, informationally efficient (Fama, 1970, 1991). Market prices
already contain most information about fundamental
value and, because the business of discovering information about the value of traded assets is extremely
competitive, there are no easy quick profits to be made,
just as there are not in any other well-established
and competitive industry. The only way to earn large
returns is by taking on additional risk.
These views are not ideological or doctrinaire
beliefs. Rather, they summarize the findings of a quarter century of careful empirical work. However, every
one of them has now been extensively revised by a
new generation of empirical research. The new findings need not overturn the cherished view that markets
are reasonably competitive and, therefore, reasonably
efficient. However, they do substantially enlarge our
view of what activities provide rewards for holding
risks, and they challenge our understanding of those
risk premiums.

Federal Reserve Bank of Chicago

Now, we know that:
1. There are assets whose average returns can
not be explained by their beta. Multifactor extensions
of the CAPM dominate the description, performance
attribution, and explanation of average returns. Multifactor models associate high average returns with a
tendency to move with other risk factors in addition
to movements in the market as a whole. (See box 1.)
2. Returns are predictable. In particular: Variables
including the dividend/price (d/p) ratio and term premium can predict substantial amounts of stock return
variation. This phenomenon occurs over business
cycle and longer horizons. Daily, weekly, and monthly
stock returns are still close to unpredictable, and technical systems for predicting such movements are still
close to useless.
Bond returns are predictable. Though the expectations model works well in the long run, a steeply
upward sloping yield curve means that expected
returns on long-term bonds are higher than on shortterm bonds for the next year. These predictions are not
guarantees—there is still substantial risk—but the
tendency is discernible.
Foreign exchange returns are predictable. If you
put your money in a country whose interest rates are
higher than usual relative to the U.S., you expect to
earn more money even after converting back to dollars.
Again, this prediction is not a guarantee—exchange
rates do vary, and a lot, so the strategy is risky.
Volatility does change through time. Times of
past volatility indicate future volatility. Volatility also
is higher after large price drops. Bond market volatility is higher when interest rates are higher, and possibly when interest rate spreads are higher as well.
3. Some mutual funds seem to outperform simple
indexes, even after controlling for risk through market
betas. Fund returns are also slightly predictable: Past
winning funds seem to do better than average in the
future, and past losing funds seem to do worse than
average in the future. For a while, this seemed to indicate that there is some persistent skill in active management. However, multifactor models explain most
fund persistence: Funds earn persistent returns by
following fairly mechanical styles, not by persistent
skill at stock selection.
Again, these statements are not dogma, but a
cautious summary of a large body of careful empirical
work. The strength and usefulness of many results
are hotly debated, as are the underlying reasons for
many of these new facts. But the old world is gone.

37

BOX 1

The CAPM and multifactor models
The CAPM uses a time-series regression to measure beta, β, which quantifies an asset’s or portfolio’s tendency to move with the market as a whole,

Rti - Rt f = ai + bim ( Rtm - Rtf ) + eit ;
t = 1, 2 ... T for each asset i.
Then, the CAPM predicts that the expected
excess return should be proportional to beta,

E ( Rti - Rtf ) = biml m for each i.
λm gives the “price of beta risk” or “market risk premium”—the amount by which expected returns
must rise to compensate investors for higher beta.
Since the model applies to the market return as
well, we can measure λm via

l m = E ( Rtm - Rt f ).

The CAPM and multifactor models
The CAPM
The CAPM proved stunningly successful in a
quarter century of empirical work. Every strategy that
seemed to give high average returns turned out to
have a high beta, or a large tendency to move with
the market. Strategies that one might have thought
gave high average returns (such as holding very volatile stocks) turned out not to have high average
returns when they did not have high betas.
Figure 1 presents a typical evaluation of the
CAPM. I examine 10 portfolios of NYSE stocks sorted
by size (total market capitalization), along with a portfolio of corporate bonds and long-term government
bonds. As the vertical axis shows, there is a sizable
spread in average returns between large stocks (lower
average return) and small stocks (higher average
return) and a large spread between stocks and bonds.
The figure plots these average returns against market
betas. You can see how the CAPM prediction fits: Portfolios with higher average returns have higher betas.
In fact, figure 1 captures one of the first significant failures of the CAPM. The smallest firms (the far
right portfolio) seem to earn an average return a few
percent too high given their betas. This is the celebrated “small-firm effect,” (Banz, 1981) and this deviation is statistically significant. Would that all failed
economic theories worked so well! However, the plot
shows that this effect is within the range that statisticians can argue about. Estimating the slope of the

38

Multifactor models extend this theory in a
straightforward way. They use a time-series multiple regression to quantify an asset’s tendency to
move with multiple risk factors FA, FB, etc.
3)

Rti - Rtf = ai + bim ( Rtm - Rtf ) + biA Ft A + biB Ft B

+ ... + eit ; t = 1, 2 ... T for each asset i.
Then, the multifactor model predicts that the
expected excess return is proportional to the betas

4) E( Rti - Rtf ) = bimlm + biAl A + biBlB + ...
for each i.
The residual or unexplained average return in
either case is called an alpha,

ai ¢ E ( Rti - Rtf ) - (biml m + biAl A + biB l B + ...).

line by fitting a cross-sectional regression (average
return against beta), shown in the colored line, rather
than forcing the line to go through the market and
Treasury bill return, shown in the black line, halves
FIGURE 1

CAPM—Mean excess returns vs. beta, version 1
mean excess returns, percent
18

Means and betas
Fitted market premium
Direct market premium

14

10

6

2

–2
0.0

0.2

0.4

0.6
0.8
betas

1.0

1.2

1.4

Notes: Average returns versus betas on the NYSE value-weighted
portfolio for ten size-sorted stock portfolios, government bonds,
and corporate bonds. Sample period 1947–96. The black line
draws the CAPM prediction by fitting the market proxy and
Treasury bill rates exactly (a time-series test) and the colored line
draws the CAPM prediction by fitting an OLS cross-sectional
regression to the displayed data points (a second-pass or crosssectional test). The small-firm portfolios are at the top right.
Moving down and to the left, one sees increasingly large-firm
portfolios and the market index. The points far down and to the
left are the government bond and Treasury bill returns.

Economic Perspectives

FIGURE 2

CAPM—Mean excess returns vs. beta, version 2
mean excess returns, percent
16

Means and betas
Direct market premium
Fitted market premium

12

8

4

0
0.0

0.2

0.4

0.6
0.8
betas

1.0

1.2

1.4

Notes: CAPM using the equally weighted NYSE as the “market
portfolio.” Otherwise, this figure is identical to figure 1.

the small-firm effect. Figure 2 uses the equally weighted
portfolio as market proxy, and this change in specification eliminates the small-firm effect, making the line of
average returns versus betas if anything too shallow
rather than too steep.
Why we expect multiple factors
In retrospect, it is surprising that the CAPM
worked so well for so long. The assumptions on which
it is built are very stylized and simplified. Asset pricing
theory recognized at least since Merton (1973, 1971)
the theoretical possibility, indeed probability, that we
should need factors, state variables or sources of
priced risk, beyond movements in the market portfolio to explain why some average returns are higher
than others. (See box 1 for details of the CAPM and
multifactor models.)
Most importantly, the average investor has a
job. The CAPM (together with the use of the NYSE
portfolio as the market proxy) simplifies matters by
assuming that the average investor only cares about
the performance of his investment portfolio. While
there are investors like that, for most of us eventual
wealth comes both from investment and from earning
a living. Importantly, events like recessions hurt the
majority of investors. Those who don’t actually lose
jobs get lower salaries or bonuses. A very limited number of people actually do better in a recession.
With this fact in mind, compare two stocks. They
both have the same sensitivity to market movements.
However, one of them does well in recessions, while
the other does poorly. Clearly, most investors prefer
the stock that does well in recessions, since its performance will cushion the blows to their other income.

Federal Reserve Bank of Chicago

If lots of people feel that way, they bid up the price of
that stock, or, equivalently, they are willing to hold it
at a lower average return. Conversely, the procyclical
stock’s price will fall or it must offer a higher average
return in order to get investors to hold it.
In sum, we should expect that procyclical stocks
that do well in booms and worse in recessions will have
to offer higher average returns than countercyclical
stocks that do well in recessions, even if the stocks
have the same market beta. We expect that another
dimension of risk—covariation with recessions—
will matter in determining average returns.1
What kinds of additional factors should we look
for? Generally, asset pricing theory specifies that
assets will have to pay high average returns if they
do poorly in “bad times”—times in which investors
would particularly like their investments not to perform
badly and are willing to sacrifice some expected return
in order to ensure that this is so. Consumption (or,
more generally, marginal utility) should provide the
purest measure of bad times. Investors consume less
when their income prospects are low or if they think
future returns will be bad. Low consumption thus
reveals that this is indeed a time at which investors
would especially like portfolios not to do badly, and
would be willing to pay to ensure that wish. Alas,
efforts to relate asset returns to consumption data
are not (yet) a great success. Therefore, empirically
useful asset pricing models examine more direct measures of good times or bad times. Broad categories
of such indicators are
1. The market return. The CAPM is usually
included and extended. People are unhappy if the
market crashes.
2. Events, such as recessions, that drive investors’ noninvestment sources of income.
3. Variables, such as the p/d ratio or slope of
the yield curve, that forecast stock or bond returns
(called “state variables for changing investment
opportunity sets”).
4. Returns on other well-diversified portfolios.
One formally justifies the first three factors by
stating assumptions under which each variable is related to average consumption. For example, 1) if the
market as a whole declines, consumers lose wealth
and will cut back on consumption; 2) if a recession
leads people to lose their jobs, then they will cut back
on consumption; and, 3) if you are saving for retirement, then news that interest rates and average stock
returns have declined is bad news, which will cause
you to lower consumption. This last point establishes
a connection between predictability of returns and the
presence of additional risk factors for understanding

39

the cross-section of average returns. As pointed out
by Merton (1971), one would give up some average
return to have a portfolio that did well when there
was bad news about future market returns.
The fourth kind of factor—additional portfolio
returns—is most easily defended as a proxy for any of
the other three. The fitted value of a regression of any
pricing factor on the set of all asset returns is a portfolio that carries exactly the same pricing information as
the original factor—a factor-mimicking portfolio.
It is vital that the extra risk factors affect the
average investor. If an event makes investor A worse
off and investor B better off, then investor A buys
assets that do well when the event happens and investor B sells them. They transfer the risk of the event,
but the price or expected return of the asset is unaffected. For a factor to affect prices or expected returns,
it must affect the average investor, so investors collectively bid up or down the price and expected return of
assets that covary with the event rather than just transferring the risk without affecting equilibrium prices.
Inspired by this broad direction, empirical researchers have found quite a number of specific factors that
seem to explain the variation in average returns across
assets. In general, empirical success varies inversely
with theoretical purity.
Small and value/growth stocks
The size and book to market factors advocated
by Fama and French (1996) are one of the most popular additional risk factors.
Small-cap stocks have small market values (price
times shares outstanding). Value (or high book/market)
stocks have market values that are small relative to
the value of assets on the company’s books. Both
categories of stocks have quite high average returns.
Large and growth stocks are the opposite of small
and value and seem to have unusually low average
returns. (See Fama and French, 1993, for a review.)
The idea that low prices lead to high average returns
is natural.
High average returns are consistent with the
CAPM, if these categories of stocks have high sensitivity to the market, high betas. However, small and
especially value stocks seem to have abnormally
high returns even after accounting for market beta. Conversely, growth stocks seem to do systematically worse
than their CAPM betas suggest. Figure 3 shows this
value–size puzzle. It is just like figure 1, except that the
stocks are sorted into portfolios based on size and
book/market ratio2 rather than size alone. The highest
portfolios have three times the average excess return
of the lowest portfolios, and this variation has nothing
at all to do with market betas.

40

FIGURE 3

Mean excess returns vs. market beta,
Fama–French portfolios
mean excess returns
1.25
1.00
0.75
0.50
0.25
0.00
0.0

0.2

0.4

0.6
0.8
1.0
market beta

1.2

1.4

Notes: Average monthly returns versus market beta for 25 stock
portfolios sorted on the basis of size and book/market ratio.

In figure 4, I connect portfolios of different sizes
within the same book/market category (panel A). Variation in size produces a variation in average returns
that is positively related to variation in market betas,
as shown in figure 1. In panel B, I connect portfolios
that have different book/market ratios within size categories. Variation in book/market ratio produces a
variation in average return that is negatively related to
market beta. Because of this value effect, the CAPM is
a disaster when confronted with these portfolios.
To explain these facts, Fama and French (1993,
1996) advocate a multifactor model with the market
return, the return of small less big stocks (SMB), and
the return of high book/market less low book/market
stocks (HML) as three factors. They show that variation in average returns of the 25 size and book/market
portfolios can be explained by varying loadings (betas)
on the latter two factors.
Figure 5 illustrates Fama and French’s results.
As in figure 4, the vertical axis is the average returns
of the 25 size and book/market portfolios. Now, the
horizontal axis is the predicted values from the Fama–
French three-factor model. The points should all lie
on a 45 degree line if the model is correct. The points
lie much closer to this prediction in figure 5 than in
figures 3 and 4. The worst fit is for the growth stocks
(lowest line, panel A), for which there is little variation
in average return despite large variation in size beta
as one moves from small to large firms.
What are the size and value factors?
One would like to understand the real, macroeconomic, aggregate, nondiversifiable risk that is proxied
by the returns of the HML and SMB portfolios. Why

Economic Perspectives

FIGURE 4

Mean excess returns vs. market beta, varying size and book/market ratio
A. Changing size within book/market category

B. Changing book/market within size category

mean excess return
1.25

mean excess return
1.25

1.00

1.00

0.75

0.75

0.50

0.50

0.25

0.25

0.00
0.0

0.2

0.4

0.6
0.8
1.0
market beta

1.2

0.00
0.0

1.4

0.2

0.4

0.6
0.8
1.0
market beta

1.2

1.4

Notes: Average returns versus market beta for 25 stock portfolios sorted on the basis of size and book/market ratio.
The points are the same as figure 3. In panel A, lines connect portfolios as size varies within book/market categories;
in panel B, lines connect portfolios as book/market ratio varies within size categories.

are investors so concerned about holding stocks that
do badly at the times that the HML (value less growth)
and SMB (small-cap less large-cap) portfolios do badly,
even though the market does not fall? The answer to
this question is not yet totally clear.
Fama and French (1995) note that the typical value
stock has a price that has been driven down due to
financial distress. The stocks of firms on the verge
of bankruptcy have recovered more often than not,
which generates the high average returns of this

strategy.3 This observation suggests a natural interpretation of the value premium: In the event of a credit
crunch, liquidity crunch, or flight to quality, stocks in
financial distress will do very badly, and this is precisely when investors least want to hear that their portfolio is losing money. (One cannot count the “distress”
of the individual firm as a risk factor. Such distress
is idiosyncratic and can be diversified away. Only
aggregate events that average investors care about
can result in a risk premium.)

FIGURE 5

Mean excess return vs. three-factor model predictions
A. Changing size within book/market category

B. Changing book/market within size category

actual mean excess return, E(R i – R f )
1.2

actual mean excess return, E(R i – R f )
1.2

0.9

0.9

0.6

0.6

0.3

0.3

0.0
0.0

0.2
0.4
0.6
0.8
1.0
1.2
predicted, β i , m E(R m – R f ) + βi , h E(HML) +β i ,s E(SMB)

0.0
0.0
0.2
0.4
0.6
0.8
1.0
1.2
predicted, β i , m E(R m – R f) + β i ,h E(HML) +β i ,sE(SMB)

Notes: Average returns versus market beta for 25 stock portfolios sorted on the basis of size and book/market ratio versus
predictions of Fama–French three-factor model. The predictions are derived by regressing each of the 25 portfolio returns, R it ,
on the market portfolio, R mt, and the two Fama–French factor portfolios, SMBt (small minus big) and HMLt (high minus low
book/market). (See equation 4 in box 1.)

Federal Reserve Bank of Chicago

41

Heaton and Lucas’s (1997) results add to this
story for the value effect. They note that the typical
stockholder is the proprietor of a small, privately held
business. Such an investor’s income is, of course,
particularly sensitive to the kinds of financial events
that cause distress among small firms and distressed
value firms. Therefore, this investor would demand a
substantial premium to hold value stocks and would
hold growth stocks despite a low premium.
Liew and Vassalou (1999), among others, link value
and small-firm returns to macroeconomic events. They
find that in many countries, counterparts to HML and
SMB contain supplementary information to that contained in the market return for forecasting gross domestic product (GDP) growth. For example, they report
a regression
GDPt→ t+4 = a + 0.065 MKTt–4→ t
+ 0.058 HMLt–4→ t + εt+4,
where GDPt→ t+4 denotes the following year’s GDP
growth and MKTt–4→t and HMLt–4→t denote the previous
year’s return on the market index and HML portfolio.
Thus, a 10 percent HML return raises the GDP forecast
by 0.5 percentage points. (Both coefficients are significant with t-statistics of 3.09 and 2.83, respectively.)
The effects are still under investigation. Figure 6
plots the cumulative return on the HML and SMB
portfolios; a link between these returns and obvious
macroeconomic events does not jump out. Both portfolios have essentially no correlation with the market
return, though HML does seem to move inversely with
large market declines. HML goes down more than the
market in some business cycles, but less in others.
On the other hand, one can ignore Fama and
French’s motivation and regard the model as an arbitrage pricing theory (APT) following Ross (1976). If
the returns of the 25 size and book/market portfolios
could be perfectly replicated by the returns of the
three-factor portfolios—if the R2 values in the timeseries regressions of the 25 portfolio on the three
factors were 100 percent—then the multifactor model
would have to hold exactly, in order to preclude arbitrage opportunities. To see this, suppose that one of
the 25 portfolios—call it portfolio A—gives an average
return 5 percent above the average return predicted
by the Fama–French model, and its R2 is 100 percent.
Then, one could short a combination of the threefactor portfolios, buy portfolio A, and earn a completely
riskless profit. This logic is often used to argue that
a high R2 should imply an approximate multifactor
model. If the R2 were only 95 percent, then an average
return 5 percent above the factor model prediction

42

FIGURE 6

Cumulative returns on market portfolios
cumulative return
8

6

HML
Market

4

SMB
2

0
1945

‘55

‘65

‘75

‘85

‘95

Notes: Cumulative returns on the market RMRF, SMB, and
HML portfolios. The SMB return is formed by R TB
+ aSMBt ;
t
a = σ(RMRF)/σ(SMB). In this way it is a return that can be
cumulated rather than a zero-cost portfolio, and its standard
deviation is equal to that of the market return. HML is adjusted
similarly. The vertical axis is the log base 2 of the cumulative
return or value of $1 invested at the beginning of the sample
period. Thus, each time a line increases by 1 unit, the value
doubles.

would imply that the strategy long portfolio A and
short a combination of the three-factor portfolios
would earn a very high average return with very little,
though not zero, risk—a very high Sharpe ratio.
In fact, the R2 values of Fama and French’s (1993)
time-series regressions are all in the 90 percent to 95
percent range, so extremely high risk prices for the
residuals would have to be invoked for the model not
to fit well. Conversely, given the average returns from
HML and SMB and the failure of the CAPM to explain
those returns, there would be near-arbitrage opportunities if value and small stocks did not move together
in the way described by the Fama–French model.
One way to assess whether the three factors
proxy for real macroeconomic risks is by checking
whether the multifactor model prices additional portfolios, especially portfolios whose ex-post returns are
not well explained by the factors (portfolios that do
not have high R2 values in time-series regressions).
Fama and French (1996) find that the SMB and HML
portfolios comfortably explain strategies based on
alternative price multiples (price/earnings, book/market),
five-year sales growth (this is the only strategy that
does not form portfolios based on price variables),
and the tendency of five-year returns to reverse. All
of these strategies are not explained by CAPM betas.
However, they all also produce portfolios with high
R2 values in a time-series regression on the HML and
SMB portfolios. This is good and bad news. It might

Economic Perspectives

mean that the model is a good APT, and that the size
and book/market characteristics describe the major
sources of priced variation in all stocks. On the other
hand, it might mean that these extra ways of constructing portfolios just haven’t identified other sources of
priced variation in stock returns. (Fama and French,
1996, also find that HML and SMB do not explain
momentum, despite high R2 values. I discuss this
anomaly below.) The portfolios of stocks sorted by
industry in Fama and French (1997) have lower R2
values, and the model works less well.
A final concern is that the size and book/market
premiums seem to have diminished substantially in
recent years. The sharp decline in the SMB portfolio
return around 1980 when the small-firm effect was first
popularized is obvious in figure 6. In Fama and French’s
(1993) initial samples, 1960–90, the HML cumulative
return starts about one-half (0.62) below the market
and ends up about one-half (0.77) above the market.
On the log scale of the figure, this corresponds to Fama
and French’s report that the HML average return is
about double (precisely, 20.62+0.77 = 2.6 times) that of
the market. However, over the entire sample of the
plot, the HML portfolio starts and ends at the same
place and so earns almost exactly the same as the
market. From 1990 to now, the HML portfolio loses
about one-half relative to the market, meaning an investor in the market has increased his money one and a
half times as much as an HML investor. (The actual
number is 0.77 so the market return is 20.77 = 1.71 times
better than the HML return.)
Among other worries, if the average returns decline
right after publication it suggests that the anomalies
may simply have been overlooked by a large fraction
of investors. As they move in, prices go up further,
helping the apparent anomaly for a while. But once a
large number of investors have moved in to include
small and value stocks in their portfolios, the anomalous high average returns disappear.
However, average returns are hard to measure.
There have been previous ten- to 20-year periods in
which small stocks did very badly, for example the
1950s, and similar decade-long variations in the HML
premium. Also, since SMB and HML have a beta of
essentially zero on the market, any upward trend is a
violation of the CAPM and says that investors can
improve their overall mean–variance tradeoff by taking
on some of the HML or SMB portfolio.
Macroeconomic factors
I focus on the size and value factors because they
provide the most empirically successful multifactor
model and have attracted much industry as well as

Federal Reserve Bank of Chicago

academic attention. Several authors have used macroeconomic variables as factors. This procedure examines
directly whether stock performance during bad macroeconomic times determines average returns. Jagannathan and Wang (1996) and Reyfman (1997) use
labor income; Chen, Roll, and Ross (1986) look at
industrial production and inflation among other variables; and Cochrane (1996) looks at investment growth.
All these authors find that average returns line up
with betas calculated using the macroeconomic indicators. The factors are theoretically easier to motivate,
but none explains the value and size portfolios as
well as the (theoretically less solid, so far) size and
value factors.
Merton’s (1973, 1971) theory says that variables
which predict market returns should show up as factors that explain cross-sectional variation in average
returns. Campbell (1996) is the lone test I know of to
directly address this question. Cochrane (1996) and
Jagannathan and Wang (1996) perform related tests in
that they include “scaled return” factors, for example,
market return at t multiplied by d/p ratio at t – 1; they
find that these factors are also important in understanding cross-sectional variation in average returns.
The next step is to link these more fundamentally
determined factors with the empirically more successful value and small-firm factor portfolios. Because of
measurement difficulties and selection biases, fundamentally determined macroeconomic factors will never
approach the empirical performance of portfolio-based
factors. However, they may help to explain which portfolio-based factors really work and why.
Predictable returns
The view that risky asset returns are largely unpredictable, or that prices follow “random walks,” remains
immensely successful ( Malkiel, 1990, is a classic and
readable introduction). It is also widely ignored.
Unpredictable returns mean that if stocks went
up yesterday, there is no exploitable tendency for them
to decline today because of “profit taking” or to continue to rise today because of “momentum.” “Technical”
signals, including analysis of past price movements
trading volume, open interest, and so on are close to
useless for forecasting short-term gains and losses.
As I write, value funds are reportedly suffering large
outflows because their stocks have done poorly in
the last few months, leading fund investors to move
money into blue-chip funds that have performed better (New York Times Company, 1999). Unpredictable
returns mean that this strategy will not do anything
for investors’ portfolios over the long run except rack
up trading costs. If funds are selling stocks, then

43

contrarian investors must be buying them, but unpredictable returns mean that this strategy can not
improve performance either. If one can not systematically make money, one can not systematically
lose money either.
As discussed in the introduction, researchers
once believed that stock returns (more precisely, the
excess returns on stocks over short-term interest
rates) were completely unpredictable. It now turns
out that average returns on the market and individual
securities do vary over time and that stock returns
are predictable. Alas for would-be technical traders,
much of that predictability comes at long horizons
and seems to be associated with business cycles and
financial distress.
Market returns
Table 1 presents a regression that forecasts returns. Low prices—relative to dividends, book value,
earnings, sales, or other divisors—predict higher subsequent returns. As the R2 values in table 1 show,
these are long-horizon effects: Annual returns are
only slightly predictable and month-to-month returns
are still strikingly unpredictable, but returns at five-year
horizons seem very predictable. (Fama and French, 1989,
is an excellent reference for this kind of regression).
The results at different horizons are reflections
of a single underlying phenomenon. If daily returns
are very slightly predictable by a slow-moving variable, that predictability adds up over long horizons.
For example, you can predict that the temperature in
Chicago will rise about one-third of a degree per day
in spring. This forecast explains very little of the day
to day variation in temperature, but tracks almost all
of the rise in temperature from January to July. Thus,
the R2 rises with horizon.
Precisely, suppose that we forecast returns with
a forecasting variable x, according to
1)

Rt +1 - R

2)

x t +1 = c + rxt + d t +1 .

TB
t +1

= a + bx t + e t +1

Small values of b and R in equation 1 and a large
coefficient ρ in equation 2 imply mathematically that
the long-horizon regression as in table 1 has a large
regression coefficient b and large R2.
This regression has a powerful implication: Stocks
are in many ways like bonds. Any bond investor understands that a string of good past returns that pushes the price up is bad news for subsequent returns.
Many stock investors see a string of good past returns
and become elated that we seem to be in a “bull
2

44

market,” concluding future stock returns will be good
as well. The regression reveals the opposite: A string
of good past returns which drives up stock prices
is bad news for subsequent stock returns, as it is
for bonds.
Long-horizon return predictability was first documented in the volatility tests of Shiller (1981) and
LeRoy and Porter (1981). They found that stock prices vary far too much to be accounted for by changing expectations of subsequent cash flows; thus
changing discount rates or expected returns must
account for variation in stock prices. These volatility
tests turn out to be almost identical to regressions
such as those in table 1 (Cochrane, 1991).
Momentum and reversal
Since a string of good returns gives a high price,
it is not surprising that individual stocks that do
well for a long time (and reach a high price) subsequently do poorly, and stocks that do poorly for a
long time (and reach a low price, market value, or market to book ratio) subsequently do well. Table 2, taken
from Fama and French (1996) confirms this hunch.
(Also, see DeBont and Thaler, 1985, and Jegadeesh
and Titman, 1993.)
The first row in table 2 tracks the average monthly
return from the reversal strategy. Each month, allocate
all stocks to ten portfolios based on performance
from year –5 to year –1. Then, buy the best-performing portfolio and short the worst-performing portfolio.
This strategy earns a hefty –0.74 percent monthly
return.4 Past long-term losers come back and past
winners do badly. Fama and French (1996) verify that
these portfolio returns are explained by their threefactor model. Past winners move with value stocks,
TABLE 1

OLS regression of excess returns on
price/dividend ratio
Horizon k
1 year
2 years
3 years
5 years

b

Standard error

R2

–1.04
–2.04
–2.84
–6.22

0.33
0.66
0.88
1.24

0.17
0.26
0.38
0.59

Notes: OLS regressions of excess returns (value-weighted
NYSE–Treasury bill rate) on value-weighted price/dividend
ratio.
TB
RtVW
–t +k - Rt –t +k = a + b( Pt / Dt ) + et +k .

Rt→t+k indicates the k year return. Standard errors use GMM
to correct for heteroskedasticity and serial correlation.

Economic Perspectives

results of table 2. The key is the large
standard deviation of individual stock
Average monthly returns, reversal and momentum strategies
returns, typically 40 percent or more on
Portfolio
Average
an annual basis. The average return of
Strategy
Period
formation return, 10–1
the best performing decile of a normal
(months)
(monthly %)
distribution is 1.76 standard deviations
above the mean,5 so the winning momenReversal
July 1963–Dec. 1993
60–13
–0.74
tum portfolio went up about 80 percent
Momentum July 1963–Dec. 1993
12–2
+1.31
Reversal
Jan. 1931–Feb. 1963
60–13
–1.61
in the previous year and the typical losMomentum Jan. 1931–Feb. 1963
12–2
+0.38
ing portfolio went down about 60
percent. Only a small amount of continuNotes: Each month, allocate all NYSE firms to 10 portfolios
based on their performance during the “portfolio formation months”
ation will give a 1 percent monthly return
interval. For example, 60–13 forms portfolios based on returns from 5
when multiplied by such large past
years ago to 1 year, 1 month ago. Then buy the best-performing decile
portfolio and short the worst-performing decile portfolio.
returns. To be precise, the monthly indiSource: Fama and French (1996, table 6).
vidual stock standard deviation is about
40% / 12 £ 12%. If the R2 is 0.0025,
the standard deviation of the predictand so inherit the value stock premium. (To compare
able part of returns is .0025 œ 12% £ 0.6%. Hence,
the strategies, the table always buys the winners and
the decile predicted to perform best will earn
shorts the losers. In practice, of course, you buy the
176
. œ 0.6% £ 1% above the mean. Since the strategy
losers and short the winners to earn +0.71 percent
buys the winners and shorts the losers, an R2 of
monthly average return.)
0.0025 implies that one should earn a 2 percent monthly
The second row of table 2 tracks the average
return by the momentum strategy.
monthly return from a momentum strategy. Each
We have known at least since Fama (1965) that
month, allocate all stocks to ten portfolios based on
monthly and higher frequency stock returns have
performance in the last year. Now, the winners conslight, statistically significant predictability with R2
tinue to win and the losers continue to lose, so that
about 0.01. Campbell, Lo, and MacKinlay (1997, table
buying the winners and shorting the losers generates
2.4) provide an updated summary of index autocorrelaa positive 1.31 percent monthly return.
tions (the R2 is the squared autocorrelation), part of
Momentum is not explained by the Fama–French
which I show in table 3. Note the correlation of the equal(1996) three-factor model. The past losers have low
ly weighted portfolio, which emphasizes small stocks.6
prices and tend to move with value stocks. Hence, the
However, such small, though statistically signifimodel predicts that they should have high average
cant, high-frequency predictability has thus far failed
returns, not low average returns.
to yield exploitable profits after one takes into account
Momentum stocks move together, as do value
transaction costs, thin trading of small stocks, and
and small stocks, so a “momentum factor” works to
high short-sale costs. The momentum strategy for ex“explain” momentum portfolio returns (Carhart, 1997).
ploiting this correlation may not work in practice for
This step is so obviously ad hoc (that is, an APT facthe same reasons. Momentum does require frequent
tor that will only explain returns of portfolios orgatrading. The portfolios in table 2 are re-formed every
nized on the same characteristic as the factor rather
than a proxy for macroeconomic risk) that most peoTABLE 3
ple are uncomfortable adding it. We obviously do not
want to add a new factor for every anomaly.
First-order autocorrelation, CRSP value- and
Is momentum really there, and if so, is it exploitable
equally weighted index returns
after transaction costs? One warning is that it does not
Frequency
Portfolio
Correlation ρ1
seem stable over subsamples. The third and fourth
lines in table 2 show that the momentum effect essenDaily
Value-weighted
0.18
tially disappears in the earlier data sample, while reversal
Equally weighted
0.35
is even stronger in that sample.
Monthly
Value-weighted
0.043
Momentum is really just a new way of looking at
Equally weighted
0.17
an old phenomenon, the small apparent predictability
of monthly individual stock returns. A tiny regression
Note: Sample 1962–94.
Source: Campbell, Lo, and MacKinlay (1997).
R2 for forecasting monthly returns of 0.0025 (0.25 percent) is more than adequate to generate the momentum
TABLE 2

Federal Reserve Bank of Chicago

45

month. Annual winners and losers will not change
that often, but the winning and losing portfolio must
be turned over at least once per year. In a quantitative
examination of this effect, Carhart (1997) concludes
that momentum is not exploitable after transaction
costs are taken into account. Moskowitz and Grinblatt
(1999) note that most of the apparent gains from the
momentum strategy come from short positions in
small illiquid stocks. They also find that a large part
of momentum profits come from short positions taken
in November. Many investors sell losing stocks toward
the end of December to establish tax losses. By shorting illiquid losing stocks in November, an investor
can profit from the selling pressure in December. This
is also an anomaly, but it seems like a glitch rather than
a central principle of risk and return in asset markets.
Even if momentum and reversal are real and as
strong as indicated by table 2, they do not justify much
of the trading based on past results that many investors seem to do. To get the 1 percent per month

momentum return, one buys a portfolio that has typically gone up 80 percent in the last year, and shorts a
portfolio that has typically gone down 60 percent.
Trading between stocks and fund categories such as
value and blue-chip with smaller past returns yields
at best proportionally smaller results. Since much of
the momentum return seems to come from shorting
small illiquid stocks, mild momentum strategies may
yield even less. And we have not quantified the substantial risk of momentum strategies.
Bonds
The venerable expectations model of the term
structure specifies that long-term bond yields are
equal to the average of expected future short-term bond
yields (see box 2). For example, if long-term bond yields
are higher than short-term bond yields—if the yield
curve is upward sloping—this means that short-term
rates are expected to rise in the future. The rise in future short-term rates means that investors can expect

BOX 2

Bond definitions and expectations hypothesis
Let pt( N ) denote the log of the N year discount
bond price at time t. The N period continuously
1
compounded yield is defined by yt( N ) = - pt( N ) .
N
The continuously compounded holding period
return is the selling price less the buying price,
hprt(+N1 ) = pt(+N1-1) - pt( N ) . The forward rate is the rate
at which an investor can contract today to borrow
money N – 1 years from now, and repay that money
N years from now. Since an investor can synthesize a forward contract from discount bonds, the
forward rate is determined from discount bond
prices by

f t ( N ) = pt( N –1) – pt( N ) .
The “spot rate” refers, by contrast with a forward rate, to the yield on any bond for which the
investor take immediate delivery. Forward rates
are typically higher than spot rates when the yield
curve rises, since the yield is the average of intervening forward rates,
yt( N ) =

1 (1)
( f t + f t ( 2 ) + f t ( 3) + ... + f t ( N ) ).
N

The expectations hypothesis states that the
expected log or continuously compounded return
should be the same for any bond strategy. This
statement has three mathematically equivalent
expressions:

46

1. The forward rate should equal the expected
value of the future spot rate,

f t( N ) = Et ( yt(+1)N -1 ).
2. The expected holding period return should be
the same on bonds of any maturity

Et (hprt(+N1 ) ) = Et ( hprt(+M1 ) ) = yt(1) .
3. The long-term bond yield should equal the
average of the expected future short rates,
yt( N ) =

1
Et ( yt(1) + yt(1+)1 + ... + yt(+1)N -1 ).
N

The expectations hypothesis is often amended to allow a constant risk premium of undetermined sign in these equations. Its violation is then
often described as evidence for a “time-varying
risk premium.”
The expectations hypothesis is not quite the
same thing as risk-neutrality, because the expected
log return is not equal to the log expected return.
However, the issues here are larger than the difference between the expectations hypothesis and
strict risk-neutrality.

Economic Perspectives

TABLE 4

Zero-coupon bond returns
Maturity Average holding
N
period return
1
2
3
4
5

5.83
6.15
6.40
6.40
6.36

Standard
error
0.42
0.54
0.69
0.85
0.98

Standard
deviation
2.83
3.65
4.66
5.71
6.58

Note: Continuously compounded one-year holding period
returns on zero-coupon bonds of varying maturity. Annual
data from CRSP 1953–97.

the same rate of return whether they hold a long-term
bond to maturity or roll over short-term bonds with
initially low returns and subsequent higher returns.
As with the CAPM and the view that stock returns
are independent over time, a new round of research
has significantly modified this traditional view of
bond markets.
Table 4 calculates the average return on bonds
of different maturities. The expectations hypothesis
seems to do pretty well. Average holding period returns
do not seem very different across bond maturities,
despite the increasing standard deviation of longermaturity bond returns. The small increase in average
returns for long-term bonds, equivalent to a slight
average upward slope in the yield curve, is usually
excused as a “liquidity premium.” Table 4 is just the
tip of an iceberg of successes for the expectations
model. Especially in times of significant inflation and
exchange rate instability, the expectations hypothesis
has done a very good first-order job of explaining the
term structure of interest rates.

However, if there are times when long-term bonds
are expected to do better and other times when shortterm bonds are expected to do better, the unconditional
averages in table 4 could still show no pattern. Similarly, one might want to check whether a forward rate
that is unusually high forecasts an unusual increase
in spot rates.
Table 5 updates Fama and Bliss’s (1987) classic
regression tests of this idea. Panel A presents a regression of the change in yields on the forward-spot
spread. (The forward-spot spread measures the slope
of the yield curve.) The expectations hypothesis predicts a slope coefficient of 1.0, since the forward rate
should equal the expected future spot rate. If, for
example, forward rates are lower than expected future
spot rates, traders can lock in a borrowing position
with a forward contract and then lend at the higher
spot rate when the time comes.
Instead, at a one-year horizon we find slope coefficients near zero and a negative adjusted R2. Forward rates one year out seem to have no predictive
power whatsoever for changes in the spot rate one
year from now. On the other hand, by four years out,
we see slope coefficients within one standard error of
1.0. Thus, the expectations hypothesis seems to do
poorly at short (one-year) horizons, but much better
at longer horizons.
If the expectations hypothesis does not work at
one-year horizons, then there is money to be made—
one must be able to foresee years in which short-term
bonds will return more than long-term bonds and vice
versa, at least to some extent. To confirm this implication, panel B of table 5 runs regressions of the one-year
excess return on long-term bonds on the forwardspot spread. Here, the expectations hypothesis predicts a coefficient of zero: No signal (including the

TABLE 5

Forecasts based on forward-spot spread
A.

N

Change in yields

Intercept

Standard
error,
intercept

Slope

B.
Standard
error,
slope

Adjusted
R2

Holding period returns

Intercept

Standard
error,
intercept

Slope

Standard
error,
slope

Adjusted
R2

1

0.10

0.3

–0.10

0.36

–0.020

–0.1

0.3

1.10

0.36

0.16

2

–0.01

0.4

0.37

0.33

0.005

–0.5

0.5

1.46

0.44

0.19

3

–0.04

0.5

0.41

0.33

0.013

–0.4

0.8

1.30

0.54

0.10

4

–0.30

0.5

0.77

0.31

0.110

–0.5

1.0

1.31

0.63

0.07

Notes: OLS regressions, 1953–97 annual data. Panel A estimates the regression y – y = a +b ( ft
– y ) + ε t+N
and panel B estimates the regression hpr (N)
–yt(1) = a + b (f t( N+1) – y (1)
) + ε t+1, where y denotes the N-year bond yield at
t+1
t
date t; f t( N) denotes the N-period ahead forward rate; and hpr (Nt+)1 denotes the one-year holding period return at date t + 1
on an N-year bond. Yields and returns in annual percentages.
(1)
t+ n
(N)
t

Federal Reserve Bank of Chicago

(1)
t

(N +1)

(1)
t

47

forward-spot spread) should be able to tell you that
this is a particularly good time for long bonds versus
short bonds, as the random walk view of stock prices
says that no signal should be able to tell you that
this is a particularly good or bad day for stocks versus
bonds. However, the coefficients in panel B are all
about 1.0. A high forward rate does not indicate that
interest rates will be higher one year from now; it
seems to indicate that investors will earn that much
more by holding long-term bonds.7
Of course, there is risk. The R2 values are all 0.1–
0.2, about the same values as the R 2 from the d/p regression at a one-year horizon, so this strategy will
often go wrong. Still, 0.1–0.2 is not zero, so the strategy does pay off more often than not, in violation of
the expectations hypothesis. Furthermore, the forward-spot spread is a slow-moving variable, typically
reversing sign once per business cycle. Thus, the R2
builds with horizon as with the d/p regression, peaking in the 30 percent range (Fama and French, 1989).
Foreign exchange

making tidy sums borrowing at 5 percent in dollars to
lend at 20 percent in local currencies. This suggests
that traders were anticipating a 15 percent devaluation, or a smaller chance of a larger devaluation, which
is exactly what happened. Many observers attribute
high nominal interest rates in troubled economies to
“tight monetary policy” aimed at defending the currency. In reality, high nominal rates reflect a large probability of inflation and devaluation—loose monetary
policy—and correspond to much lower real rates.
Still, does a 5 percent interest rate differential
correspond to a 5 percent expected depreciation, or
does some of it represent a high expected return from
holding debt in that country’s currency? Furthermore, while expected depreciation is clearly a large
part of the interest rate story in high-inflation economies, how does the story play out in economies like
the U.S. and Germany, where inflation rates diverge
little but exchange rates still fluctuate a large amount?
The first row of table 6 (from Hodrick, 2000, and
Engel, 1996) shows the average appreciation of the
dollar against the indicated currency over the sample
period. The dollar fell against the deutschemark, yen,
and Swiss franc, but appreciated against the pound
sterling. The second row gives the average interest
rate differential—the amount by which the foreign
interest rate exceeds the U.S. interest rate.9 According to the expectations hypothesis, these two numbers should be equal—interest rates should be
higher in countries whose currencies depreciate
against the dollar.

Suppose interest rates are higher in Germany
than in the U.S. Does this mean that one can earn
more money by investing in German bonds? There
are several reasons that the answer might be no.
First, of course, is default risk. Governments have
defaulted on bonds in the past and may do so again.
Second, and more important, is the risk of devaluation. If German interest rates are 10 percent and U.S.
interest rates are 5 percent, but the euro falls 5 percent relative to the dollar during the year, you make
no more money holding the German bonds
despite their attractive interest rate. Since
TABLE 6
lots of investors are making this calculaForward discount puzzle
tion, it is natural to conclude that an interest rate differential across countries on
DeutschePound
Swiss
bonds of similar credit risk should reveal
mark
sterling
Yen
franc
an expectation of currency devaluation.
Mean appreciation
–1.8
3.6
–5.0 –3.0
The logic is exactly the same as that of the
Mean interest differential
–3.9
2.1
–3.7 –5.9
expectations hypothesis in the term strucb, 1975–89
–3.1
–2.0
–2.1 –2.6
ture. Initially attractive yield or interest rate
R2
.026
.033
.034 .033
differentials should be met by an offsetting
b, 1976–96
–0.7
–1.8
–2.4 –1.3
event so that you make no more money
b, 10-year horizon
0.8
0.6
0.5
–
on average in one maturity or currency
Notes: The first row gives the average appreciation of the dollar against the
versus another.8
indicated currency, in percent per year. The second row gives the average
interest differential—foreign interest rate less domestic interest rate,
As with the expectations hypothesis
measured as the forward premium—the 30-day forward rate less the spot
in the term structure, the expected depreciexchange rate. The third through sixth rows give the coefficients and R2 in a
regression of exchange rate changes on the interest differential,
ation view still constitutes an important
first-order understanding of interest rate
st+1 – st = a + b ( rtf – rtd) + ε t+1,
differentials and exchange rates. For examwhere s = log spot exchange rate, r f = foreign interest rate, and
r d = domestic interest rate.
ple, interest rates in east Asian currencies
Source: Hodrick (2000), Engel (1996), and Meredith and Chinn (1998).
were very high on the eve of the recent
currency tumbles, and many banks were

48

Economic Perspectives

The second row shows roughly the expected pattern. Countries with steady long-term inflation have
steadily higher interest rates and steady depreciation.
The numbers in the first and second rows are not
exactly the same, but exchange rates are notoriously
volatile so these averages are not well measured. Hodrick (2000) shows that the difference between the
first and second rows is not statistically different
from zero. This fact is analogous to the evidence in
table 4 that the expectations hypothesis works well
on average for U.S. bonds.
As in the case of bonds, however, we can ask
whether times of temporarily higher or lower interest
rate differentials correspond to times of above- and
below-average depreciation as they should. The third
and fifth rows of table 6 update Fama’s (1984) regression tests. The number here should be +1.0 in each
case—1 percentage point extra interest differential
should correspond to 1 percentage point extra expected depreciation. On the contrary, as table 6 shows,
a higher than usual interest rate abroad seems to lead
to further appreciation. This is the forward discount
puzzle. See Engel (1996) and Lewis (1995) for recent
surveys of the avalanche of academic work investigating whether this puzzle is really there and why.
The R2 values shown in table 6 are quite low.
However, like d/p and the term spread, the interest differential is a slow-moving forecasting variable, so the
return forecast R2 builds with horizon. Bekaert and
Hodrick (1992) report that the R2 rises to the 30 percent
to 40 percent range at six-month horizons and then
declines. That’s high, but not 100 percent; taking
advantage of any predictability strategy is quite risky.
The puzzle does not say that one earns more by
holding bonds from countries with higher interest
rates than others. Average inflation, depreciation,
and interest rate differentials line up as they should.
The puzzle does say that one earns more by holding
bonds from countries whose interest rates are higher
than usual relative to U.S. interest rates (and vice
versa). The fact that the “usual” rate of depreciation
and interest differential changes through time will, of
course, diminish the out-of-sample performance of
these trading rules.
One might expect that exchange rate depreciation
works better for long-run exchange rates, as the expectations hypothesis works better for long-run interest
rate changes. The last row of table 6, taken from
Meredith and Chinn (1998) verifies that this is so.
Ten-year exchange rate changes are correctly forecast
by the interest differentials of ten-year bonds.

Federal Reserve Bank of Chicago

Mutual funds
Studying the returns of funds that follow a specific strategy gives us a way to assess whether that
strategy works in practice, after transaction costs
and other trading realities are taken into account.
Studying the returns of actively managed funds tells
us whether the time, talent, and effort put into picking
securities pays off. Most of the literature on evaluating
fund performance is devoted to the latter question.
A large body of empirical work, starting with
Jensen (1969), finds that actively managed funds, on
average, underperform the market index. I use data
from Carhart (1997), whose measures of fund performance account for survivor bias. Survivor bias arises
because funds that do badly go out of business.
Therefore, the average fund that is alive at any point
in time has an artificially good track record.
As with the stock portfolios in figure 1, the fund
data in figure 7 show a definite correlation between
beta and average return: Funds that did well took on
more market risks. A cross-sectional regression line is
a bit flatter than the line drawn through the Treasury
bill and market return, but this is a typical result of
measurement error in the betas. (The data are annual,
and many funds are only around for a few years, contributing to beta measurement error.) The average fund
underperforms the line connecting Treasury bills and
the market index by 1.23 percent per year (that is, the
average alpha is –1.23 percent).
The wide dispersion in fund average returns in
figure 7 is a bit surprising. Average returns vary across
funds almost as much as they do across individual
stocks. This fact implies that the majority of funds
are not holding well-diversified portfolios that would
reduce return variation, but rather are loading up on
specific bets.
Initially, the fact that the average fund underperforms the market seems beside the point. Perhaps the
average fund is bad, but we want to know whether
the good funds are any good. The trouble is, we must
somehow distinguish skill from luck. The only way
to separate skill from luck is to group funds based
on some ex-ante observable characteristic, and then
examine the average performance of the group. Of
course, skillful funds should have done better, on
average, in the past, and should continue to do better
in the future. Thus, if there is skill in stock picking,
we should see some persistence in fund performance.
However, a generation of empirical work found no persistence at all. Funds that did well in the past were no
more likely to do well in the future.

49

French (1993) report that the HML portfolio alone gives nearly double the market
Average returns of mutual funds vs. market betas
Sharpe ratio—the same average return at
half the standard deviation. Why don’t
average return
funds cluster around a risk/reward line
0.25
= value
significantly above the market’s?
= neutral
Of course, we should not expect all
0.20
= growth
funds to cluster around a higher risk/re0.15
ward tradeoff. The average investor
holds the market, and if funds are large
0.10
enough, so must the average fund. Index
funds, of course, will perform like the in0.05
dex. Still, the typical actively managed
0.00
fund advertises high mean and, perhaps,
low variance. No fund advertises cutting
–0.05
average returns in half to spare investors
exposure to nonmarket sources of risk.
–0.10
Such funds, apparently aimed at mean–
variance investors, should cluster
–0.50
0.00
0.50
1.00
1.50
around the highest risk/reward tradeoff
beta
available from mechanical strategies
Notes: Average returns of mutual funds over the Treasury bill rate versus their
(and more, if active management does
market betas. Sample consists of all funds with average total net assets
greater than $25 million and more than 25 percent of their assets in stocks,
any good). Most troubling, funds who
in the Carhart (1996) database. Data sample 1962–96. The average excess
return is computed as E (R – R ) = a + β x 9%. a and β are computed from a
say they follow value strategies don’t
time-series regression of fund annual excess returns on market annual excess
returns over the life of the fund. The o, +, and x labels in the figure sort funds
outperform the market either. For example,
into thirds based on their regression coefficient h on the Fama–French value
Lakonishok, Shleifer, and Vishny (1992,
(HML) portfolio. The breakpoints are h = –0.084, 0.34. The dashed line gives
the fit of a cross-sectional ordinary least squares regression of a on β ; The
table 3) find that the average value fund
solid line connects the Treasury bill (β and excess return = 0) and the market
return (β = 1, excess return = 9%).
underperforms the S&P500 by 1 percent
just like all the others.
We can resolve this contradiction if
we
think
that
fund
managers were simply unaware of
Since the average fund underperforms the market,
the
possibilities
offered
by our new facts, and so
and fund returns are not predictable, we conclude
(despite
the
advertising)
were not really following
that active management does not generate superior
them.
That
seems
to
be
the
implication of figure 7,
performance, especially after transaction costs and
which
sorts
funds
by
their
HML
beta. One would exfees. This fact is surprising. Professionals in almost
pect
the
high-HML
beta
funds
to
outperform the marany field do better than amateurs. One would expect
ket
line.
But
the
cutoff
for
the
top
one-third of funds
that a trained experienced professional who spends
is
only
a
HML
beta
of
0.3,
and
even
that may be high
all day reading about markets and stocks should be
(many
funds
don’t
last
long,
so
betas
are poorly meaable to outperform simple indexing strategies. Even if
sured;
the
distribution
of
measured
betas
is wider
entry into the industry is so easy that the average
than
the
actual
distribution).
Thus,
the
“value
funds”
fund does not outperform simple indexes one would
were
really
not
following
the
“value
strategy”
that
expect a few stars to outperform year after year, as
earns the HML returns; if they were doing so they
good teams win championship after championship.
would have HML betas of 1.0. Similarly, Lakonishok,
Alas, the contrary fact is the result of practically every
Shleifer, and Vishny’s (1992) documentation of value
investigation, and even the anomalous results docufunds’ underperformance reveals that their market
ment very small effects.
beta is close to 1.0. These results imply that value
Funds and value
funds are not really following a value strategy, since
Given the value, small-firm, and predictability
their returns correlate with the market portfolio and not
effects, the idea that funds cluster around the market
the value portfolio.
line is quite surprising. All of these new facts imply
Interestingly, the number of value and small-cap
inescapably that there are simple, mechanical stratefunds (as revealed by their betas, not their marketing
gies that can give a risk/reward ratio greater than that
claims) is increasing quickly. Before 1990, 14 percent
of buying and holding the market index. Fama and
of funds had measured SMB betas greater than 1.0,
FIGURE 7

i

f

i

i

i

i

i

50

i

Economic Perspectives

and 12 percent had HML betas greater than 1.0. In
the full sample, both numbers have doubled to 22
percent and 23 percent. This trend suggests that
funds will, in the future, be much less well described
by the market index.
The view that funds were unaware of value strategies, and are now moving quickly to exploit them,
can explain why most funds still earn near the market
return, rather than the higher value return. However,
this view contradicts the view that the value premium
is an equilibrium risk premium, that is, that everyone
knew about the value returns but chose not to invest
all along because they feared the risks of value strategies. If it is not an equilibrium risk premium, it won’t
last long.

individual fund returns. This result verifies that mutual
fund performance is persistent.
Perhaps the funds that did well took on more
market risks, raising their betas and, hence, average
returns in the following year. The third column in table
7 shows that this is not the case. The cross-sectional
variation in fund average returns has nothing to do
with market betas. Just as in the case of individual
stock returns, we have to understand fund returns
with multifactor models, if at all.
The last column of table 7 presents alphas (intercepts, the part of average return not explained by the
model) from a model with four factors—the market,
the Fama–French HML and SMB factors, and a momentum factor, PR1YR, that is long NYSE stocks that did
well in the last year and short NYSE stocks that did
Persistence in fund returns
poorly in the last year. In general, one should object
The fund counterpart to momentum in stock reto the inclusion of so many factors and such ad-hoc
turns has been more extensively investigated than
factors. However, this is a performance attribution
the value and size effects. Fund returns have also
rather than an economic explanation use of a multibeen found to be persistent. Since such persistence
factor model. We want to know whether fund perforcan be interpreted as evidence for persistent skill in
mance, and persistence in fund performance in
picking stocks, it is not surprising that it has attractparticular, is due to persistent stock-picking skill or
ed a great deal of attention, starting with Hendricks,
to mechanical strategies that investors could just as
Patel, and Zeckhauser (1993).
easily follow on their own, without paying the manTable 7, taken from Carhart (1997), shows that a
agement costs associated with investing through a
portfolio of the best-performing one-thirtieth of funds
fund. For this purpose, it does not matter whether the
last year outperforms a portfolio of the worst-perform“factors” represent true, underlying sources of macroing one-thirtieth of funds by 1 percent per month
economic risks.
(column 2). This is about the same size as the momenThe alphas in the last column of table 7 are almost
tum effect in stocks, and similarly results from a small
all about 1 percent to 2 percent per year negative. Thus,
autocorrelation plus a large standard deviation in
Carhart’s model explains that the persistence in fund
performance is due to persistence in the
underlying stocks, not persistent stockTABLE 7
picking skill. These results support the
Portfolios of mutual funds formed on previous year’s return
old conclusion that actively managed
funds underperform mechanical indexing
Average
4-factor
Last year rank
return
CAPM alpha
alpha
strategies. There is some remaining puzzling persistence, but it is all in the large
( - - - - - - - - - - - - - - - - - - - percent -- - - - - - - - - - - - - - - - - -)
negative alphas of the bottom one-tenth
1/30
0.75
0.27
–0.11
to bottom one-thirtieth of performers,
1/10
0.68
0.22
–0.12
which lose money year after year. Car5/10
0.38
–0.05
–0.14
hart also shows that the persistence of
9/10
0.23
–0.21
–0.20
fund performance is due to momentum in
10/10
0.01
–0.45
–0.40
the underlying stocks, rather than mo30/30
–0.25
–0.74
–0.64
mentum funds. If, by good luck, a fund
Notes: Each year, mutual funds are sorted into portfolios based on the
happened to pick stocks that went up
previous year’s return. The rank column gives the rank of the selected
portfolio. For example, 1/30 is the best performing portfolio when funds are
last year, the portfolio will continue to
divided into 30 categories. Average return gives the average monthly return
go up a bit this year.
in excess of the T-bill rate of this portfolio of funds for the following year.
Four-factor alpha gives the average return less the predictions of a
In sum, the new research does nothmultifactor model that uses the market, the Fama–French HML and SMB
ing
to
dispel the disappointing view of
portfolios, and portfolio PR1YR which is long NYSE stocks that did well in
the last year and short NYSE stocks that did poorly in the last year.
active management. However, we disSource: Carhart (1997).
cover that passively managed “style”

Federal Reserve Bank of Chicago

51

portfolios can earn returns that are not explained by
the CAPM.
Catastrophe insurance
A number of prominent funds have earned very
good returns (and others, spectacular losses) by following strategies such as convergence trades and
implicit put options. These strategies may also reflect
high average returns as compensation for nonmarket
dimensions of risk. They have not been examined at
the same level of detail as the value and small-cap
strategies, so I offer a possible interpretation rather
than a documented one.
Convergence trades take strong positions in very
similar securities that have small price differences.
For example, a 29.5-year Treasury bond typically
trades at a slightly higher yield (lower price) than a
30-year Treasury bond. (This was the most famous
bet placed by LTCM. See Lewis, 1999.) A convergence
trade puts a strong short position on the expensive
security and a strong long position on the cheap security. This strategy is often mislabeled an “arbitrage.”
However, the securities are similar, not identical. The
spread between 29.5- and 30-year Treasury bonds
reflects the lower liquidity of the shorter maturity
and the associated difficulty of selling it in a financial
panic. It is possible for this spread to widen. Nonetheless, panics are rare, and the average returns in
all the years when they do not happen may more than
make up for the spectacular losses when they do.
Put options protect investors from large price
declines. The volatility smile in put option prices
reflects the surprisingly high prices of such options,
compared with the small probability of large market
collapses (even when one calibrates the probability
directly, rather than using the log-normal distribution
of the Black–Scholes formula). Writers of out-of-themoney puts collect a fee every month; in a rare market
collapse they will pay out a huge sum, but if the probability of the collapse is small enough, the average
returns may be quite good.
All of these strategies can be thought of as catastrophe insurance (Hsieh and Fung, 1999). Most of
the time they earn a small premium. Once in a great
while they lose a lot, and they lose a lot in times of
financial catastrophe, when most investors are really
anxious that the value of their investments not evaporate. Therefore, it is economically plausible that
these strategies can earn positive average returns,
even when we account for stock market risk via the
CAPM and we correctly measure the small probabilities
of large losses.
The difficulty in empirically estimating the true
average return of such strategies, of course, is that

52

rare events are rare. Many long samples will give a
false sense of security because “the big one” that
justifies the premium happened not to hit.
The value, yield curve, and foreign exchange
strategies I survey above also exhibit features of
catastrophe insurance. Value stocks may earn high
returns because distressed stocks will all go bankrupt
in a financial panic. Buying bonds of countries with
high interest rates leaves one open to the small chance
of a large devaluation, and such devaluation is
especially likely to happen in a global financial panic.
Similarly, buying long-term bonds in the depth of a
recession when the yield curve is upward sloping
may expose one to a small risk of a large inflation.
If these interpretations bear out, they also suggest that the premiums—the average returns from
holding stocks sensitive to HML or from following
the bond and foreign exchange strategies—may be
overstated in the data. The markets have had an unusually good 50 years, and devastating financial
panics have not happened.
Implications of the new facts
While the list of new facts appears long, similar
patterns show up in every case. Prices reveal slowmoving market expectations of subsequent returns,
because potential offsetting events seem sluggish
or absent. The patterns suggest that investors can
earn substantial average returns by taking on the
risks of recession and financial stress. In addition,
there is a small positive autocorrelation of highfrequency returns.
The effects are not completely new. We have
known since the 1960s that high-frequency returns
are slightly predictable, with R2 of 0.01 to 0.1 in daily
to monthly returns. These effects were dismissed
because there didn’t seem to be much one could do
about them. A 51/49 bet is not very attractive, especially if there is any transaction cost. Also, the increased
Sharpe ratio (mean excess return/standard deviation)
from exploiting predictability is directly related to the
forecast R2, so a tiny R2, even if exploitable, did not
seem important. Now, we have a greater understanding of the potential importance of these effects and
their economic interpretations.
For price effects, we now realize that the R2 rises
with horizon when the forecasting variables are slowmoving. Hence, a small R2 at short horizons can mean a
really substantial R2 in the 30 percent to 50 percent
range at longer horizons. Also, the nature of these
effects suggests the kinds of additional sources of
priced risk that theorists had anticipated for 20 years.
For momentum effects, the ability to sort stocks and

Economic Perspectives

funds into momentum-based portfolios means that very
small predictability times portfolios with huge past returns gives important subsequent returns, though it is
not totally clear that this amplification of the small predictability really does survive transaction costs.
Price-based forecasts
If expected returns rise, prices are driven down,
since future dividends or other cash flows are discounted at a higher rate. A “low” price, then, can reveal a market expectation of a high expected or required return.10
Most of our results come from this effect. Low
price/dividend, price/earnings, or price/book values
signal times when the market as a whole will have high
average returns. Low market value (price times shares)
relative to book value signals securities or portfolios
that earn high average returns. The “small-firm” effect derives from low prices—other measures of size
such as number of employees or book value alone
have no predictive power for returns (Berk, 1997).
The “five-year reversal” effect derives from the fact
that five years of poor returns lead to a low price. A
high long-term bond yield means that the price of longterm bonds is “low,” and this seems to signal a time
of good long-term bond returns. A high foreign interest rate means a low price on foreign bonds, and this
seems to indicate good returns on the foreign bonds.
The most natural interpretation of all these effects
is that the expected or required return—the risk premium—on individual securities as well as the market
as a whole varies slowly over time. Thus we can track
market expectations of returns by watching price/dividend, price/earnings, or book/market ratios.

Absent offsetting events
In each case, an apparent difference in yield
should give rise to an offsetting movement, but does
not seem to do so. Something should be predictable
so that returns are not predictable, and it is not. Figure
8 provides a picture of the results in table 5. Suppose
that the yield curve is upward sloping as in panel A.
What does this mean? If the expectations model were
true, the forward rates plotted against maturity would
translate one for one to the forecast of future spot
rates in panel B, as plotted in the black line marked
“Expectations model.” A high long-term bond yield
relative to short-term bond yields should not mean a
higher expected long-term bond return. Subsequent
short rates should rise, cutting off the one-period
advantage of long-term bonds and raising the multiyear advantage of short-term bonds.
In figure 8, panel b, the colored line marked “Estimates” shows the actual forecast of future spot interest rates from the results in table 5. The essence of the
phenomenon is sluggish adjustment of the short
rates. The short rates do eventually rise to meet the
forward rate forecasts, but not as quickly as the forward rates predict they should. Short-term yields
should be forecastable so that returns are not forecastable. In fact, yields are almost unforecastable,
so, mechanically, bond returns are. The roughly 1.0
coefficients in panel B of table 5 mean that a 1 percentage point increase in the forward rate translates
into a 1 percentage point increase in expected return.
It seems that old fallacy of confusing bond yields
with their expected returns for the first year contains
a grain of truth.

FIGURE 8

Yield curve and forecast one-year interest rates
A. Current yield curve

B. Forecast one-year interest rates

percent
7.0

forecast interest rate
7.0

Forward rate

6.5

Expectations
model

6.5

Yield

6.0

6.0

Estimates

5.5

5.5
5.0

5.0

4.5

4.5
1

2

3
maturity, years

4

5

1

2

3
4
time in the future, years

5

Notes: Assuming that the current yield curve is as shown in panel A, the black line in panel
B gives the forecast from the expectations hypothesis, in which case forward rates today are
the forecast of future spot rates. The colored line in panel B gives the actual forecast of future
spot rates, constructed from the estimates in table 5.

Federal Reserve Bank of Chicago

53

In the same way, a high dividend yield on a stock
or portfolio should mean that dividends grow more
slowly over time, or, for individual stocks, that the firm
has taken on more market risk and will have a higher
market beta. These tendencies seem to be completely
absent. Dividend/price ratios do not seem to forecast
dividend growth and, hence, (mechanically) they
forecast returns. The one-year coefficient in table 1 is
very close to 1.00, meaning that a 1 percentage point
increase in the dividend yield translates into a 1 percentage point increase in return. It seems that the old
fallacy of confusing increased dividend yield with increased total return does contain a grain of truth.
A high foreign interest rate relative to domestic
interest rates should not mean a higher expected return. We should see, on average, an offsetting depreciation. But here, the coefficients are even larger than
1.0. An interest rate differential seems to predict a further appreciation. It seems that the old fallacy of
confusing interest rate differentials across countries
with expected returns, forgetting about depreciation,
also contains a grain of truth.
Economic interpretation
The price-based predictability patterns suggest
a premium for holding risks related to recession and
economy-wide financial distress. Stock and bond predictability are linked: The term spread (forward-spot,
or long yield–short yield) forecasts stock returns as
well as bond returns (Fama and French, 1989). Furthermore, the term spread is one of the best variables
for forecasting business cycles. It rises steeply at the
bottom of recessions and is inverted at the top of a
boom. Return forecasts are high at the bottom of a
business cycle and low at the top of a boom. Value
and small-cap stocks are typically distressed. Empirically successful economic models of the recession and
distress premiums are still in their infancy (Campbell
and Cochrane, 1999, is a start), but the story is at
least plausible and the effects have been expected
by theorists for a generation.
To make this point come to life, think concretely
about what you have to do to take advantage of the
predictability strategies. You have to buy stocks or
long-term bonds at the bottom, when stock prices are
low after a long and depressing bear market, in the
bottom of a recession or the peak of a financial panic.
This is a time when few people have the guts or the
wallet to buy risky stocks or risky long-term bonds.
Looking across stocks rather than over time, you have
to invest in value or small-cap companies, with years
of poor past returns, poor sales, or on the edge of
bankruptcy. You have to buy stocks that everyone
else thinks are dogs. Then, you have to sell stocks

54

and long-term bonds in good times, when stock prices
are high relative to dividends, earnings, and other
multiples and the yield curve is flat or inverted so
that long-term bond prices are high. You have to sell
the popular growth stocks, with good past returns,
good sales, and earnings growth.
You have to sell now, and the stocks that you
should sell are the blue-chips that everyone else seems
to be buying. In fact, the market timing strategies said
to sell long ago; if you did so, you would have missed
much of the runup in the Dow past the 6,000 point. Value stocks too have missed most of the recent market
runup. However, this shouldn’t worry you—a strategy
that holds risks uncorrelated with the market must underperform the market close to half of the time.
If this feels uncomfortable, what you’re feeling
is risk. If you’re uncomfortable watching the market
pass you by, perhaps you don’t really only care about
long-run mean and variance; you also care about
doing well when the market is doing well. If you want
to stay fully invested in stocks, perhaps you too feel
the time-varying aversion to or exposure to risk that
drives the average investor to stay fully invested
despite low prospective returns.
This line of explanation for the foreign exchange
puzzle is still a bit farther off (see Engel, 1996, for a
survey; Atkeson, Alvarez, and Kehoe, 1999, offer a
recent stab at an explanation). The strategy leads investors to invest in countries with high interest rates.
High interest rates are often a sign of monetary instability or other economic trouble, and thus may mean
that the investments are more exposed to the risks of
global financial stress or a global recession than are
investments in the bonds of countries with low interest
rates, which are typically enjoying better times.
Return correlation
Momentum and persistent fund performance
explained by a momentum factor are different from the
price-based predictability results. In both cases, the
underlying phenomenon is a small predictability of
high-frequency returns. The price-based predictability strategies make this predictability important by
showing that, with a slow-moving forecasting variable,
the R2 builds over horizon. Momentum, however, is
based on a fast-moving forecast variable—the previous year’s return. Therefore, the R2 declines rather
than building with horizon. Momentum makes the
small predictability of high-frequency returns significant in a different way, by forming portfolios of extreme
winners and losers. The large volatility of returns means
that the extreme portfolios will have extreme past
returns, so only a small continuation of past returns
gives a large current return.

Economic Perspectives

It would be appealing to understand momentum
as a reflection of slowly time-varying average expected
returns or risk premiums, like the price-based predictability strategies. If a stock’s average return rises for
a while, that should make returns higher both today
and tomorrow. Thus, a portfolio of past winners will
contain more than its share of stocks that performed
well because their average returns were higher, along
with stocks that performed well due to luck. The
average return of such a portfolio should be higher
tomorrow as well.
Unfortunately, this story has to posit a substantially different view of the underlying process for
varying expected returns than is needed to explain
everything else. The trouble is that a surprise increase
in expected returns means that prices will fall, since
dividends are now discounted at a greater rate. This
is the phenomenon we have relied on to explain why
low price/dividend, price/earnings, book/market,
value, and size forecast higher subsequent returns.
Therefore, positive correlation of expected returns
typically yields a negative correlation of realized
returns. To get a positive correlation of realized returns
out of slow expected return variation, you have to
imagine that an increase in average returns today is
either highly correlated with a decrease in expected
future dividend growth or with a decrease in expected
returns in the distant future (an impulse response
that starts positive but is negative at long horizons).
Campbell, Lo, and MacKinlay (1997) provide a quantitative exposition of these effects.
Furthermore, momentum returns have not yet
been linked to business cycles or financial distress in
even the informal way that I suggested for price-based
strategies. Thus, momentum still lacks a plausible
economic interpretation. To me, this adds weight to
the view that it isn’t there, it isn’t exploitable, or it
represents a small illiquidity (tax-loss selling of small
illiquid stocks) that will be quickly remedied once a
few traders understand it.
Remaining doubts
The size of all these effects is still somewhat
in question. It is always hard to measure average
returns of risky strategies. The standard formula
s / T for the standard error of the mean, together
with the high volatility σ of any strategy, means that
one needs 25 years of data to even start to measure
average returns. With σ = 16 percent (typical of the
index), even T = 25 years means that one standard
error is 16/5 ≅ 3 percent per year, and a two-standard
error confidence interval runs plus or minus 6 percentage points. This is not much smaller than the average
returns we are trying to measure. In addition, all of

Federal Reserve Bank of Chicago

these facts are highly influenced by the small probability of rare events, which makes measuring average
returns even harder.
Finally, viewed the right way, we have very few
data points with which to evaluate predictability. The
term premium and interest rate differentials only change
sign with the business cycle, and the dividend/price
ratio only crosses its mean once every generation.
The history of interest rates and inflation in the U.S.
is dominated by the increase, through two recessions,
to a peak in 1980 and then a slow decline after that.
Many of the anomalous risk premiums seem to
be declining over time. Figure 6 shows the decline in
the HML and SMB premiums, and the same may be
true of the predictability effects. The last three years
of high market returns have cut the estimated return
predictability from the dividend/price ratio in half.
This fact suggests that at least some of the premium
the new strategies yielded in the past was due to the
fact that they were simply overlooked.
Was it really clear to average investors in 1947 or
1963 (the beginning of the data samples) that stocks
would earn 9 percent over bonds, and that the strategy of buying distressed small stocks would double
even that return for the same level of risk? Would average investors have changed their portfolios with this
knowledge? Or would they have stayed pat, explaining that these returns are earned as a reward for risk
that they were not willing to take? Was it clear that
buying stocks at the bottom in the mid-1970s would
yield so much more than even that high average return? If we interpret the premiums measured in sample as true risk premiums, the answer must be yes.
If the answer is no, then at least some part of the
premium was luck and will disappear in the future.
Since the premiums are hard to measure, one is
tempted to put less emphasis on them. However, they
are crucial to our interpretation of the facts. The
CAPM is perfectly consistent with the fact that there
are additional sources of common variation. For example, it was long understood that stocks in the same
industry move together; the fact that value or small
stocks also move together need not cause a ripple.
The surprise is that investors seem to earn an average
return premium for holding these additional sources
of common movement, whereas the CAPM predicts
that (given beta) they should have no effect on a
portfolio’s average returns.
The behavior of funds also suggests the “overlooked strategy” interpretation. As explained earlier,
fund returns still cluster around the market line. It turns
out that very few fund returns actually followed the
value or other return-enhancing strategies. However,
the number of small, value, and related funds—funds

55

that actually do follow the strategies—has increased
dramatically in recent years. It might be possible to
explain this in a way consistent with the idea that investors knew the premiums were there all along, but
such an argument is obviously strained.
Conclusion
In sum, it now seems that investors can earn a
substantial premium for holding dimensions of risk
unrelated to market movements, such as recessionrelated or distress-related risk. Investors earn these
premiums by following strategies, such as value and

growth, market-timing possibilities generated by return
predictability, dynamic bond and foreign exchange
strategies, and maybe even a bit of momentum. The
exact size of the premiums and the economic nature
of the underlying risks is still a bit open to question,
but researchers are unlikely to go back to the simple
view that returns are independent over time and that
the CAPM describes the cross section.
The next question is, What should investors do
with this information? The article, “Portfolio advice
for a multifactor world,” also in this issue, addresses
that question.

NOTES
The market also tends to go down in recessions; however
recessions can be unusually severe or mild for a given level of
market return. What counts here is the severity of the recession
for a given market return. Technically, we are considering
betas in a multiple regression that includes both the market
return and a measure of recessions. See box 1.
1

2

I thank Gene Fama for providing me with these data.

The rest of the paragraph is my interpretation, not Fama and
French’s. They focus on the firm’s financial distress, while I
focus on the systematic distress, since idiosyncratic distress
cannot deliver a risk price.
3

Fama and French do not provide direct measures of standard deviations for these portfolios. One can infer, however, from the
betas, R2 values, and standard deviation of the market and factor
portfolios that the standard deviations are roughly one to two
times that of the market return, so Sharpe ratios of these strategies are comparable to that of the market return in sample.
4

The index autocorrelations suffer from some upward bias
since some stocks do not trade every day. Individual stock
autocorrelations are generally smaller, but are enough to
account for the momentum effect.
6

Panel B is really not independent evidence, since the coefficients in panels A and B of table 5 are mechanically linked. For
example, 1.14 + (–0.14) = 1.0, and this holds as an accounting
identity. Fama and Bliss (1987) call them “complementary
regressions.”
7

As with bonds, the expectations hypothesis is slightly different
from pure risk neutrality since the expectation of the log is
not the log of the expectation. Again, the size of the phenomena we study swamps this distinction.
8

The data are actually the spread between the forward exchange
rate and the spot exchange rate, but this quantity must equal
the interest rate differential in order to preclude arbitrage.
9

This effect is initially counterintuitive. One might suppose
that a higher average return would attract investors, raising
prices. But the higher prices, for a given dividend stream, must
reduce subsequent average returns. High average returns persist,
in equilibrium, when investors fear the increased risks of an
asset and try to sell, lowering prices.
10

5

We are looking for


I rf (r)dr
E(r|r › x) = x
,
I x f (r)dr

where x is defined as the top one-tenth cutoff,
 f ( r ) dr = 1 .

I x

10

With a normal distribution, x = 1.2816σ and E(r|r≥x) = 1.755σ.

REFERENCES

Atkeson, Andrew, Fernando Alvarez, and Patrick
Kehoe, 1999, “Volatile exchange rates and the forward
premium anomaly: A segmented asset market view,”
University of Chicago, working paper.
Banz, R. W., 1981, “The relationship between return
and market value of common stocks,” Journal of
Financial Economics, Vol. 9, No. 1, pp. 3–18.
Bekaert, Geert, and Robert J. Hodrick, 1992, “Characterizing predictable components in excess returns

56

on equity and foreign exchange markets,” Journal of
Finance, Vol. 47, No. 2, June, pp. 467–509.
Berk, Jonathan, 1997, “Does size really matter?,”
Financial Analysts Journal, Vol. 53, September/
October, pp. 12–18.
Campbell, John Y., 1996, “Understanding risk and return,” Journal of Political Economy, Vol. 104, No. 2,
April, pp. 298–345.

Economic Perspectives

Campbell, John Y., and John H. Cochrane, 1999, “By
force of habit: A consumption-based explanation of
aggregate stock market behavior,” Journal of Political
Economy, Vol. 107, No. 2, April, pp. 205–251.

Fama, Eugene F., and Robert R. Bliss, 1987, “The
information in long-maturity forward rates,” American
Economic Review, Vol. 77, No. 4, September,
pp. 680–692.

Campbell, John Y., Andrew W. Lo, and A. Craig
MacKinlay, 1997, The Econometrics of Financial
Markets, Princeton, NJ: Princeton University Press.

Fama, Eugene F., and Kenneth R. French, 1997, “Industry costs of equity,” Journal of Financial Economics,
Vol. 43, No. 2, February, pp. 153–193.

Carhart, Mark M., 1997, “On persistence in mutual
fund performance,” Journal of Finance, Vol. 52, No.
1, March, pp. 57–82.

, 1996, “Multifactor explanations of assetpricing anomalies,” Journal of Finance, Vol. 51, No. 1,
March, pp, 55–84.

Chen, Nai-Fu, Richard Roll, and Stephen A. Ross,
1986, “Economic forces and the stock market,” Journal
of Business, Vol. 59, No. 3, July, pp. 383–403.

, 1995, “Size and book-to-market factors in
earnings and returns,” Journal of Finance, Vol. 50,
No.1, March, pp. 131–155.

Cochrane, John H., 1997, “Where is the market going?
Uncertain facts and novel theories,” Economic
Perspectives, Federal Reserve Bank of Chicago, Vol.
21, No. 6, November/December, pp. 3–37.

, 1993, “Common risk factors in the returns
on stocks and bonds,” Journal of Financial Economics, Vol. 33, No. 1, February, pp. 3–56.

, 1996, “A cross-sectional test of an investment-based asset pricing model,” Journal of Political
Economy, Vol. 104, No. 3, June, pp. 572–621.
, 1991, “Volatility tests and efficient markets: Review essay,” Journal of Monetary Economics; Vol. 27, No. 3, June, pp. 463-485.
Daniel, Kent, David Hirshleifer, and Avanidhar
Subrahmanyam, 1998, “Investor psychology and
security market under- and overreactions,” Journal
of Finance, Vol. 3, No. 6, December, pp. 1839–1885.
DeBondt, Werner F. M., and Richard H. Thaler,
1985, “Does the stock market overreact?,” Journal
of Finance, Vol. 40, No. 3, pp. 793–805.
Engel, Charles, 1996, “The forward discount anomaly
and the risk premium: A survey of recent evidence,”
Journal of Empirical Finance, Vol. 3, pp. 123–192.
Fama, Eugene F. 1991, “Efficient markets II,” Journal
of Finance, Vol. 46, No. 5, December, pp. 1575–1617.
, 1984, “Forward and spot exchange
rates,” Journal of Monetary Economics, Vol. 14,
No. 3, November, pp. 319–338.
, 1970, “Efficient capital markets: A review
of theory and empirical work,” Journal of Finance,
Vol. 25, No. 2, May, pp. 383–417.
, 1965, “The behavior of stock market prices,”
Journal of Business, Vol. 38, No. 1, pp. 34–105.

Federal Reserve Bank of Chicago

, 1989, “Business conditions and expected
returns on stocks and bonds,” Journal of Financial
Economics, Vol. 25, No. 1, November, pp. 23–49.
Heaton, John, and Deborah Lucas, 1997, “Portfolio
choice and asset prices: The importance of entrepreneurial risk,” Northwestern University, manuscript.
Hendricks, Darryll, Jayendu Patel, and Richard
Zeckhauser, 1993, “Hot hands in mutual funds:
Short-term persistence of performance,” Journal of
Finance, Vol. 48, No. 1, March, pp. 93–130.
Hodrick, Robert, 2000, International Financial
Management, Englewood Cliffs, NJ: Prentice-Hall,
forthcoming.
Hsieh, David, and William Fung, 1999, “Hedge fund
risk management,” Duke University, working paper.
Jagannathan, Ravi, and Zhenyu Wang, 1996, “The
conditional CAPM and the cross-section of expected
returns,” Journal of Finance, Vol. 51, No. 1, March,
pp. 3–53.
Jegadeesh, Narasimham, and Sheridan Titman,
1993, “Returns to buying winners and selling losers:
Implications for stock market efficiency,” Journal of
Finance, Vol. 48, No. 1, March, pp. 65–91.
Jensen, Michael C., 1969, “The pricing of capital
assets and evaluation of investment portfolios,”
Journal of Business, Vol. 42, No. 2, April, pp. 167–247.

57

Lakonishok, Josef, Andrei Shleifer, and Robert W.
Vishny, 1992, “The structure and performance of the
money management industry,” Brookings Papers on
Economic Activity: Microeconomics 1992, Washington, DC, pp. 339–391
LeRoy, Stephen F., and Richard D. Porter, 1981, “The
present-value relation: Tests based on implied variance bounds,” Econometrica, Vol. 49, No. 3, May, pp.
555–574.
Lewis, Karen, K., 1995, “Puzzles in international
financial markets,” in Handbook of International
Economics, Vol. 3, G. Grossman and K. Rogoff (eds.),
Amsterdam, New York, and Oxford: Elsevier Science
B.V, pp. 1913–1971.

, 1971, “Optimum consumption and portfolio rules in a continuous time model,” Journal of
Economic Theory, Vol. 3, No. 4, pp. 373–413.
, 1969, “Lifetime portfolio selection under
uncertainty: The continuous time case,” Review of
Economics and Statistics, Vol. 51, No. 3, August,
pp. 247–257.
Moskowitz, Tobias, and Mark Grinblatt, 1999, “Tax
loss selling and return autocorrelation: New evidence,”
University of Chicago, working paper.
, 1998, “Do industries explain momentum?,” University of Chicago, CRSP working paper,
No. 480.

Lewis, Michael, 1999, “How the eggheads cracked,”
New York Times Magazine, January 24,
pp. 24–42.

New York Times Company, 1999, “Mutual funds
report: What’s killing the value managers?,” New
York Times, April 4, Section 3, p. 29.

Liew, Jimmy, and Maria Vassalou, 1999, “Can book-tomarket, size and momentum be risk factors that predict economic growth?,” Columbia University,
working paper.

Reyfman, Alexander, 1997, “Labor market risk and
expected asset returns,” University of Chicago,
Ph.D. thesis.

MacKinlay, A. Craig, 1995, “Multifactor models do
not explain deviations from the CAPM,” Journal of
Financial Economics, Vol. 38, No. 1, pp. 3–28.
Malkiel, Burton, 1990, A Random Walk Down Wall
Street, New York: Norton.
Markowitz, H., 1952, “Portfolio selection,” Journal
of Finance, Vol. 7, No. 1, March, pp. 77–99.
Meredith, Guy, and Menzie D. Chinn, 1998, “Longhorizon uncovered interest rate parity,” National Bureau of Economic Research, working paper, No. 6797.
Merton, Robert C., 1973, “An intertemporal capital
asset pricing model,” Econometrica, Vol. 41, No. 5,
September, pp. 867–887.

58

Ross, S. A., 1976, “The arbitrage theory of capital
asset pricing,” Journal of Economic Theory, Vol. 13,
No. 3, December, pp. 341–360.
Samuelson, Paul A., 1969, “Lifetime portfolio selection by dynamic stochastic programming,’’ Review
of Economics and Statistics, Vol. 51, No. 3, August,
pp. 239–246.
Sargent, Thomas J., 1993, Bounded Rationality in
Macroeconomics, Oxford: Oxford University Press.
Shiller, Robert J., 1981, “Do prices move too much
to be justified by subsequent changes in dividends?,” American Economic Review, Vol. 71, No. 3,
June, pp. 421–436.

Economic Perspectives

Portfolio advice for a multifactor world

John H. Cochrane

Introduction and summar
y
summary
A companion article in this issue, “New facts in
finance,” summarizes the revolution in how financial
economists view the world. Briefly, there are strategies
that result in high average returns without large betas,
that is, with no strong tendency for the strategy’s
returns to move up and down with the market as a
whole. Multifactor models have supplanted the capital
asset pricing model (CAPM) in describing these phenomena. Stock and bond returns, once thought to be
independent over time, turn out to be predictable at
long horizons. All of these phenomena seem to reflect
a premium for holding macroeconomic risks associated
with the business cycle and for holding assets that
do poorly in times of financial distress. They also all
reflect the information in prices—high prices lead to
low returns and low prices lead to high returns.
The world of investment opportunities has also
changed. Where once an investor faced a fairly straightforward choice between managed mutual funds, index
funds, and relatively expensive trading on his own
account, now he must choose among a bewildering
variety of fund styles (such as value, growth, small
cap, balanced, income, global, emerging market, and
convergence), as well as more complex claims of active
fund managers with customized styles and strategies,
and electronic trading via the Internet. (Msn.com’s
latest advertisement suggests that one should sign
up in order to “check the hour’s hottest stocks.” Does
a beleaguered investor really have to do that to earn
a reasonable return?) The advertisements of investment advisory services make it seem important to tailor
an investment portfolio from this bewildering set of
choices to the particular circumstances, goals, and
desires of each investor.
What should an investor do? An important current of academic research investigates how portfolio
theory should adapt to our new view of the financial

Federal Reserve Bank of Chicago

world. In this article, I summarize this research and I
distill its advice for investors. In particular, which of
the bewildering new investment styles seem most
promising? Should you attempt to time stock, bond,
or foreign exchange markets, and if so how much?
To what extent and how should an investment portfolio be tailored to your specific circumstances? Finally,
what can we say about the future investment environment? What kind of products will be attractive to
investors in the future, and how should public policy
react to these financial innovations?
I start by reviewing the traditional academic portfolio advice, which follows from the traditional view
that the CAPM is roughly correct and that returns are
not predictable over time. In that view, all investors
(who do not have special information) should split
their money between risk-free bonds and a broadbased passively managed index fund that approximates
the “market portfolio.” More risk-tolerant investors
put more money into the stock fund, more risk averse
investors put more money into the bond fund, and
that is it.
The new academic portfolio advice reacts to the
new facts. An investor should hold, in addition to the
market portfolio and risk-free bonds, a number of passively managed “style” funds that capture the broad
(nondiversifiable) risks common to large numbers

John H. Cochrane is the Sigmund E. Edelstone
Professor of Finance in the Graduate School of
Business at the University of Chicago, a consultant to
the Federal Reserve Bank of Chicago, and a research
associate at the National Bureau of Economic Research
(NBER). The author thanks Andrea Eisfeldt for research
assistance and David Marshall, John Campbell, and
Robert Shiller for comments. The author’s research is
supported by the Graduate School of Business and by
a grant from the National Science Foundation,
administered by the NBER.

59

of investors. In addition to the overall level of risk
aversion, his exposure to or aversion to the various
additional risk factors matters as well. For example,
an investor who owns a small steel company should
shade his investments away from a steel industry
portfolio, or cyclical stocks in general; a wealthy investor with no other business or labor income can afford
to take on the “value” and other stocks that seem to
offer a premium in return for potentially poor performance in times of financial distress. The stock market
is a way of transferring risks; those exposed to risks
can hedge them by proper investments, and those
who are not exposed to risks can earn a premium by
taking on risks that others do not wish to shoulder.
Since returns are somewhat predictable, investors
can enhance their average returns by moving their
assets around among broad categories of investments.
However, the market timing signals are slow-moving,
and I show that the uncertainty about the nature and
strength of market timing effects dramatically reduces
the optimal amount by which investors can profit
from them.
I emphasize a cautionary fact: The average investor must hold the market. You should only vary from
a passive market index if you are different from everyone else. It cannot be the case that every investor
should tilt his portfolio toward “value” or other highyield strategies. If everybody did it, the phenomenon
would disappear. Thus, if such strategies will persist
at all, it must be the case that for every investor who
should take advantage of them, there is another investor who should take an unusually low position,
sacrificing the good average returns for a reduction
in risk. It cannot be the case that every investor
should “market time,” buying when prices are low
and selling when prices are high. If everyone did it,
that phenomenon would also disappear. The phenomenon can only persist if, for every investor who should
enhance returns by such market-timing, there is another investor who is so exposed to or averse to the
time-varying risks that cause return predictability,
that he should “buy high and sell low,” again earning
a lower average return in exchange for avoiding risks.
We have only scratched the surface of asset
markets’ usefulness for sharing risks. As often in economics, what appears from the outside to be greedy
behavior is in fact socially useful.
The traditional view
The new portfolio theory really extends rather
than overturns the traditional academic portfolio theory. Thus, it’s useful to start by reminding ourselves
what the traditional portfolio theory is and why. The
traditional academic portfolio theory, starting from

60

Markowitz (1952) and expounded in every finance
textbook, remains one of the most useful and enduring
bits of economics developed in the last 50 years.
Two-fund theorem
The traditional advice is to split your investments
between a money-market fund and a broad-based,
passively managed stock fund. That fund should concentrate on minimizing fees and transaction costs,
period. It should avoid the temptation to actively
manage its portfolio, trying to chase the latest hot
stock or trying to foresee market movements. An index
fund or other approximation to the market portfolio
that passively holds a bit of every stock is ideal.
Figure 1 summarizes the analysis behind this
advice. The straight line gives the mean-variance
frontier—the portfolios that give the highest mean
return for every level of volatility. Every investor
should pick a portfolio on the mean-variance frontier.
The upward-curved lines are indifference curves that
represent investors’ preference for more mean return
and less volatility. The indifference curve to the lower
left represents a risk-averse investor, who chooses a
portfolio with less mean return but also less volatility;
the indifference curve to the upper right represents a
more risk-tolerant investor who chooses a portfolio
with more mean return but also higher risk.
This seems like a lot of person-specific portfolio
formation. However, every portfolio on the mean-variance frontier can be formed as a combination of a
risk-free money-market fund and the market portfolio
of all risky assets. Therefore, every investor need only
hold different proportions of these two funds.
Bad portfolio advice
The portfolio advice is not so remarkable for what
it does say, which given the setup is fairly straightforward, as for what it does not say. Compared with
common sense and much industry practice, it is radical advice.
One might have thought that investors willing to
take on a little more risk in exchange for the promise
of better returns should weight their portfolios to
riskier stocks, or to value, growth, small-cap, or other
riskier fund styles. Conversely, one might have thought
that investors who are willing to forego some return
for more safety should weight their portfolios to safer
stocks, or to blue-chip, income, or other safer fund
styles. Certainly, some professional advice in deciding
which style is suited for an investor’s risk tolerance,
if not a portfolio professionally tailored to each investor’s circumstances, seems only sensible and prudent.
The advertisements that promise “we build the portfolio that’s right for you” cater to this natural and
sensible-sounding idea.
Economic Perspectives

Figure 1 proves that nothing of the sort is true.
All stock portfolios lie on or inside the curved risky
asset frontier. Hence, an investor who wants more
return and is willing to take more risk than the market
portfolio will do better by borrowing to invest in the
market—including the large-cap, income, and otherwise safe stocks—than by holding a portfolio of riskier stocks. An investor who wants something less
risky than the market portfolio will do better by splitting his investment between the market and a moneymarket fund than by holding only safe stocks, even
though his stock portfolio will then contain some of
the small-cap, value, or otherwise risky stocks. Everyone holds the same market portfolio; the only decision
is how much of it to hold.
The two-fund theorem in principle still allows for
a good deal of customized portfolio formation and
active management if investors or managers have
different information or beliefs. For example, if an investor knows that small-cap stocks are ready for a
rebound, then the optimal (or tangency) portfolio that
reflects this knowledge will be more heavily weighted
toward small-cap stocks than the market portfolio held
by the average investor. All the analysis of figure 1
holds, but this specially constructed tangency portfolio goes in the place indicated by the market portfolio in the figure. However, the empirical success of
market efficiency, and the poor performance of professional managers relative to passive indexation, strongly
suggests that these attempts will not pay off for most
investors. For this reason, the standard advice is to
hold passively managed funds that concentrate on
minimizing transaction costs and fees, rather than a
carefully constructed tangency portfolio that reflects
an investor’s or manager’s special insights. However,
a quantitative portfolio management industry tries
hard to mix information or beliefs about the behavior

of different securities with the theory of figure 1 (for
example, see Black and Litterman, 1991).
The two-fund theorem leaves open the possibility
that the investor’s horizon matters as well as his risk
aversion. What could be more natural than the often
repeated advice that a long-term investor can afford
to ride out all the market’s short-term volatility, while
a short-term investor should avoid stocks because he
may have to sell at the bottom rather than wait for the
inevitable recovery after a price drop? The fallacy
lies in the “inevitable” recovery. If returns are close
to independent over time (like a coin flip), and prices
are close to a random walk, a price drop makes it no
more likely that prices will rise more in the future. This
view implies that stocks are not safer in the long run,
and the stock/bond allocation should be independent
of investment horizon.
This proposition can be shown to be precisely
true in several popular mathematical models of the
portfolio decision. If returns are independent over
time, then the mean and variance of continuously
compounded returns rises in proportion to the horizon:
The mean and variance of ten-year returns are ten
times those of one-year returns, so the ratio of mean
to variance is the same at all horizons. More elegantly,
Merton (1969) and Samuelson (1969) show that an
investor with a constant relative risk aversion utility
who can continually rebalance his portfolio between
stocks and bonds will always choose the same stock/
bond proportion regardless of investment horizon,
when returns are independent over time.

Taking the advice
This advice has had a sizable impact on portfolio
practice. Before this advice was widely popularized in
the early 1970s, the proposition that professional active
management and stock selection could outperform
blindly holding an index seemed selfevident, and passively managed index
FIGURE 1
funds were practically unknown. They
have exploded in size since then. The
Mean-variance frontier, optimal portfolios
remaining actively managed funds clearly
and two-fund theorem
feel the need to defend active management
average return E(R)
Mean-variance
in the face of the advice to hold passive
frontier
Investors want
index funds and the fact that active managers selected on any ex-ante basis underRisky-asset
Optimal portfolios
frontier
perform indexes ex-post.
Market portfolio
The one input to the optimal portfoOriginal assets
lio advice is risk tolerance, and many providers of investment services have started
Rf
thinking about how to measure risk tolerance using a series of questionnaires.
This is the trickiest part of the convenvolatility σ (R)
tional advice, in part since conventional

Federal Reserve Bank of Chicago

61

measures of risk tolerance often seem quite out of
whack with risk aversion displayed in asset markets.
(This is the equity premium puzzle; see Cochrane,
1997, for a review.) However, the basic question is
whether one is more risk tolerant or less risk tolerant
than the average investor. This question is fairly easy
to conceptualize and can lead to a solid qualitative,
if not quantitative, answer.
One might object to the logical inconsistency of
providing portfolio advice based on a view of the
world in which everyone is already following such
advice. (This logic is what allowed me to identify the
mean-variance frontier with the market portfolio.)
However, this logic is only wrong if other investors
are systematically wrong. If some investors hold too
much of a certain stock, but others hold too little of
it, market valuations are unaffected and the advice to
hold the market portfolio is still valid.
New portfolio theory
Multiple factors: An N-fund theorem
Figure 2 shows how the simple two-fund theorem
of figure 1 changes if there are multiple sources of
priced risk. This section is a graphical version of Fama’s
(1996) analysis. Much of the theory comes from
Merton (1969, 1971, 1973).
To keep the figure simple I consider just one additional factor. For concreteness, think of an additional
recession factor. Now, investors care about three
attributes of their portfolios: 1) They want higher

average returns; 2) they want lower standard deviations or overall risk; and 3) they are willing to accept
a portfolio with a little lower mean return or a little
higher standard deviation of return if the portfolio
does not do poorly in recessions. In the context of
figure 2, this means that investors are happier with
portfolios that are higher up (more mean), more to the
left (less standard deviation), and farther out (lower
recession sensitivity). The indifference curves of
figure 1 become indifference surfaces. Panel A of figure 2 shows one such surface curving upwards.
As with figure 1, we must next think about what
is available. We can now calculate a frontier of portfolios based on their mean, variance, and recession
sensitivity. This frontier is the multifactor efficient
frontier. A typical investor then picks a point as shown
in panel A of figure 2, which gives him the best possible
portfolio—trading off mean, variance, and recession
sensitivity—that is available. Investors want to hold
multifactor efficient, rather than mean-variance efficient, portfolios. As the mean-variance frontier of figure 1 is a hyperbola, this frontier is a revolution of a
hyperbola. The appendix summarizes the mathematics
behind this figure.
Panel B of figure 2 adds a risk-free rate. As the
mean-variance frontier of figure 1 was the minimal V
shape emanating from the risk-free rate (Rf) that includes
the hyperbolic risky frontier, now the multifactor efficient frontier is the minimum cone that includes the
hyperbolic risky multifactor efficient frontier, as shown.

FIGURE 2

Portfolio theory in a multifactor world
A. No risk-free rate

B. Risk-free rate
E(R)

E(R)

σ (R)

σ (R)
β

β

Notes: Panel A shows an indifference surface and optimal portfolio in the case with no risk-free rate.
The dot marks the optimal portfolio where the indifference surface touches the multifactor efficient
frontier. Panel B shows the set of multifactor efficient portfolios with a risk-free rate. The two coneshaped surfaces intersect on the black line with two dots. The two dots are the market portfolio and an
additional multifactor-efficient portfolio; all multifactor-efficient portfolios on the outer cone can be
reached by combinations of the risk-free rate, the market, and the extra multifactor-efficient portfolio.

62

Economic Perspectives

As every point of the mean-variance frontier of
figure 1 can be reached by some combination of two
funds—a risk-free rate and the market portfolio—now
every point on the multifactor efficient frontier can
be reached by some combination of three multifactor
efficient funds. The most convenient set of portfolios
is the risk-free rate (money-market security), the market
portfolio (the risky portfolio held by the average investor), and one additional multifactor efficient portfolio
on the tangency region as shown in panel B of figure
2. It is convenient to take this third portfolio to be a
zero-cost, zero-beta portfolio, so that it isolates the
extra dimension of risk.
Investors now may differ in their desire or ability
to take on recession-related risk as well as in their tolerance for overall risk. Thus, some will want portfolios
that are farther in and out, while others will want portfolios that are farther to the left and right. They can
achieve these varied portfolios by different weights in
the three multifactor efficient portfolios, or three funds.
Implications for mean-variance investors
The mean-variance frontier still exists—it is the
projection of the cone shown in figure 2 on the meanvariance plane. As the figure shows, the average investor is willing to give up some mean or accept more
variance in order to reduce the recession-sensitivity
of his portfolio. The average investor must hold the
market portfolio, so the market return is no longer
on the mean-variance frontier.
Suppose, however, that you are concerned only
with mean and variance—you are not exposed to the
recession risk, or the risks associated with any other
factor, and you only want to get the best possible mean
return for given standard deviation. If so, you still
want to solve the mean-variance problem of figure 1,
and you still want a mean-variance efficient portfolio.
The important implication of a multifactor world is that
you, the mean-variance investor, should no longer
hold the market portfolio.
You can still achieve a mean-variance efficient
portfolio just as in figure 1 by a combination of a
money market fund and a single tangency portfolio,
lying on the upper portion of the curved risky-asset
frontier. The tangency portfolio now takes stronger
positions than the market portfolio in factors such as
value or recession-sensitive stocks that the average
investor fears.
Predictable returns
The fact that returns are in fact somewhat predictable modifies the standard portfolio advice in three
ways. It introduces horizon effects, it allows markettiming strategies, and it introduces multiple factors

Federal Reserve Bank of Chicago

via hedging demands (if expected returns vary over
time, investors may want to hold assets that protect
them against this risk).
Horizon effects
Recall that when stock returns are independent
over time (like coin flips), the allocation between stocks
and bonds does not depend at all on the investment
horizon, since mean returns (reward) and the variance
of returns (risk) both increase in proportion to the
investment horizon. But if returns are predictable, the
mean and variance may no longer scale the same way
with horizon. If a high return today implies a high return
tomorrow—positive serial correlation—then the variance of returns will increase with horizon faster than
does the mean return. In this case, stocks are worse
in the long run. If a high return today implies a lower
return tomorrow—negative serial correlation or mean
reversion—then the variance of long-horizon returns
is lower than the variance of one-period returns times
the horizon. In this case, stocks are more attractive
for the long run.1 For example, if the second coin flip
is always the opposite of the first coin flip, then two
coin flips are much less risky than they would be if
each flip were independent, and a “long-run coin
flipper” is more likely to take the bet.
Which case is true? Overall, the evidence suggests that stock prices do tend to come back slowly
and partially after a shock, so return variances at
horizons of five years and longer are about one-half
to two-thirds as large as short-horizon variances suggest. Direct measures of the serial correlation of stock
returns, or equivalent direct measures of the mean
and variance of long-horizon returns, depend a lot on
the period studied and the econometric method. Multivariate methods give somewhat stronger evidence.
Intuitively, the price/dividend (p/d) ratio does not
explode. Hence, the long-run variance of prices must
be the same as the long-run variance of dividends,
and this extra piece of information helps to measure the
long-run variance of returns. (Cochrane and Sbordone,
1986, and Cochrane, 1994, use this idea. See Campbell,
Lo, and MacKinlay, 1997, for a summary of these issues
and of the extensive literature.)
How big are the horizon effects? Barberis (1999)
calculates optimal portfolios for different horizons
when returns are predictable. Figure 3 presents some
of his results.
We start with a very simple setup: The investor
allocates his portfolio between stocks and bonds and
then holds it without rebalancing for the indicated
horizon. His objective is to maximize the expected utility of wealth at the indicated horizon. The flat line in
figure 3 shows the standard result: If returns are not

63

predictable, then the allocation to stocks does not
depend on horizon.
The top (black) line in figure 3 adds the effects of
return predictability on the investment calculation. The
optimal allocation to stocks increases sharply with
horizon, from about 40 percent for a monthly horizon
to 100 percent for a ten-year horizon. To quantify the
effects of predictability, Barberis uses a simple model,
1)
2)

Rt +1 - RtTB
+1 = a + bx t + e t +1

Allocation to stocks, different
investment horizons
percent allocation to stocks
100

80

Predictable
Predictable,
uncertain parameters

60

x t +1 = c + rxt + d t +1 ,

using the d/p ratio for the forecasting variable x.
(Whether or not one includes returns in the right
hand side makes little difference.) Barberis estimates
significant mean-reversion: In Barberis’s regressions,
the implied standard deviation of ten-year returns is
23.7 percent, just more than half of the 45.2 percent
value implied by the standard deviation of monthly
returns. Stocks are indeed safer in the long run, and
the greater allocation to stocks shown in figure 3 for
a long-run investor reflects this fact.
Uncertainty about predictability
This calculation ignores the fact that we do not
know how predictable returns really are. One could
address this fact by calculating standard errors for
portfolio computations; and such standard errors do
indicate substantial uncertainty. However, standard
error uncertainty is symmetric—returns might be more
predictable than we think or they might be less predictable. This measure of uncertainty would say that we
are just as likely to want an even greater long-run
stock allocation as we are to shade the advice back
to a constant allocation.
Intuitively, however, uncertainty about predictability should lead us to shade the advice back toward
the standard advice. Standard errors do not capture
the uncertainties behind this (good) intuition for at
least two reasons.
First, the predictability captured in Barberis’s
regression of returns on dividend/price ratios certainly
results to some extent from data-dredging. Thousands
of series were examined by many authors, and we have
settled on the one or two that seem to predict returns
best in sample. The predictability will obviously be
worse out of sample, and good portfolio advice should
account for this bias. Standard errors take the set of
forecasting variables and the functional form as given.
Second, the portfolio calculation assumes that
the investor knows the return forecasting process
perfectly. Standard errors only reflect the fact that we
do not know the return forecasting process, so we

64

FIGURE 3

40

Unpredictable
Unpredictable, uncertain parameters

20
0

2

4
6
horizon, years

8

10

Notes: The investor maximizes the utility of terminal wealth via
a buy-and-hold investment in stocks versus bonds. The investor
has constant relative risk aversion utility with a risk aversion
coefficient of 10. The top calculation (black) includes predictable
returns modeled by a regression on d/p ratios (equation 1). The
second calculation (color) includes predictable returns and the
effects of parameter uncertainty. The third calculation (black dash)
assumes unpredictable returns, and no parameter uncertainty. The
bottom calculation (color dash) assumes unpredictable returns, but
adds parameter uncertainty. All distributions are conditional on a
dividend/price ratio equal to its historical mean.
Source: Barberis (1999).

are unsure about what investors want to do.2 What
we would like to do is to solve a portfolio problem in
which investors treat uncertainty about the forecastability of returns as part of the risk that they face, along
with the risks represented by the error terms of the
statistical model. Kandel and Stambaugh (1996) and
Barberis (1999) tackle this important problem.
Figure 3 also gives Barberis’s calculations of the
effects of parameter uncertainty on the stock/bond
allocation problem. (See box 1 for a description of how
these calculations are made.) The lowest (dotted) line
considers a simple case. The investor knows, correctly,
that returns are independent over time (not predictable) but he is not sure about the mean return. Without parameter uncertainty, this situation gives rise to
the constant stock allocation—the flat line. Adding
parameter uncertainty lowers the allocation to stocks
for long horizons; it declines from 34 percent to about
28 percent at a ten-year horizon.
The reason is simple. If the investor sees a few
good years of returns after making the investment,
this raises his estimate of the actual mean return and,
thus, his estimate of the returns over the rest of the
investment period. Conversely, a few bad years will
lower his estimate of the mean return for remaining
years. Thus, learning about parameters induces a

Economic Perspectives

BOX 1

How to include model uncertainty in portfolio problems
A statistical model, such as equations 1 and 2,
tells us the distribution of future returns once we
know the parameters θ, f (Rt +1 q, x1 , x2 , ..., xt ), where
xt denotes all the data used (returns, d/p, etc.).
We would like to evaluate uncertainty by the
distribution of returns conditional only on the history available to make guesses about the future,
f ( Rt +1 x 1 , x2 ... x t ). We can use Bayesian analysis
to evaluate this concept. If we can summarize the
information about parameters given the historical
data as f (q x1 , x2 .... xt ), then we can find the distribution we want by
f ( Rt +1 x1 , x2 ... xt ) =

I f (R

t +1

q) f (q x1 , x2 ... xt )d q.

f ( x1 , x2 ... xt q) via the standard law for conditional probabilities,
f ( q x 1 , x2 ... xt ) =
f ( x1 , x2 ... xt ) =

I

f ( x1 , x2 ... xt q ) f (q )
f ( x1 , x2 ... x t )

f ( x1 , x2 ... xt q ) f ( q ) d q.

Barberis (1999), Kandel and Stambaugh
(1996), Brennan, Schwartz, and Lagnado (1997)
use these rules to compute f ( Rt +1 x 1 , x2 ... xt ),
and solve portfolio problems with this distribution over future returns.

In turn, we can construct f (q x1 , x2 .... xt ),
from a prior f (q) and the likelihood function

positive correlation between early returns and later
returns. Positive correlation makes long-horizon returns
more than proportionally risky and reduces the optimal allocation to stocks.
The colored line in figure 3 shows the effects of
parameter uncertainty on the investment problem,
when we allow return predictability as well. As the
figure shows, uncertainty about predictable returns
cuts the increase in stock allocation from one to ten
years in half. In addition to the positive correlation of
returns due to learning about their mean mentioned
above, uncertainty about the true amount of predictability adds to the risk (including parameter risk) of
longer horizon returns.
Market timing
Market-timing strategies are the most obvious
implication of return predictability. If there are times
when expected returns are high and other times when
they are low, you might well want to hold more stocks
when expected returns are high, and fewer when expected returns are low. Of course, the crucial question is,
how much market-timing should you engage in? Several authors have recently addressed this technically
challenging question.
Much of the difficulty with return predictability
(as with other dynamic portfolio questions) lies in computing the optimal strategy—exactly how should you
adjust your portfolio as the return prediction signals
change? Gallant, Hansen, and Tauchen (1990) show
how to measure the potential benefits of market-timing
without actually calculating the market-timing strategy.

Federal Reserve Bank of Chicago

The mean–standard deviation tradeoff or Sharpe
ratio—the slope of the frontier graphed in figure 1—
is a convenient summary of any strategy. If the riskfree rate is constant and known, the square of the
maximum unconditional Sharpe ratio is the average
of the squared conditional Sharpe ratios. (The appendix details the calculation.) Since we take an average
of squared conditional Sharpe ratios, volatility in
conditional Sharpe ratios—time-variation in expected
returns or return volatility—is good for an investor
who cares about the unconditional Sharpe ratio. By
moving into stocks in times of high Sharpe ratio and
moving out of the market in times of low Sharpe ratio,
the investor does better than he would by buying
and holding. Furthermore, the best unconditional
Sharpe ratio is directly related to the R2 in the return-forecasting regression.
The buy-and-hold Sharpe ratio has been about
0.5 on an annual basis in U.S. data—stocks have
earned an average return of about 8 percent over
Treasury bills, with a standard deviation of about 16
percent. Table 1 presents a calculation of the increased
Sharpe ratio one should be able to achieve by markettiming, based on regressions of returns on dividend/
price ratios. (I use the regression estimates from table
1 of “New facts in finance.”)
As table 1 indicates, market-timing should be a
great benefit. Holding constant the portfolio volatility,
market-timing should raise average returns by about
two-fifths at an annual horizon and it should almost
double average returns at a five-year horizon.

65

TABLE 1

Maximum unconditional Sharpe ratios
Horizon k (years)
Buy & hold
1
2
3
5

R2

Annualized
Sharpe ratio

0.17
0.26
0.38
0.59

0.50
0.71
0.72
0.78
0.95

Notes: Maximum unconditional Sharpe ratios available
from market-timing based on regressions of value-weighted
NYSE index returns on the dividend/price ratio. The table
reports annualized Sharpe ratios corresponding to each R2.
The formula is

S*
k

= 05
. 2+

R2
/ 1 - R 2 and is derived in
k

the appendix.

FIGURE 4

Optimal allocation to stocks

fraction wealth in equities

Optimal market-timing:
An Euler equation approach
Brandt (1999) presents a clever way to estimate a
market-timing portfolio rule without solving a model.
Where standard asset pricing models fix the consumption or wealth process and estimate preference parameters, Brandt fixes the preference parameters (as one
does in a portfolio question) and estimates the portfolio decision, that is, he estimates the optimal consumption or wealth process.3 This calculation is very
clever because it does not require one to specify a
statistical model for the stock returns (like equations
1–2), and it does not require one to solve
the economic model.
Figure 4 presents one of Brandt’s results. The figure shows the optimal allocation to stocks as a function of investment
horizon and of the dividend/price ratio,
which forecasts returns. There is a mild
2.5
horizon effect, about in line with Barberis’s
results of figure 3 without parameter uncer2.0
tainty: Longer term investors hold more
1.5
stocks. There is also a strong market-timing
effect. The fraction of wealth invested in
1.0
stocks varies by about 200 percentage
0.5
points for all investors. For example, long0.0
term investors vary from about 75 percent
48
to 225 percent of wealth invested in stocks
as the d/p ratio rises from 2.8 percent to
5.5 percent.

who desire lifetime consumption4 rather than portfolio
returns at a fixed horizon. They model the time-variation in expected stock returns via equation 1 on d/p
ratios. Their investors live only off invested wealth
and have no labor income or labor income risk. Thus,
these investors are poised to take advantage of business cycle related variation in expected returns.
As one might expect, the optimal investment strategy takes strong advantage of market-timing possibilities. Figure 5 reproduces Campbell and Vicera’s optimal
allocation to stocks as a function of the expected
return, forecast from d/p ratios via equation 1. A risk
aversion coefficient of 4 implies that investors roughly want to be fully invested in stocks at the average
expected excess return of 6 percent, so this is a sensible risk aversion value to consider. Then, as the d/p
ratio ranges from minus two to plus one standard
deviations from its mean, these investors range from
–50 percent in stocks to 220 percent in stocks. This
is aggressive market-timing indeed.
Figure 6 presents the calculation in a different
way: It gives the optimal allocation to stocks over
time, based on dividend/price ratio variation over
time. The high d/p ratios of the 1950s suggest a
strong stock position, and that strong position profits
from the high returns of the late 1950s to early 1960s.
The low d/p ratios of the 1960s suggest a much smaller
position in stocks, and this smaller position avoids
the bad returns of the 1970s. The high d/p ratios of
the 1970s suggest strong stock positions again, which
benefit from the good return of the 1980s; current

Optimal market-timing: A solution
Campbell and Vicera (1999) actually
calculate a solution to the optimal markettiming question. They model investors

66

26
hor
izo
n (m24
ont 12
hs)

0

2.8

3.3

3.9
e
divid

4.9
4.4
ratio
e
ric
nd/p

5.5

Notes: Optimal allocation to stocks as a function of horizon and dividend yield.
Source: Brandt (1999).

Economic Perspectives

FIGURE 5

Optimal allocation to stocks

Campbell and Vicera also present achieved utility
calculations that mirror the lesson of table 1: Failing
to time the market seems to impose a large cost.

allocation to stocks, percent
220

Doubts
One may be understandably reluctant to take on
quite
such strong market-timing positions as indicated
140
by
figures
5 and 6, or to believe table 1 that market
100
timing
can
nearly double five-year Sharpe ratios. In
60
particular,
one
might question advice that would have
20
meant missing the dramatic runup in stock values of
–20
the late 1990s. Rather than a failure of nerve, perhaps
–60
such reluctance reveals that the calculations do not
–100
yet include important considerations and, therefore,
–5
–4
–2
0
2
4
6
8
10
12
overstate the desirable amount of market-timing and
log expected gross excess return, percent
its benefits.
Notes: Optimal allocation to stocks as a function of the expected
return implied by a regression that forecasts stock returns from
First, the unconditional Sharpe ratio as reported
dividend/price ratios. The line extends from a d/p ratio two
in
table
1 for, say, five-year horizons answers the quesstandard deviations above its mean (low expected returns) to
one standard deviation below its mean (high expected returns).
tion,
“Over
very long periods, if an investor follows
Risk aversion is 4.0.
Source: Campbell and Vicera (1999).
the best possible market-timing strategy and evaluates
his portfolio based on five-year returns, what Sharpe
ratio does he achieve?” It does not answer the question, “Given today’s d/p, what is the best Sharpe raunprecedented high prices suggest the lowest stock
tio you can achieve for the next five years by
positions ever. The optimal allocation to stocks again
following market-timing signals?” The latter question
varies wildly, from 0 (now) to over 300 percent.
characterizes the return distribution conditional on toCampbell and Vicera’s calculations are, if anything,
day’s d/p. It is harder to evaluate; it depends on the
conservative compared with others in this literature.
initial d/p level, and it is lower, especially for a slowOther calculations, using other utility functions, somoving signal such as d/p.
lution techniques, and calibrations of the forecasting
To see the point, suppose that the d/p ratio is
process often produce even more aggressive marketdetermined
on day one, is constant thereafter, and intiming strategies. For example, Brennan, Schwartz, and
dicates
high
or low returns in perpetuity. Conditional
Lagnado (1997) make a similar calculation with two
on
the
d/p
ratio,
one cannot time the market at all. But
additional forecasting variables. They report marketsince
the
investor
will invest less in stocks in the lowtiming strategies that essentially jump back and forth
return
state
and
more
in the high-return state, he will
between constraints at 0 percent in stocks and 100
unconditionally
time
the
market (that is, adjust his
percent in stocks.
portfolio based on day one information)
and this gives him a better date-zero (unconditional) Sharpe ratio than he would
FIGURE 6
obtain by fixing his allocation at date zero.
Optimal allocation to stocks based on dividend/price ratio
This fact captures the intuition that there
allocation to stocks, percent
is a lot more money to be made from a 50
400
percent R2 at a daily horizon than at a five320
year horizon, where the calculations in
table 1 are not affected by the persistence
240
of the market-timing signal. Campbell and
Vicera’s (1999) utility calculations are also
160
based on the unconditional distribution,
80
so the optimal degree and benefit of market-timing might be less, conditional on
0
1940
1950
1960
1970
1980
1990
2000
the observed d/p ratio at the first date.
Notes: Risk aversion γ = 4.00 (black line) and γ = 20.00 (colored dashed line).
Second, there are good statistical
Source: Campbell and Vicera (1999).
reasons to think that the regressions
overstate the predictability of returns.
180

Federal Reserve Bank of Chicago

67

1) Figure 6 emphasizes one reason: The d/p ratio signal
has only crossed its mean four times in the 50 years
of postwar history. You have to be very patient to
profit from this trading rule. Also, we really have only
four postwar data points on the phenomenon. 2) The
dividend/price ratio was selected, in sample, among
hundreds of potential forecasting variables. It has
not worked well out of sample—the last two years of
high market returns with low d/p ratios have cut the
estimated predictability in half! 3) The model imposes
a linear specification, where the actual predictability
is undoubtedly better modeled by some unknown
nonlinear function. In particular, the linear specification implies negative expected stock returns at many
points in the sample, and one might not want to take
this specification seriously for portfolio construction.
4) The d/p ratio is strongly autocorrelated, and estimates of this autocorrelation are subject to econometric
problems. For this reason, long-horizon return properties inferred from a regression such as equations 1 and
2 are often more dramatic and apparently more precisely measured than direct long-horizon estimates.
The natural next step is to include this parameter
uncertainty in the portfolio problem, as I did above
for the case of independent returns. While this has
not been done yet in a model with Campbell and Vicera’s (1999) level of realism (and for good reasons—
Campbell and Vicera’s non-Bayesian solution is
already a technical tour de force), Barberis (1999)
makes such calculations in his simpler formulation. He
uses a utility of terminal wealth and no intermediate
trading, and he forces the allocation to stocks to be
less than 100 percent.
Figure 7 presents Barberis’s (1999) results.5 As
the figure shows, uncertainty about the parameters of

the regression of returns on d/p almost eliminates the
usefulness of market-timing.
Third, it is uncomfortable to note that fund returns
still cluster around the (buy-and-hold) market Sharpe
ratio (see figure 7 of “New facts in finance”). Here is
a mechanical strategy that supposedly earns average
returns twice those of the market with no increase in
risk. If the strategy is real and implementable, one
must argue that funds simply failed to follow it.
Market-timing, like value, does requires patience
and the willingness to stick with a portfolio that departs
from the indexing crowd. For example, a market-timer
following Campbell and Vicera’s rules in figures 5 and
6 would have missed most of the great runup in stocks
of the last few years. Fund managers who did that are
now unemployed. On the other hand, if an eventual
crash comes, the market timer will look wise.
Finally, one’s reluctance to take such strong
market-timing advice reflects the inescapable fact
that getting more return requires taking on more, or
different, kinds of risk. A market-timer must buy at
the bottom, when everyone else is in a panic; he
must sell at the top (now) when everyone else is feeling flush. His portfolio will have a greater mean for a
given level of variance over very long horizons, but
it will do well and badly at very different times from
everyone else’s portfolio. He will often underperform
a benchmark.
Hedging demands
Market-timing addresses whether you should
change your allocation to stocks over time as a return
signal rises or falls. Hedging demands address whether
your overall allocation to stocks, or to specific
portfolios, should be higher or lower as a result of

FIGURE 7

Allocation to stocks as a function of dividend/price ratio, with parameter uncertainty
A. Risk aversion coefficient 10

B. Risk aversion coefficient 20

allocation to stocks, percent
100

allocation to stocks, percent
100

No uncertainty

80

80

No uncertainty
60

60

Parameter uncertainty

40

40
20

20
0

Parameter uncertainty

0
2.06

3.75
d/p, percent

5.43

2.06

3.75
d/p, percent

5.43

Notes: The colored line ignores parameter uncertainty, as in Campbell and Vicera (1999).
The black line includes parameter uncertainty, as in Barberis (1999). Data sample is in months (523).

68

Economic Perspectives

return predictability, in order to protect you against
reinvestment risk.
A long-term bond is the simplest example. Suppose
you want to minimize the risk of your portfolio ten
years out. If you invest in apparently safe short-term
risk-free assets like Treasury bills or a money-market
fund, your ten-year return is in fact quite risky, since
interest rates can fluctuate. You should hold a tenyear (real, discount) bond. Its price will fluctuate a lot
as interest rates go up and down, but its value in ten
years never changes.
Another way of looking at this situation is that,
if interest rates decline, the price of the ten-year bond
will skyrocket; it will skyrocket just enough so that,
reinvested at the new lower rates, it provides the same
ten-year return as it would have if interest rates had
not changed. Changes in the ten-year bond value
hedge the reinvestment risk of short-term bonds. If
lots of investors want to secure the ten-year value
of their portfolios, this will raise demand for ten-year
bonds and lower their prices.
In general, the size and sign of a hedging demand
depend on risk aversion and horizon and, thus, will
be different for different investors. If the investor is
quite risk averse—infinitely so in my bond example—
he wants to buy assets whose prices go up when expected returns decline. But an investor who is not so
risk averse might want to buy assets whose prices
go up when expected returns rise. If the investor is
sitting around waiting for a good time to invest, and
is willing to pounce on good (high expected return)
investments, he would prefer to have a lot of money
to invest when the good opportunity comes around.
It turns out that the dividing line in the standard
(CRRA) model is logarithmic utility or a risk aversion
coefficient of 1—investors more risk averse than this
want assets whose prices go up when expected returns decline, and vice versa. Most investors are undoubtedly more risk averse than this, but not
necessarily all investors. Horizon matters as well. A
short horizon investor cares nothing about reinvestment risk and, therefore, has zero hedging demand.
In addition, the relationship between price and
expected returns is not so simple for stocks as for
bonds and must be estimated statistically. The predictability evidence reviewed above suggests that
high stock returns presage lower subsequent returns.
High returns drive up price/dividend, price/earnings,
and market/book ratios, all of which have been strong
signals of lower subsequent returns. Therefore, stocks
are a good hedge against their own reinvestment
risk—they act like the long-term assets that they are.
This consideration raises the attractiveness of stocks

Federal Reserve Bank of Chicago

for typical (risk aversion greater than 1) investors.
Precisely, if the two-fund analysis of figure 1 suggests
a certain split between stocks and short-term bonds
for a given level of risk aversion and investment horizon, then return predictability, a long horizon, and
typical risk aversion greater than 1 will result in a
higher fraction devoted to stocks. Again, exactly how
much more one should put into stocks in view of this
consideration is a tough question.
(In this case, the hedging demand reduces to
much the same logic as the horizon effects described
above. The market portfolio is a good hedge against
its own reinvestment risk, and so its long horizon
variance is less than its short horizon variance would
suggest. More generally, hedging demands can tilt a
portfolio toward stocks whose returns better predict
and, hence, better hedge the expected return on the
market index, but this long-studied possibility from
Merton [1971a, 1971b] has not yet been implemented
in practice.)
Campbell and Vicera’s (1999) calculations address
this hedging demand as well as market-timing demand,
and figure 5 also illustrates the strength of the hedging demand for stocks. Campbell and Vicera’s investors want to hold almost 30 percent of their wealth in
stocks even if the expected return of stocks is no
greater than that of bonds. Absent the hedging motive,
of course, the optimal allocation to stocks would be
zero with no expected return premium. Almost a 2 percent negative stock return premium is necessary to
dissuade Campbell and Vicera’s investors from holding
stocks. At the average (roughly 6 percent) expected return, of the roughly 130 percent of wealth that the risk
aversion 4 investors want to allocate to stocks, nearly
half is due to hedging demand. Thus, hedging demands
can importantly change the allocation to stocks.
However, hedging demand works in opposition
to the usual effects of risk aversion. Usually, less risk
averse people want to hold more stocks. However,
less risk averse people have lower or even negative
hedging demands, as explained above. It is possible
that hedging demand exactly offsets risk aversion;
everybody holds the same mean allocation to stocks.
This turns out not to be the case for Campbell and
Vicera’s numerical calibration; less risk averse people
still allocate more to stocks on average, but the effect
depends on the precise specification.
Choosing a risk-free rate
Figure 1 describes a portfolio composed of the
market portfolio and the risk-free rate. But the riskfree rate is not as simple as it once was either. For a
consumer or an institution6 with a one-year horizon,

69

one-year bonds are risk-free, while for one with a tenyear horizon, a ten year zero-coupon bond is risk-free.
For a typical consumer, whose objective is lifetime consumption, an interest-only strip (or real level annuity) is
in fact the risk-free rate, since it provides a riskless
coupon that can be consumed at each date. Campbell
and Vicera (1998) emphasize this point. Thus, the
appropriate bond portfolio to mix with risky stocks in
the logic of figure 1 is no longer so simple as a shortterm money market fund.
Of course, these comments refer to real or indexed bonds, which are only starting to become easily
available. When only nominal bonds are available,
the closest approximation to a risk-free investment
depends additionally on how much interest rate variability is due to real rates versus nominal rates. In the
extreme case, if real interest rates are constant and
nominal interest rates vary with inflation, then rolling
over short-term nominal bonds carries less long-term
real risk than holding long-term nominal bonds. In the
past, inflation was much more variable than real interest
rates in the U.S., so the fact that portfolio advice paid
little attention to the appropriate risk-free rate may
have made sense. We seem to be entering a period
in which inflation is quite stable, so real interest rate
fluctuations may dominate interest rate movements.
In this case, longer term nominal bonds become more
risk-free for long-term investors, and inflation-indexed
bonds open up the issue in any case. Once again,
new facts are opening up new challenges and opportunities for portfolio formation.
Notes of caution
The new portfolio theory can justify all sorts of
interesting new investment approaches. However,
there are several important qualifications that should
temper one’s enthusiasm and that shade portfolio advice back to the traditional view captured in figure 1.
The average investor holds the market
The portfolio theory that I have surveyed so far
asks, given multiple factors or time-varying investment
opportunities, How should an investor who does not
care about these extra risks profit from them? This
may result from intellectual habit, as the past great
successes of portfolio theory addressed such investors, or it may come from experience in the money
management industry, where distressingly few investors
ask about additional sources of risk that multifactor
models and predictable returns suggest should be
a major concern.
Bear in mind, however, that the average investor
must hold the market portfolio. Thus, multiple factors
and return predictability cannot have any portfolio

70

implications for the average investor. In addition, for
every investor who should follow a value strategy or
time the market for the extra returns offered by those
extra risks, there must be an investor who should follow the exact opposite advice. He should follow a
growth strategy or sell stocks at the bottom and buy
at the top, because he is unusually exposed to or
averse to the risks of the value or market-timing strategies in his business or job. He knows that he pays a
premium for not holding those risks, but he rationally
chooses this course just as we all choose to pay a
premium for home insurance.
Again, dividend/price, price/earnings, and book/
market ratios forecast returns, if they do, because the
average investor is unwilling to follow the value and
market-timing strategies. If everyone tries to time the
market or buy more value stocks, the premiums from
these strategies will disappear and the CAPM, random
walk view of the market will reemerge. Market-timing
can only work if it involves buying stocks when
nobody else wants them and selling them when everybody else wants them. Value and small-cap anomalies
can only work if the average investor is leery about
buying financially distressed and illiquid stocks. Portfolio advice to follow these strategies must fall on
deaf ears for the average investor, and a large class
of investors must want to head in exactly the other
direction. If not, the premiums from these strategies
will not persist.
One can see a social function in all this: The stock
market acts as a big insurance market. By changing
weights in, say, recession-sensitive stocks, people
whose incomes are particularly hurt by recessions
can purchase insurance against that loss from people
whose incomes are not hurt by recessions. They pay
a premium to do so, which is why investors are willing
to take on the recession-related risk.
The quantitative portfolio advice is all aimed at
the providers of insurance, which may make sense if
the providers are large wealthy investors or institutions.
But for each provider of insurance, there must be a
purchaser, and his portfolio must take on the opposite characteristics.
Are the effects real or behavioral, and will they last?
So far, I have emphasized the view that the average
returns from multifactor or market-timing strategies are
earned as compensation for holding real, aggregate
risks that the average investor is anxious not to hold.
This view is still debated. Roughly half of the academic studies that document such strategies interpret
them this way, while the other half interpret them as
evidence that investors are systematically irrational.
This half argues that a new “behavioral finance”

Economic Perspectives

should eliminate the assumption of rational consumers
and investors that has been at the core of all economics
since Adam Smith, in order to explain these asset
pricing anomalies.
For example, I have followed Fama and French’s
(1993, 1996) interpretation that the value effect exposes
the investor to systematic risks associated with economywide financial distress. However, Lakonishok,
Shleifer, and Vishny (1994) interpret the same facts as
evidence for irrationality: Investors flock to popular
stocks and away from unpopular stocks. The prices
of the unpopular stocks are depressed, and their average returns are higher as the fad slowly fades. Fama
and French point out that the behavioral view cannot
easily account for the comovement of value stocks;
the behavioral camp points out that the fundamental
risk factor is still not determined.
Similarly, the predictability of stock returns over
time is interpreted as waves of irrational exuberance
and pessimism as often as it is interpreted as timevarying, business cycle related risk or risk aversion.
Those who advocate an economic interpretation
point to the association with business cycles (Fama
and French, 1989) and to some success for explicit
models of this association (Campbell and Cochrane,
1999); those who favor the irrational investors view
point out that the rational models are as yet imperfect.
While this academic debate is entertaining, how
does it affect a practical investor who is making a
portfolio decision? At a basic level, it does not. If you
are not exposed to the risk a certain investment represents, it does not matter why other investors shy away
from holding it.
Analogously, to decide what to buy at the grocery
store, you only have to know how you feel about
various foods and what their prices are. You do not
have to understand the economic determinants of
food prices: You do not need to know whether a sale on
tomatoes represents a “real” factor like good weather in
tomato growing areas, or whether it represents an
“irrational” fear or fad.
Will they last?
Investments do not come with average returns
as clearly marked as grocery prices, however. Investors
have to figure out whether an investment opportunity that did well in the past will continue to do well.
This is one reason that it is important to understand
whether average returns come from real or irrational
aversion to risk.
If it is real, it is most likely to persist. If a high
average return comes from exposure to risk, well understood and widely shared, that means all investors
understand the opportunity but shrink from it. Even if

Federal Reserve Bank of Chicago

the opportunity is widely publicized, investors will
not change their portfolio decisions, and the relatively
high average return will remain.
On the other hand, if it is truly irrational, or a
market inefficiency, it is least likely to persist. If a high
average return strategy involves no extra exposure to
real risks and is easy to implement (it does not incur
large transaction costs), that means that the average
investor will immediately want to invest when he hears
of the opportunity. News travels quickly, investors
react quickly, and such opportunities vanish quickly.
Recent work in behavioral finance tries to document a way that irrational phenomena can persist
in the face of the above logic. If an asset-pricing
anomaly corresponds to a fundamental, documented,
deeply formed aspect of human psychology, then the
average investor may not pounce on the strategy the
minute he hears of it, and the phenomenon may last
(DeBondt and Thaler, 1985, and Daniel, Hirshleifer,
and Subrahmanyam, 1998). For example, many people
systematically overestimate the probability that airplanes crash, and make wrong decisions resulting
from this belief, such as choosing to drive instead.
No amount of statistics changes this view. Most such
people readily admit that a fear of flying is “irrational”
but persist in it anyway. If an asset-pricing anomaly
results from such a deep-seated misperception of
risk, then it could in fact persist.
A final possibility is that the average return premiums are the result of narrowly held risks. This view
is (so far) the least stressed in academic analysis. In
my opinion, it may end up being the most important.
It leads to a view that the premiums will be moderately persistent. Catastrophe-insurance enhanced bonds
provide a good example of this effect. These bonds
pay well in normal times, but either part of the principal
or interest is pledged against a tranche of a property
reinsurance contract. Thus, the bonds promise an
average return of 10 percent to 20 percent (depending
on one’s view of the chance of hurricanes). However,
the risk of hurricane damage is uncorrelated with anything else, and hence it is perfectly diversifiable.
Therefore, catastrophe bonds are an attractive opportunity. Before the introduction of catastrophe bonds,
there was no easy way for the average investor or
fund to participate in property reinsurance. As more
and more investors and funds hold these securities,
the prices will rise and average returns will fall. Once
the risks are widely shared, every investor (at least
those not located in hurricane-prone areas) will hold
a little bit of the risk and the high average returns will
have vanished.

71

The essential ingredients for this story are that
the risk is narrowly shared; the high average returns
only disappear when the risk is widely shared (it cannot be arbitraged away by a few savvy investors);
and an institutional change (the introduction, packaging, and marketing of catastrophe-linked bonds)
is required before it all can happen.
This story gives a plausible interpretation of many
of the anomalies I document above. Small-cap stocks
were found in about 1979 to provide higher returns
than was justified by their market (β) risk. Yet at that
time, most funds did not invest in such stocks, and individual investors would have had a hard time forming a portfolio of small-cap stocks without losing all
the benefits in the very illiquid markets for these
stocks. The risks were narrowly held. After the popularization of the small-cap effect, many small-cap funds
were started, and it is now easy for investors to hold
such stocks. As the risk has been more widely shared,
the average returns seem to have fallen.
The value effect may be amenable to a similar interpretation. Before about 1990, as I noted earlier, few
funds actually followed the high-return strategy of
buying really distressed stocks or shorting the popular growth stocks. It would be a difficult strategy for
an individual investor to follow, requiring courage and
frequent trading of small illiquid stocks. Now that the
effect is clear, value funds have emerged that really
do follow the strategy, and the average investor can
easily include such an offering in his portfolio. The
risk is becoming widely shared, and its average return
seems to be falling.
Even average returns on the stock market as a
whole (the equity premium) may follow the same story,
since participation has increased a great deal through
the invention of index funds, low-commission brokerages, and tax-sheltered retirement plans.
This story does not mean that the average returns
corresponding to such risks will vanish. They will
decline, however, until the markets have established
an equilibrium, in which every investor has bought as
much of the risk as he likes. In this story, one would
expect a large return as investors discover each strategy and bid prices up to their equilibrium levels. This
may account for some of the success of small and
value stocks observed in the literature, as well as
some of the stunning success of the overall market
in recent years.
Inconsistent advice
Unfortunately, the arguments that a factor will
persist are all inconsistent with aggressive portfolio
advice. If the premium is real, an equilibrium reward

72

for holding risk, then the average investor knows
about it but does not invest because the extra risk exactly counteracts the extra average return. If more
than a minuscule fraction of investors are not already
at their best allocations, then the market has not
reached equilibrium and the premiums will change.
If the risk is irrational, then by the time you and
I know about it, it’s gone. An expected return corresponding to an irrational risk premium has the strongest portfolio implications—everyone should do
it—but the shortest lifetime. Thus, this view is also
inconsistent with the widespread usefulness of portfolio advice.
If the average return comes from a behavioral
aversion to risk, it is just as inconsistent with widespread portfolio advice as if were real. We can not all
be less behavioral than average, just as we can not
all be less exposed to a risk than average. The whole
argument for behavioral persistence is that the average
investor would not change his portfolio, just as the
average traveler would not quickly adjust his traveling
behavior to fear the cab ride out to the airport more
than the flight. Thus, the advice must be useless to
the vast majority of investors. If most people, on seeing the strategy, can be persuaded to act differently
and buy, then it is an irrational risk and will disappear.
If it is real or behavioral and will persist, then this
necessarily means that very few people will follow
the portfolio advice.
If the average return comes from a narrowly held
risk, one has to ask what institutional barriers keep
investors from sharing this risk more widely. Simple
portfolio advice may help a bit—most investors still
do not appreciate the risk/return advantages of stocks
overall, small-caps, value stocks, market-timing, and
aggressive liquidity trades. But by and large, a risk
like this needs packaging, securitizing, and marketing
more than advice. Then there will be a period of high
average returns to the early investors, followed by
lower returns, but still commoditization of the product with fees for the intermediaries.
Economic logic
The issue of why the risk gives an average return
premium is also important to decide whether the opportunity is really there. It is not that easy to establish the
average returns of stocks and dynamic portfolio
strategies. There are many statistical anomalies that
vanish quickly out of sample. Figuring out why a
strategy carries a high average return is one of the
best ways to ensure that the high average return is
really there in the first place. Anything that is going
to work has a real economic function. A story such

Economic Perspectives

as “I don’t care much about recessions; the average
investor does; hence it makes good sense for me to
buy extra amounts of recession sensitive stocks since
I am selling insurance to the others at a premium”
makes a strategy much more plausible than the output
of some statistical black box.
Conclusion
Practical application of portfolio theory
How does an investor who is trying patiently to
sort through the bewildering variety of investment
opportunities use all the new portfolio theory? It’s
best to follow a step by step procedure, starting with
a little introspection.
1. What is your overall risk tolerance? As before,
you must first figure out to what extent you are willing to trade off volatility for extra average returns, to
determine an appropriate overall allocation to risky
versus risk-free assets. While this question is hard to
answer in the abstract, you only need to know whether
you are more or less risk tolerant than the average investor. (Honestly, now—everyone wants to say they
are a risk taker.) The overall market is about 60 percent
stocks and 40 percent bonds, so average levels of risk
aversion, whatever they are, wind up at this value.
2. What is your horizon? This question is first
of all important for figuring out what is the relevant
risk-free asset. Longer term investors can hold longer
term bonds despite their poor one-year performance,
especially in a low-inflation environment. Second, we
have seen that stocks are somewhat safer for “longrun” investors.
3. What are your risks? Would you be willing to
trade some average return in order to make sure that
your portfolio does well in particular circumstances?
For example, an investor who owns a small company
would not want his investment portfolio to do poorly
at the same time that his industry suffers a downturn,
that there is a recession, or a credit crunch, or that the
industries he sells to suffer a downturn. Thus, it makes
good sense for him to avoid stocks in the same industry
or downstream industries, or stocks that are particularly sensitive to recessions or credit crunches, or
even to short them if possible. This strategy would
make sense even if these stocks give high average
returns, like the value portfolios. Similarly, he should
avoid high yield bonds that will all do badly in a credit
crunch. If the company will do poorly in response to
increases in interest rates, oil prices or similar events,
and if the company does not hedge these risks, then
the investor should take positions in interest rate sensitive or oil-price sensitive securities to offset those

Federal Reserve Bank of Chicago

risks as well. We’re just extending the principles behind
fire and casualty insurance to investment portfolios.
This logic extends beyond the kind of factors
(size, book to market, and so on) that have attracted
academic attention. It applies to any identifiable movement in asset portfolios. For example, industry portfolios are not badly explained by the CAPM, as they all
seem to have about the same average return. Therefore, they do not show up in multifactor models. However, shorting your industry portfolio protects you
against the risks of your occupation. In fact, factors
that do not carry unusual risk premiums are even better opportunities than the priced factors that attract
attention, since you buy insurance at zero premium.
This was always true, even in the CAPM, unpredictable return view. I think that the experience with multifactor models just increases our awareness of how
important this issue is.
4. What are not your risks? Next, figure out what
risks you do not face, but that give rise to an average
return premium in the market because most other investors do face these risks. For example, an investor who
has no other source of income beyond his investment
portfolio does not particularly care about recessions.
Therefore, he should buy extra amounts of recessionsensitive stocks, value stocks, high yield bonds, etc.,
if these strategies carry a credible high average return.
This action works just like selling insurance, in return
for a premium. This is the type of investor for whom
all the portfolio advice is well worked out.
In my opinion, too many investors think they are
in this class. The extra factors and time-varying returns
would not be there (and will quickly disappear in the
future) if lots of people were willing and able to take
them. The presence of multiple factors wakes us up to
the possibility that we, like the average investor, may
be exposed to extra risks, possibly without realizing it.
5. Apply the logic of the multifactor-efficient
frontier. Figure 2 now summarizes the basic advice.
After thinking through which risk factors are good to
hold, and which ones you are already too exposed to;
after thinking through what extra premiums you are
likely to get for taking on extra risks, you can come to
a sensible decision about which risks to take and
which to hedge.
6. Do not forget, the average investor holds the
market. If you’re pretty much average, all this thought
will lead you right back to holding the market index.
To rationalize anything but the market portfolio, you
have to be different from the average investor in some
identifiable way. The average investor sees some risk
in value stocks that counteracts their attractive average
returns. Maybe you should too! Right now the average

73

investor is feeling very wealthy and risk-tolerant,
therefore stock prices have risen to unprecedented
levels and expected stock returns look very low. It’s
tempting to sell, but perhaps you’re feeling pretty
wealthy and risk-tolerant as well.
7. Of course, avoid taxes and snake oil. The marketing of many securities and funds is not particularly clear on the nature of the risks. There is no reliable
extra return without risk. The economic reasoning in
this article should be useful to figure out exactly what
type of risk a specific fund or strategy is exposed to,
and then whether it is appropriate for you. The average actively managed fund still underperforms its style
benchmark, and past performance has almost no information about future performance.
The most important piece in traditional portfolio
advice applies as much as ever: Avoid taxes and
transaction costs. The losses from churning a portfolio and paying needless short-term capital gain, inheritance, and other taxes are larger than any of the
multifactor and predictability effects I have reviewed.
Tax issues are much less fun but more important to
the bottom line.

A big insurance market
It is tempting to think of asset markets like a
racetrack, but they are in reality a big insurance market.
Value funds seem to provide extra returns to their investors by buying distressed stocks on the edge of
bankruptcy. Long-Term Capital Management was, it
seems, providing catastrophe insurance by intermediating liquid assets that investors like into illiquid
assets that were vulnerable to a liquidity crunch.
Who better to provide catastrophe insurance than rich
investors with no other labor income or other risk exposure? Once again, we are reminded that Adam
Smith’s invisible hand guides self-interested decisions
to socially useful ends, often in mysterious ways.
However, asset markets could be better insurance
markets. Both new and old portfolio advice implies
that the typical investor should hold a stock position
that is short his company, industry, or other easily
hedgeable kinds of risk. Many managers and some
senior employees must hold long positions in their
own companies, for obvious incentive reasons. But
there is no reason that this applies to union pension
funds, for example. A little marketing and help from
policy should make funds that hedge industry-specific risks to labor income much more attractive vehicles.

APPENDIX

The mean of the portfolio return is

Multifactor portfolio mathematics
This section summarizes algebra in Fama (1996). The
big picture is that we still get a hyperbolic region
since betas are linear functions of portfolio weights
just like means.
The problem is, minimize the variance of a portfolio given a value for the portfolio mean and its beta
on some factor. Let

w "#  R
w #
R
w=
;R=
##
!w $ ! R

1

1

2

2

M

M

M

N

"# 1" b
## ; 1 = 1## ; b = b
## 1## b
$ !$ !

N

1, F
2, F

M
N ,F

"#
## .
##
$

E ( R p ) = E ( wŠR) = wŠE ( R) = w ŠE .
The last equality just simplifies notation. The beta of
the portfolio on the extra factor is

b p = w Šb.
The variance of the portfolio return is

var( R p ) = w ŠVw,
where V is the variance-covariance matrix of returns.
The problem is then
min
w

Then the portfolio return is

R = w ŠR ;
p

the condition that the weights add up to 1 is
1 = 1Š w.

74

1
w Š Vw s. t. w ŠE = m ; w Š1 = 1; w Šb = b p .
2

The Lagrangian is

3

8

1
w Š Vw - l 0 ( w Š E - m ) - l 1 ( w Š1 - 1) - l 2 w Š b - b p .
2

The first order conditions with respect to w give

1

6

w = V -1 El 0 + 1l 1 + bl 2 = V -1 Al

Economic Perspectives

where

where

A= E 1 b

s * = max E ( R - R f ) / s ( R - R f )

l = l 0 l1 l 2 Š

denotes the unconditional Sharpe ratio, and

st = max E t ( R - R f ) / s t ( R - R f )

d = m 1 b p Š.
Plugging this value of w into the constraint
equations

A′ w = δ ,

denotes the conditional Sharpe ratio.
The technique exploits ideas from Gallant, Hansen,
and Tauchen (1990). I exploit Hansen and Jagannathan’s
(1991) theorem that for any excess return Z and discount factor m such that 0 = E(mZ), we have

AV Aλ = δ
−1

we get

E (Z )
s( Z )

λ = ( A′ V A) δ
−1

−1

w = V −1 A( A′V −1 A)−1δ .

var ( R p ) = w ŠVw = d Š ( AŠV -1 A) -1 d.

max

Or, writing out the sum of the matrix notation,

Var( R ) = m 1 b

p

-1

( A ŠV A)

-1

E (m)

,

 m "#
1 #.
! b #$

The variance is a quadratic function of the mean
return and of the desired beta on additional factors.
That’s why we draw cup-shaped frontiers. As with
the mean-variance case, the multifactor efficient frontier is a revolution of a hyperbola. If V is a second
moment matrix, to handle a risk-free rate,

var ( R p ) = d Š( AŠ V -1 A) -1 A ŠV -1 SV -1 A( A ŠV -1 A) -1 d,
where Σ now denotes the return variance-covariance
matrix.
Finding the benefits of a market timing
strategy without computing the strategy
I show first that the squared maximum unconditional
Sharpe ratio is the average of the squared conditional
Sharpe ratios when the riskfree rate is constant,

 E  = s(m*) ,
 s  E (m*)

where m* solves

m* = arg min s (m) s.t.
{m}

Et (mt +1Zt +1 ) = 0; Et (mt +1 ) = 1/ Rtf .

p

s *2 = E st2 ,

s(m)

and equality is attained for some choice of m. Thus,
the maximal unconditional Sharpe ratio is

The portfolio variance is then

p

‹

Gallant, Hansen, and Tauchen show how to
solve this problem in quite general situations. They
phrase their result as a “lower bound on discount
factor volatility” but given E(Z)/σ(Z) ≤ σ(m)/E(m), one
can read the maximum slope of the unconditional meanvariance frontier (Sharpe ratio) available from markettiming portfolios. To keep the calculation transparent
and simple, I specialize to the case of a constant and
observed real risk-free rate Rf = 1/Et(m). Then, the
unconditional squared Sharpe ratio is the average
of the conditional squared Sharpe ratios,

s 2 ( m)
E ( m)

2

=

s[ Et (m) 2 ] + E[ s 2t ( m)]
E ( m)

2

=E

 s (m)  .
 E (m) 
2
t

2

t

Next, I show that when we forecast stock returns
with a regression such as equation 1, and interest
rates and the conditional variance of the error term
are constant, then the best unconditional Sharpe
ratio is related to the regression R2 by
s* =

s02 + R 2
1 - R2

,

where s0 = E ( R - R f ) / s( R - R f ) denotes the unconditional buy-and-hold Sharpe ratio.

Federal Reserve Bank of Chicago

75

If the conditional Sharpe ratio is generated by
a single asset (the market), and a linear model with
constant error variance,
Z t +1 = EZ + b( xt - Ex ) + e t +1 ,

then,

 s (m)  =  E (Z)  =  EZ +b( x - Ex)  ,
 E (m)   s (Z)   s 
2

and

2

2

t

t

t

t

t

e

 EZ + b( x − Ex)  2 
 σ t2 (m) 
t
E
= E 
 
σ
 Et (m)2 

ε


=

( E ( Z )) 2 + b 2σ 2 ( x)
σ ε2

=

( E ( Z ))2
b 2σ 2 ( x)
+
2
2
(1 − R )σ ( Z ) (1 − R 2 )σ 2 ( Z )

=

1
( EZ ) 2
R2
+
2
2
(1 − R ) σ ( Z ) (1 − R 2 )

=

1
(1 − R 2 )

 E ( Z )  2

+ R2  .


 σ ( Z ) 


The last line demonstrates s* = s02 + R 2 / 1- R2 .
To obtain the annualized Sharpe ratios reported in
table 1, I divide by the square root of horizon, since
mean returns roughly scale with horizon and standard deviations roughly scale with the square root
of horizon.

76

Economic Perspectives

NOTES
To be precise, these statements refer to the conditional serial
correlation of returns. It is possible for the conditional serial
correlations to be non-zero, resulting in conditional variances
that increase with horizon faster or slower than linearly, while
the unconditional serial correlation of returns is zero. Conditional distributions drive portfolio decisions.

portfolio. For example, in the simplest case of a one-period
investment problem, consumption equals terminal wealth.
Equation 3 then becomes

This effort falls in a broader inquiry in economics. Once we
recognize that people are unlikely to have much more data
and experience than economists, we have to think about economic models in which people learn about the world they live
in through time, rather than models in which people have so
much history that they have learned all there is to know about
the world. See Sargent (1993) for a review of learning in macroeconomics.

Brandt uses this condition to estimate the portfolio allocation
α. He extends the technique to multiperiod problems and problems in which the allocation decision depends on a forecasting
variable, that is, market-timing problems.

1

2

The standard first-order condition for optimal consumption
and portfolio choice is
3

3)

E[(ct +1 ) -g Z t +1 ] = 0,

where c denotes consumption, Z denotes an excess return, and
γ is a preference parameter. We usually take data on c and Z,
estimate γ, and then test whether the condition actually does
hold across assets. In a portfolio problem, however, we know
the preference parameter γ, but we want to estimate the

E [(aR f + (1 - a ) Rtm+1 ) - g Zt +1 ] = 0.

That is, Campbell and Vicera model investors’ objectives by
a utility function, max E Ít b t u(ct ) rather than a desire for
wealth at some particular date, max Eu(W T ).
4

I thank Nick Barberis for providing this figure. While it is not
in Barberis (1999), it can be constructed from results given in
that paper.
5

Of course, in theory, institutions, as such, should not have
preferences, as their stockholders or residual claimants can
unwind any portfolio decisions they make. This is the famous
Modigliani-Miller theorem. In practice, institutions often
make portfolio decisions as if they were individuals, and people
purveying portfolio advice will run into many such institutions.
6

REFERENCES

Barberis, Nicholas, 1999, “Investing for the long run
when returns are predictable,” Journal of Finance,
forthcoming.
Black, Fischer, and Robert Litterman, 1991, “Global
asset allocation with equities, bonds, and currencies,”
Goldman Sachs Fixed Income Research.
Brandt, Michael W., 1999, “Estimating portfolio and
consumption choice: A conditional Euler equations
approach,” Journal of Finance, October, forthcoming.
Brennan, Michael J., Eduardo S. Schwartz, and
Roland Lagnado, 1997, “Strategic asset allocation,”
Journal of Economic Dynamics and Control, Vol. 21,
No. 7, pp. 1377–1403.

, 1998, “Who should buy long term bonds?,”
Harvard University, manuscript.
Cochrane, John H., and Argia M. Sbordone, 1988,
“Multivariate estimates of the permanent components
in GNP and stock prices,” Journal of Economic
Dynamics and Control, Vol. 12, No. 2, pp. 255–296.
Cochrane, John H., 1997, “Where is the market going? Uncertain facts and novel theories,” Economic
Perspectives, Federal Reserve Bank of Chicago, Vol.
21, No. 6, November/December, pp. 3–37.
, 1994, “Permanent and transitory components of GNP and stock prices,” Quarterly Journal
of Economics, Vol. 109, February, pp. 241–266.

Campbell, John Y., and John H. Cochrane, 1999, “By
force of habit: A consumption-based explanation of
aggregate stock market behavior,” Journal of Political
Economy, Vol. 107, No. 2, April, pp. 205–251.

Daniel, Kent, David Hirshleifer, and Ananidhar
Subrahmanyam, 1998, “Investor psychology and
security market under- and overreactions,” Journal
of Finance, Vol. 53, No. 6, pp. 1839–1885.

Campbell, John Y., Andrew W. Lo, and A. Craig
MacKinlay, 1996, The Econometrics of Financial
Markets Princeton, NJ: Princeton University Press.

DeBondt, Werner F. M., and Richard H. Thaler, 1985,
“Does the stock market overreact?,” Journal of
Finance, Vol. 40, No. 3, pp. 793–808

Campbell, John Y. and Luis M. Vicera, 1999, “Consumption and portfolio decisions when expected
returns are time varying,” Quarterly Journal of
Economics, forthcoming.

Fama, Eugene F., 1996, “Multifactor portfolio efficiency and multifactor asset pricing,” Journal of Financial
and Quantitative Analysis, Vol. 31, No. 4, December,
pp. 441–465.

Federal Reserve Bank of Chicago

77

Fama, Eugene F., and Kenneth R. French, 1993,
“Common risk factors in the returns on stocks and
bonds,” Journal of Financial Economics, Vol. 33,
No. 1, February, pp. 3–56.
, 1989, “Business conditions and expected
returns on stocks and bonds,” Journal of Financial
Economics, Vol. 25, No. 1, pp. 23–49.
Gallant, A. Ronald, Lars Peter Hansen, and George
Tauchen, 1990, “Using conditional moments of asset
payoffs to infer the volatility of intertemporal marginal
rates of substitution,” Journal of Econometrics, Vol.
45, No. 1/2, pp. 141–179.
Kandel, Shmuel, and Robert F. Stambaugh, 1996, “On
the predictability of stock returns: An asset allocation perspective,” Journal of Finance, Vol. 51, No. 2,
June, pp. 385–424.
Kim, Tong Suk, and Edward Omberg, 1996, “Dynamic
nonmyopic portfolio behavior,” Review of Financial
Studies, Vol. 9, No. 1, Winter, pp. 141–161.

Markowitz, H., 1952, “Portfolio selection,” Journal
of Finance, Vol. 7, No. 1, pp. 77–91.
Merton, Robert C., 1973, “An intertemporal capital
asset pricing model,” Econometrica, Vol. 41, No. 5,
pp. 867–887.
, 1971, “Optimum consumption and portfolio rules in a continuous time model,” Journal of
Economic Theory, Vol. 3, No. 4, pp. 373–413.
, 1969, “Lifetime portfolio selection under
uncertainty: The continuous time case,” Review of
Economics and Statistics, Vol. 51, No. 3, August,
pp. 247–257.
Samuelson, Paul A., 1969, “Lifetime portfolio selection by dynamic stochastic programming,” Review
of Economics and Statistics, Vol. 51, No. 3, August,
pp. 239–246.
Sargent, Thomas J., 1993, Bounded Rationality in
Macroeconomics, Oxford: Oxford University Press.

Lakonishok, Josef, Andrei Shleifer, and Robert W.
Vishny, 1994, “Contrarian investment, extrapolation
and risk,” Journal of Finance, Vol. 49, No. 5, December, pp. 1541–1578.

78

Economic Perspectives