View original document

The full text on this page is automatically extracted from the file linked above and may contain errors and inconsistencies.

Flation
William Poole and Robert H. Rasche
lation”—not inflation, not deflation—is
lifted from the title of a book by Abba P.
Lerner.1 For the past 35 years in the
United States and, indeed, in most of the world,
policymakers and the public in general have been
focused on the issue of inflation—that is, the continual upward drift in prices of the overwhelming
fraction of goods and services produced in the
economy. Sometimes the drift was more of a gallop.
For most of this period, the upward trend was also
characteristic of the prices of assets such as land,
houses, and equities. Inflation, prevalent though
it has been in our recent economic experience,
has not been the norm for most of U.S. history. In
the early 1930s, exactly the opposite experience
occurred: deflation, or a continual downward drift
in the prices of goods, services, and assets.
Deflation has a frightening history. Simultaneously with the deflation of the early 1930s, the U.S.
unemployment rate soared to about 25 percent in
1933 at the depth of the Great Depression. Although
deflation ended in 1933, the damage to the economy
was so great that poor economic conditions persisted
until the United States became involved in World
War II in 1941. Moreover, the economic history of
the 1990s in Japan is characterized by deflation. The
Japanese economy has stagnated, and unemployment there has risen today to levels not seen in over
40 years. From these and other episodes around
the world, many people associate deflation with
“hard times.”
The purpose of this analysis is not to get into a
discussion of whether a little deflation is compatible
with prosperity, although within limits it may be.
The more important point is that, without question,
substantial deflation is inconsistent with prosperity.
Thus, deflation is every bit as serious an issue as
inflation; however, the U.S. economy today does
not run any significant risk of deflation.

“F

William Poole is the president and Robert H. Rasche is a senior vice
president and director of research at the Federal Reserve Bank of St.
Louis. This article is adapted from a speech of the same title presented
before the International Mass Retail Association Leadership Forum,
January 21, 2002. The authors appreciate comments provided by
colleagues at the Federal Reserve Bank of St. Louis. The views expressed
do not necessarily reflect official positions of the Federal Reserve
System.

© 2002, The Federal Reserve Bank of St. Louis.

Obviously, not everyone agrees with this judgment. Based on a few recent observations of monthto-month price changes, some commentators have
used the “D” word to express their concern about
the current state of the U.S. economy. The objective
in this paper is to explore this subject and, we hope,
make a contribution to public understanding of the
issue.
First, we will elaborate on what we believe is
the appropriate objective for Federal Reserve policy.
Second, we will explain the generally accepted definitions of inflation and deflation, and discuss the
fundamental sources of these phenomena. Third, we
will review aspects of price behavior in our economy
and discuss how data should be interpreted to determine the inflationary or deflationary state of the
economy. Finally, although the issue concerns the
behavior of the aggregate price level, we will examine some particular sectoral price changes to help
better understand the aggregate price level.

THE APPROPRIATE POLICY OBJECTIVE
FOR THE FEDERAL RESERVE
Our monetary policy framework is this. First
and foremost, the central bank must maintain a
commitment to price stability. An operational definition of price stability is an environment in which
the inflation rate, properly measured and averaged
over several years, is zero. All of our inflation data
are subject to measurement errors. Experts in such
measurements generally agree that current price
indexes, despite statisticians’ best efforts, still leave
inflation measures that have some upward bias.
Hence, in terms of the various inflation indexes, we
can say that price stability prevails when broad price
indexes exhibit small positive average values for
measured inflation and that year-to-year fluctuations
around that average are well contained.
If the price level comes unstuck, yielding inflation or deflation, all sorts of other problems will
arise. Nevertheless, within the confines of the goal
of price stability, the central bank has some flexibility to lean against fluctuations in output and employment. However, the central bank ought not to pursue
the goal of stabilizing economic activity so aggressively that it runs any substantial risk of compromising the goal of price stability.
Finally, in leaning against fluctuations in growth
and employment, the central bank ought not to
1

Lerner, Abba P. Flation. New York: Quadrangle Books, 1972.

N OV E M B E R / D E C E M B E R 2 0 0 2

1

Poole and Rasche

have goals for levels of the economy’s growth and
unemployment rates per se. Within a wide range,
no one knows what the economy’s equilibrium rate
of growth is or what rate of unemployment will
clear the labor market in the long run. The biggest
risks of a major monetary policy mistake occur if a
central bank attempts to target the levels of real
variables.
Achieving the objective of price stability, as
defined above, will yield a highly stable economy.
When the market has confidence in Fed policy, shortrun changes—that is, over a few months or even a
few quarters—in the rate of inflation or deflation will
tend to be self-reversing rather than self-reinforcing.

THE DEFINITION AND SOURCES OF
INFLATION AND DEFLATION
At the beginning of the great inflation of 1965-80,
there was a wide disparity of professional opinion
about the fundamental source of inflation or deflation in an economy. One proposition came to be
known as the “monetarist view.” This view held that
sustained inflation or deflation was always a monetary phenomenon; that is, that the only source of
long-run positive or negative trends in the general
level of prices in an economy is the creation of an
excess or insufficient supply of money balances
relative to the growth of the productive capacity of
that economy. Milton Friedman of the University of
Chicago was the most publicly visible proponent
of this proposition. The Federal Reserve Bank of St.
Louis, in particular the president of the Bank at that
time, Darryl Francis, and the Research staff were
vocal advocates of this proposition in the policy
arena during the late 1960s and early 1970s. A reading of the Memorandum of Discussion of the Federal
Open Market Committee (FOMC) for this period
makes clear that there were sharp debates over these
issues. The FOMC is the Fed’s main monetary policymaking body, and the public record of that period
shows that Darryl Francis was a vigorous advocate
of the monetarist view.
The proposition that the central bank is the
source of ongoing inflation or deflation was a distinct minority view 35 years ago. In the FOMC, Darryl
Francis was usually the only one expressing this
view. The development of economic theory and the
economic history of the past three decades have
produced a major change in both professional
thinking and public attitudes toward the sources of
inflation and deflation. Economists are now largely
2

N OV E M B E R / D E C E M B E R 2 0 0 2

REVIEW

in agreement that if the central bank does not achieve
the goal of price stability, no one else can. Many
central banks around the world, starting with the
Reserve Bank of New Zealand in 1990, have acknowledged this responsibility and have adopted explicit
numeric inflation targets.
This view also spread into public thinking about
inflation in the United States. Paul Volcker, former
chairman of the Board of Governors of the Federal
Reserve System, is widely credited for the disinflation
that occurred in the United States in the early 1980s.
Chairman Greenspan is applauded for the additional
progress in the 1990s that brought the U.S. inflation
rate to the lowest level in almost 40 years.
Today the Federal Reserve accepts its responsibility for the trend rate of inflation. However, a central
bank is not responsible for month-to-month wiggles
in the inflation statistics. Nor should a central bank
attempt to react to short-run variations, since the
sources of such noise are beyond its control and
likely to average out over a period of a few months
or at most a couple of years. One obvious reason
for not reacting to short-run developments is that
an unknown part of these changes in the reported
inflation rate is purely measurement error, or statistical noise.
Professional opinion has also changed about the
source of deflation in the 1930s. It is now widely
acknowledged that, at a minimum, the intensity of
the Great Depression was magnified by the failure
of the Federal Reserve to provide sufficient liquidity
to the economy in the face of widespread bank failures. The Federal Reserve in turn has learned from
that experience. When the U.S. economy has been
threatened by liquidity crises in recent years—such
as the stock market crash of 1987, the Asian crises
and Russian default of 1998, and the terrorist attack
of September 11, 2001—the Fed has moved rapidly
to inject large amounts of liquidity into the economy. Liquidity crises have been averted, inflation
has remained low and stable, and deflation has not
occurred.
Experience elsewhere has not been as benign.
Over the period from 1981 through 1990, the
Japanese economy grew at an annual rate of 3.7
percent and the inflation rate (measured by the
gross domestic product [GDP] price index) averaged
1.5 percent per year. The situation in Japan in the
1990s has been remarkably different. The Japanese
economy has struggled in and out of recession, and
real growth from 1991 to 2000 averaged only 1.1
percent. Over the same period, very low inflation

FEDERAL RESERVE BANK OF ST. LOUIS

has turned into deflation. From 1991 to 1996, the
Japanese consumption deflator rose at an average
annual rate of only 0.5 percent; for 1996 to 2000,
the rate was –0.2 percent. Asset prices fell dramatically. The decline of the Nikkei equity price index
from a value of close to 40,000 in late 1989 to its
recent level of less than 10,000 is common knowledge. What is not as well known outside Japan is
that land and real estate prices over the past decade
have experienced equally dramatic declines as those
seen in equity markets. In April 1993 an index of
housing prices in Japan stood at 42.35 million yen.
By April 2001 it had fallen to 36.52 million yen, an
annual average rate of decline of 1.7 percent.2 The
index of residential land prices reached a peak in
March 1991 of 109.7 and fell to 81.7 by September
2001, an annual average rate of decline of 2.4 percent. The decline in commercial land prices was
even larger. From a peak of 111.7 in September 1991,
the index of these prices fell to 49.1 in September
2001, an annual average rate of decline of 5.6 percent.3 In terms of the impact on Japan’s output and
employment, the large deflation of asset prices was
probably more important than the gentle deflation
of goods prices.
What is responsible for the incredible difference in the performance of the Japanese economy
between the 1980s and 1990s? Japan’s money stock
(using Japan’s own preferred measure, M2+CDs)
grew at an average annual rate of 7.9 percent from
1981 through 1990, but only at 2.3 percent per year
over the decade from 1991 through 2000. A conclusion consistent with research on this issue is
that the ongoing stagnation and deflation that the
Japanese economy has experienced in the past
decade is likely related to an insufficient supply of
liquidity by the Bank of Japan. Slow money growth
is not the whole story, but is certainly a significant
part of it.

RECENT PRICE BEHAVIOR IN THE
U.S. ECONOMY
Public discussion of inflation in the United
States generally is focused on the consumer price
index (CPI) published monthly by the Bureau of
Labor Statistics. The monthly change in the overall
CPI is the so-called “headline” inflation number.
The CPI is very visible; it has been widely reported
for years and is used to construct cost-of-living
adjustments in union wage contracts and Social
Security benefits.

Poole and Rasche

Sometimes reference is also made to a “core”
inflation rate, usually measured by the CPI excluding
prices of food and energy products. The rationale
for excluding food and energy prices is that they can
be quite volatile, and hence longer-term inflation
trends can be obscured when they are included.
Starting in 2000, the FOMC chose to focus on a
different measure of inflation: changes in the price
index for personal consumption expenditures in the
national income accounts. This measure of inflation,
which for convenience we will call the “consumption
price index,” is reported monthly by the Bureau of
Economic Analysis of the Department of Commerce.
Although this index receives less public attention
than the CPI, it is preferred by the FOMC because
the methodology used in its construction reduces
the measurement bias relative to that in the CPI;
also, the coverage of goods and services in this index
is believed to better represent consumption patterns.
For example, prices of medical services are included
in the CPI only to the extent that such services are
paid directly by consumers. Prices of all medical
services are included in the consumption price
index whether those services are paid for directly
by consumers or are paid for on behalf of consumers
by third parties such as insurance companies.
In recent years, inflation as measured by the
consumption price index has been lower than that
measured by the CPI.4 Although the following discussion will refer primarily to the consumption price
index, no important issues depend on whether the
focus is on that index or the CPI.
What should we expect to observe in an economy where price stability prevails? If it were possible to measure the average level of prices with little
or no bias in such an economy, then over a period
of time an average measured inflation rate very close
to zero should be observed. From month-to-month
or quarter-to-quarter, positive or negative changes
of the inflation index will occur, but over time these
would average out to about zero.
2

The Housing Loan Progress Association. “Price Survey of the Housing
Market.” <http://jin.jcic.or.jp/stat/stats/>.

3

National Land Agency.

4

In August 2002 the Bureau of Labor Statistics introduced a new measure
of consumer prices—the chained consumer price index for all urban
consumers (C-CPI-U). Monthly data are available from December 1999.
The objective of the new index is to reduce the substitution bias that
is present in the CPI-U. Between December 1999 and December 2000
(the only period for which final estimates of the C-CPI-U are available),
the inflation rate measured by the C-CPI-U differs from that measured
by the consumption price index by only 0.1 percent.

N OV E M B E R / D E C E M B E R 2 0 0 2

3

Poole and Rasche

What about prices of individual goods and services under such conditions? There would likely be
a dispersion of changes in the prices of individual
goods and services around zero. In fact, prices of
some goods and services could be continually falling,
while prices of other goods and services could be
continually rising. It is perfectly normal to experience
divergent trends of individual prices under conditions of overall price stability. Thus, trends in the
prices of individual goods or services cannot be used
to judge whether an economy is experiencing inflation or deflation.
An important influence on inflation data in the
United States over the past three years has been the
behavior of energy prices on world markets. In 1998,
energy prices collapsed as world demand dropped
dramatically in response to the crises in Asian economies. Petroleum inventories rose unexpectedly
and major producers, including OPEC nations, cut
production to stabilize prices and adjust inventories.
In 1999 and 2000, energy prices rose sharply as
economic activity boomed in the United States and
other major industrialized economies at a time when
world inventories of oil were particularly low. Leading up to 2002, as the U.S. economy sank into recession and economic growth slowed in Europe, energy
demand growth slowed and energy prices on world
markets fell again.
The average inflation rate over the four years
1994 through 1997 was 2.7 percent per year as
measured by the consumption price index. The
average inflation rate over the four years from 1998
through 2001 was 1.7 percent per year. The core
inflation component of the consumption price index
has fallen from 2.1 percent in the earlier period to
1.6 percent in the latter period. The conclusion
from these observations is that there has been a
small reduction in trend inflation, whether measured
by the total or the core consumption price index,
over the past four years.
No estimates of the biases in the index are so
large as to suggest that the true rate of inflation is
now negative—that is, the U.S. economy is not in a
deflationary situation. What, then, is the origin of
the “deflation threat” that has been featured in some
economic and newspaper commentaries? Some of
these discussions appear to concentrate unduly on
particular prices and on short-run data collected in
the immediate aftermath of the September 11 terrorist attacks. The change in the price index for personal
consumption expenditures for September 2001
compared with August 2001 was reported at –0.4
4

N OV E M B E R / D E C E M B E R 2 0 0 2

REVIEW

percent. The decline is attributable to falling energy
prices and to a statistical artifact of the decision
made by the Bureau of Economic Analysis in measuring insurance claim payments as a result of the
September 11 attacks. The December 2001 consumption price index showed a decline of 0.2 percent
for the month and led to further press speculation
about deflation. Again, it is necessary to emphasize
that a focus on very short-term movements in the
price indexes can lead to misinterpretation of the
underlying trends of inflation or deflation in an
economy.

CHANGES IN RELATIVE PRICES
One of the great strengths of the U.S. economy
is that prices of individual goods and services fluctuate freely. These price changes allow markets to
signal how our productive resources can be allocated
most efficiently. The disparity among inflation rates
for particular goods and services over longer periods
of time is significant. From 1980 to 2000, the overall consumption price index rose 95 percent. Consider price behavior in a half-dozen categories within
overall personal consumption expenditures. Prices
of personal computers and peripheral equipment
stand out: such prices are estimated to have fallen
by 99 percent since 1980. Note that despite this
dramatic price decline, people do not talk about the
computer industry suffering from deflation. This is
a growth industry, driven by dramatic innovations
and increases in efficiency.
Prices of durable goods are estimated to have
increased by 20 percent since 1980, considerably
slower than the general inflation over this period.
Prices of nondurables are estimated to have increased
by 65 percent since 1980; nondurable goods prices
have risen more than durable goods prices, but still
considerably less than the overall rate of inflation.
Prices of food and beverages are estimated to have
increased 79 percent since 1980, somewhat slower
than the overall rate of inflation.
Consider some examples at the other extreme.
Since 1980, prices of tobacco and smoking products
are estimated to have increased 480 percent and
prices of medical services by 197 percent. In the
tables that show prices by various sectors, wide
differences in experience such as those mentioned
here can be seen.
Are falling prices, or prices that increase slowly
relative to the general rate of inflation, indicative of
“hard times” for particular industries? Sometimes,
but certainly not always. Consider personal comput-

FEDERAL RESERVE BANK OF ST. LOUIS

ers and consumer electronics in general (the latter
is included in the durable goods component of the
consumption price index). These are goods that have
demonstratively high income and price elasticities.
What that means is that the amounts consumers
buy increase a lot as incomes rise and/or prices fall.
Over time, as consumer incomes have increased
and prices have fallen, the size of the market for
these high-elasticity products has increased dramatically. Color TVs, camcorders, VCRs, DVDs, and personal computers, to name a few such products, are
all now common household items in the United
States. Many consumers can remember when these
products were either unknown or owned by relatively
few households.
This is an important point: expansion of the
markets for certain products occurred simultaneously with a fall in prices. Price deflation for these
goods was not inconsistent with prosperity in the
industries producing them. Indeed, declining prices
were essential to expanding these markets. The
fall in prices was the result of rapid productivity
increases from innovations in the production of
these items and/or their components. Firms found
it profitable to cut prices and expand production.
Workers in these industries found their improved
productivity rewarded in higher wages. Consumers,
workers, and shareholders all have benefited, even
though prices have fallen substantially over time.
High-demand elasticities are a critical element
in such success stories. In contrast, consider markets
for basic agricultural products in the United States.
Productivity improvement in U.S. agricultural production over the years has been tremendous. Prices
of these products have also fallen relative to goods
in general over the long run. However, both income
and price elasticities for agricultural products are
relatively low. Hence, economic growth and declining prices have not produced large increases in
consumption. As a result, fewer and fewer workers
have been required over time to produce more than
enough output to satisfy both domestic and foreign
demand. Farms have gone out of business, the number of people engaged in agricultural production
has decreased, and in recent years farm income
has been sustained by large “emergency” farm
appropriations out of the federal budget. Because
of the low price and income elasticities for agricultural goods, deflation in this industry means hard
times for many farmers.
Health care provides a really interesting case
of relative price changes. In part, the rapid rate of
price increase here represents innovation in the

Poole and Rasche

form of new products and/or improved procedures.
Such price changes really reflect significant quality
improvements. Ideally such quality improvements
would be incorporated into the measurement of a
standardized unit of medical services. With some
consumer durables, such as automobiles, statisticians
have been quite successful in measuring quality
improvement. In other areas, capturing quality
change into the measurement of a standard unit of
output is difficult if not impossible.
As an example, consider laparoscopic surgery
to remove the gall bladder. Not that long ago, gall
bladder surgery required a substantial period of
hospitalization, during which patient activity levels
were significantly restricted. Today, with laparoscopic
surgery, the length of the hospital stay is much
shorter and patient discomfort much less. Moreover,
the patient can resume reasonably normal activity,
including going to work, after a short postoperative
period. The patient and/or a third-party payer may
pay the surgeon substantially more today to remove
the gall bladder than 35 years ago, but does this
increase mean that the price properly measured is
dramatically higher? A well-constructed price index
might adjust for the reduction in the pecuniary cost
of confinement—fewer hospital days—from the
improved technology. However, it is unlikely that
any price index would reflect the improved quality
of the procedure represented by the reduced nonpecuniary costs of confinement and the shorter
recovery time now available. Hence the reported
change in the price index for such a procedure
certainly overstates the true rate of price change.

FLATION AND THE FED
The Fed’s goal is to maintain low and steady
inflation, so that expectations of changes in inflation
do not enter importantly in the decisions businesses
and households make. Using several different measures of inflation expectations, it is clear that longterm expected inflation has changed little in recent
years. There is no evidence that changing inflation
expectations figure importantly in economic decisions at this time.
Substantial variability in prices of individual
goods is consistent with stability in the overall inflation rate. The variability serves to allocate and reallocate resources across different sectors of the
economy, according to changes in consumer tastes
and differential trends in productivity advancement.
Simply put, it is normal that some industries are
growing while others are contracting.
N OV E M B E R / D E C E M B E R 2 0 0 2

5

Poole and Rasche

A common business problem is to determine a
successful pricing strategy. One aspect of pricing
strategy is directly relevant to this discussion. When
a firm cuts prices to stimulate sales, it may not be
successful if its customers believe that even deeper
price cuts are around the corner. An expectation of
falling prices may, temporarily, reduce rather than
increase sales. It is for this reason that generalized
deflation can be so dangerous to the economy. A
widespread expectation of falling prices may lead
to declining demand across much of the economy
as people wait for lower prices in the future. Declining demand may force layoffs, which further depress
household and business confidence. Conversely,
inflation expectations can lead to rising demands
and anticipatory buying.
Many analysts seem to view low inflation and
high employment as competing goals. That is certainly not the only possible scenario. Maintaining
low and stable inflation contributes mightily to overall economic stability. Consider the situation in the
weeks following the terrorist attacks of September 11,
2001, when the economic outlook was highly
uncertain. The auto industry was successful in selling
a record number of cars in October 2001 through
price cuts in the form of zero-interest financing. If
consumers had reacted by expecting even deeper
price cuts and had delayed purchases, the situation
in early 2002 would have been very different. Overall, consumers view price cuts in today’s environment as a buying opportunity, not as a forecast of
further price cuts to come.
Clearly, the stability in the overall price environment—stability in longer-run expectations—is what
allows temporary price cuts to work to boost sales
and is an important element in stabilizing the general
economy. The current U.S. situation does not match
cases in the United States and elsewhere that historically have been associated with ongoing deflation.
The Federal Reserve pursued an expansionary monetary policy throughout 2001 that has contributed
to restoring equilibrium to the U.S. economy. What
policy actions will be appropriate going forward will
have to be determined as evidence arrives on the
strength and durability of the economic expansion.
We must be vigilant, but today it is likely that we
enjoy flation—no “in” and no “de.”

6

N OV E M B E R / D E C E M B E R 2 0 0 2

REVIEW

A Case Study of a
Currency Crisis: The
Russian Default of 1998
Abbigail J. Chiodo and Michael T. Owyang

A

currency crisis can be defined as a speculative attack on a country’s currency that can
result in a forced devaluation and possible
debt default. One example of a currency crisis
occurred in Russia in 1998 and led to the devaluation
of the ruble and the default on public and private
debt.1 Currency crises such as Russia’s are often
thought to emerge from a variety of economic conditions, such as large deficits and low foreign reserves.
They sometimes appear to be triggered by similar
crises nearby, although the spillover from these contagious crises does not infect all neighboring economies—only those vulnerable to a crisis themselves.
In this paper, we examine the conditions under
which an economy can become vulnerable to a
currency crisis. We review three models of currency
crises, paying particular attention to the events leading up to a speculative attack, including expectations
of possible fiscal and monetary responses to impending crises. Specifically, we discuss the symptoms
exhibited by Russia prior to the devaluation of the
ruble. In addition, we review the measures that were
undertaken to avoid the crisis and explain why those
steps may have, in fact, hastened the devaluation.
The following section reviews the three generations of currency crisis models and summarizes the
conditions under which a country becomes vulnerable to speculative attack. The third section examines
the events preceding the Russian default of 1998 in
the context of a currency crisis. The fourth section
applies the aforementioned models to the Russian
crisis.

CURRENCY CRISES: WHAT DOES
MACROECONOMIC THEORY SUGGEST?
A currency crisis is defined as a speculative
attack on country A’s currency, brought about by
Abbigail J. Chiodo is a senior research associate and Michael T. Owyang
is an economist at the Federal Reserve Bank of St. Louis. The authors
thank Steven Holland, Eric Blankmeyer, John Lewis, and Rebecca
Beard for comments and suggestions and Victor Gabor at the World
Bank for providing real GDP data.

© 2002, The Federal Reserve Bank of St. Louis.

agents attempting to alter their portfolio by buying
another currency with the currency of country A.2
This might occur because investors fear that the
government will finance its high prospective deficit
through seigniorage (printing money) or attempt to
reduce its nonindexed debt (debt indexed to neither
another currency nor inflation) through devaluation.
A devaluation occurs when there is market pressure to increase the exchange rate (as measured by
domestic currency over foreign currency) because
the country either cannot or will not bear the cost
of supporting its currency. In order to maintain a
lower exchange rate peg, the central bank must buy
up its currency with foreign reserves. If the central
bank’s foreign reserves are depleted, the government
must allow the exchange rate to float up—a devaluation of the currency. This causes domestic goods
and services to become cheaper relative to foreign
goods and services. The devaluation associated with
a successful speculative attack can cause a decrease
in output, possible inflation, and a disruption in
both domestic and foreign financial markets.3
The standard macroeconomic framework
applied by Fleming (1962) and Mundell (1963) to
international issues is unable to explain currency
crises. In this framework with perfect capital mobility, a fixed exchange rate regime results in capital
flight when the central bank lowers interest rates
and results in capital inflows when the central bank
raises interest rates. Consequently, the efforts of the
monetary authority to change the interest rate are
undone by the private sector. In a flexible exchange
rate regime, the central bank does not intervene in
the foreign exchange market and all balance of payment surpluses or deficits must be financed by
private capital outflows or inflows, respectively.
The need to explain the symptoms and remedies
of a currency crisis has spawned a number of models
designed to incorporate fiscal deficits, expectations,
and financial markets into models with purchasing
power parity. These models can be grouped into
three generations, each of which is intended to
explain specific aspects that lead to a currency crisis.
1

Kharas, Pinto, and Ulatov (2001) provide a history from a fundamentalsbased perspective, focusing on taxes and public debt issues. We
endeavor to incorporate a role for monetary policy.

2

The speculative attack need not be successful to be dubbed a currency
crisis.

3

Burnside, Eichenbaum, and Rebelo (2001) show that the government
has at its disposal a number of mechanisms to finance the fiscal costs
of the devaluation. Which policy is chosen determines the inflationary
effect of the currency crisis.

N OV E M B E R / D E C E M B E R 2 0 0 2

7

REVIEW

Chiodo and Owyang

First-Generation Models

(1997), and others are particularly useful in explaining self-fulfilling contagious currency crises. One
possible scenario suggested by these models involves
a devaluation in one country affecting the price level
(and therefore the demand for money) or the current
account by a reduction of exports in a neighboring
country. In either case, devaluation in a neighboring
country becomes increasingly likely.
Eichengreen, Rose, and Wyplosz (1997) find
that a correlation exists between the likelihood of
default across countries. That is, the probability of
a speculative attack in country A increases when
its trading partner, country B, experiences an attack
of its own. They estimate that a speculative attack
somewhere in the world increases the probability
of a domestic currency crisis by about 8 percent.
The spillover from one currency crisis into neighboring countries can be attributed to a number of different scenarios. First, an economic event, such as a
war or an oil price shock, that is common to a geographical area or a group of trading partners can
affect those economies simultaneously; in addition,
an individual shock can be transmitted from one
country to another via trade. Second, a devaluation
or default in one country can raise expectations of
the likelihood of a devaluation in other countries.
Expectations can rise either because countries are
neighboring trade partners or because they have
similar macroeconomic policies or conditions (e.g.,
high unemployment or high government debt). Since
the crises are self-fulfilling, these expectations make
the likelihood of devaluation increase as well. Lastly,
a devaluation can be transmitted via world financial
markets to other susceptible countries. Any combination of scenarios can serve as an explanation of
the apparent international linkages that are responsible for the spread of speculative attacks from one
country to another.

The first-generation models of a currency crisis
developed by Krugman (1979) and Flood and Garber
(1984) rely on government debt and the perceived
inability of the government to control the budget
as the key causes of the currency crisis. These models
argue that a speculative attack on the domestic
currency can result from an increasing current
account deficit (indicating an increase in the trade
deficit) or an expected monetization of the fiscal
deficit. The speculative attack can result in a sudden
devaluation when the central bank’s store of foreign
reserves is depleted and it can no longer defend the
domestic currency. Agents believe that the government’s need to finance the debt becomes its overriding concern and eventually leads to a collapse
of the fixed exchange rate regime and to speculative
attacks on the domestic currency.
Krugman presents a model in which a fixed
exchange rate regime is the inevitable target of a
speculative attack. An important assumption in the
model is that a speculative attack is inevitable. The
government defends the exchange rate peg with its
store of foreign currency. As agents change the composition of their portfolios from domestic to foreign
currency (because rising fiscal deficits increase the
likelihood of devaluation, for example), the central
bank must continue to deplete its reserves to stave
off speculative attacks. The crisis is triggered when
agents expect the government to abandon the peg.
Anticipating the devaluation, agents convert their
portfolios from domestic to foreign currency by buying foreign currency from the central bank’s reserves.
The central bank’s reserves fall until they reach the
critical point when a peg is no longer sustainable
and the exchange rate regime collapses. The key
contribution of the first-generation model is its
identification of the tension between domestic fiscal
policy and the fixed exchange rate regime.4
While the first-generation models help explain
some of the fundamentals that cause currency crises,
they are lacking in two key aspects. First, the standard first-generation model requires agents to suddenly increase their estimates of the likelihood of
a devaluation (perhaps through an increase in
expected inflation). Second, they do not explain
why the currency crises spread to other countries.

The literature on contagious currency crises
has helped clarify the spread of devaluations and
their magnitudes. However, the first two generations
of models have not provided a policy recommendation for the central bank in the face of a crisis. Indeed,
Krugman’s first-generation model suggests that a
crisis cannot be thwarted—that once a devaluation
is expected, it is inevitable. Thus, third-generation

Second-Generation Models

4

The second-generation models suggested by
Obstfeld (1994), Eichengreen, Rose, and Wyplosz
8

N OV E M B E R / D E C E M B E R 2 0 0 2

Third-Generation Models

Obstfeld (1986) outlines a multiple equilibrium model in which a
currency crisis is brought about when government policy (financing
a deficit through seignorage, for example) causes agents to expect a
crisis and push the economy to a bad equilibrium.

FEDERAL RESERVE BANK OF ST. LOUIS

currency crisis models suggested by Krugman (1999)
and Aghion, Bacchetta, and Banarjee (2000, 2001)
examine the effects of monetary policy in a currency
crisis.
These models argue that fragility in the banking
and financial sector reduces the amount of credit
available to firms and increases the likelihood of a
crisis. They suggest that a currency crisis is brought
on by a combination of high debt, low foreign
reserves, falling government revenue, increasing
expectations of devaluation, and domestic borrowing constraints. Firms’ access to domestic loans is
constrained by assuming they can borrow only a
portion of their wealth (somewhat similar to requiring the firm to collateralize all domestic loans). In
these lending-constrained economies, the credit
market does not clear: interest rates rise, but not
enough to compensate investors for the increase in
perceived default risk. Increasing the domestic interest rate, then, does not raise the supply of domestic
lending in the normal fashion. Moral hazard, a firm’s
ability to take its output and default on its loan, forces
banks to restrict lending. Therefore, increasing
the interest rate reduces the amount of loans as it
increases firms’ incentive to default.
These third-generation models offer a role for
monetary policy (aside from the decision to abandon
the exchange rate peg) through a binding credit
constraint in an imperfect financial market. If firms’
leverage in the domestic market is substantially
reduced, they may be forced to accumulate a large
amount of foreign-denominated debt. When, in
domestic markets, the amount of available lending
depends on the nominal interest rate, the central
bank can deepen a crisis by further reducing firms’
ability to invest. The typical prescription for a currency crisis is to raise interest rates and raise the
demand for domestic currency.5 However, in the
third-generation models, an interest rate increase
can greatly affect the amount of lending and further
restrict firms’ access to financial capital. In cases
where lending is highly sensitive to the interest rate,
an increase in the nominal interest rate can be
detrimental, altering the productive capacity of the
economy by stifling investment. The perceived drop
in output puts additional pressure on the exchange
rate, perhaps through actual or expected tax revenue,
exacerbating the crisis. In this situation, an alternative strategy for the central bank is warranted: it is
actually beneficial to lower the interest rate to spur
investment.6
These three generations of models suggest four

Chiodo and Owyang

factors that can influence the onset and magnitude
of a currency crisis. Domestic public and private
debt, expectations, and the state of financial markets
can, in combination with a pegged exchange rate,
determine whether a country is susceptible to a
currency crisis and also determine the magnitude
and success of a speculative attack. In the next
section, we provide an example of a recent currency
crisis, keeping these four factors in mind.

THE RUSSIAN DEFAULT: A BRIEF
HISTORY
After six years of economic reform in Russia,
privatization and macroeconomic stabilization had
experienced some limited success. Yet in August
1998, after recording its first year of positive economic growth since the fall of the Soviet Union, Russia
was forced to default on its sovereign debt, devalue
the ruble, and declare a suspension of payments by
commercial banks to foreign creditors. What caused
the Russian economy to face a financial crisis after
so much had been accomplished? This section
examines the sequence of events that took place in
Russia from 1996 to 1998 and the aftermath of the
crisis. (For a timeline, see Table 1.)

1996 and 1997
Optimism and Reform. In April 1996, Russian
officials began negotiations to reschedule the payment of foreign debt inherited from the former
Soviet Union. The negotiations to repay its sovereign
debt were a major step toward restoring investor
confidence. On the surface, 1997 seemed poised
to be a turning point toward economic stability.

• The trade surplus was moving toward a
balance between exports and imports (see
Figure 1).
• Relations with the West were promising: the
World Bank was prepared to provide expanded
assistance of $2 to $3 billion per year and the
International Monetary Fund (IMF) continued
to meet with Russian officials and provide aid.
• Inflation had fallen from 131 percent in 1995
to 22 percent in 1996 and 11 percent in 1997
(see Figure 2).
• Output was recovering slightly.
5

Flood and Jeanne (2000) argue that increasing domestic currency
interest rates can act only to speed devaluation.

6

The expansionary monetary policy in this case is assumed not to be
inflationary since it only alleviates liquidity constraints.

N OV E M B E R / D E C E M B E R 2 0 0 2

9

REVIEW

Chiodo and Owyang

Table 1
A Timeline of Russian Events
April 1996

Negotiations with the Paris and London Clubs for repayment of Soviet debt begin.

1997

Trade surplus moving toward balance.
Inflation around 11 percent.
Oil selling at $23/barrel.
Analysts predict better credit ratings for Russia.
Russian banks increase foreign liabilities.
Real wages sagging.
Only 40 percent of workforce being paid fully and on time.
Public-sector deficit high.

September/October 1997

Negotiations with Paris and London Clubs completed.

November 11, 1997

Asian crisis causes a speculative attack on the ruble.
CBR defends the ruble, losing $6 billion.

December 1997

Year ends with 0.8 percent growth.
Prices of oil and nonferrous metal begin to drop.

February 1998

New tax code submitted to the Duma.
IMF funds requested.

March 23, 1998

Yelstin fires entire government and appoints Kiriyenko.
Continued requests for IMF funds.

April 1998

Another speculative attack on the ruble.

April 24, 1998

Duma finally confirms Kiriyenko’s appointment.

Early May 1998

Dubinin warns government ministers of impending debt crisis, with reporters in the
audience.
Kiriyenko calls the Russian government “quite poor.”

May 19, 1998

CBR increases lending rate from 30 percent to 50 percent and defends the ruble
with $1 billion.

Mid May 1998

Lawrence Summers not granted audience with Kiriyenko.
Oil prices continue to decrease.
Oil and gas oligarchs advocate devaluation of ruble to increase value of their exports.

May 23, 1998

IMF leaves Russia without agreement on austerity plan.

May 27, 1998

CBR increases the lending rate again to 150 percent.

Summer 1998

Russian government formulates and advertises anti-crisis plan.

July 20, 1998

IMF approves an emergency aid package (first disbursement to be $4.8 billion).

August 13, 1998

Russian stock, bond, and currency markets weaken as a result of investor fears of
devaluation; prices diminish.

August 17, 1998

Russian government devalues the ruble, defaults on domestic debt, and declares a
moratorium on payment to foreign creditors.

August 23-24, 1998

Kiriyenko is fired.

September 2, 1998

The ruble is floated.

December 1998

Year ends with a decrease in real output of 4.9 percent.

NOTE: CBR, Central Bank of Russia.

10

N OV E M B E R / D E C E M B E R 2 0 0 2

FEDERAL RESERVE BANK OF ST. LOUIS

Chiodo and Owyang

Figure 1

Figure 2

Russian Merchandise Trade Balance

CPI Inflation
Percent Change over Previous Year

US$ Millions
30,000

Percent
Exports

250

25,000
200
20,000
150
Imports

15,000

100
10,000
50
5,000
1994:Q1

1996:Q1

1998:Q1

2000:Q1

0
Dec
1995

SOURCE: CBR.

Mar
1997

Jun
1998

Sep
1999

Dec
2000

SOURCE: IMF.

• A narrow exchange rate band was in place
keeping the exchange rate between 5 and 6
rubles to the dollar (see Figure 3).
• And oil, one of Russia’s largest exports, was
selling at $23 per barrel—a high price by
recent standards. (Fuels made up more than
45 percent of Russia’s main export commodities in 1997.)
In September 1997, Russia was allowed to join
the Paris Club of creditor nations after rescheduling
the payment of over $60 billion in old Soviet debt
to other governments. Another agreement for a
23-year debt repayment of $33 billion was signed
a month later with the London Club. Analysts predicted that Russia’s credit ratings would improve,
allowing the country to borrow less expensively.
Limitations on the purchase of government securities
by nonresident investors were removed, promoting
foreign investment in Russia. By late 1997, roughly
30 percent of the GKO (a short-term government
bill) market was accounted for by nonresidents. The
economic outlook appeared optimistic as Russia
ended 1997 with reported economic growth of 0.8
percent.
Revenue, Investment, and Debt. Despite the
prospects for optimism, problems remained. On
average, real wages were less than half of what they
were in 1991, and only about 40 percent of the
work force was being paid in full and on time. Per
capita direct foreign investment was low, and regu-

Figure 3
Exchange Rate
Ruble/US$
35
30
25
20
15
10
5
0
Dec
1995

Mar
1997

Jun
1998

Sep
1999

Dec
2000

SOURCE: IMF (end of period data).

lation of the natural monopolies was still difficult
due to unrest in the Duma, Russia’s lower house
of Parliament. Another weakness in the Russian
economy was low tax collection, which caused the
N OV E M B E R / D E C E M B E R 2 0 0 2

11

REVIEW

Chiodo and Owyang

public sector deficit to remain high. The majority
of tax revenues came from taxes that were shared
between the regional and federal governments,
which fostered competition among the different
levels of government over the distribution. According to Shleifer and Treisman (2000), this kind of
tax sharing can result in conflicting incentives for
regional governments and lead them to help firms
conceal part of their taxable profit from the federal
government in order to reduce the firms’ total tax
payments. In return, the firm would then make
transfers to the accommodating regional government. This, Shleifer and Treisman suggest, may
explain why federal revenues dropped more rapidly
than regional revenues.
Also, the Paris Club’s recognition of Russia as a
creditor nation was based upon questionable qualifications. One-fourth of the assets considered to
belong to Russia were in the form of debt owed to
the former Soviet Union by countries such as Cuba,
Mongolia, and Vietnam. Recognition by the Paris
Club was also based on the old, completely arbitrary
official Soviet exchange rate of approximately 0.6
rubles to the dollar (the market exchange rate at
the time was between 5 and 6 rubles to the dollar).
The improved credit ratings Russia received from
its Paris Club recognition were not based on an
improved balance sheet. Despite this, restrictions
were eased and lifted and Russian banks began
borrowing more from foreign markets, increasing
their foreign liabilities from 7 percent of their assets
in 1994 to 17 percent in 1997.
Meanwhile, Russia anticipated growing debt
payments in the coming years when early credits
from the IMF would come due. Policymakers faced
decisions to decrease domestic borrowing and
increase tax collection because interest payments
were such a large percentage of the federal budget.
In October 1997, the Russian government was counting on 2 percent economic growth in 1998 to compensate for the debt growth. Unfortunately, events
began to unfold that would further strain Russia’s
economy; instead of growth in 1998, real GDP
declined 4.9 percent.
The Asian Crisis. A few months earlier, in
the summer of 1997, countries in the Pacific Rim
experienced currency crises similar to the one that
eventually affected Russia. In November 1997, after
the onset of this East Asian crisis, the ruble came
under speculative attack. The Central Bank of Russia
(CBR) defended the currency, losing nearly $6 billion
(U.S. dollars) in foreign-exchange reserves. At the
12

N OV E M B E R / D E C E M B E R 2 0 0 2

same time, non-resident holders of short-term
government bills (GKOs) signed forward contracts
with the CBR to exchange rubles for foreign currency, which enabled them to hedge exchange rate
risk in the interim period.7 According to Desai
(2000), they did this in anticipation of the ruble losing value, as Asian currencies had. Also, a substantial
amount of the liabilities of large Russian commercial
banks were off-balance-sheet, consisting mostly
of forward contracts signed with foreign investors.
Net obligations of Russian banks for such contracts
were estimated to be at least $6 billion by the first
half of 1998. Then another blow was dealt to the
Russian economy: in December 1997, the prices
of oil and nonferrous metal, up to two-thirds of
Russia’s hard-currency earnings, began to drop.

1998
Government, Risk, and Expectations. With
so many uncertainties in the Russian economy,
investors turned their attention toward Russian
default risk. To promote a stable investment environment, in February 1998, the Russian government
submitted a new tax code to the Duma, with fewer
and more efficient taxes. The new tax code was
approved in 1998, yet some crucial parts that were
intended to increase federal revenue were ignored.
Russian officials sought IMF funds but agreements
could not be reached. By late March the political
and economic situation had become more dire, and,
on March 23, President Yeltsin abruptly fired his
entire government, including Prime Minister Viktor
Chernomyrdin. In a move that would challenge
investor confidence even further, Yeltsin appointed
35-year-old Sergei Kiriyenko, a former banking and
oil company executive who had been in government
less than a year, to take his place.
While fears of higher interest rates in the United
States and Germany made many investors cautious,
tensions rose in the Russian government. The executive branch, the Duma, and the CBR were in conflict.
Prompted by threats from Yeltsin to dissolve Parliament, the Duma confirmed Kiriyenko’s appointment
on April 24 after a month of stalling. In early May,
during a routine update, CBR chair Sergei Dubinin
warned government ministers of a debt crisis within
the next three years. Unfortunately, reporters were
in the audience. Since the Asian crisis had heightened
investors’ sensitivity to currency stability, Dubinin’s
7

The requirement of forward contracts was the CBR’s way of preventing
runs on its foreign currency reserves.

FEDERAL RESERVE BANK OF ST. LOUIS

restatement of bank policy was misinterpreted to
mean that the Bank was considering a devaluation
of the ruble. In another public relations misunderstanding, Kiriyenko stated in an interview that tax
revenue was 26 percent below target and claimed
that the government was “quite poor now.” In actuality, the government was planning to cut government
spending and accelerate revenue, but these plans
were never communicated clearly to the public.
Instead, people began to expect a devaluation of
the ruble.
Investors’ perceptions of Russia’s economic
stability continued to decline when Lawrence
Summers, one of America’s top international-finance
officials, was denied a meeting with Kiriyenko while
in Russia. An inexperienced aide determined that
Summers’s title, Deputy Secretary of the Treasury,
was unworthy of Kiriyenko’s audience and the two
never met. At the same time, the IMF left Russia,
unable to reach an agreement with policymakers
on a 1998 austerity plan. Word spread of these incidents, and big investors began to sell their government bond portfolios and Russian securities,
concerned that relations between the United States
and Russia were strained.
Liquidity, Monetary Policy, and Fiscal Policy.
By May 18, government bond yields had swelled
to 47 percent. With inflation at about 10 percent,
Russian banks would normally have taken the
government paper at such high rates. Lack of confidence in the government’s ability to repay the
bonds and restricted liquidity, however, did not
permit this. As depositors and investors became
increasingly cautious of risk, these commercial
banks and firms had less cash to keep them afloat.
The federal government’s initiative to collect more
taxes in cash lowered banks’ and firms’ liquidity.8
Also, in 1997, Russia had created a U.S.-style treasury system with branches, which saved money
and decreased corruption, yet also decreased the
amount of cash that moved through banks. The
banks had previously used these funds to buy bonds.
Also, household ruble deposits increased by only
1.3 billion in 1998, compared with an increase of
29.8 billion in 1997.
The CBR responded by increasing the lending
rate to banks from 30 to 50 percent, and in two days
used $1 billion of Russia’s low reserves to defend
the ruble. (Figure 4 shows the lending rate.) However,
by May 27, demand for bonds had plummeted so
much that yields were more than 50 percent and
the government failed to sell enough bonds at its

Chiodo and Owyang

Figure 4
Lending Rate
Percent
160
140
120
100
80
60
40
20
0
Jul
1996

Feb
1997

Nov
1997

Mar
1998

Jun
1998

Mar
2000

Nov
2000

SOURCE: CBR.

weekly auction to refinance the debt coming due.
Meanwhile, oil prices had dropped to $11 per
barrel, less than half their level a year earlier. Oil
and gas oligarchs were advocating a devaluation of
the ruble, which would increase the ruble value of
their exports. In light of this, the CBR increased the
lending rate again, this time to 150 percent. CBR
chairman Sergei Dubinin responded by stating
“When you hear talk of devaluation, spit in the eye
of whoever is talking about it” (quoted in Shleifer
and Treisman, 2000, p. 149).
The government formed and advertised an anticrisis plan, requested assistance from the West, and
began bankruptcy processes against three companies with large debts from back taxes. Kiriyenko
met with foreign investors to reassure them. Yeltsin
made nightly appearances on Russian television,
calling the nation’s financial elite to a meeting at
the Kremlin where he urged them to invest in Russia.
In June the CBR defended the ruble, losing $5 billion
in reserves.
Despite all of the government efforts being made,
there was widespread knowledge of $2.5 to $3 billion
8

As a result of a 1998 elimination of tax-offsets paper issued by government agencies to pay for goods and services, the receipts of which
could be used to decrease their tax duties, banks and companies were
forced to provide more cash to pay their taxes, thus lowering their
liquidity.

N OV E M B E R / D E C E M B E R 2 0 0 2

13

REVIEW

Chiodo and Owyang

was down 65 percent with a small number of shares
actually traded. From January to August the stock
market had lost more than 75 percent of its value,
39 percent in the month of May alone. (Figure 5
shows the Russian stock market’s boom and bust.)
Russian officials were left with little choice. On
August 17 the government floated the exchange rate,
devalued the ruble, defaulted on its domestic debt,
halted payment on ruble-denominated debt (primarily GKOs), and declared a 90-day moratorium on
payment by commercial banks to foreign creditors.

Figure 5
The Russian Stock Market
Daily Observations (Sept. 1, 1995=100)
600

500

400

300

The Aftermath
200

100

0
Jan
1996

Dec
1996

Dec
1997

Dec
1998

Dec
1999

Dec
2000

Nov
2001

SOURCE: <http://red-stars.com/financial>.

in loans from foreign investors to Russian corporations and banks that were to come due by the end
of September. In addition, billions of dollars in ruble
futures were to mature in the fall. In July the IMF
approved additional assistance of $11.2 billion, of
which $4.8 billion was to be disbursed immediately.
Yet between May and August, approximately $4
billion had left Russia in capital flight, and in 1998
Russia lost around $4 billion in revenue due to sagging oil prices. After losing so much liquidity, the
IMF assistance did not provide much relief.
The Duma, in an effort to protect natural monopolies from stricter regulations, eliminated crucial
parts of the IMF-endorsed anti-crisis program before
adjourning for vacation. The government had hoped
that the anti-crisis plan would bring in an additional
71 billion rubles in revenue. The parts that the Duma
actually passed would have increased it by only 3
billion rubles. In vain, lawmakers requested that the
Duma reconvene, lowering investors’ confidence
even further.
Default and Devaluation. On August 13, 1998,
the Russian stock, bond, and currency markets
collapsed as a result of investor fears that the
government would devalue the ruble, default on
domestic debt, or both. Annual yields on rubledenominated bonds were more than 200 percent.
The stock market had to be closed for 35 minutes
as prices plummeted. When the market closed, it
14

N OV E M B E R / D E C E M B E R 2 0 0 2

Russia ended 1998 with a decrease in real output of 4.9 percent for the year instead of the small
growth that was expected. The collapse of the ruble
created an increase in Russia’s exports while imports
remained low (see Figure 1). Since then, direct
investments into Russia have been inconsistent at
best. Summarized best by Shleifer and Treisman
(2000), “the crisis of August 1998 did not only undermine Russia’s currency and force the last reformers
from office…it also seemed to erase any remaining
Western hope that Russia could successfully reform
its economy.”
Some optimism, however, still persists. Figure
6 shows Russian real GDP growth, which grew 8.3
percent in 2000 and roughly 5 percent in 2001—
lower but still positive. Imports trended up in the
first half of 2001, helping to create a trade balance.
At the same time, consumer prices grew 20.9 percent
and 21.6 percent in 2000 and 2001, respectively,
compared with a 92.6 percent increase in 1999.
Most of the recovery so far can be attributed to the
import substitution effect after the devaluation; the
increase in world prices for Russia’s oil, gas, and
commodity exports; monetary policies; and fiscal
policies that have led to the first federal budget surplus (in 2000) since the formation of the Russian
Federation.

HOW DO THE THEORIES EXPLAIN
THE RUSSIAN CRISIS?
As discussed earlier, four major factors influence
the onset and success of a speculative attack. These
key ingredients are (i) an exchange rate peg and a
central bank willing or obligated to defend it with a
reserve of foreign currency, (ii) rising fiscal deficits
that the government cannot control and therefore
is likely to monetize (print money to cover the deficit),
(iii) central bank control of the interest rate in a

FEDERAL RESERVE BANK OF ST. LOUIS

fragile credit market, and (iv) expectations of devaluation and/or rising inflation. In this section we discuss these aspects in the context of the Russian
devaluation. We argue that an understanding of all
three generations of models is necessary to evaluate
the Russian devaluation. Krugman’s (1979) firstgeneration model explains the factors that made
Russia susceptible to a crisis. The second-generation
models show how contagion and other factors can
change expectations to trigger the crisis. The thirdgeneration models show how the central bank can
act to prevent or mitigate the crisis.

The Exchange Rate and the Peg
When the ruble came under attack in November
1997 and June 1998, policymakers defended the
ruble instead of letting it float. The real exchange
rate did not vary much during 1997. Clearly a primary component of a currency crisis in the models
described here is the central bank’s willingness to
defend an exchange rate peg. Prior to August 1998,
the Russian ruble was subject to two speculative
attacks. The CBR made efforts both times to defend
the ruble. The defense was successful in November
1997 but fell short in the summer of 1998. Defending the ruble depleted Russia’s foreign reserves.
Once depleted, the Russian government had no
choice but to devalue on August 17, 1998.

Revenue, Deficits, and Fiscal Policy
Russia’s high government debt and falling revenue contributed significantly to its susceptibility
to a speculative attack. Russia’s federal tax revenues
were low because of both low output and the opportunistic practice of local governments helping firms
conceal profits. The decrease in the price of oil also
lowered output, further reducing Russia’s ability to
generate tax revenue. Consequently, Russia’s revenue
was lower than expected, making the ruble ripe for
a speculative attack. In addition, a large amount of
short-term foreign debt was coming due in 1998,
making Russia’s deficit problem even more serious.
Krugman’s first-generation model suggests that a
government finances its deficit by printing money
(seigniorage) or depleting its reserves of foreign
currency. Under the exchange rate peg, however,
Russia was unable to finance through seigniorage.
Russia’s deficit, low revenue, and mounting interest
payments put pressure on the exchange rate. Printing rubles would only have increased this pressure
because the private sector would still have been able

Chiodo and Owyang

Figure 6
Real GDP Growth
Quarterly Change from Previous Year
Percent
11
10
8
6
4
2
0
–2
–4
–6
–8
–10
1995:Q1

1996:Q2

1997:Q3

1998:Q4

2000:Q1

2001:Q2

SOURCE: Russian Statistics Committee and International
Bank for Reconstruction and Development staff estimates.

to trade rubles for foreign currency at the fixed rate.
Thus, whether directly through intervention in the
foreign currency market or indirectly by printing
rubles, Russia’s only alternative under the fixed
exchange rate regime was to deplete its stock of
foreign reserves.

Monetary Policy, Financial Markets,
and Interest Rates
During the summer of 1998, the Russian economy was primed for the onset of a currency crisis.
In an attempt to avert the crisis, the CBR intervened
by decreasing the growth of the money supply and
twice increasing the lending rate to banks, raising
it from 30 to 150 percent. Both rate hikes occurred
in May 1998, the same month in which the Russian
stock market lost 39 percent of its value. The rise
in interest rates had two effects. First, it exacerbated
Russia’s revenue problems. Its debt grew rapidly as
interest payments mounted. This put pressure on
the exchange rate because investors feared that
Russia would devalue to finance its non-denominated
debt. Second, high government debt prevented firms
from obtaining loans for new capital and increasing
the interest rate did not increase the supply of lending capital available to firms. At the same time, forN OV E M B E R / D E C E M B E R 2 0 0 2

15

REVIEW

Chiodo and Owyang

eign reserves held by the CBR were so low that the
government could no longer defend the currency
by buying rubles.

Expectations
Three components fueled the expectations of
Russia’s impending devaluation and default. First,
the Asian crisis made investors more conscious of
the possibility of a Russian default. Second, public
relations errors, such as the publicized statement
to government ministers by the CBR and Kiriyenko’s
refusal to grant Lawrence Summers an audience,
perpetuated agents’ perceptions of a political crisis
within the Russian government. Third, the revenue
shortfall signaled the possible reduction of the public
debt burden via an increase in the money supply.
This monetization of the debt can be associated
with a depreciation either indirectly through an
increase in expected inflation or directly in order
to reduce the burden of ruble-denominated debt.
Each of these three components acted to push the
Russian economy from a stable equilibrium to one
vulnerable to speculative attack.

CONCLUSION
In this paper we investigate the events that lead
up to a currency crisis and debt default and the
policies intended to avert it. Three types of models
exist to explain currency crises. Each model explains
some factor that has been hypothesized to cause a
crisis. After reviewing the three generations of currency crisis models, we conclude that four key
ingredients can trigger a crisis: a fixed exchange
rate, fiscal deficits and debt, the conduct of monetary
policy, and expectations of impending default. Using
the example of the Russian default of 1998, we show
that the prescription of contractionary monetary
policy in the face of a currency crisis can, under
certain conditions, accelerate devaluation. While
we believe that deficits and the Asian financial crisis
contributed to Russia’s default, the first-generation
model proposed by Krugman (1979) and Flood and
Garber (1984) and the second-generation models
proposed by Obstfeld (1984) and Eichengreen, Rose,
and Wyplosz (1997) do not capture every aspect of
the crisis. Specifically, these models do not address
the conduct of monetary policy. It is therefore necessary to incorporate both the first-generation
model’s phenomenon of increasing fiscal deficits
and the third-generation model’s financial sector
fragility. We conclude that the modern currency
16

N OV E M B E R / D E C E M B E R 2 0 0 2

crisis is a symptom of an ailing domestic economy.
In that light, it is inappropriate to attribute a single
prescription as the prophylactic or cure for a currency crisis.

REFERENCES
Aghion, Philippe; Bacchetta, Philippe and Banerjee, Abhijit.
“A Simple Model of Monetary Policy and Currency Crises.”
European Economic Review, May 2000, 44(4-6), pp. 728-38.
___________; ___________ and ___________. “Currency
Crises and Monetary Policy in an Economy with Credit
Constraints.” European Economic Review, June 2001, 45(7),
pp. 1121-50.
Ahrend, Rudiger. “Foreign Direct Investment Into Russia—
Pain Without Gain? A Survey of Foreign Direct Investors.”
Russian Economic Trends, June 2000, 9(2), pp. 26-33.
Burnside, Craig; Eichenbaum, Martin, and Rebelo, Sergio.
“On The Fiscal Implications of Twin Crises.” Working
Paper No. 8277, National Bureau of Economic Research,
May 2001.
Desai, Padma. “Why Did the Ruble Collapse in August 1998?”
American Economic Review: Papers and Proceedings, May
2000, 90(2), pp. 48-52.
Economist. “Surplus to Requirements.” 8 July 2000, p. 79.
Eichengreen, Barry; Rose, Andrew and Wyplosz, Charles.
“Contagious Currency Crisis.” March 1997.
<http://www.haas.berkeley.edu/~arose/>.
Fischer, Stanley. “The Russian Economy at the Start of
1998.” U.S.-Russian Investment Symposium, Harvard
University, Cambridge, MA, 9 January 1998.
Flemming, Marcus. “Domestic Financial Policies Under
Fixed and Under Floating Exchange Rates.” IMF Staff
Papers, 9 November 1962.
Flood, Robert P. and Garber, Peter M. “Collapsing Exchange
Rate Regimes: Some Linear Examples.” Journal of
International Economics, August 1984, 17(1-2), pp 1-13.
___________ and Jeanne, Olivier. “An Interest Rate Defense
of a Fixed Exchange Rate?” Working Paper WP/00/159,
International Monetary Fund, October 2000.
Kharas, Homi; Pinto, Brian and Ulatov, Sergei. “An Analysis
of Russia’s 1998 Meltdown: Fundamentals and Market
Signals.” Brookings Papers on Economic Activity, 2001,
0(1), pp. 1-67.

FEDERAL RESERVE BANK OF ST. LOUIS

Chiodo and Owyang

Krugman, Paul. “A Model of Balance-of-Payment Crises.”
Journal of Money, Credit, and Banking, August 1979, 11(3),
pp. 311-25.
___________. “Balance Sheets, the Transfer Problem, and
Financial Crises.” International Tax and Public Finance,
November 1999, 6(4), pp. 459-72.
Malleret, Thierry; Orlova, Natalia and Romanov, Vladimir.
“What Loaded and Triggered the Russian Crisis?” PostSoviet Affairs, April-June 1999, 15(2), pp. 107-29.
Mudell, R.A. “Capital Mobility and Stabilization Policy Under
Fixed and Flexible Exchange Rates.” Canadian Journal of
Economics, November 1963.
Obstfeld, Maurice. “Rational and Self-Fulfilling Balance-ofPayments Crises.” American Economic Review, March 1986,
76(1), pp. 72-81.
___________. “The Logic of Currency Crises.” Cahiers
Economiques et Monetaires, Banque de France, 1994, 43,
pp. 189-213.
Popov, A. “Lessons of the Currency Crisis in Russia and in
Other Countries.” Problems of Economic Transition, May
2000, 43(1), pp. 45-73.
Russian Economic Trends. Various months.
Shleifer, Andre and Treisman, Daniel. Without A Map:
Political Tactics and Economic Reform in Russia. Cambridge,
MA: MIT Press, 2000.
Velasco, Andrés. “Financial Crises in Emerging Markets.”
National Bureau of Economic Research Reporter, Fall
1999, pp.17-19.

N OV E M B E R / D E C E M B E R 2 0 0 2

17

Chiodo and Owyang

18

N OV E M B E R / D E C E M B E R 2 0 0 2

REVIEW

Asset Mispricing,
Arbitrage, and Volatility
William R. Emmons and Frank A. Schmid
fter nearly four decades, academic economists continue to debate financial-market
efficiency as vigorously as ever.1 The original
theoretical arguments put forward in favor of efficient markets were based on the notion of stabilizing speculation in the form of arbitrage (Friedman,
1953). Simply put, arbitrage is “the simultaneous
purchase and sale of the same, or essentially similar,
security in two different markets for advantageously
different prices” (Sharpe and Alexander, 1990). In
theory, a perfectly hedged trading position of this
sort could be executed at no cost (as the short-sale
proceeds are used to finance the long position).
Vigilant traders on the look-out for just such arbitrage opportunities would ensure that no one could
consistently “beat the market”—the hallmark of
efficient markets theory.
The academics’ logical case for efficient markets
boils down to a pair of simple rhetorical questions:
Why would utility-maximizing traders leave unexploited any profitable opportunities (after adjusting
properly for risk)? And if no risk-adjusted “free
lunches” exist, how could market prices be predictable enough to make money? For several decades,
empirical evidence piled up both for and against
market efficiency. As of the early 1990s, neither
side could claim total vindication. As the 1990s
progressed, however, the weight of the evidence
seemed to tip toward those who claimed asset prices
were, at least to some extent, predictable (Campbell,
Lo, and MacKinlay, 1997, Chaps. 2 and 7).
The academic asset-pricing literature today is
dominated by attempts to explain why and to what
extent the price movements of financial assets are
predictable. One potential explanation of stockreturn predictability is that markets are efficient
(“no free lunch”) but expected returns are timevarying, perhaps being linked to the business cycle.
For example, expected returns may be highest when
economic risks are perceived to be high, such as at
or near the bottom of a business cycle. Conversely,

A

William R. Emmons is an economist and Frank A. Schmid is a senior
economist at the Federal Reserve Bank of St. Louis. William V. Bock
provided research assistance.

© 2002, The Federal Reserve Bank of St. Louis.

expected returns may be lowest when economic
risks are perceived to be low, at or near a businesscycle peak. Thus, the simple random-walk model
of stock returns may be false, but a relevant notion
of market efficiency survives because high returns
are earned only by taking large amounts of risk. A
different type of explanation of return predictability
rejects market efficiency and focuses on market
imperfections of various sorts, such as incomplete
stock-market participation by households, significant
transactions costs, changes in investor sentiment,
or limited wealth and liquidity resources to conduct
arbitrage (as in the current article).2
Whatever its economic explanation, mounting
evidence of return predictability leads Campbell, Lo,
and MacKinlay (1997, p. 24) to suggest that it is time
for financial economists to focus their attention
on the “relative efficiency” of a market instead of
continuing the all-or-nothing battle of attrition that
is characteristic of much of the earlier market efficiency literature.
As we now understand more clearly, the original
case for efficient markets probably leaned too heavily on the notion of risk-free, cost-free arbitrage to
eliminate all profitable trading strategies immediately. In real markets, arbitrage is neither as easy
nor as effective as economists once had assumed.
For one thing, financial markets are not complete
and frictionless, so arbitrage in general is risky and
costly. In addition, it is not realistic to assume that
the number of informed arbitrageurs or the supply
of financial resources they have to invest in arbitrage
strategies is limitless.
This article builds on an important and insightful recent model of arbitrage by professional traders
who need—but lack—wealth of their own to trade
(Shleifer and Vishny, 1997). Professional arbitrageurs
must convince wealthy but uninformed investors
to entrust them with investment capital in order to
exploit mispricing and push the market back toward
the ideal of efficiency. Unfortunately, arbitrageurs
cannot prove that they recognize the intrinsic (or
“fundamental”) values of the assets they claim are
mispriced. Even worse, it is possible the assets will
1

For early statements of the theory of efficient markets and the unpredictability of asset-price movements, see Fama (1965), Muth (1960),
or Samuelson (1965). For a recent summary of the evidence for return
predictability and its implications for efficient-markets theory, see
Campbell, Lo, and MacKinlay (1997, Chap. 2).

2

Ironically, Keynes (1936, Chap. 12) clearly foreshadowed the recent
interest in investor sentiment and liquidity for understanding stock
market behavior, but was forgotten for decades as the efficient-markets
hypothesis dominated the academic discussion.

N OV E M B E R / D E C E M B E R 2 0 0 2

19

REVIEW

Emmons and Schmid

become even more mispriced before reverting
eventually to their intrinsic values. Having incurred
losses, the outside investors may demand their
money back at this point even though the expected
profit of staying invested actually has increased.
Thus, market efficiency may depend ultimately
on the successful resolution of a principal-agent
problem that exists between informed but wealthconstrained arbitrageurs and uninformed wealthy
investors. The resulting degree of market efficiency
may change over time and differ across markets,
and it could depend importantly on factors such as
the outside investors’ use of performance-based
(“feedback”) strategies when deciding on the possible termination of ongoing investment mandates.
After developing a simple model of wealthconstrained professional arbitrage that departs in
several important aspects from the canonical Shleifer
and Vishny (1997) model, we calibrate our model
to illustrate its qualitative features. We show that the
existence of professional arbitrageurs mitigates—
but cannot eliminate—mispricing in the market
relative to intrinsic values, regardless of how sensitive the outside investors are to arbitrageurs’ past
performance in deciding whether to remain invested
with them. We also show that arbitrage dampens
the unconditional volatility of asset returns, which
we measure as the expected value of squared returns.
Most importantly, the presence of arbitrageurs limits
both the degree of increased mispricing and level
of volatility during a financial crisis, which we define
as a period of heightened volatility and acute shortage of liquidity.3 This result points out that professional arbitrageurs tend to stabilize markets even
when they are wealth-constrained. Other papers
show that investors who use “positive feedback”
trading strategies—such as portfolio insurers—tend
to destabilize markets (Grossman and Zhou, 1996).
We analyze a three-date (two-period) model of
an aspiring professional arbitrageur (or “convergence
trader” in the language of Kyle and Xiong, 2001,
and Xiong, 2001) who must obtain financing from
investors less informed than he is about the intrinsic value of a financial asset—that is, its liquidation
value at the end of the second period. In addition to
these two types of individuals, there are noise traders
who have wealth to invest but who misperceive the
asset’s intrinsic value. It is the noise traders who
drive the asset’s price away from the intrinsic value.
The investors provide the arbitrageur with funds
to invest in an underpriced asset at the outset of the
model. The price is observed again at the end of the
20

N OV E M B E R / D E C E M B E R 2 0 0 2

first period, at which time the investors may “roll
over” their funds with the arbitrageur or demand
their money back if they have lost confidence in his
ability. The asset will assume its intrinsic value at the
end of the second period with certainty, although
only the arbitrageur knows in advance what that
value is. Consequently, the two-period return on the
arbitrageur’s private information would be both positive and risk-free if he could be assured of financing.
Our set-up highlights the fact that a two-period
risk-free arbitrage nevertheless can be risky over a
one-period horizon in the presence of noise traders
and financial constraints on the arbitrageur. The
risk arises because the arbitrageur needs outside
investors, and these outside investors might revise
their beliefs about the arbitrageur’s talent at the
interim date, based on the return the arbitrageur
achieved in the first period. If the investors downwardly revise their beliefs about the arbitrageur’s
abilities because the fund lost money due to a deepening of the mispricing, they might withdraw their
money precisely when the expected return on the
arbitrage is at its maximum. One implication is that
the arbitrageur will invest “strategically”—that is,
he will not invest as much initially as he would in a
world without wealth constraints—in order to hedge
against the possibility of being unable to exploit
even greater mispricing should it occur one period
ahead. Of course, this is not a new finding; for papers
with similar results, see Grossman and Vila (1992),
Shleifer and Vishny (1997), or Gromb and Vayanos
(2001).4 Our paper’s contributions in this respect
3

Myron Scholes (2000) suggests that the global financial crisis of 199798 was characterized by an increase in volatility, especially in equity
markets, and a flight to liquidity (that is, a preference by many investors
for assets whose liquidity was expected to be good). The crisis was
accentuated by the “negotiated bankruptcy” of Long-Term Capital
Management (LTCM), a hedge fund in which Scholes himself was a
partner. According to Scholes, prior to the crisis, LTCM “was in the
business of supplying liquidity” and therefore its demise worsened
the crisis by eliminating the liquidity it had been supplying. A theoretical
analysis relevant to this episode is Xiong (2001).

4

The first rigorous investigations of the multi-period investment problem
were Merton (1971, 1973) and Breedon (1979). Merton concluded that
a trader should keep a constant fraction of his wealth invested in the
risky asset at all times. The fraction depends on the asset’s expected
return and risk and the investor’s degree of risk aversion. Grossman
and Vila (1992) added leverage and solvency constraints to the dynamic
trader’s problem. Their trader optimally commits more wealth to the
risky asset the shorter is the investment horizon and the further from
the leverage constraint (not just today but prospectively in the future)
the trader finds himself. Campbell and Viceira (1999) is a recent examination of the problem under the assumption that the investor is aware
that the probability distributions from which asset returns are drawn
change over time.

FEDERAL RESERVE BANK OF ST. LOUIS

Emmons and Schmid

Figure 1
Timeline of the Model
t1

t2

t3

Noise traders’
misperception, S1 ,
in place.

Noise traders’
misperception
might deepen to S 2 .

Noise traders’
misperception
corrects (S3 = 0).

Arbitrageur allocates
financial resources
to asset and cash.

If asset price reverts
to intrinsic value, V,
arbitrageur liquidates.

Hedge fund
winds down.

are a more realistic objective function for the arbitrageur and a set-up in which the arbitrageur’s trading significantly affects the asset’s price. Our model
generates interior solutions and we provide calibrated illustrations of the model’s results. While
Shleifer and Vishny (1997) assume that the arbitrageur maximizes assets under management, we
assume that he maximizes his income. The arbitrageur’s income is determined by an incentive
scheme that resembles real-world contracts of hedge
fund managers.

THE MODEL
There are three types of agents in the model.
Noise traders have wealth but misperceive the intrinsic value of a financial asset. Professional arbitrageurs
have no wealth or borrowing capacity but know the
intrinsic value of the financial asset. Investors have
wealth but no insight into the financial asset’s intrinsic value. Unlike noise traders, investors know that
they cannot recognize the asset’s intrinsic value.
All parties are risk-neutral.
The investors may provide the arbitrageur with
funds to invest in an underpriced asset at the outset
of the model (see Figure 1). We refer to this arrangement as a hedge fund. Noise traders misperceive
the intrinsic value of at least one financial asset in
the economy, which generates arbitrage opportunities that so-called “long-short” investment strategies
seek to exploit. Asset mispricing implies that there
are relatively overpriced and relatively underpriced
assets, which means that a portfolio that is long on
the relatively undervalued asset and short on the
relatively overvalued asset trades below intrinsic
value.
We treat a market-neutral long-short portfolio
as a single, complex financial asset. Arbitrage is the

process of acquiring a long-short portfolio and holding it until its price returns to the portfolio’s intrinsic
value. The long-short portfolio that any arbitrageur
might hold defines a market segment of a larger
arbitrage industry. We assume that arbitrageurs are
highly skilled people who pursue proprietary trading
strategies and therefore enjoy a monopoly in their
segment. For simplicity only, we make the assumption that the operating costs in the arbitrage industry
are zero.
The risk-free rate of return, and therefore the
opportunity cost of capital, is zero. For simplicity,
we assume that risky assets trading at fair value—
including the stock market index—also have an
expected return of zero. This implies that there are
no priced systematic risk factors in the economy,
that is, there is no equity risk premium.
The asset trades at three moments in time, t
(t=1,2,3). We capture the influence of the noise
traders’ misperceptions of the intrinsic value of the
asset at times t1 and t2 with the parameters S1 and
S2, respectively. There is no fundamental risk in the
model because the price of the asset will revert to
the intrinsic value at a known date (t3) with certainty
(so S3=0).
The supply of the financial asset is unity. Noise
traders’ demand for the financial asset at time t
(t=1,2,3) is expressed as
V − St
(1)
, 0 ≤ St<V,
QNt =
pt
where pt is the price of the financial asset and St is
the misperception of the noise traders about the
intrinsic value of the financial asset. Because the
financial asset in question is a long-short portfolio
whose value is underestimated by the noise traders,
the noise traders demand less than one unit of the
N OV E M B E R / D E C E M B E R 2 0 0 2

21

REVIEW

Emmons and Schmid

financial asset. Without misperception (S=0), the
noise traders would be willing to absorb the unit
supply of the asset or, in other words, the asset would
trade at the intrinsic value ( pt=V ).
The arbitrageur is compensated in two ways in
accord with actual practice—via an up-front “management fee” and an after-the-fact performancebased “incentive fee.”5 At the beginning of each
period, he receives a fraction (α ) of the assets under
management, and at the end of the period he receives
a fraction ( β ) of any positive return on the portfolio.
This corresponds to compensation structures in
real-world hedge funds, where managers typically
collect α=1 percent or α=2 percent of the equity
capital, plus β=20 percent of any positive return
on the fund’s equity. We assume that the arbitrageur
invests his entire fee income in the fund. This is
because he recognizes the profitability of the fund’s
activities.
The variable Ft denotes the total financial
resources available to the arbitrageur at time t
(t=1,2,3). The value of F1 is exogenous, while the
quantities F2 and F3 are determined in the model.
The startup capital, F1, is provided solely by the
investors, while the arbitrageur acquires the share
α in F1 immediately as part of his compensation.
The arbitrageur acquires additional equity at t2 in
the amount of a fraction α of the outsiders’ share
in F2. Furthermore, the arbitrageur acquires equity
in the fund through capital gains on his equity
position and through his share β in the capital
gains on the outsiders’ equity. The quantity F3 is the
fund’s liquidation value. Note that the arbitrageur
is both the general equity partner of the fund and
its manager, receiving compensation from outside
investors (limited partners) according to the fee
schedule described above.
We assume that the fund raises equity capital
only at the outset—at t1. This assumption prevents
the arbitrageur from diluting initial investors’ equity
stakes later on. Remember that the arbitrageur’s
compensation depends not only on the return on
but also on the amount of the outsiders’ equity
capital under management. The arbitrageur therefore might have an incentive to raise fresh capital
at t2, particularly if he expects low returns in the
second period. This would dilute the fund’s existing
investors’ equity stakes. Thus, we assume (in keeping
with typical hedge-fund arrangements) that the fund
closes to new and existing investors after raising
the initial capital. Reinvested capital gains are conse22

N OV E M B E R / D E C E M B E R 2 0 0 2

quently the sole source of additional equity capital
in the second period.
At time t2, the price of the asset either reverts
to V or it does not. If the asset price is V at t2, the
arbitrageur liquidates the fund and holds cash until
t3. If the asset price does not equal V at t2, the arbitrageur invests aggressively—albeit not all of the
fund’s cash—in the underpriced asset. This portfolio
then generates a risk-free return because the asset
price rises to V at t3 with certainty.
The arbitrageur’s (that is, the hedge fund’s)
demand for the asset at the interim date, t2, is given
by
D
QA2 = 2 , 0 ≤ D2 ≤ F2 ,
(2)
p2
where D2 is the amount of the hedge fund’s demand
in dollars. The amount F2 – D2 ≥ 0 is held in cash.
Because total demand aggregated across noise
traders and the arbitrageur must equal the asset
supply of one unit (QN2+QA2=1), the price of the
financial asset at t2 is determined by combining (1)
and (2):
(3)

p2 = V − S2 + D2 , 0 ≤ D2 < S2 .

The condition D2<S2 implies that the asset still
trades at a discount to the intrinsic value at t2: p2<V.
This assumption recognizes the arbitrageur’s incentive not to bid up the price all the way to intrinsic
value immediately—an implication of the fact that
the arbitrageur will be compensated during the last
period for achieving a positive return on investment.
As shown by Grossman and Vila (1992), the
arbitrageur does not want to invest all of F1 in the
asset at t1, either. After all, the asset may become
even more underpriced at t2, in which event he will
want to increase his investment (“double up”). With
D1 denoting the amount the arbitrageur invests in
the asset at t1, we have
D
QA1 = 1 ,
(4)
p1
which implies the initial asset price will be
5

Amin and Kat (2001) report that hedge funds typically charge their
investors a fixed annual “management fee” of 1 or 2 percent of assets
under management plus an “incentive fee” of 15 to 25 percent of the
fund’s realized annual return. The incentive fee is waived if a particular
“hurdle rate” has not been achieved, which can be a fixed number or
a reference rate such as the T-bill rate plus or minus a spread. Most
funds also apply a “high-water mark” provision, which requires the
fund to make up any past losses before the incentive fee is paid. We
assume the hurdle rate is zero, and we exclude any high-water mark
for the sake of simplicity.

FEDERAL RESERVE BANK OF ST. LOUIS

(5)

p1 = V − S1 + D1, D1 < S1.

The condition D1<S1 implies p1<V, which again
captures the fact that the arbitrageur will not bid
the price all the way up to the asset’s intrinsic value
because of the incentives built into his compensation schedule.
The investors have prior beliefs about the arbitrageur’s talent in exploiting possible asset mispricing, but are not perfectly informed. Investors update
their beliefs about the arbitrageur’s talent using a
simple Bayesian learning rule, which is based solely
on the arbitrageur’s past performance. When past
returns are poor, investors don’t know for sure
whether the poor returns are due to a random error
(noise), a deepening of noise trader misperception
(bad luck), or truly inferior investment talent. Pulling
some of their money from the hedge fund after the
asset mispricing has deepened—that is, when the
expected return on the long-short portfolio is
highest—is the investor’s rational response to the
problem of inferring the arbitrageur’s (unobservable)
talent from data that are ambiguous (that is, observationally equivalent under more than one possible
economic structure).
The investor’s rule of updating his beliefs about
the arbitrageur’s talent implies that, if the hedge
fund loses money during the first period, the fund
faces withdrawals at the interim date, t2. Specifically,
we assume that the withdrawals at t2 are a multiple
of the hedge fund’s posted gross return (that is,
before management fees) at t2, denoted R2, should
this return be negative. Remember that, while
investors can withdraw capital, they cannot inject
additional funds. Thus, the supply of funds in the
second period is the following 6:
(6)

Emmons and Schmid

The arbitrageur knows that—despite a temporary
deepening of the mispricing—the price of the asset
will revert to intrinsic value at t3 for certain, so he
will keep his own money invested, come what may.
Our multiplicative feedback rule provides the
arbitrageurs with what may be a more realistic incentive structure than the linear feedback rule in Shleifer
and Vishny (1997). Our feedback rule does not penalize small negative returns quite as severely for a
high degree of responsiveness, γ, as is the case in
Shleifer and Vishny. For a responsiveness coefficient
of γ =5, for instance, a gross return in the first period,
R2, of –1 percent reduces the fund’s equity capital
by approximately 4 percentage points. A 5 percent
loss, on the other hand, leaves the fund with approximately 77 percent of its equity capital at the beginning of the next period. We provide more results
from the model below.
The gross return of the hedge fund in the first
period, R2, is given by
p
p − p1
( F1 − D1 ) + D1 ⋅ 2 − F1 D1 ⋅ 2
p1
p1
R2 =
=
.
(7)
F1
F1
The fund’s first-period return consists of its return
on the financial asset, normalized by the total funds
available for investment.
For simplicity, we assume a specific form of
uncertainty about noise trader sentiment at t2, S2.
With probability 1 – q (0<q<1), noise traders recognize the true value of the asset, which implies S2=0.
In this case, the arbitrageur liquidates at t2 and holds
cash until t3. Then the arbitrageur’s assets under
management at t3 would amount to
(8)
where

 F ⋅ α ⋅ (1 + R ) + F ⋅ (1 − α ) ⋅ (1 + R )γ ,γ > 1, if − 1 ≤ R < 0
2
1
2
2
F2 =  1
⋅
1
+
,
if
0
,
F
R
R
>
2)
2
 1 (

where γ is a parameter that determines the responsiveness of the investor to past performance. For
γ =1, poor first-period returns do not shake the
confidence of investors in the arbitrageur’s talent.
At the other extreme, responsiveness that becomes
unboundedly large implies that even a small firstperiod loss is multiplied into a huge withdrawal of
funds. Note that the outside investors may withdraw
only what is theirs. This means that, even if the outsiders pull all of their money, the arbitrageur’s equity
stake remains and the fund can stay in business.

F3S2 = 0 = F2S2 =0 ≡ F1 ⋅ (1 + R2S2 = 0 ),

R2S2 = 0 =

( F1 − D1 ) + D1 ⋅
F1

V
− F1
p1

.

On the other hand, noise trader misperception
deepens to S2 with probability q, S2=S>S1(>0). If
noise traders continue to misperceive the intrinsic
value of the asset, the hedge fund’s assets at t3 will
amount to the following:
6

Some hedge funds have “lock-up” periods of one to three years, while
others allow investors to withdraw money with only a few weeks’
notice. As a result of the poor quality of investors’ information about
the arbitrageur’s talent, the arbitrageur’s past performance often is a
major determinant of the resources he receives to manage, regardless
of the actual arbitrage opportunities available to him.

N OV E M B E R / D E C E M B E R 2 0 0 2

23

REVIEW

Emmons and Schmid

F3S2 = S =

(9)

=
where
R2S2 = S

V
p2S2 = S
V
p2S2 = S

misperception deepen in the first period, while
MF2S2=0 is the fee income if the asset price reverts
to intrinsic value. The arbitrageur also captures capital gains on the equity he builds from the reinvested
management fees. The expected value of the capital
gains, CG, equals

⋅ D2 + ( F2S2 = S − D2 )
⋅ D2 + F1 ⋅ (1 + R2S2 = S )γ − D2 ,

(14)

p S2 = S
− F1
( F1 − D1 ) + D1 ⋅ 2
p1
=
.
F1

CG = q ⋅ ( R2S2 = S ⋅ MF1 ⋅ [1 + R3S2 = S ] + R3S2 = S ⋅ MF2S2 = S )
+ (1 − q ) ⋅ R2S2 = 0 ⋅ MF1 .

THE ARBITRAGEUR’S OPTIMIZATION
PROGRAM
The arbitrageur’s total income consists of
management fees and capital gains on reinvested
management fees. The expected value of the management fees, MF, equals the sum of the expected
values of the management fees collected at t1(MF1),
at t2(MF2), and at t3(MF3). The expected value of the
capital gains is CG. The arbitrageur’s maximization
problem therefore is
Max {MF1 + MF2 + MF3 + CG},

(10)

D1 , D2

where the management fees are
MF1 = α ⋅ F1,

(11)

MF2 = MF2S2 = S + MF2S2 = 0, and

(12)
(13)

MF3 = β ⋅ q ⋅ R3S2 = S
⋅( F2S2 = S − (1 + R2S2 = S ) ⋅ MF1 − MF2S2 = S )

and where
MF2S2 = S = α ⋅ q ⋅ ( F2S2 = S − α ⋅ [1 + R2S2 = S ] ⋅ F1
− β ⋅ max{0, R2S2 = S} ⋅ (1 − α ) ⋅ F1 )
+ β ⋅ q ⋅ max{0, R2S2 = S} ⋅ (1 − α ) ⋅ F1 ,
MF2S2 = 0 = α ⋅ (1 − q ) ⋅ ( F2S2 = 0 − α ⋅ [1 + R2S2 = 0 ] ⋅ F1
− β ⋅ R2S2 = 0 ⋅ (1 − α ) ⋅ F1 )
+ β ⋅ (1 − q ) ⋅ R2S2 = 0 ⋅ (1 − α ) ⋅ F1 , and

R3S2 = S =

( F2S2 = S − D2S2 = S ) + D2S2 = S ⋅
F2S2 = S

V
p2S2 = S

− F2S2 = S
.

The quantity MF2S2=S represents the income the
arbitrageur collects at t2 should the noise traders’
24

N OV E M B E R / D E C E M B E R 2 0 0 2

The arbitrageur’s choice variables are D1( ≤ F1 )
and D2( ≤ F2S2=S ), which are the amounts the arbitrageur invests in the asset at t1 and t2, respectively.
Unless the asset reverts to intrinsic value at t2( p2=V ),
the t2 price of the asset given in equation (3) is a
function of the t2 choice variable, D2. Similarly, the
t1 price of the asset given in equation (5) is a function
of the choice variable, D1.

SOLUTION TO THE MAXIMIZATION
PROBLEM
We solve the maximization problem numerically. We hold constant all of the following: V=1;
F1=S1=0.2; S2=0.4; q=1– q=0.5; α=0.02; and
β=0.2. Note that F1=S1=0.2 means that the arbitrageur has sufficient buying power to eliminate the
t1 mispricing entirely if so desired. Also, note that
0.4=S2>S1=0.2 means that noise trader misperception may deepen between t1 to t2—that is, the asset
may become even more mispriced. For the values
chosen for S1, S2, and q, noise trader misperception,
S, is as likely to double as it is to vanish. Thus, the
expected value of noise trader misperception in the
second period, q·S2, equals the noise trader misperception observed in the first period, S1.
We vary γ , the responsiveness to past performance of fund withdrawals, from γ=1 (no responsiveness by the investors to past investment performance,
that is, no withdrawals) to γ =20 (extreme responsiveness) with a step length of unity. We use a grid
search method to solve the maximization problem.
This involves varying D1 and D2 independently in
very small increments within their bounds, 0 ≤ Di ≤
Fi ( i=1,2), to find the maximum of the objective
function.
The findings of the grid search are displayed
in Figures 2 through 5. The first important point to
make concerns the extent to which the presence
of the hedge fund affects asset mispricing. Figure 2
shows that the mispricing is less pronounced in each
period than it would be without the hedge fund.

FEDERAL RESERVE BANK OF ST. LOUIS

Emmons and Schmid

Figure 2

Figure 3
Effect of Investor Responsiveness on
Asset Price Volatility

0.90

0.090

0.88

0.085
Unconditional Volatility

Asset Prices p1, E(p2)

Effect of Investor Responsiveness on
Asset Prices

0.86
0.84
p1

0.82

E(p2)

0.080
0.075
0.070
0.065

0.80

0.060

0.78
0

2

4

6

8

10

12

14

16

18

20

Responsiveness of Cash Flows (γ )

Remember that, without arbitrage, the first-period
price, p1, and the expected value of the second-period
price, E[p2], both would equal 0.8 (shown as a dashed
line). On the other hand, without noise traders, the
asset would trade at unit value in both periods (not
shown). The hedge fund almost halves the difference
between the expected value of the second-period
price, E[ p2] (shown as solid circles), and the asset’s
intrinsic, unit value. In fact, the degree of investor
responsiveness, γ, has little bearing on E[ p2], which
approaches the value of approximately 0.8873
(shown as a solid horizontal line) as γ approaches
infinity. By comparison, the degree of responsiveness
has a strong impact on the first-period price, p1
(shown as open boxes). This is because the arbitrageur treads even more cautiously when putting on
this trade in the first period when he knows that the
investors penalize negative returns with sizeable
withdrawals. In fact, the higher is γ, the more cash
the arbitrageur holds in the first period, and therefore,
the lower is p1. As the degree of investor responsiveness, γ, goes to infinity, the amount the arbitrageur
invests in the first period goes to zero and, consequently, the first-period price, p1, converges to 0.8—
the value the asset would adopt if there were no
hedge fund in the market (shown as a dashed line).
Thus we conclude that the hedge fund pushes the
price of the asset (or its respective expected value)
toward the intrinsic, unit value in both periods. This
is our first main finding.

0

2

4

6

8

10

12

14

16

18

20

Responsiveness of Cash Flows (γ )

Figure 3 shows the unconditional volatility of
the asset’s returns for various degrees of investor
responsiveness, γ. The unconditional volatility is
calculated as the expected value of the squared
returns over the two periods. For low values of
investor responsiveness, volatility increases as γ
increases. For high values of responsiveness, a further increase in γ reduces volatility monotonically.
As γ goes to infinity, volatility approaches a level
(as shown by the solid line) that is lower than the
volatility level at γ =1 (as indicated by the leftmost
symbol), which is the benchmark case of unwavering investor confidence in the hedge fund manager.
The reason for this “volatility hump” lies in the existence of two opposite effects. All else equal, the
higher γ is, the bigger is the drop in the asset’s price
from t1 to t2 should the noise traders’ misperception
deepen. On the other hand, the higher γ is, the lower
is the price of the asset at t1 because the arbitrageur
puts less money to work. For low values of investor
responsiveness, the volatility-increasing effect dominates. For increasingly higher values of γ, this effect
becomes progressively weaker until it vanishes for
an infinitely large degree of investor responsiveness.
It is important to note that the hedge fund
greatly reduces asset price volatility, regardless of
the degree of investor responsiveness. The unconditional volatility without the hedge fund runs at
0.5694 (not shown), which is a multiple of the volatility that we observe even at the degree of responsiveN OV E M B E R / D E C E M B E R 2 0 0 2

25

REVIEW

Emmons and Schmid

Figure 4

Figure 5

Effect of Investor Responsiveness on
Asset Return When Misperception
Deepens

Effect of Investor Responsiveness on
Arbitrageur’s Profit
0.0142

Arbitrageur’s Expected Profit

First-Period Asset Return for S=S2

–0.025

–0.050

–0.075

–0.100

0.0140
0.0138
0.0136
0.0134
0.0132

0.0130
0

–0.125
0

2

4

6

8

10

12

14

16

18

20

2

6

4

8

10

12

14

16

18

20

Responsiveness of Cash Flows (γ )

Responsiveness of Cash Flows (γ )

ness that generates the highest level of volatility.
Thus, we conclude that the hedge fund unambiguously reduces unconditional volatility. This is our
second main finding.
Another way to look at the impact of arbitrage
on volatility is to ask how the market behaves when
asset mispricing deepens. Such an event—if severe—
might cause, or occur alongside, a financial crisis.
Figure 4 shows, for the case of a deepening noise
trader misperception of the asset’s intrinsic value,
the first-period asset return as a function of investor
responsiveness. The absolute value of the percentage
decline of the asset price increases with investor
responsiveness, γ. For an infinitely high value of γ,
the arbitrageur holds cash in the first period and
then invests aggressively at t2, although he does not
invest all the cash available. The horizontal line in
Figure 4 signifies the first-period return for this
borderline case of an infinite degree of responsiveness. Note that, without a hedge fund, the first-period
return would amount to a negative 25 percent (not
shown), which is more than twice as much (in absolute value) as what is observed even with a degree
of responsiveness of zero (that is, γ equal to one).
Hence, we conclude that the presence of a hedge
fund dampens volatility in the event of a deepening
of noise trader misperception, as might occur in a
financial panic. This is our third main finding.
Finally, we are interested in the question of how
investor responsiveness affects the arbitrageur’s
26

N OV E M B E R / D E C E M B E R 2 0 0 2

profit, that is, his incentive to set up a hedge fund
and engage in arbitrage. Figure 5 shows the arbitrageur’s profit as a function of γ. Not surprisingly,
the profit of the arbitrageur decreases monotonically
with increased investor responsiveness to past performance. The monotonic decline in the profitability
of arbitrage with increasing investor responsiveness
to past performance is a manifestation of the fact
that liquidating a hedge portfolio when the expected
return from arbitrage is highest is counterproductive—that is, it runs against “the nature of the trade.”

CONCLUSION
Even financially constrained professional
arbitrageurs may be able to exploit asset mispricing
if they can link up with rational but uninformed
investors. To achieve this goal, the two parties must
overcome—at least to a degree—the problem of
asymmetric information about the arbitrageur’s
talent. The result of such an endeavor is a hedge
fund that goes long on (comparatively) underpriced
assets and short on (comparatively) overpriced assets.
As a byproduct, the impacts of noise trader misperceptions on asset prices and volatility are reduced.
This holds for any degree of responsiveness to past
performance (“feedback”) of the investors’ confidence in the arbitrageur’s talent.
This article builds on the dynamic-investment
literature that reaches back at least to Merton (1971).

FEDERAL RESERVE BANK OF ST. LOUIS

Shleifer and Vishny (1997) provided an insightful
model of wealth-constrained arbitrageurs that can
be, and has been, extended in several directions.
We add several realistic features to the professional
arbitrageur’s problem in the canonical model, including the ability to build an equity stake in his hedge
fund over time, and a potentially more realistic multiplicative (rather than linear) investor feedback rule.
Like Shleifer and Vishny, we assume that the hedge
fund can influence the market price. Hedge funds
do, in fact, sometimes move market prices because
they operate in specialized market segments that
have limited liquidity. It is also true, however, that
hedge funds alone cannot prevent asset-price volatility or occasional mispricing—which might deepen
before it eventually corrects.

Emmons and Schmid

Working paper, London Business School and Massachusetts
Institute of Technology, October 2001.
Grossman, Sanford J. and Vila, Jean-Luc. “Optimal Dynamic
Trading with Leverage Constraints.” Journal of Financial
and Quantitative Analysis, June 1992, 27(2), pp. 151-68.
___________ and Zhou, Zhongquan. “Equilibrium Analysis
of Portfolio Insurance.” Journal of Finance, September
1996, 51(4), pp. 1379-1403.
Keynes, John Maynard. The General Theory of Employment,
Interest, and Money. New York: Harcourt, Brace & World,
1936.
Kyle, Albert S. and Xiong, Wei. “Contagion as a Wealth
Effect.” Journal of Finance, August 2001, 56(4), pp. 1401-40.

REFERENCES
Amin, Guarav S. and Kat, Harry M. “Hedge Fund Performance
1990-2000: Do the ‘Money Machines’ Really Add Value?”
Working paper, University of Reading (UK), 15 May 2001.
Breedon, Douglas T. “An Intertemporal Asset Pricing Model
with Stochastic Consumption and Investment Opportunities.” Journal of Financial Economics, September 1979,
7(3), pp. 265-96.
Campbell, John Y.; Lo, Andrew and MacKinlay, A. Craig.
The Econometrics of Financial Markets. Princeton, N.J.:
Princeton University Press, 1997.

Merton, Robert C. “Optimal Consumption and Portfolio
Rules in a Continuous-Time Model.” Journal of Economic
Theory, December 1971, 3(4), pp. 373-413.
___________. “An Intertemporal Capital Asset Pricing Model.”
Econometrica, September 1973, 41(5), pp. 323-61.
Muth, J. “Optimal Properties of Exponentially Weighted
Forecasts.” Journal of the American Statistical Association,
1960, 55, pp. 299-306.
Samuelson, Paul. “Proof That Properly Anticipated Prices
Fluctuate Randomly.” Industrial Management Review,
1965, 6, pp. 41-49.

___________ and Viceira, Luis M. “Consumption and
Portfolio Decisions When Expected Returns Are Time
Varying.” Quarterly Journal of Economics, May 1999, 114(2),
pp. 433-95.

Scholes, Myron S. “Crisis and Risk Management.” American
Economic Review Papers and Proceedings, May 2000, 90(2),
pp. 17-21.

Fama, Eugene F. “The Behavior of Stock Market Prices.”
Journal of Business, 1965, 38, pp. 34-105.

Sharpe, William P. and Alexander, Gordon. Investments.
4th Edition. Englewood Cliffs, N.J.: Prentice Hall, 1990.

Friedman, Milton. “The Case for Flexible Exchange Rates,”
in Essays in Positive Economics. Chicago: University of
Chicago Press, 1953.

Shleifer, Andrei and Vishny, Robert W. “The Limits of
Arbitrage.” Journal of Finance, March 1997, 52(1), pp. 35-55.

Gromb, Denis and Vayanos, Dimitri. “Equilibrium and Welfare
in Markets with Financially Constrained Arbitrageurs.”

Xiong, Wei. “Convergence Trading with Wealth Effects: An
Amplification Mechanism in Financial Markets.” Journal
of Financial Economics, November 2001, 62(2), pp. 247-92.

N OV E M B E R / D E C E M B E R 2 0 0 2

27

Emmons and Schmid

28

N OV E M B E R / D E C E M B E R 2 0 0 2

REVIEW

Regime-Dependent
Recession Forecasts
and the 2001 Recession
Michael J. Dueker
usiness recessions, as a major source of
nondiversifiable risk, impose high costs on
society. Since firms cannot obtain “recession
insurance,” they try to foresee recessions and reduce
their exposure ahead of time. Consequently, forecasting business cycle turning points has remained
an important endeavor. Of course, the difficulty is
deriving reliable methods to forecast business cycle
turning points. Previous studies found that accurate
recession forecasts remain elusive (Filardo, 1999;
Del Negro, 2001; Chin, Geweke, and Miller, 2000).
Forecasts of economic output are not a good substitute for forecasts of business cycle turning points
because less than 20 percent of all months pertain
to recessions. Hence, empirical models of output
growth focus largely on explaining variation in
output growth during economic expansions, since
this variation accounts for the lion’s share of the
sample variance.
Throughout the 1990s, recession forecasting
models relied exclusively on the 1990-91 recession
for out-of-sample confirmation (Estrella and Mishkin,
1998; Birchenhall et al., 1999; Friedman and Kuttner,
1998). Out-of-sample confirmation is particularly
important for recession forecasting because recessions are infrequent, making it tempting to overfit
specific episodes in sample. In general, recession
forecasting models failed to predict the 1990-91
recession out of sample. The occurrence of the 2001
recession raises the question: Was the 1990-91 recession uniquely difficult to predict or is recession forecasting a failed enterprise? If recession forecasting
models were to repeat in 2001 their dismal performance from the 1990-91 recession, then doubts
about such models would mount with justification.
In this article, I examine the out-of-sample forecasts
from recession forecasting models with three levels
of sophistication. All three models concur with the
previous finding that the 1990-91 recession was

B

Michael J. Dueker is a research officer at the Federal Reserve Bank of
St. Louis. Mrinalini Lhila and John Zhu provided research assistance.

© 2002, The Federal Reserve Bank of St. Louis.

hard to predict. The simplest of the three models
largely misses the 2001 recession, but two regimeswitching models come quite close to predicting
the onset date of the recession six months ahead
of time. One innovation to recession forecasting
introduced here is to allow the critical probability
level for a recession to be predicted to depend on
the current state of a Markov switching process—
hence, regime-dependent recession forecasts.
In this way, the forecasts presented here respond
to the criticism that economists equivocate too much
when it comes to their recession forecasts. When
recession forecasts are expressed as probability
statements, it is tempting to claim ex post that the
ex ante probability of recession from the forecasting
model was “close enough” to either one or zero to be
considered a correct forecast. For example, if the
model suggested that a recession would occur with
a probability of 35 percent, then after the fact the
model builder could try to justify either outcome:
If a recession ensued, the model builder could cite
the jump in probability from zero to 35 percent as
evidence that a large shift toward recession was
detected; if no recession ensued, the model builder
could say that the 35 percent probability was far from
100 percent and did not indicate recession. To avoid
such ambiguity, economists are often asked to make
specific calls as to whether the economy will or will
not be in recession six months from now. A yes/no
recession signal comes from comparing the forecasted probability of recession with a critical value
determined prior to the out-of-sample forecasting.
The dependent variable I use to separate business
recessions from expansions is based on the business
cycle turning points defined by the National Bureau
of Economic Research (NBER). As in Birchenhall et al.
(1999), I use the composite index of leading indicators (CLI) from the Conference Board as the explanatory variable in the probit forecasting models. The
CLI receives much attention as a harbinger of future
business cycle conditions. It has ten components:
manufacturing hours, consumer expectations, stock
prices, initial unemployment claims, building permits,
the money supply, the spread between short- and
long-term interest rates on government securities,
vendor performance, manufacturing orders for
consumer goods, and manufacturing orders for
capital goods. Figure 1 shows the upward trend in
the CLI and its three-month percentage changes.
The decreased volatility in the three-month change
since the early 1980s is symptomatic of decreased
volatility in the business cycle.
N OV E M B E R / D E C E M B E R 2 0 0 2

29

REVIEW

Dueker

Figure 1

THE SIMPLE PROBIT MODEL

Index of Leading Indicators

I study three probit forecasting methods. One is
the simple probit model and the other two are based
on a Markov switching probit from Dueker (1997).
A probit model can be used to predict a qualitative
variable such as a recession indicator, Rt, where

Index
120

3-Month Change in Index
20

100

15

80

10

60

5

40

0

20

–5

0
1959

–10
1965

1971

1977

1983

1989

1995

2001

The 1990-91 and 2001 recessions can be used
to evaluate the out-of-sample performance of the
model and the leading indicators. I find that the
1990-91 recession is the anomaly in that the recession signal emanating from the leading indicators
largely misses it. The 2001 recession, in contrast,
was largely predictable from the leading indicators
six months ahead.
The probit model takes the monthly movements
in the leading indicators and translates them into
precise probability statements regarding the likelihood of recession. When dealing with qualitative
events such as recessions, however, it is often desirable to define a specific recession yes/no signal from
the probit probabilities. That way, the forecasting
method either correctly “calls” the recession or it
does not. Birchenhall et al. (1999) use this approach
to say that a recession is signaled if the logit probability of recession exceeds a critical value. If one
chooses the critical value to maximize the in-sample
fit, then out-of-sample confirmation of recession
forecasting models becomes particulary important,
given Robert Lucas’s dictum: Beware of econometricians bearing free parameters. The critical value that
defines a recession signal is an example of what
Lucas calls a free parameter because—although it
is not an inherent part of the econometric model—
it combines with the model estimates to suggest that
the model fits the data well. Out-of-sample confirmation helps ensure that the free parameters are not
simply overfitting the in-sample data.
30

N OV E M B E R / D E C E M B E R 2 0 0 2

Rt=1 if the economy is in NBER recession in period t
=0 otherwise.
One way to think of a probit model is to assume that
a normally distributed latent variable, R*, lies behind
the recession indicator:
(1)

Rt* = − c0 + − c1 Xt − k + ut ,

where u is a normally distributed error team and X
is the leading indicator explanatory variable lagged
k periods—the forecasting horizon. A probability
of recession is associated with each possible value
of the latent variable, where the latent variable is
assumed to be negative during expansions and positive during recessions. In this case, the forecasted
probability of recession is
(2)

Prob ( Rt =1) = 1 − Φ (c0 + c1 Xt − k ) ,

where Φ(.) is the cumulative standard normal density
function.
The log-likelihood function for a simple probit
model is
(3)

(

L = ∑ Rt lnProb Rt=1 Xt − k
t

(

)

)

+ (1 − Rt ) ln Prob Rt=0 Xt − k .

COEFFICIENT VARIATION VIA MARKOV
SWITCHING IN THE PROBIT MODEL
The log-likelihood function in equation (3) highlights the assumption in the probit model that the
recession outcomes conditional on available information are independently distributed from month
to month. This assumption is questionable unless
the econometric model allows for considerable
serial correlation in the recession probabilities. As
in Dueker (1997), one way to achieve this degree of
serial correlation is to introduce serial correlation
in the model’s coefficients by making them subject
to Markov switching.
The simplest interpretation of Markov switching
coefficients in a probit model is that they capture
time variation in the variance of the error term u
from equation (1). In the low-variance regime the
variance would still be normalized to one, but in

FEDERAL RESERVE BANK OF ST. LOUIS

Dueker

(

the high-variance regime the variance would be
greater than one:
Prob ( Rt =1| High Variance )

Prob St = 0 Rt −1 , Xt −2

= 1 − Φ (c0 / σ + c1 / σ Xt − k ), σ > 1 .

St equals 0 or 1
Prob (St=0|St–1=0)=p
Prob (St=1|St–1=1)=q.

(4)

In this way, the coefficients take on either of two
values and thereby change the magnitude of the
shock needed to induce a recession:
(5)

Rt=1 if ut >c0 ( St ) + c1( St ) + c1( St ) Xt − k

(6)

(
)
Prob ( S = 0 R , X ) Prob ( R = 0 S = 0, X )
∑1 Prob ( S = s R , X ) Prob ( R = 0 S = s, X )
Prob St = 0 Rt = 0, Xt −1 =
t −1

s=0

t

t −2

t −1

t

t −2

t −1

t

t

t

t −1

)

The probability in equation (7) is called the oneperiod-ahead prior probability because it is not conditional on the recession outcome at time t. For a
forecast horizon of several months, we need to use
the transition probabilities to derive a k-period-ahead
probability of the state variable:
(8)

(St = 0 R

Prob

 Prob


(

)

′
Xt − k ,
=
St = 1 Rt − k , Xt − k 
t −k ,

(
(

)

)
)

Prob S = 0 R , X
′
t −k
t −k
t −k ,
 ,
G
Prob S = 1 R , X 
t −k
t −k
t −k 

k

where G is the transition matrix of the Markov state
variable.
For forecasting k periods ahead, one finds the
best-fitting model by maximizing the corresponding
likelihood function:
(9)

(

 1
∑ Prob St=s Rt − k , Xt − k
∑ Rt ln  s=0

t=1
 Prob R = 1 St = s, Xt − k

t

(

The transition probabilities, p and q, indicate
the persistence of the states and determine the
unconditional probability of the state St=0 to be
(1 – q )/(2 – p – q ). Since the state is unobserved and
must be inferred as a probability, allowance for two
states means that the expected values of the coefficients can lie anywhere between the high and low
values corresponding to the two states. In the estimation, Bayes’ rule is used to obtain filtered probabilities of the states in order to sum over possible
values of the unobserved states and evaluate the
likelihood function, as in Hamilton (1990):

t

(

+ (1 − q ) Prob St −1 = 1 Rt −1, Xt −2 .

T

=0 if ut ≤ c0 ( St ) + c1( St ) Xt − k .

)

= p Prob St −1 = 0 Rt −1 , Xt −2

(7)

Variance switching implies that the coefficients are
restricted to change by the same percentage between
regimes. I do not impose this condition because I
do not want to restrict the signs of the coefficients,
for example, to be the same in both regimes. Nevertheless, conditional heteroskedasticity helps motivate
why the coefficients might be subject to regime
switching, since volatility is one aspect of the economy that does vary across the business cycle. Because
it is not the only aspect of the economy that varies
across the business cycle, however, we keep the
model more general by not restricting the regime
switching to variance switching.
In a Markov switching model, the parameters
change values according to an unobserved binary
state variable, St, which follows a Markov process:

(

)

)

(

)




 1
∑ Prob St=s Rt − k , Xt − k
+(1− Rt ) ln  s=0

 Prob Rt = 0 St = s, Xt − k

(

) .

) 

In this forecasting exercise, the forecaster is
assumed to know whether the economy is currently
in recession when forecasting whether the economy
will be in recession six months from now. This
assumption is somewhat problematic when forecasting from the early stages of a recession before
the NBER has officially declared that the economy
entered a recession. For example, when forecasting whether the economy would be in recession in
October 2001, it was probably not clear that the
economy was in recession in April 2001. The NBER
did not announce that the recession had started in
March 2001 until November 26, 2001. On the other
hand, forecasts of the onsets of recessions are not
likely to be clouded by this assumption. In forecasting whether the economy would be in recession in
N OV E M B E R / D E C E M B E R 2 0 0 2

31

REVIEW

Dueker

Figure 2

Figure 3

Probability of Recession Conditional on
Regime S=0

Six-Month-Ahead Probability of Regime S=0
Probability

Probability

In Sample

1

1

Out of Sample

0.9
0.8

0.8
0.7

0.6

0.6
0.5

0.4

0.4
0.3

0.2

0.2
0.1
0
1962

1968

1974

1980

1986

1992

1998

NOTE: Shaded bars indicate recessions.

0
1962

1968

1974

1980

1986

1992

1998

NOTE: Shaded bars indicate recessions.

March or April 2001, no one believed in September
or October 2000 that the economy was already in
recession. There was no confusion about the current
state of the economy. Similarly, in January 1990
everyone knew that the economy was still in an
expansion when forecasting whether a recession
would start by July 1990. For this reason, one might
pay special attention to how the model predicts the
onset dates of recessions.

The explanatory variable, X, is the three-month
percentage change in the leading indicators index
shown in Figure 1. The minimum value of the change
in the leading indicators index is about –5. If we plug
this minimum value and the coefficient estimates
from Table 1 into equation (5), we see that a recession
is essentially never predicted in the state where S=1.
Given the values of c0(S=1) and c1(S=1), a standard
normal shock, u, greater than 2.7 would have to
occur at the minimum value of the leading indicators
for a recession to occur in that state. In contrast,
recessions are often implied in the regime where
S=0, as shown in Figure 2. (Note that in this and
the following charts, the coefficient estimates are
also applied to the out-of-sample data from January
1990 to December 2001.) The unconditional probability of the regime S=0, however, is only 0.25 and
it is forecasted less often than regime S=1, as seen
in Figure 3. Combining Figures 2 and 3, we see that
six-month-ahead forecasts of a high probability of
S=0 amount to a forecast of recession. Figure 4 combines the two explicitly by plotting the probability
of recession after summing across the two states:
(10)

FORECAST RESULTS

Prob Rt = 1 Xt − k

Table 1
Coefficients from the Markov Switching Probit
Equation

Coefficient value

c 0 (S=0)

0.140 (0.233)

c 1 (S=0)

1.415 (0.336)

c 0 (S=1)

2.397 (0.730)

c 1 (S=1)

–0.062 (0.351)

p

0.950 (0.012)

q

0.984 (0.010)

NOTE: Standard errors are in parentheses.

The Markov switching probit model was estimated with data from May 1960 to December 1989.
32

N OV E M B E R / D E C E M B E R 2 0 0 2

(

1

(

)

)

(

)

= ∑ Prob St=s Rt − k , Xt − k Prob Rt = 1 St = s, Xt − k .
s=0

FEDERAL RESERVE BANK OF ST. LOUIS

Dueker

A standard approach—which I call model 1—
to deriving explicit recession signals from the forecasted probability of recession, Prob (Rt=1|Xt –k ),
is to choose a critical value, m, such that a recession
is signaled if

(

)

Prob Rt = 1 Xt − k − m >0 .

(11)

(
)
(Prob(R = 1 S = 0, X ) − m )
+ Prob ( S = 1 R , X )
(Prob (R =1 S =1, X ) − m ) >0.

Prob St = 0 Rt − k , Xt − k
t

t −k

t

t −k

t −k

t

t

t

(

)

Out of Sample

0.8

0.6

0

t −k

0.2

0
1962

1

(
+ Prob ( S =1 R

)
)m .

Prob Rt =1 Xt − k >Prob St = 0 Rt − k , Xt − k m0
t

In Sample

0.4

Alternatively, one can rewrite this recession
signal as
(13)

Probability of Recession from
Six-Month-Ahead Forecasts
1

A key innovation in this paper is to recognize
that one can derive an alternative recession signal—
called model 2—from regime-specific critical values,
m0 and m1, such that a recession is signaled if
(12)

Figure 4

t − k , Xt − k

1

Either way, two critical values are used, where the
weight given to each depends on the regime probabilities. As shown in Figure 2, the probability of recession is relatively high on average in the regime where
S=0, with an average probability of 0.34. In contrast,
the average probability of recession in the regime
where S=1 is not much above zero. Given the difference between the average probabilities of recession
in the two regimes, it seems desirable to have separate critical probability levels for each regime, as in
equation (12).
I chose critical probability levels based on the
in-sample period through 1989 and examined how
well they work in the out-of-sample period. The
criterion I used was the greatest number of correct
signals, where one point was given to a correct signal
during a recession and half a point to a correct signal
during an expansion. This point scheme puts greater
emphasis on not missing recessions versus supplying false recession signals during expansions. The
impetus for this asymmetry in the point scheme is
the belief that most firms would be more willing to
pay for recession insurance than for a contract that
would indemnify them in the case where the economy performed above expectations when a recession
was forecast.

1968

1974

1980

1986

1992

1998

NOTE: Shaded bars indicate recessions.

For the in-sample period through 1989, I found
that values of m=0.28 (for model 1) and m0=0.43
and m1=0.135 (for model 2) gave the greatest number of correct signals. It makes sense that the optimal
value of m would lie between optimal m0 and m1,
since it is trying to fill both roles. The critical probability level m0 lies above the average probability of
recession conditional on S=0 (0.34), so that one
predicts a recession less than half of the time that
S=0.
Figure 5 shows the fit of the recession-signaling
model, where the signal is based on model 2 with
the two regime-dependent critical values. With forecasts from model 2, recessions are generally not
missed and the only notable false signal occurred
in the 1966 slowdown. Figure 6, in turn, shows the
fit of model 1—the signaling procedure that uses
only one critical value, as in equation (11). Here, some
of the recessions are projected to start earlier and
end later than they did and there are many more
false recession signals during expansions.
Similarly, Figure 7 shows the fit of the signal
from the simple probit model from equation (2). The
optimal critical value, m, for the 1960-89 period is
0.24. This approach, model 3, generated even more
and longer-lasting false recession signals than the
Markov switching model with one critical value
(model 1). Based on these results, we do not look
N OV E M B E R / D E C E M B E R 2 0 0 2

33

REVIEW

Dueker

Figure 5

Figure 7

Fit of Recession-Signaling Model

Fit of Recession Signal from Simple Probit
Model (Model 3)

(Model 2: Two Critical Values)
Out of Sample

In Sample

1962

1968

1974

1980

1986

1992

1998

Out of Sample

In Sample

1962

1968

1974

1980

1986

1992

1998

NOTE: Actual recessions (top) and six-month-ahead recession
signals (bottom).

NOTE: Actual recessions (top) and six-month-ahead recession
signals (bottom).

Figure 6

Figure 8

Fit of Recession-Signaling Model

1990-91 Recession and Signal

(Model 1: One Critical Value)

(Signals from Models 1 and 2 Coincide: One or Two
Critical Values)
Out of Sample

In Sample

1962

1968

1974

1980

1986

1992

1998

NOTE: Actual recessions (top) and six-month-ahead recession
signals (bottom).

further at the predictions from the simple probit
model.
In comparing the two signals from the Markov
switching models, a closer look at the two out-ofsample recessions will help determine whether the
34

N OV E M B E R / D E C E M B E R 2 0 0 2

Jan 1990

Jul 1990

Jan 1991

Jul 1991

NOTE: Actual recession (top) and six-month-ahead recession
signal (bottom).

use of two critical values in model 2 as free parameters to fit the in-sample data resulted in overfitting.
Figure 8 zooms in on the 1990 recession and shows
that the recession signals from models 1 and 2 are
identical and they both miss the 1990 recession in

FEDERAL RESERVE BANK OF ST. LOUIS

Dueker

Figure 9

Figure 10

2001 Recession and Signal

2001 Recession and Signal

(Model 1: One Critical Value)

(Model 2: Two Critical Values)

Jan 2000

Jul 2000

Jan 2001

Jul 2001

NOTE: Actual recession (top) and six-month-ahead recession
signal (bottom).

the sense that the six-month-ahead signal does not
kick in until at least six months too late. This result
confirms previous findings that the 1990 recession
was difficult to predict out of sample.
Fortunately, the performance of both signal
approaches is better in the 2001 recession. Figure 9
shows that the signal from model 1 with one critical
value predicted a recession onset in April 2001 using
information through October 2000. Thus, this signal
did not miss the actual onset date of March 2001
by much. The approach with two critical values—
model 2—does slightly worse during the 2001 recession. Figure 10 shows that this signal needed data
through November 2000 to predict a recession onset
date of May 2001. In addition, it incorrectly projected
August 2001 as an expansion month. Nevertheless,
one has to keep Figures 5 and 6 in mind before concluding that the signal derived from one critical value
is better out of sample than the signal derived from
two critical values. Model 2 gave fewer false recession signals than model 1 during the long economic
expansion of the 1990s, as seen by comparing
Figures 5 and 6.
Looking out from the December 2001 data,
both signaling approaches based on the Markov
switching model—with either one or two critical
values—predict that the recession would have ended
by January 2002. Later, the NBER will officially date
the end of the recession and then the accuracy of

Jan 2000

Jul 2000

Jan 2001

Jul 2001

NOTE: Actual recession (top) and six-month-ahead recession
signal (bottom).

the model’s trough prediction will be known.

SUMMARY AND CONCLUSIONS
This article looks at forecasting the 1990 and
2001 recessions out of sample and shows that 1990
appears to be an anamoly in terms of being difficult
to predict. Thus, one should not conclude based on
the 1990 recession that recession forecasting is a
failed enterprise. This article also responds to the
exhortation economists receive to provide unequivocal predications about whether or not the economy
will be in recession in six months. To translate from
a probability of recession to a yes/no recession signal,
one compares the probability with a critical value.
One innovation in recession signaling that I pursue
here is to have regime-specific critical values when
the recession probability comes from a regimeswitching model. This method of deriving a recession
signal reduces the number of false recession signals
outside of recession, without impairing the ability
to signal the recessions that occur.

REFERENCES
Birchenhall, Chris R.; Jessen, Hans; Osborn, Denise R. and
Simpson, Paul. “Predicting U.S. Business-Cycle Regimes.”
Journal of Business and Economic Statistics, July 1999,
17(3), pp. 313-23.
Chin, Dan M.; Geweke, John F. and Miller, Preston J. “Predicting

N OV E M B E R / D E C E M B E R 2 0 0 2

35

Dueker

Turning Points.” Federal Reserve Bank of Minneapolis
Staff Report No. 267, June 2000.
Del Negro, Marco. “Turn, Turn, Turn: Predicting Turning
Points in Economic Activity.” Federal Reserve Bank of
Atlanta Economic Review, Second Quarter 2001, 86(2),
pp. 1-12.
Dueker, Michael. “Strengthening the Case for the Yield Curve
as a Predictor of U.S. Recessions.” Federal Reserve Bank
of St. Louis Review, March/April 1997, 79(2), pp. 41-51.
Estrella, Arturo and Mishkin, Frederic S. “Predicting U.S.
Recessions: Financial Variables as Leading Indicators.”
Review of Economics and Statistics, February 1998, 80(1),
pp. 45-61.
Filardo, Andrew J. “How Reliable Are Recession Prediction
Models?” Federal Reserve Bank of Kansas City Economic
Review, Second Quarter 1999, 84(2), pp. 35-55.
Friedman, Benjamin M. and Kuttner, Kenneth N. “Indicator
Properties of the Paper-Bill Spread: Lessons from Recent
Experience.” Review of Economics and Statistics,
February 1998, 80(1), pp. 34-44.
Hamilton, James D. “Analysis of Time Series Subject to
Changes in Regime.” Journal of Econometrics, July/August
1990, 45(1/2), pp. 39-70.

36

N OV E M B E R / D E C E M B E R 2 0 0 2

REVIEW

Investment-Specific
Technology Growth:
Concepts and Recent
Estimates
Michael R. Pakko

T

he rapid pace of productivity growth since the
mid-1990s has been attributed to improvements in technology, particularly in the areas
of information processing and communications.
From e-mail and cell phones to inventory management and robotic manufacturing techniques, new
ways of doing business—facilitated by the use of
new types of capital equipment—have transformed
the workplace.
However, traditional growth theory and growth
accounting techniques—which emphasize the role
of disembodied, neutral technological progress—are
deficient in explaining the phenomenon of productivity growth driven by technology that is embodied
in new types of capital equipment. Consequently,
models of “investment specific” technological
progress have gained prominence as a framework
for evaluating the role of capital-embodied growth.
This article outlines a general model of investment-specific technological change, presents some
new estimates, and examines the role that this type
of technological progress has in explaining and
predicting recent and prospective productivity
growth trends.

GROWTH THEORY WITH INVESTMENTSPECIFIC TECHNOLOGY
The idea that technology can be manifested in
new, more efficient types of capital equipment has
a long history in economics, dating at least to the
“embodiment controversy” of Solow and Jorgenson
in the 1960s.1 The rapid advancement of informationprocessing and communications technologies has
renewed interest in the issue, inspiring the development of general equilibrium models that include
investment-specific technological progress.
Michael R. Pakko is a senior economist at the Federal Reserve Bank
of St Louis. Rachel Mandal, Mrinalini Lhila, and Athena Theodorou
provided research assistance.

© 2002, The Federal Reserve Bank of St. Louis.

In this section, I describe a simple neoclassical
growth framework—based on the model of
Greenwood, Hercowitz, and Krussell (1997)—that
incorporates this idea. In addition to balanced,
neutral technological progress, the model includes
a source of technological change that is associated
with improvement in the quality of investment goods
that becomes embodied in the productive capital
stock. The model differs slightly from Greenwood,
Hercowitz, and Krussell in two respects: First, the
model in this paper treats equipment and nonresidential structures as two components of a single,
composite capital good. In addition, the model
described below includes a convex production
possibilities frontier.
Our interest is in explaining economic growth—
a sustained increase in economic activity per capita.
Hence, attention will focus on “steady state” growth
paths in which all variables increase at constant
(though possibly differing) rates.

A Growth Model with Two Types of
Technological Change
A simple model that incorporates both types of
technological change can be described as follows:
The household sector is modeled as a representative
agent who directly controls the production technology and owns the capital stock. Households supply
labor inelastically to the production sector and make
consumption-saving decisions by maximizing a
stream of discounted utility over consumption:
(1)

∞

t
∑ β u (ct ) ,

t =0

where β<1 is a constant discount factor and ct is
(per capita) consumption. The momentary utility
function is assumed to be of the constant relative
risk aversion (CRRA) form u(ct )=ct 1–σ/1– σ.
Technology is typically incorporated directly
into the production function: Output is produced
using capital, labor, and the current state of technology. The production function is Cobb-Douglas and
technology is specified in labor-augmenting form2:
(2)

Yt = Ktα ( z t Lt )1− α .

1

See, for example, Solow (1960) and Jorgenson (1966). Hercowitz (1998)
describes the 1960s controversy in the context of contemporary
models of investment-specific technology like the one described here.

2

Note that with a simple transformation of variables, z̆=z1– α, the production function can be written in the alternative form Y=z̆·F(K,L),
where z̆ is an index of total factor productivity.

N OV E M B E R / D E C E M B E R 2 0 0 2

37

REVIEW

Pakko

Figure 1
Production Possibilities Frontier
Investment

x/c

A

Consumption

Here, z t is the technology index that directly
enhances the productivity of labor, Lt, and indirectly
that of capital, Kt, in the production of output, Yt.
Note that equation (2) can be written in terms of
labor productivity as
(2′)

yt = ktα z t1− α ,

where the ratio of capital and output to labor are
represented by lower case variables; that is, k=K/L
is the capital/labor ratio and y=Y/L is labor productivity.
Writing equation (2) in terms of log-differences,
productivity growth can be expressed as
(3)

g y = (1 − α ) gz + α gk ,

where gi , for example, denotes the growth rate of
variable i. The “growth accounting” equation (3)
shows that productivity growth can be decomposed
into components representing “total factor productivity,” (1–α )gz, and “capital deepening,” α gz .3
Investment-specific technology enters the model
through the capital accumulation equation,
(4)

(5)

N OV E M B E R / D E C E M B E R 2 0 0 2

k tα z t1− α = H (ct , xt ) ,

and to the capital accumulation equation (4).
3

Using the notation from the previous footnote, the growth accounting
equation (3) can be written explicitly in terms of the total factor productivity variable, gy=gz̆ +α gk.

4

The accumulation equation is often written so that there is a one-period
time to build; that is, capital at time t depends on investment at time
t–1. The specification in this paper simplifies the exposition of capital
growth and emphasizes the flow concept of investment.

5

Formally, the H(·) function is assumed to be homothetic. The nonlinear
PPF can be thought of as shorthand for a more detailed model in which
consumption goods and investment goods are produced in separate
sectors, with costly transfer of factors between sectors.

k t = (1 − δ )kt −1 + qt xt ,

which states that the current productive capital stock
consists of undepreciated capital from the previous
period plus net investment, qt x t.4 In equation (4)
physical investment measured in consumption units,
38

x t, is enhanced by an index of the quality of newly
produced capital goods, qt . The product qt x t represents investment as measured in efficiency units.
The improvement in the quality of capital goods
reflected in increasing values of qt is the driving force
behind investment-specific technological change.
In the subsequent analysis of the growth properties of these two types of technology, we will assume
that the economy’s opportunities for producing
consumption goods and investment goods is characterized by a nonlinear production possibilities frontier, H(ct , x t ), that is concave and invariant to scale.5
Figure 1 illustrates the tradeoff summarized by H(≅).
For a given level of technology and existing capital,
the economy is capable of producing any combination of consumption and investment lying on or
below the production possibilities frontier (PPF).
Points that lie outside the frontier are not feasible
given the current state of technology, while points
inside the frontier imply inefficient underutilization
of resources. The optimal production combination
will therefore lie on the frontier itself. The slope of
the PPF at any given point shows the trade-off
between consumption goods and investment goods—
that is, their relative price.
The durability of capital goods means that
investment produces a stream of consumption goods
into the future. Hence, the location of the optimal
point on the PPF will depend on household preferences for substituting consumption between the
present and the future (which, given the separable
CRRA form of utility assumed, is time invariant in
this model).
This combination of consumption and investment can be found from the representative agent’s
problem of maximizing utility (1) subject to the overall resource contraint,

FEDERAL RESERVE BANK OF ST. LOUIS

Pakko

For a given level of technology, the representative
household’s maximization problem yields an optimal investment/consumption ratio, implying a
specific equilibrium such as that shown as point A
in Figure 1. The slope of the dotted line shows the
price of consumption goods relative to investment
goods implied by this equilibrium.

Figure 2
Balanced Growth
Investment

Case 1: Neutral Technological Change
and Balanced Growth
Suppose that the sole source of technology
growth is zt, the index of labor-augmenting technological progress. For given quantities of labor and
preexisting capital, an increase in zt shifts the production possibilities frontier outward, as shown in
Figure 2.
Because the expansion takes the form of a
radial outward shift, both consumption and investment expand at the same rate as total output; that
is, gc=gx=gy.6 Moreover, with q constant, the capital
accumulation equation (4) implies that capital grows
at the same rate as investment, gk=gx . Hence, this
type of growth is often referred to as “balanced,”
based on “neutral” technological progress. With
investment and consumption growing at the same
rate, the economy’s growth path will be characterized
by a constant x/c ratio, as shown by the growth path
running through points A and B in Figure 2. Along
this growth path, the slope of the PPF, representing
the relative price of consumption and investment
goods, is also constant.
From the growth accounting relationship (3),
the shift in the PPF includes the direct effect of the
increase in zt (represented in Figure 2 as movement
point A′ ) as well as a component associated with
capital growth (accounting for the remaining shift
to point B in Figure 2). However, the role of capital
deepening for this type of technological expansion
is distinctly secondary. The direct effect of technology growth is an expansion of investment, which
gives rise to a commensurate growth rate of capital.
Indeed, substituting the relationship gk=gy into the
growth accounting equation (3), we find that the
rate of output growth (as well as of consumption,
investment, and capital growth) is equal to the rate
of labor-augmenting technical progress. Although
the growth accounting decomposition shows a role
for capital deepening, there is no sense in which
technological progress is “embodied” in capital
growth. Rather, the capital component represents a
passive response to “disembodied” technological

x/c
B
A

A'

Consumption

progress and does not comprise a truly independent
source of economic expansion.7

Case 2: Investment-Specific
Technological Change and CapitalEmbodied Growth
Growth associated with investment-specific
technological progress differs from neutral technology growth in several respects. First, note that qt does
not appear directly in the economy’s resource constraint, (5). Instead, the investment-specific technology index appears in the capital-accumulation
equation and therefore operates through the capitaldeepening component of the growth accounting
equation.
Investment-specific technological progress can
be illustrated using a modified PPF framework, as
shown in Figure 3. The vertical axis now measures
“effective” investment, qt x t, incorporating the notion
of improvement in the quality of investment goods.
In Figure 3, the direct effect of an increase in qt is
6

This outcome is ensured by the scale-invariance property that is
implied by the assumed homotheticity of utility and the PPF function.

7

Capital deepening does play an important role in the adjustment
dynamics of the model. That is, when the economy is not on its steadystate growth path (King and Rebelo, 1993) or is in the transition
between steady-state paths (Pakko, 2002b), capital deepening is the
mechanism that moves the economy toward long-run equilibrium.

N OV E M B E R / D E C E M B E R 2 0 0 2

39

REVIEW

Pakko

in qt has the effect of increasing the effective capital
stock. In fact, when improvement in the quality of
capital goods is accounted for, the growth rate of
the capital stock will be the sum of the growth rates
of physical investment and quality improvement,

Figure 3
Investment-Specific Growth
Effective Investment

(6)

q' . x/c

A

q . x/c

B

A

Consumption

shown by the rotation of the PPF to the dashed line
passing through point A′ . This twist in the PPF represents a change in the tradeoff between consumption
and capital accumulation. The movement from
point A to point A′ represents no change in c or x
(or their ratio), but is simply a measure of the growth
of “effective” capital that is made possible by the
increase in q.
For this reason, the selection of an appropriate
numeraire is important. If output were to be measured in terms of constant prices, the shift in the
PPF attributable to the increase in q would imply
that output had risen for fixed inputs of labor and
capital. Hence, growth accounting would incorrectly
attribute part of the change to an increase in neutral
technology, z. This mismeasurement would be even
more severe if total output were measured in units
of investment goods.
When the consumption good is taken as numeraire, total real income—as measured in consumption
units along the horizontal axis—is left unchanged
by the direct effect of growth in qt. Appropriate
measurement of investment-specific versus neutral
technology growth therefore requires that the data
be expressed in consumption units. In practice, this
means that for growth accounting in the presence
of investment-specific technical progress, nominal
output and investment data should be deflated by a
consumption price index.8
From the accumulation equation, an increase
40

N OV E M B E R / D E C E M B E R 2 0 0 2

gk = gx + gq .

Hence, the indirect impact of investment-specific
technology growth will be reflected in effective
capital stock growth that shifts the PPF in Figure 3
outward. As was the case for neutral technological
progress, the growth component of investmentspecific technological progress will be represented
by a radial outward shift of the PPF that is characterized by a constant xt /ct ratio and a common growth
trend for output, consumption, and physical
investment.
Substituting equation (6) and the relationship
gy=gx into (3), we obtain a relationship between
productivity growth and the two sources of technology growth:
α
g y = gz +
gq .
(7)
1− α
In the presence of investment-specific technological progress, total economic growth will be equal
to the rate of labor-augmenting technical change
plus a component reflecting improvement in the
quality of capital goods. Hence, investment-specific
growth represents a channel through which technological progress is manifested through “embodiment”
in productive capital.
Two features of the growth path passing through
points A and B in Figure 3 are important for evaluating the role of investment-specific technology in the
data. First, investment—when properly measured
to include improvements in the quality of new
capital goods—is predicted to grow faster than consumption along a steady-state growth path. In addition, the nature of the change in the tradeoff between
consumption and investment, represented by the
twist in the PPF, implies that the relative price of
investment goods should be falling over time relative
to consumption goods.
Figure 4 shows that these trends are, in fact, a
characteristic of the data in the National Income and
Product Accounts (NIPA).9 The ratio of investment
8

Greenwood, Hercowitz, and Krusell (1997) emphasize this point.

9

In Figure 4, “investment” corresponds to total nonresidential fixed
investment and “consumption” is measured as nondurables plus services less housing services.

FEDERAL RESERVE BANK OF ST. LOUIS

to consumption has risen persistently over the past
half-century and has appeared to accelerate sharply
in the past decade. Simultaneously, the price of
investment relative to consumption has followed a
clear downward trend since at least the late 1950s,
with the rate of decline increasing since the 1980s.

Pakko

Figure 4
NIPA Investment and Consumption:
Relative Prices and Quantities
Indices (1996=100)
180

ESTIMATES OF INVESTMENT-SPECIFIC
TECHOLOGICAL CHANGE
The data presented in Figure 4 suggest that
investment-specific technology growth has been
an important feature of post-WWII trends in productivity growth. In order to quantitatively evaluate the
role of investment-specific technology, however, it
is important to carefully examine the issue of quality
improvement for investment goods.
The measurement of quality change has always
been important in the construction of the NIPA data.
Quality characteristics of newly introduced goods
are routinely incorporated into the data using socalled “matching models” that compare the attributes
of new and existing products. In recent years, the
BEA has implemented several revisions to its methodologies in order to account for the rapid rate of innovation in information processing, communications,
and other high-tech sectors. In particular, so-called
“hedonic regression techniques” have been applied
to construct quantity and price indices that adjust
for changes in quality over time. Among the more
important applications of this approach, the BEA
incorporates hedonic indices for computer equipment and purchased software, telephone switching
equipment, cellular services, and video players,
among others.10 Moreover, the BEA has even changed
its aggregation methodology to more accurately
measure the contribution of quality change to GDP
growth: the adoption in 1996 of a chain-weighting
methodology was intended to allow aggregates to
track quality improvement better over time.
Nevertheless, some economists contend that a
significant amount of quality change goes unmeasured in the official statistics, particularly in cases
where quality improvement is more incremental.
In a seminal 1990 study, The Measurement of Durable
Goods Prices, Robert Gordon undertook to quantify
the extent of this unmeasured quality change. Drawing data from a variety of sources, including special
industry studies, Consumer Reports, and the Sears
catalog, Gordon compiled a data set of more than
25,000 price observations. He constructed qualityadjusted price indexes for 105 different product
categories, then aggregated the data to correspond

160
Price
140

120

100

80
Quantity
60

40
1947 1953 1959 1965 1971 1977 1983 1989 1995 2001

to the individual components of the BEA’s measure
of spending on producers’ durable equipment. For
each of 22 major categories of investment, Gordon
calculated “drift ratios” representing the cumulative
deviation of his adjusted price measures from the
official data. The adjusted price components were
then used to deflate nominal series, with the resulting real series aggregated to create a new qualityadjusted series for investment in durable equipment.
The bottom line of Gordon’s study was that the
official NIPA data (as constructed at the time) understated the true growth rate of real investment spending by nearly 3 percentage points per year over the
period 1947-83. This quality adjustment for real
investment spending is mirrored in the price deflator:
the finding that quality-adjusted real investment
spending is undermeasured implies that increases
in the price of investment goods have been overstated. Unfortunately, Gordon’s data set extends
only through 1983.

Quality Improvement for Equipment
and Software Investment
Previous estimates of investment-specific technology growth have been based on extrapolation of
10

Landefeld and Grimm (2000) report that 18 percent of GDP is estimated
using hedonic methods.

N OV E M B E R / D E C E M B E R 2 0 0 2

41

REVIEW

Pakko

Figure 5
Growth Rates for Equipment
and Software Investment
25
20
15
10
5
0
–5
–10
–15
1948

1954
BEA

1960

1966

1972
Pakko

1978

1984

1990

1996

Cummins-Violante

Gordon’s aggregate data for series producers’ durable
equipment. For example, Greenwood, Hercowitz,
and Krusell extended the Gordon data through 1990
by adding 1.5 percent to the growth rates of real
investment spending for all categories except computers. Hornstein (1999) invoked a similar procedure
to extend the estimate through 1997.
As BEA definitions and methodologies are
updated and as relative shares of the components
of equipment investment change over time, however,
the simple extrapolation of Gordon’s aggregate data
becomes less satisfactory. Ideally, one would like to
have extended data series at the disaggregated level
of Gordon’s original study. A less ambitious alternative is to extrapolate the drift ratios for each of
Gordon’s 22 major investment categories independently—accounting for changes in BEA definitions
and methodology—then aggregate the extrapolated
data to calculate a new, extended series.11
Two recent studies have followed variants of
this procedure. Cummins and Violante (2002) estimate a simple time-series model that relates Gordon’s
quality-adjusted estimates and official BEA time
series data for each of the individual investment
categories.12 After estimating the coefficients of
the model, Cummins and Violante extrapolate outof-sample estimates of quality-adjusted price levels
for the period 1984-2000.13 Pakko (2002a,b) uses a
simpler extrapolation technique: recognizing that
the measurement bias documented by Gordon is
42

N OV E M B E R / D E C E M B E R 2 0 0 2

larger in the earlier years of the sample period than
the latter period, the Pakko estimates are based on
a linear extrapolation of Gordon’s drift ratios for the
period 1973-83. The drift ratios were then applied
to the official BEA price data to create extended
quality-adjusted series.
Both sets of estimates were then aggregated to
create a quality-adjusted measure of equipment
investment for the period 1947-2000. Recent changes
in BEA definitions and methodology complicate this
procedure. One important innovation made in 1996
was the inclusion of software as an investment
component. Gordon’s data set did not include software, so both Pakko and Cummins and Violante
used the official BEA measure for this component.
Similarly, the BEA has devoted considerable effort
to accurately measuring quality change for computers and peripheral equipment; hence, both studies
assume that the bias found by Gordon in the vintage
data has been eliminated in contemporary time
series estimates for that component.
Figure 5 shows annual growth rates of these
quality-adjusted series for aggregate equipment and
software investment, along with the corresponding
BEA measure. The two adjusted series track each
other closely during the 1947-83 period, since both
are based on Gordon’s original data.14 The main
source of divergence between the estimates over
this period is the difference in aggregation methodologies: Cummins and Violante use the Törnqvist
index approach advocated by Gordon, while Pakko
uses the Fisher-ideal chain-weighting approach
that has subsequently been adopted by the BEA.15
During the post-1983 period, the CumminsViolante series displays more rapid growth than the
11

A disaggregated approach is preferable to a simple extrapolation of
the aggregate trend for two reasons: First, several changes in the BEA’s
definitions and methodology have, for some components, eliminated
or at least mitigated the measurement problems suggested by Gordon’s
study. In addition, the procedure of re-aggregating the quality-adjusted
components using a chain-weighting methodology allows the role of
changing expenditure shares over time to be appropriately accounted
for.

12

The model posits that the adjusted price index is a function of a constant, a time trend, current and lagged values of the BEA time series,
a cyclical indicator (lagged GDP growth), and an error term.

13

Giovanni Violante was kind enough to provide the data from Cummins
and Violante (2002).

14

The growth rates of both adjusted measures exceed the official BEA
growth rate by an average of about 2.75 percent per year over this
period.

15

The measures also differ in that the Pakko aggregate includes net
sales of scrap equipment (excluding autos), as measured by the BEA.

FEDERAL RESERVE BANK OF ST. LOUIS

Pakko

Table 1
Growth Rates and Contributions to Growth of Nonresidential Fixed Investment
Nonresidential
Source of
fixed investment
Equipment and software
Nonresidential structures
equipment
and software data 1950-2001 1950-1975 1976-2001 1950-2001 1950-1975 1976-2001 1950-2001 1950-1975 1976-2001
BEA

5.17 (100) 4.59 (100)

5.75 (100) 6.26 (80.3) 5.30 (70.9) 7.23 (87.8) 2.96 (19.7) 3.48 (29.1) 2.44 (12.2)

Adjusted 1

6.85 (100) 6.65 (100)

7.05 (100) 8.35 (80.0) 8.02 (74.0) 8.68 (85.7) 3.95 (20.0) 4.48 (26.0) 3.42 (14.3)

Adjusted 2

7.28 (100) 6.61 (100)

7.95 (100) 8.97 (81.2) 7.95 (73.8) 9.98 (87.3) 3.95 (18.8) 4.48 (26.2) 3.42 (12.7)

NOTE: Numbers in parentheses refer to percent contributions to NFI growth.

Pakko series—due largely to assumptions regarding
quality change in communications equipment. In
1997, the BEA introduced a quality-adjusted price
index for telephone switching equipment and carried
back these revisions to 1985 in the 1999 comprehensive revision of the national accounts.16 Because
this component (the largest single component in
the communications equipment category) was the
predominant source of quality bias in Gordon’s study,
Pakko considers that the updated BEA data accurately
measure quality change in that sector. On the other
hand, Cummins and Violante note that the quality
of other types of telecommunications equipment
has been improving rapidly, so they opt to use their
extrapolated estimate of quality bias from the Gordon
data set (amounting to a drift ratio of nearly 7 percent). The two studies also differ somewhat in their
treatment of automobiles, instruments and photocopy equipment, and office equipment other than
computers.17 The effect of these differences in
assumptions and methodology is that, for the 19842000 period, the Cummins-Violante series displays
an average annual growth rate that is 2.7 percent
higher than the official BEA data, while the growth
rate of the Pakko series exceeds the BEA measure
by only 1.1 percent per year.18

Incorporating Quality Change for
Nonresidential Structures
In addition to equipment and software, another
important component of the capital stock is the
structures component—accounting for approximately 35 percent of nominal nonresidential
fixed investment in the period 1948-2001. Gort,
Greenwood, and Rupert (1999) examined the measurement of quality improvement in nonresidential

structures and estimated that the official NIPA data
understates real, quality-adjusted growth by approximately 1 percent per year.
To account for this source of investment-specific
technology growth, I construct an adjusted measure
of nonresidential structures by adding 1 percentage
point to each year’s growth rate in real nonresidential
structures over the sample period of 1947-2001
(deducting 1 percent growth annually from its price
index). The resulting real investment series and price
index are then aggregated by chain-weighting with
the adjusted measures of equipment and software
spending to produce quality-adjusted decompositions
for total private nonresidential fixed investment (NFI).
Table 1 shows the growth rates for these estimates of quality-adjusted NFI, along with the contribution of equipment and software and nonresidential
structures to total growth.19 Two measures of qualityadjusted data are included: The first corresponds to
the equipment and software data from Pakko. The
second measure uses the Cummins-Violante data
series. Both measures incorporate the quality
improvement in structures suggested by Gort,
Greenwood, and Rupert.
For the period 1950-2001, equipment and software spending accounted for more than 80 percent
of the growth in total nonresidential investment.
The relative contributions to growth have not been
constant over time, however. During the first half
16

Moulten and Seskin (1999).

17

For more detail, see the appendix to Pakko (2002a).

18

These growth spreads are used to extrapolate each of the adjusted
series for the final growth observation from 2000 to 2001 (which was
not included in either of the original series).

19

The contributions to growth in Table 1 are calculated using the BEA’s
current methodology, as described in Moulten and Seskin (1999).

N OV E M B E R / D E C E M B E R 2 0 0 2

43

REVIEW

Pakko

Figure 6
Investment-Specific Technology (q)
Log Indices (Base= 1950)
1.8
1.6
1.4
1.2
1
0.8
0.6
0.4
0.2
0
–0.2
1950

1960
BEA

1970

1980

Adjusted 1

1990

2000
Adjusted 2

of the sample period, equipment and software
investment accounted for less than 75 percent of
total NFI growth, but accounted for 85 to 90 percent
during the second half of the sample.
In previous literature, estimates of investmentspecific technology growth have treated the equipment and structures components separately. The
estimates in this article use chain-weighted aggregates of both components, allowing flexibility to
account for the shifting growth-shares, suggested
by the pattern of growth contributions shown in
Table 1.

Growth Accounting with InvestmentSpecific Technical Progress
The data for quality-adjusted investment and
associated price indices form the basis for estimating
the contribution of investment-specific technology
to productivity growth. The first step is to calculate
the index of investment-specific technology, q, as
the price of consumption goods relative to (qualityadjusted) investment goods:
(8)

qt = Pc / ˜Pi ,

∼
where Pi is a quality-adjusted price index for investment and Pc is a consumption price index. Following
the practice common in previous literature, the
44

N OV E M B E R / D E C E M B E R 2 0 0 2

consumption price index used for this calculation
covers nondurables and non-housing services.20
Durable goods are excluded from the consumption
measure so as to avoid issues of quality improvement
in that component.
Figure 6 shows this measure of q for each of the
three measures of investment prices constructed
in the previous section. The data are indexed to a
base year of 1950 in order to show their cumulative
growth. The two quality-adjusted measures track
each other closely through 1983, exceeding the
growth rate of the unadjusted NIPA relative price by
an average of 1.9 percent. For the period 1984-2001,
the two adjusted series exceed the NIPA-based series
by 1.0 percent (estimate 1, Pakko) and 2.0 percent
(estimate 2, Cummins and Violante) per year.
The estimates of q, along with associated data
for real investment, x, can be used to construct
adjusted measures of the capital stock that account
for embodied technological progress. The model
suggests that real physical investment corresponds
to nominal investment deflated by the consumption
price index, Pi I/Pc. Effective investment, qx, is there∼
∼
fore given by Pi I/Pi . In the NIPA data, with Pi =Pi , this
is simply the real investment series, so the BEA’s data
for private nonresidential fixed assets is an appropriate measure of the capital stock. For each of the
adjusted investment series, qx is the quality-adjusted
real component from which a quality-adjusted
measure of the capital stock can be derived.
The procedure used to construct quality-adjusted
capital stock measures is as follows: First, I use the
accumulation equation (4) and the NIPA series for
investment and capital to back out a series of implied
depreciation factors, (1– δt ).21 These factors are then
used to construct synthetic capital-stock series using
a perpetual-inventory method—that is, by reconstructing the capital stock using equation (4) with
the quality-adjusted investment data. Starting values
for capital stocks in the base year used for these
calculations, 1950, are initialized using the accumu20

The non-housing services data are constructed by chain-weighting
PCE services with the additive inverse of the housing services component. The resulting series is then chain-weighted with nondurables
consumption.

21

The BEA constructs measures of net stocks for individual components,
then uses chain-weighted aggregation to build aggregates. The use of
these annual depreciation factors approximately adjusts for changes
in the composition of the capital stock and total depreciation that arise
from this procedure. For more information about the construction of
the BEA’s fixed-assets series, see Katz and Herman (1997).

FEDERAL RESERVE BANK OF ST. LOUIS

Pakko

Figure 7

Figure 8

Capital Stock Growth Rates

Neutral Technology (z)

Percent
9

Log Index (Base=1950)
1

8

0.9

7

0.8

6

0.7

5

0.6

4

0.5

3
0.4
2
0.3
1
0.2
0
1950

1960
BEA

1970

1980

Adjusted 1

1990

2000

0.1
0
1950

Adjusted 2

1960
BEA

lation equation (4) to relate investment/capital ratios
to the BEA data.22 The growth rates for these adjusted
capital stock series, shown in Figure 7, exceed the
official BEA measure by about 2.2 percent (Adjusted
1) and 2.5 percent (Adjusted 2) per year on average
over the entire sample period.
Completion of the growth accounting exercise
requires data for output and labor. In order for the
data to correspond to a broad measure of labor
productivity, output is taken to be gross domestic
business product.23 Business sector hours—from
the BLS labor productivity accounts—is used to
measure labor input. These data series, along with
the series for capital growth, can be used to back
out measures of labor-augmenting technological
change from the growth accounting equation (3).
Figure 8 shows measures of “neutral” technology that are derived from this procedure, where the
series are expressed in log levels relative to a 1950
base in order to illustrate cumulative growth. Each
of the series displays a clear decelleration beginning
in the early 1970s, corresponding to the widely cited
“productivity slowdown” that has prevailed for much
of the subsequent period. For the two measures of
z derived from quality-adjusted data, the slowdown
is particularly distinctive. After growth associated
with investment-specific technology has been
accounted for, the indexes of neutral technological
progress have been nearly flat since 1970.

1970

1980
Adjusted 1

1990

2000
Adjusted 2

Estimates of Neutral and InvestmentSpecific Technology
Table 2 provides a summary of the two sources
of growth, reporting the contributions to total potential growth provided by neutral and investmentspecific technological change as in equation (7).
For comparability with previous studies, the data
coverage for this decomposition begins with 1954.
Over the entire period from 1954 to 2001, the qualityadjusted measures show that the role of investmentspecific technological change has been considerable,
accounting for 60 to 68 percent of total potential
growth. Even the unadjusted NIPA data show a contribution of investment-specific technology that is 25
percent of the implied total.
The relative contributions of the two types of
technology growth have not been constant over
22

In particular, the accumulation equation implies that qx/k=
(1– δ )/(1+gk ). The properties of the adjusted investment series and
growth rates of investment-specific technology can therefore be used
to relate the initial adjusted capital stock levels to the BEA data. This
calculation yields initial values for the adjusted capital series of about
one-third the level of the official data. Results are not very sensitive
to small changes in the assumptions generating this relationship,
however.

23

In keeping with the model’s implications for price measurement,
real output is calculated by deflating nominal business sector GDP
using the consumption price deflator.

N OV E M B E R / D E C E M B E R 2 0 0 2

45

REVIEW

Pakko

Table 2
Sources of Technological Progress, 1954-2001
1954-2001 1954-1977 1978-2001
BEA
Neutral
Investment-specific
Total

1.65
0.55
2.21

2.46
0.10
2.56

0.83
1.01
1.85

Adjusted 1
Neutral
Investment-specific
Total

0.69
1.20
1.89

1.29
0.91
2.20

0.09
1.50
1.59

Adjusted 2
Neutral
Investment-specific
Total

0.57
1.39
1.97

1.33
0.93
2.26

–0.18
1.86
1.67

1.76

2.52

1.00

Actual productivity
growth

the entire sample period, however. In the first half
of the sample, 1954-1977, the quality-adjusted
measures show that investment-specific technology
contributes only about 38 percent to total potential.
For the NIPA numbers, the measured contribution
of investment-specific technology during this period
is negligible. During the second half of the sample
period, investment-specific technology overwhelmingly predominates. The second of these two adjusted
measures (derived from Cummins and Violante)
shows a negative contribution for neutral technology
growth, while the first measure (based on Pakko)
measures the contribution of neutral technology at
only about 6 percent.

Implications for Potential Productivity
Growth
The final line in Table 2 shows the actual growth
rate of output per worker for the relevant sample
periods. Over the entire period from 1954 to 2001,
all three measures of technology change overpredict
actual growth. As demonstrated in the breakdown
between the first half and the second half of the
sample, however, the overprediction is attributable
to the more recent span of years. For the period
1954-77, the two adjusted measures slightly underpredict actual productivity growth. For the period
1978-2001, all three overpredict actual growth.
Of course, the measures of potential growth
derived from the estimated technology series are
approximated measures of long-run relationships,
46

N OV E M B E R / D E C E M B E R 2 0 0 2

so it is not surprising that they do not precisely
replicate actual growth over any given finite sample.
During the period from the mid-1970s to the present,
however, the magnitude and prevalence of the
discrepancy suggest more than measurement or
approximation error.
Recent research on the economic effects of
introducing new technologies help to explain the
apparent gap between measures of technology
growth and productivity growth. The data suggest
a rather dramatic change in the pattern of technology trends: the period of slow productivity growth
in the 1970s and 1980s is associated with a change
in the composition of technological progress from
neutral to investment-specific technology.
Many economists have suggested that changes
in trend technology growth— particularly for capitalembodied technologies—are associated with long
transition periods during which productivity lags the
rate of technological advance. Indeed, both Cummins
and Violante (2002) and Pakko (2002b) focus on the
adjustment of productivity growth to technological
innovations. Cummins and Violante calculate that
the “technology gap”—the gap between the productivity of the best technology and average productivity—rose from 15 percent in 1975 to 40 percent
in 2000. This finding is in the spirit of “technology
diffusion” models (e.g., Hornstein and Krusell, 1996;
Jovanovic and MacDonald, 1994; Greenwood and
Yorukoglu, 1997; Andolfatto and MacDonald, 1998;
Hornstein, 1999), which posit that learning about
the full potential of new technologies can generate
long implementation lags as resources are channeled
into the process of adapting new technologies into
existing production structures.24 Pakko (2002b)
shows that even in the absence of explicit diffusion
lags, the adjustment of the capital stock to changes
in technology growth trends give rise to long lags
between technology and productivity—particularly
when technology growth is investment-specific.
These findings can be interpreted as suggesting that
a great deal of the potential productivity improvement has yet to be fully incorporated into measured
actual productivity growth.

CONCLUSIONS
A great deal of attention has recently been paid
to the notion that rapid technological innovation
24

Another class of general growth models addressing the adaptation of
“general purpose technologies” (e.g., Helpman, 1998) suggests similar
lags.

FEDERAL RESERVE BANK OF ST. LOUIS

has been the driving force behind recent gains in
U.S. productivity growth. However, the nature of
these technology advances—being embodied in
entirely new types of high-tech capital equipment—
is not well explained by classical growth theory.
This paper has reviewed a class of economic models
featuring “investment specific” growth that explicitly
describe a process in which new technologies are
capital-embodied.
Recent estimates of the magnitude of this type
of technology growth reported in this article suggest
that over 60 percent of potential productivity growth
over the past half-century can be attributed to
investment-specific technology. Since at least the
mid-1970s, the estimates suggest that the importance
of investment-specific technology has increased
sharply, accounting for practically all of the implied
potential productivity gains.
However, measured productivity growth has
fallen short of these estimates of potential. Recent
research on the process of adapting new technologies
to existing production frameworks gives reason for
optimism about this finding. To the extent that rapid
growth of investment-specific technological innovation has yet to be fully exploited, as the data suggest,
technology-related gains in productivity should be
expected to continue well into the future.

REFERENCES

Pakko

Rochester Conference Series on Public Policy, June 1997,
46(0), pp. 49-95.
Helpman, Elhanan, ed. General Purpose Technologies and
Economic Growth. Cambridge, MA: MIT Press, 1998.
Hercowitz, Zvi, “The ‘Embodiment’ Controversy: A Review
Essay.” Journal of Monetary Economics, February 1998,
41(1), pp. 217-24.
Hornstein, Andreas. “Growth Accounting with Technological
Revolutions.” Federal Reserve Bank of Richmond Economic
Quarterly, Summer 1999, 85(3), pp. 1-22.
___________ and Krusell, Per. “Can Technology Improvements
Cause Productivity Slowdowns?” NBER Macroeconomics
Annual 1996. Cambridge, MA: MIT Press, 1996, pp. 209-59.
Jorgenson, Dale W. “The Embodiment Hypothesis.” Journal
of Political Economy, February 1966, 74(1), pp. 1-17.
Jovanovic, Boyan and MacDonald, Glenn M. “Competitive
Diffusion.” Journal of Political Economy, February 1994,
102(1), pp. 24-52.
Katz, Arnold J. and Herman, Shelby W. “Improved Estimates
of Fixed Reproducible Tangible Wealth, 1929-95.” Survey
of Current Business, July 1997, 77(5), pp. 69-92.

Andolfatto, David and MacDonald, Glenn M. “Technology
Diffusion and Aggregate Dynamics.” Review of Economic
Dynamics, April 1998, 1(2), pp. 338-70.

King, Robert G. and Rebelo, Sergio T. “Transitional Dynamics
and Economic Growth in the Neoclassical Model.”
American Economic Review, September 1993, 83(4), pp.
908-31.

Cummins, Jason G. and Violante, Giovanni L. “InvestmentSpecific Technical Change in the United States (19472000): Measurement and Applications.” Review of Economic
Dynamics, April 2002, 5(2), pp. 243-84.

Landefeld, J. Steven and Grimm, Bruce T. “A Note on the
Impact of Hedonics and Computers on Real GDP.” Survey
of Current Business, December 2000, 80(12), pp. 17-22.

Gordon, Robert J. The Measurement of Durable Goods Prices.
Chicago: University of Chicago Press, 1990.
Gort, Michael; Greenwood, Jeremy and Rupert, Peter.
“Measuring the Rate of Technological Progress in
Structures.” Review of Economic Dynamics, January 1999,
2(1), pp. 207-30.
Greenwood, Jeremy; Hercowitz, Zvi and Krusell, Per. “LongRun Implications of Investment-Specific Technological
Change.” American Economic Review, June 1997, 87(3),
pp. 342-62.
___________ and Yorukoglu, Mehmet. “1974.” Carnegie-

Moulton, Brent R. and Seskin, Eugene P. “A Preview of the
1999 Comprehensive Revision of the National Income
and Product Accounts: Statistical Changes.” Survey of
Current Business, October 1999, 79(10), pp. 6-17.
Pakko, Michael R. “The High-Tech Investment Boom and
Economic Growth in the 1990s: Accounting for Quality.”
Federal Reserve Bank of St. Louis Review, March/April
2002a, 84(2), pp. 3-18.
___________, “What Happens When the Technology Growth
Trend Changes?: Transition Dynamics, Capital Growth
and the ‘New Economy’.” Review of Economic Dynamics,
April 2002b, 5(2), pp. 376-407.

N OV E M B E R / D E C E M B E R 2 0 0 2

47

Pakko

Solow, Robert M. “Investment and Technical Progress,” in
Kenneth J. Arrow, Samuel Karlin, and Patrick Suppes, eds.,
Mathematical Methods in the Social Sciences. Palo Alto, CA:
Stanford University Press, 1960.

48

N OV E M B E R / D E C E M B E R 2 0 0 2

REVIEW