View original document

The full text on this page is automatically extracted from the file linked above and may contain errors and inconsistencies.

ECONOMIC REVIEW

FEDERAL RESERVE BANK
OF CLEVELAND

1
http://clevelandfed.org/research/review/

ECONOMIC REVIEW
2001 Quarter 4
Vol. 37, No. 4

Estimates of Scale and Cost Efficiency
for Federal Reserve Currency Operations

2

by James Bohn, Diana Hancock, and Paul Bauer
Meeting the currency demands of depository institutions, businesses,
and consumers costs the Federal Reserve more than half a billion dollars
each year, yet very little research has been devoted to understanding what
factors affect such costs. This paper estimates a cost function in order to
obtain estimates of scale and cost efficiency for this service. Similar to
other paper-based technologies, such as checks, we find that scale
economies are achieved at a relatively low level of output, implying that
currency services are not a natural monopoly. We also provide estimates
of facility-specific marginal costs and returns-to-scale measures that
could potentially be used to improve resource allocations. Lastly, we find
that the average processing facility operates at more than 80 percent of
the efficiency of the “best-practice” facility, comparable to cost-efficiency
estimates that have been reported elsewhere for private-sector financial
institutions.

The Employment of Nations—
A Primer

27

Economic Review is published quarterly
by the Research Department of the
Federal Reserve Bank of Cleveland.
To receive copies or to be placed on the
mailing list, e-mail your request to:
4d.subscriptions@clev.frb.org
or fax it to 216/579/3050.

Economic Review is available
electronically through the Cleveland
Fed’s site on the World Wide Web:
http://www.clev.frb.org.

Research Director: Mark Sniderman
Editor:
Paul Gomme
Managing Editor: Monica CrabtreeReusser
Article Editors:
Michele Lachman,
Deborah Zorska
Design:
Michael Galka
Typography:
MAC Services

by Richard Rogerson
This paper examines low-frequency movements in employment in a
cross-section of industrialized countries for the period 1960–95, using
both aggregate and disaggregated data. It documents nine stylized facts
about cross-country variations in employment.

Monetary Policy and Asset Prices
with Imperfect Credit Markets

51

by Charles T. Carlstrom and Timothy S. Fuerst
The Modigliani–Miller theorem is fundamental to the theory of corporate
finance. One of the theorem’s immediate implications is that there is no
reason for the monetary authority to respond to asset prices. This article
posits a world in which the Modigliani–Miller theorem does not hold. The
authors assume that the amount of an entrepreneur’s external financing is
limited by the amount of collateral she holds. They examine the implications for the monetary authority in such an environment.

Opinions stated in Economic Review are
those of the authors and not necessarily
those of the Federal Reserve Bank of
Cleveland or of the Board of Governors
of the Federal Reserve System.

Material may be reprinted provided
that the source is credited. Please send
copies of reprinted material to the
managing editor.

We welcome your comments, questions,
and suggestions. E-mail us at
editor@clev.frb.org.

ISSN 0013-0281

2

Estimates of Scale and Cost
Efficiency for Federal Reserve
Currency Operations
by James Bohn, Diana Hancock, and Paul Bauer

Introduction
It is the conventional wisdom that a government monopoly in the provision of currency is
an appropriate public policy. In fact, although
Congress required the Federal Reserve to
establish a system of fees for its payment
services (check clearing and collection, wire
transfers, automated clearinghouse transfers,
and securities safekeeping) to encourage competition and to promote payment system efficiency, it did not require any charges for cash
services of a “governmental nature,” such as
the disbursement and receipt of new or fit coin
and currency.1 Accordingly, provisions in the
Monetary Control Act of 1980 require a schedule
of fees only for cash services such as coin
wrapping and transportation. Congress felt that
it could safely open such services to competition
from the private sector because the provision
of these services is not essential to the Federal
Reserve in its pursuit of its primary responsibility
“to provide the nation with currency and coin
of high quality” nor with the Federal Reserve’s
ability to “expand or contract the amount of
currency and coin in response to the public’s
demand.”2

James Bohn is an associate at the Brattle
Group, in Cambridge, Mass., Diana Hancock
is the chief of monetary and financial studies
at the Board of Governors of the Federal
Reserve System, and Paul Bauer is an
economic advisor at the Federal Reserve Bank
of Cleveland. The authors thank Jon Cameron,
Eric Severin-Lossin, and James Thomson for
their valuable comments and suggestions.
All errors remain those of the authors.

Recently, some researchers have developed
stylized theoretical models to defend this conventional wisdom, while others have begun to
question it. Two themes, which do not necessarily depend upon one another, have emerged
to support the view that a currency monopoly
is economically efficient. One theme is that
currency provision is a natural monopoly.3
This natural monopoly arises in part because
there are increasing returns to scale that result
from the cross-sectional independence of
demands for individual redemptions of currency.4 Increasing returns to scale could also
arise because fixed costs are high—sophisticated technology for soil measurement and

■ 1 During the course of congressional consideration of the
Monetary Control Act of 1980, Senator Proxmire indicated that the
Federal Reserve would not be required to charge for these services.
(Remarks of Senator William Proxmire, Congressional Record (daily
edition), March 27, 1980, p. S 3168.)
■ 2 See U.S. Congress (1980, S3168, March 27).
■ 3 See King (1983, pp. 132–33).
■ 4 Miller and Orr (1966) provide a cash-management model that
possesses this increasing-returns-to-scale property.

3

counterfeit note detection is needed to process
currency, and transporting currency is expensive. The second theme focuses on the impracticality of interest-bearing currency. It is argued
that the computation, information, and transaction costs of collecting interest on currency
can easily exceed the value of the interest that
might be collected, particularly for low-denomination notes.5 In this environment, it is argued
that competitive currency providers would have
an incentive to take actions that increase the
“quality” of currency (for example, by making
expenditures that would reduce counterfeiting
or improve the fitness of notes). To the extent
that such in-kind benefits are valued by consumers at less than their cost of production,
seigniorage (the difference between the nominal interest rate and the cost of production)
would be wastefully dissipated.6
The view that a competitive currency market
would be more efficient than a monopoly
provider is based on the argument that competitive providers would choose to offer higherquality currency services—at a higher cost—
only when the additional cost is justified by the
additional value of output.7 Implicitly, it is
assumed that competitive firms are not legally
restricted from freely employing whatever technology is most cost efficient.8 That is, the
monopolist and the competitive firm have
identical costs for providing a given quantity
and quality of currency services. Under such
circumstances, nonprice competition improves
social efficiency because a monopolist has no
incentive to provide an appropriate level of
currency services with the valued quality
enhancements.
It is unlikely, however, that a monopolist
and a competitive firm would have identical
costs for providing a given quality and quantity
of currency services. This is because monopolists tend to pursue the objective of cost minimization less effectively than do competitive
firms, being more likely to produce at an
inefficient level of output (scale inefficiency) or
to produce a given level of output with more
resources than are required (cost inefficiency).
Thus, competition among providers of currency
services may be beneficial because such
providers may be induced to be more scale and
cost efficient than a monopoly provider.
Although an empirical understanding of
currency cost considerations would enlighten
the foregoing discussion, there is a dearth of
research on such topics. First, only two studies
have analyzed the scale economies for currency
provision, Zimmerman (1981) and Dotsey
(1991). Neither of those studies is likely to

shed light on potential scale economies today,
however, because they were based on data
from periods (1977 and 1988 –90, respectively)
when the Federal Reserve used relatively
unsophisticated technologies to determine the
fitness level of notes, detect counterfeits, and
destroy unfit notes. New currency processing
technologies may have increased potential
scale economies. Second, because it is a complex task to estimate the costs of improving
currency quality, such studies have been
performed even more sporadically. Currencyquality cost estimates require quality measurements (such as soil-detection standards) and
calculations that measure changes in currency
handling costs that might arise from changes to
these qualities. These calculations are complicated even more because handling costs also
depend on a number of other intertwined
factors. For example, if currency quality is
improved (by improving the quality of existing
notes or adding additional anticounterfeiting
features, for example), handling costs may rise
due to factors such as a higher destruction rate
for unfit currency, additional printing costs for
the new notes, a shorter life span for these
notes, and possibly the need to process additional volume.9 Finally, there are no studies
that estimate the cost efficiency of Federal
Reserve currency operations. For Federal
Reserve payment services that do compete
with private sector providers, cost-efficiency
estimates are available (Bauer and Hancock

■ 5 See Fama (1981, 1983), White (1987), and Sumner (1993).
■ 6 See, for example, Fama (1983) and Sumner (1993, 2000).
■ 7 See White and Boudreaux (2000, p. 152).
■ 8 Lacker (1996) has discussed the case in which competitive
private issuers have an industry cost curve everywhere higher than that of
a monopoly issuer because of legal restrictions on the competitors’
technology choice.
■ 9 Although the following example moves currency quality in the
opposite direction, in 1991 the Federal Reserve System estimated that
it would cost about $1.2 million per year (in 1990 dollars) if it lowered
the minimum quality threshold for recirculated notes by one step on a
16-step measurement scale, but such costs would be offset by lower
annual printing costs of about $7.6 million per year, which would be
absorbed by the Bureau of Engraving and Printing. Over a 10-year period,
the discounted net present value of savings from a one-step decrease in
the minimum quality threshold for U.S. currency was estimated to be
$82.9 million. See Board of Governors of the Federal Reserve System
(1991, pp. 1–14).

4

[1993], Bauer and Ferrier [1996]). Such estimates
are comparable to those for private-sector
financial institutions reported elsewhere (such
as Bauer, Berger, and Humphrey [1993] and
Berger, Hancock, and Humphrey [1993]).
This paper explores the Federal Reserve’s
cost of providing currency services, using
quarterly data from 1991 to 1996. Whereas
previous studies (Zimmerman [1981] and Dotsey
[1991]) were concerned only with whether a
Federal Reserve currency processing and handling facility was producing an efficient level
of output (scale efficiency), this study also
considers cost inefficiency (deviations from the
cost frontier), which is sometimes interpreted
as X-inefficiency. Cost inefficiency occurs when
actual costs are higher than the minimum
possible for a given level of output, either
because more of each input is employed than
is required by the production process or
because inputs are employed in suboptimal
proportions.
We construct and compare several different
models of currency operations to test the
robustness of our results. A unique aspect of
this study is that we have detailed information
about the equipment that was used at each
facility. This information included the number
of processing machines of each type, an implementation schedule for new equipment installations, and usage statistics that indicated the
amount of “downtime” each piece of equipment experienced at each facility. These data
allow us to account in our cost-function specifications for the unavoidable transition costs
associated with the introduction of a new
technology. Such adjustment costs could significantly affect the scale and cost efficiency
of each facility as it deployed the new technology.10 In addition, because the cost functions
we estimate depend not only on costs associated with keeping fit currency in circulation
but also on the costs associated with destroying
unfit currency, the parameter estimates could
possibly be used to forecast changes in processing costs associated with changing the
quality of currency in circulation.11
The next section describes the currency
operations of the Federal Reserve System.
Section II discusses the cost-function specifications we use and the econometric techniques
we employ to estimate the cost frontier and
facility-level returns to scale and cost efficiencies. Section III describes the data for output
and input quantities as well as input prices.
Section IV presents our estimates of returns
to scale, marginal costs of different outputs,
and facility-level cost efficiencies for

Federal Reserve currency processing and
handling facilities. Section V concludes and
discusses some policy implications of our
empirical findings.

I. Background
The Federal Reserve’s involvement in the
provision of paper currency dates back to its
founding in 1913. The provisions of the Federal
Reserve Act relating to currency were designed
to provide for an “elastic currency” that could
expand and contract as the public’s demand
for cash increased and decreased.12 Prior to
that time, the kinds of currency in circulation
had proved incapable of meeting the needs for
additional amounts that developed from time
to time on a seasonal and cyclical basis as well
as in periods of financial crisis.
The first Federal Reserve notes were issued
in late 1914. By statute, these notes were
obligations of the United States, a first lien on
the assets of the issuing Federal Reserve Bank,
backed 100 percent by discounted commercial
notes and bills and a 40 percent gold reserve.13
These notes were redeemable in gold or
lawful money.
In 1920, the Appropriations Act authorized
the Secretary of the Treasury to transfer to the
Federal Reserve Banks and branches various
functions that were performed in connection
with the issue, exchange, and replacement of
U.S. paper currency and coins and the receipt
for redemption of national bank notes and
Federal Reserve Bank notes.14 This assumption
of duties by the Federal Reserve Banks and
branches led to an improvement in the overall
quality of U.S. currency and an improvement
with respect to the supply of notes of desired
denominations.15

■ 10 See Hancock, Humphrey, and Wilcox (1999).
■ 11 As noted above, such cost estimates would play a role in
estimating the social costs that would be incurred to improve currency
quality.
■ 12 See Booth (1989).
■ 13 In 1934, the Gold Reserve Act stopped redemption of currency
for gold. In 1968, the requirement of gold reserves against Federal
Reserve notes was eliminated. Also in 1968, the redemption of currency
for silver was discontinued.
■ 14 See U.S. Treasury Department, Circular No. 55.
■ 15 These improvements were noted in the 1921 Annual Report of
the Board of Governors of the Federal Reserve System.

5

In 1980, Congress again expanded the role
of Federal Reserve Banks with respect to the
distribution of currency and coin. The Monetary
Control Act of 1980 (MCA) authorized Federal
Reserve Banks and branches to distribute
available supplies of coin and currency to
depository institutions, including not only
member and nonmember banks, but also savings
banks, savings and loan associations, and credit
unions. This change dramatically increased the
number of end points served by the Reserve
Banks. In addition, the MCA required that
transportation of coin and currency as well as
coin-wrapping services be provided according
to a schedule of fees established by the Board
of Governors of the Federal Reserve System.
The growing number of institutions with
access to Federal Reserve cash services, in
conjunction with the limited facilities of
Reserve Banks, led to the adoption of “uniform
cash service standards” in 1984. These standards
defined normal service to each depository
institution as once per week; indicated that
Reserve Banks would accept deposits of
reusable currency and coin when a depository
institution accumulates a surplus that cannot be
reasonably stored or disposed of by direct
exchange with other depository institutions;
and minimized or eliminated, where practicable, “cross-shipments” (the deposit of excess
fit currency and reorder of the same denomination within five business days).16
Today, Federal Reserve Banks and branches
have significant currency and coin responsibilities. Most Federal Reserve Banks and branches
maintain facilities for currency processing,
handling, and distribution. These facilities are
specially constructed, high-security areas that
receive new notes from the Bureau of Engraving
and Printing and used notes from depository
institutions with excess currency holdings.
Used notes are deposited in the form of straps
(bundles of one hundred notes) and blocks
(bundles of ten straps). Straps and blocks are
manually counted in the receiving area,
catalogued, and stored in a vault for a short
period of time, generally between 10 and 15
days, although deposits of $1 notes often
exceed 30 days.
Used notes are counted and verified as
genuine on high-speed currency sorters. This
equipment is fairly sophisticated and performs
many tasks. First, the packaging material on
each strap is removed. Second, sensors determine which notes are fit for circulation. A note
may be deemed unfit because of its physical
condition: It may be torn or have holes in it,
be too soiled, or no longer have a sufficiently

crisp texture. Unfit notes are destroyed using
on-line shredders that are attached to the highspeed equipment. Fit notes are repackaged by
the sorter into straps and blocks for workers to
return to the vault. Third, counterfeit notes or
those that cannot be read by the high-speed
equipment are sent to workers who manually
examine each note and pass it through a lowspeed machine that, along with the high-speed
machines, reconciles the account of the
depositing depository institution.
Over the last 20 years, the Federal Reserve
has adopted increasingly sophisticated
machines that are better able to determine the
fitness level of notes and detect counterfeits.
The first generation of high-speed currency
sorters, Currency Verification, Counting, and
Sorting (CVCS) machines, were installed during
the 1980s. These machines were gradually
replaced by a second generation of high-speed
currency sorters, Banknote Processing System
(BPS) 3000 machines, which were installed
during the 1990s. This second generation
of machines was subsequently upgraded with
new software and hardware to improve
performance.
Currency, whether fit or new, enters circulation when withdrawn by a depository
institution. Depository institutions cannot
specifically request new currency from the
Federal Reserve. Rather, orders are filled with
the first available currency from the vault.
Federal Reserve offices handle a considerable
amount of currency each year (table 1). In
1999, for example, more than 7 billion notes
were destroyed and more than 9.5 billion new
notes were put into circulation.

II. Estimation
Techniques
A variety of econometric techniques have been
developed for estimating cost frontiers. Such
techniques employ a specific flexible functional
form for the cost function, and each one
imposes some additional assumptions about
the statistical properties of the inefficiency
terms. In this paper, we use two different

■ 16 These standards have since been modified. See the Federal
Register notice dated April 30, 1996, for a detailed discussion of the
uniform cash access policies, which became effective May 1, 1998.

6

T A B L E

1

Currency in Circulation, New Notes Issued, and
Notes Destroyed, 1999 (millions of pieces)
Dollar
denomination

Notes in
circulation

New notes
issued a

Notes
destroyed

3,767

1

7,536

3,865

2

602

25

4

5

1,799

866

694

10

1,620

788

616

20

5,804

2,744

1,736

50

1,294

426

183

100

3,862

879

257

Total

22,516

9,593

7,257

a. Does not include additions to inventory at Reserve Banks.
SOURCE: Annual Report, Budget Review, Board of Governors of the Federal
Reserve System (2000).

functional forms for the cost function. This
allows us to examine the robustness of our
results to different assumptions with respect
to the functional form of the cost function. In
addition, we use two different frontier estimation techniques to ascertain how robust our
facility-level efficiency rankings are to alternative assumptions about the statistical properties
of the inefficiency terms.
The first functional form for the cost function
we consider is the translog,
J

1
(1) ln Cit = β0 +  βj ln yjit + _
2
j=1
L

+  δl ln w lit +
l=1

1
+_
2

J

J

lar quarter and zero otherwise and is meant to
allow for seasonal effects. The Tit terms allow
for adjustment costs associated with the installation of second-generation equipment (BPS
3000 machines). Since such costs were not
necessarily linear over time, we define three
adjustment-cost variables, T1it , T2 it , and T3it .
The first two, T1it (T2 it ), are indicator variables
that equal one in the first (second) quarter that
a machine was installed, and zero otherwise.
The third, T3it , is a time-trend variable that
begins three quarters after the new machine
was installed. Finally, u and v represent deviations from the cost frontier due to facility-level
cost efficiency (u ≥ 0) and statistical noise (v).17
The translog cost function is a second-order
approximation to any function about a point of
expansion and has been used extensively in
recent years. Unfortunately, as one moves
away from the point of expansion, the approximation becomes less precise.18
Because we derive our estimates of scale
economies and marginal costs from the relationship between outputs and cost, we must
be careful to specify the relationship with a
sufficient amount of flexibility to ensure that
the true underlying relationship between cost
and output can be revealed. To explore
whether the translog is flexible enough, we
also employ a “hybrid-translog” cost function.
This cost function includes the terms of the
translog model (equation [1]) together with
first-, second- and third-order trigonometric
terms of the Fourier functional form.19 More
formally, the hybrid-translog takes the form

J

  βjk ln yjit ln ykit

j=1 k=1

L

  δjl ln yjit ln wlit

j=1 l=1

L M

N

 lm lnwlit lnwmit + χn ln enit

l=1 m=1

n=1

1996

4

t=1992

q=1

+  t Dt +  ηq Q q
3

+  j Tjit +uit + vit ,
j=1

where i and t indicate that an observation is for
facility i at time t, y is a vector of output
quantities, w is a vector of input prices, and e
is a vector of N environmental variables (these
are described in more detail later). D is a set of
indicator variables that equals one in a particular year and zero otherwise and is meant
to allow for technical change. Q is a set of
indicator variables that equals one in a particu-

■ 17 Estimates of ui , the facility-specific cost-efficiency term, need
to be interpreted with care. These estimates are conditional on assumptions about the functional form for the cost function and its error terms.
■ 18 Generally, the expansion point used is the mean of the data.
■ 19 The hybrid-translog functional form has been used by Gallant
(1982) and Berger, Leusner, and Mingo (1997).

7

J

(2) ln Cit = β0 +  βj ln yjit
j=1

J

1
+_
2

L

  βjk lny jit ln ykit +  δl ln wlit

j =1 k=1
J

+

J

l=1

L

  δjl ln yjit lnwlit

j=1 l=1

L M

1
+ _   lm lnwlit lnwmit
2 l =1m=1
N

+  χn ln enit
n=1

1996

4

t=1992

q=1

+  t Dt +  ηq Q q
3

+  j Tjit
j=1

j

J

+ [j cos zjit +j sinzjit ]
j=1
J

J

+   [jk cos(zjit +zkit ) +jk sin(zjit +zkit )]
j=1 k=1
J

explore how sensitive our results are to differing assumptions. The generalized least squares
(GLS) estimation technique uses the longitudinal aspect of the data and avoids assuming
a specific distribution for the inefficiency
term.21 Basically, this technique uses repeated
observations over time to identify firm-specific,
time-invariant inefficiencies, that is, uit=ui .
Iterative, seemingly unrelated regression techniques are used to estimate the system of cost
and input share equations using longitudinal
data. The inefficiency terms are calculated by
using the average of the residuals for each
facility i, 
ˆi . The most efficient currency handling facility in the sample is assumed to be
fully efficient, and the inefficiency of every
other facility i is measured by the proportionate
increase in predicted costs above the predicted
cost of the most efficient facility, or 
ˆi – min 
ˆj .

J

J

+   [jkl cos(zjit +z kit +z lit )
j=1 k=1 l=1

+ jkl sin(zjit +z kit +z lit )] +u it +vit ,
where the zjit are the output quantities normalized over the range (0, 2 π ). Equations (1) and
(2) are estimated along with corresponding
equations for input shares, with the usual
restrictions on symmetry and linear homogeneity imposed.
Although the currency processing machines
themselves are quite intricate, the production
technology for currency processing, as revealed
by the cost function, may be relatively simple.
If this hypothesis is correct, then the translog
functional form should adequately characterize
currency processing. One risk associated with
using the hybrid-translog functional form is
that the data may be overfitted. That is, statistical noise in the data may be absorbed into
the cost function. By estimating both functional
forms, we can obtain a deeper understanding
of the relationship between output and costs.
We use two different econometric techniques to measure facility-level cost efficiencies
for Federal Reserve currency processing and
handling operations because empirical measurements of cost efficiency in the financial
services industry often vary significantly
depending on the methodology employed.20
By estimating the model two ways, we can

For GLS estimators to be consistent, the density
of the inefficiency disturbances must be
nonzero in the neighborhood of (0, ω) for
some ω> 0. In other words, as the number of
facilities in the sample increases, the probability
that a facility lies near the frontier approaches
one. Our efficiency measure is calculated using
the log difference in average residuals between
the facility on the cost frontier and the actual
average residual of facility i , or exp(– ( 
ˆi –
min 
ˆj ). This measure is bounded between zero
j

and one, with the most efficient facility having
an efficiency value equal to one.
The maximum likelihood technique (MLE)
identifies inefficiency primarily by the skewness in residuals, rather than by persistence
over time.22 Statistical noise, v, is assumed to
be normally distributed, while inefficiency, u,
is assumed to be half-normally distributed. Our
MLE measure of inefficiency is calculated as
the conditional mean, E (u|u+v).23 Because the
MLE technique requires a priori assumptions
concerning the distribution of the inefficiency
terms while the GLS technique does not, we
tend to favor the GLS results. Both cost-

■ 20 See, for example, Ferrier and Lovell (1990), Bauer, Berger,
and Humphrey (1993), and Berger (1993).
■ 21 See Schmidt and Sickels (1984).
■ 22 The maximum likelihood technique we employ is based on
Bauer, Ferrier, and Lovell (1987).
■ 23 This technique was developed by Jondrow, Lovell, Materov,
and Schmidt (1982).

8

efficiency measures are presented below to
ascertain how robust our facility-level
efficiency rankings are to alternative assumptions about the statistical properties of the
inefficiency terms.
In practice, managers at each facility may
have little ability to vary the amount of equipment and building inputs used over the short
run. Adding vault space, for example, would
require considerable modifications to a facility
because of the high security required for such
space. Also, high-speed currency processing
machines are relatively unique, and when they
were last purchased in 1990, enough were purchased to meet anticipated demand through
2003. If an input, l, is, for all practical purposes,
fixed, then it is appropriate to use the input
quantity, qlit , rather than the input price, wlit , in
the cost function. This substitution would be
appropriate regardless of the functional form
chosen for the cost function or the estimation
technique employed. Returns-to-scale measures,
of course, depend on whether inputs are fixed
or variable. When all inputs are variable, returns
to scale (RTS) can be measured using the cost
elasticity ( l δ lnC/δ lnql ). When K inputs are
fixed over the short run, the fixed-input version
of the RTS measure is:
J

 [ δ lnC ( yit , wit , qkit )/δ lnyjit ]

(3) RTS =

j=1
K

1 –  [ δ lnC ( yit ,wit ,qkit )/δ ln qkit ]
k =1

If the RTS measure is greater than one, then
costs increase more than proportionately with
output, which implies that the facility is operating
in a region where there are diseconomies of
scale. If, however, the RTS measure is less than
one, then the facility is operating in a region
where there are increasing returns to scale.
Only when the RTS measure equals one, is the
facility operating with constant returns to scale.

III. Data
We collected quarterly data (1991 to 1996,
inclusive) on total costs, output volumes, input
prices, input quantities, and environmental
variables for 37 Federal Reserve currency
processing and handling facilities. We chose
this time period because it was a fairly tranquil
period for currency operations. Before 1991,
some Federal Reserve facilities used an assortment of currency processing machines in
addition to the first generation of high-speed

processors, the CVCSs. After 1996, new note
designs were introduced—the $50 note in
October 1997 and the $20 note in September
1998.24 Also, as the century date change drew
near, the amount of currency outstanding
increased substantially.25 In all, the sample
included 886 observations.26
The primary data source is annual functional
cost accounting records from the Federal
Reserve’s Planning and Control System (PACS),
which are collected to monitor costs and
improve resource allocation within the System.
These data are supplemented by other cost data,
machine counts, and usage statistics; data from
occasional Federal Reserve surveys; and price
index information from the Bureau of Economic
Analysis and the Bureau of Labor Statistics.
Total costs for each currency facility, Cit ,
include the direct costs that arise from highspeed currency processing; off-line currency
verification, destruction, and cancellation; and
currency paying and receiving operations.
Associated with these cost-generating activities
are three distinct and measurable outputs. We
include in the output vector, yit , the number of
fit notes generated by the high-speed currency
processing operations, yfit ; the number of notes
destroyed either on-line by the high-speed
machines or off-line at the reconciliation
stations, ydit 27; and the total number of transactions with depository institutions, that is,
the sum of the number of incoming shipments
of currency received at facility i and the
number of outgoing orders for currency filled
by facility i, ynit .
Inputs used in currency operations are classified into buildings, B, labor, L, equipment, E,
and materials, M. Buildings’ share of total
currency costs averages only about 16 percent

■ 24 The new design for the $100 note was introduced in March
1996. Because the average life of a $100 note is about 8.5 years, however,
the number of $100 notes received by the Federal Reserve in 1996 (about
1 billion notes) was relatively small compared to the number of smallerdenomination notes received (about 22.6 billion notes). Thus, the
introduction of the new design for the $100 note is unlikely to have
materially affected the Federal Reserve's currency operations in 1996.
■ 25 Total currency in circulation increased about 22 percent in
1999, although it had only grown about 7 percent per annum in the
previous four years.
■ 26 Although the Helena, Montana, facility operated during the two
quarters when the new BPS 3000 machines were installed, there are no
reported volumes for this period.
■ 27 Off-line destruction is undoubtedly more labor intensive and
consequently more expensive, but these items make up less than half of
1 percent of notes destroyed.

9

because the interest expenses associated with
the acquisition of buildings are not included in
the cost-accounting framework. For the price
of building services, wB , we use an annual
square-foot replacement cost, adjusted by the
depreciation rate, for the location of each
building with currency operations.28 The quantity of building services, qB , is proxied by the
actual number of square feet of space occupied by currency operations at each facility.
Over the 1991–96 period, approximately
52 percent of total costs for currency operations
are attributable to labor expenses. The price of
labor, wL , is constructed using data on expenditures for labor, including salaries, retirement
and other benefits, and the number of
employee hours spent in currency operations.
Equipment expenditures consist of depreciation and maintenance expenses. For the price
of equipment, wE , we use the price index for
currency processing equipment reported in the
Producer Price Index Detailed Report. For the
quantities of equipment in operation at each
facility during each quarter, qCVCS and qBPS , we
use actual machine counts for CVCS and BPS
3000 machines, respectively.
Expenditures on materials, including currency straps and packaging, computer support,
printing and duplicating, and other centrally
provided support costs, account for about
11 percent of total costs. The price for material
inputs, wM , is constructed from material expenditure shares and various price indexes for
components of materials from the Survey of
Current Business 29 and the Producer Price
Index Detailed Report, using a Tornquist
approximation to a Divisia index.30
Environmental variables are used to control
for the “quality” of incoming shipments and
the operating environment at each facility. To
measure the quality of incoming shipments we
use several proxies: the number of depositor
errors per 100,000 notes deposited; the proportion of incoming notes that are $1 notes; the
proportion of notes that are rejected by highspeed equipment; and the mean shipment size.
A higher value for any of these proxies is
expected to increase costs at the facility.
Depositor errors occur when a strap has more
or fewer than 100 notes of the same denomination. When such errors occur, the depositor’s
reserve account must be reconciled to reflect
the correct deposit balance.31 Small-denomination notes, such as $1 notes, have a relatively
short life compared to large-denomination
notes. For example, the average life of a
$1 note is 1.5 years, while the average life of a
$50 note is 5 years. Consequently, we can use

the proportion of incoming $1 notes in a
shipment to proxy for the shipment’s overall
quality. The proportion of notes rejected by
high-speed equipment would directly affect
the costs of currency handling and processing
because rejects are handled manually by an
operator on a low-speed machine. Also, a large
average shipment size may create a temporary
backlog for processing work.
The operating environment is controlled for
by variables that capture usage intensity of the
equipment. Usage intensity is proxied by the
proportion of scheduled operating time that
processing equipment is “down”; the proportion of notes processed on BPS 3000 machines;
and the mean throughput of processing equipment (measured in notes per hour). In addition, indicator variables are used to account for
temporary cost increases that resulted from an
unusual event that would affect only one Federal Reserve processing facility. One indicator
variable allows for higher costs during the
period when the New York office was testing
the new BPS 3000 machines (between
1992:IIIQ and 1994:IVQ). Because that office
was the first to use the BPS 3000 machines, it is
reasonable to assume that its transition costs
may have been higher than those who adopted
the machines later. Another indicator variable
is used to accommodate higher costs during
the period when the new Dallas branch was
constructed (1995:IQ through 1996:IVQ).

IV. Empirical
Results
The two forms of the cost function, the
translog and the hybrid-translog, are each
specified with different combinations of fixed
and variable inputs. In one specification, all
inputs are variable, in another, equipment
inputs are fixed, and in another, equipment

■ 28 For these purposes, data on replacement costs for buildings
are taken from the 17th edition of Means Square Foot Costs
(Means, 1996).
■ 29 See U.S. Department of Commerce (1990–97).
■ 30 The Tornquist index is constructed using the rates of growth
in the prices in each category. These rates of growth are weighted by the
average proportionate shares of materials expenses to each category
over adjoining periods. The base year is 1990.
■ 31 Note that the Reserve Banks absorb the differences on $1 note
deposits, so there are no depositor errors reported for $1 notes.

10

T A B L E

2

Average Returns to Scale for
Federal Reserve Currency Operations
Functional form/estimation technique
Cost-function specification

GLS/translog

GLS/hybrid-translog

MLE/translog

All inputs variable

0.79

0.78

0.77

Equipment fixed

0.81

0.76

0.80

Equipment and
buildings fixed

0.83

0.78

0.81

Arithmetic mean

Weighted mean
All inputs variable

0.93

0.91

0.91

Equipment fixed

0.97

0.86

0.93

Equipment and
buildings fixed

0.98

0.78

0.94

SOURCE: Authors’ calculations. Weighted means use the number of notes
processed as the weighting measure.

and buildings are fixed. In all specifications,
labor and materials inputs are treated as variable. Cost frontiers are estimated using the GLS
technique for both cost-function specifications
and the MLE technique for the tranlog costfunction specification. These frontiers are
denoted as GLS/translog, GLS/hybrid-translog,
and MLE/translog.32

Returns to Scale
For every cost-function specification we
estimate, the average-cost curve implied by the
cost frontier is U-shaped with a fairly flat portion
at the bottom of the U. The implied average
cost curve descends rapidly up to 100 million
notes per quarter and then is fairly flat for most
of the remaining observed output range.33
This shape is fairly typical for the average cost
curve of paper-based payment technologies.34
The minimum efficient scale (MES) for
currency depends on whether equipment and
buildings are treated as fixed inputs. For
example, using the GLS/translog model, the
MES is 260 million notes per quarter when all
inputs are variable, 680 million notes per
quarter when equipment is fixed, and 250 million notes per quarter when equipment and
buildings are fixed.35 While this range for the
MES is quite large, the difference in average
costs for the largest and smallest MES output
levels is not economically meaningful because

the implied average cost function is so flat
through the region of 250 million notes per
quarter to 650 million notes per quarter.
Table 2 presents average RTS measures for
Federal Reserve System currency and handling
operations using several cost-function specifications and the two frontier-estimation techniques. In this table, both the arithmetic mean
and a weighted mean of facility RTS measures
are presented in the top and bottom panels,
respectively. Weights each quarter vary across
Federal Reserve facilities and are equal to the
volume of notes processed at each facility.
Strikingly, neither the cost-function specification nor the frontier-estimation technique
greatly affects the average RTS measures. The
measures are fairly robust to whether equipment and buildings are considered fixed. And,
with one exception—the GLS/hybrid-translog
model specified with fixed buildings and
equipment—such measures are fairly robust to
whether a translog or hybrid-translog functional form is used. Regardless of the costfunction specification or frontier-estimation
technique chosen, the mean of the unweighted
RTS is smaller than the mean of the weighted
RTS measure. This happens because, even
though there are substantial unexploited potential scale economies at many facilities in the
System, the vast majority of notes are processed
and handled at the facilities that have already
achieved constant returns to scale.
Focusing on the first and last quarters of the
estimation period, 1991:IQ and 1996:IVQ,
respectively, we calculate facility-specific
returns-to-scale measures. Table 3 presents
these returns-to-scale measures for each office
that is estimated using the GLS/translog model
with all inputs specified as variable. The

■ 32 We did not estimate an MLE/hybrid-translog model because
the computation time required to achieve convergence would have been
unacceptably long on account of the large number of additional
parameters required for this model.
■ 33 Most facilities with output less than 100 million notes per
quarter have only one high-speed currency processing machine.
■ 34 Zimmerman (1981) and Dotsey (1991) report cost-function
estimates for Federal Reserve currency operations that are consistent
with a U-shaped average cost curve. Humphrey (1981), Bauer and
Hancock (1993), and Bauer and Ferrier (1996) also report cost-function
estimates for Federal Reserve check processing operations that are
consistent with a U-shaped average cost curve.
■ 35 MES estimates from the GLS/translog model are representative of MES estimates from the GLS/hybrid-translog and MLE/
translog models.

11

T A B L E

3

Returns-to-Scale Measures for Federal Reserve
Currency Processing and Handling Facilities

Federal Reserve office

1991:IQ estimate

1991:IQ t-test

1996:IVQ estimate

1996:IVQ t-test

FR 1

0.800

3.199

0.837

2.652

FR 2

0.933

1.109

0.937

1.046

FR 3

0.691

4.337

0.826

2.801

FR 4

1.006

–0.080

1.030

0.366

FR 5

0.843

1.760

0.594

4.031

FR 6

1.066

0.822

1.052

0.723

FR 7

0.990

0.133

1.122

1.529

FR 8

0.765

4.087

0.821

3.189

FR 9

0.793

3.105

0.874

2.135

FR 10

0.761

3.777

0.804

3.425

FR 11

0.712

4.621

0.810

3.346

FR 12

0.838

2.804

0.815

2.304

FR 13

0.497

6.692

0.619

5.579

FR 14

0.404

6.648

0.493

6.417

FR 15

0.621

5.255

0.795

3.649

FR 16

0.735

3.351

0.687

3.324

FR 17

0.613

5.903

0.680

5.252

FR 18

0.484

6.708

0.689

5.021

FR 19

1.208

2.404

1.141

1.502

FR 20

0.641

5.240

0.701

4.860

FR 21

0.687

4.866

0.734

4.313

FR 22

0.744

2.621

0.676

2.534

FR 23

0.778

2.953

0.787

2.575

FR 24

0.724

4.671

0.715

4.923

FR 25

0.779

3.863

0.815

2.817

FR 26

1.038

0.393

1.042

0.347

FR 27

0.657

5.399

0.690

4.859

FR 28

0.460

6.682

0.583

5.596

FR 29

0.916

1.145

0.946

0.796

FR 30

0.831

2.997

0.787

3.595

FR 31

0.572

5.441

0.554

6.461

FR 32

0.973

0.359

0.922

1.286

FR 33

0.572

6.323

0.509

6.572

FR 34

0.650

5.361

0.759

3.658

FR 35

1.062

0.848

1.013

0.166

FR 36

0.592

5.504

0.707

3.829

FR 37

0.648

5.152

0.754

4.349

SOURCE: Authors’ calculations. The t-test statistic is for a “two-tail” test of the hypothesis that the estimated coefficient is equal to 1.

t-statistics in the table test the hypothesis that
the RTS estimate differs from one. At the
beginning of the sample, 27 facilities have
statistically significant increasing returns to
scale, 9 facilities exhibit constant returns to
scale, and 1 facility has significant decreasing

returns to scale. At the end of the sample, 28
facilities have statistically significant increasing
returns to scale and 10 exhibit constant returns
to scale. Although the number of notes received
by the Federal Reserve System increased by
about 20 percent over this period (figure 1),

12

F I G U R E 1
Federal Reserve System Notes Paid and
Received Annually (in thousands of notes)

Notes received
40
$100

$50

$20

$10

$5

$2

$1

Notes paid
40
$100

30

30

20

20

10

10

0

$50

$20

$10

$5

$2

$1

0
1990 1991 1992 1993 1994 1995 1996 1997 1998 1999 2000

1990 1991 1992 1993 1994 1995 1996 1997 1998 1999 2000

SOURCE: Board of Governors of the Federal Reserve System.

the introduction of higher-speed processing
machines enabled each facility to achieve similar returns to scale as before their introduction.
Facilities with insufficient processing volume to
achieve scale efficiency in 1991 continued to
have insufficient processing volume to achieve
scale efficiency in 1996. It remains the case,
however, that scale efficiency for Federal
Reserve currency operations improved on
average because no facility operated with
statistically significant decreasing returns to
scale, and far fewer facilities operated with an
RTS measure less than 0.70 by the end of the
estimation period.
Our RTS findings suggest that there was
considerable excess capacity in Federal Reserve
currency operations. This excess capacity has
been used to accommodate the public’s
growing demand for currency, which increases
demand for currency processing, including the
destruction of unfit currency. However, to the
extent that use of the new Sacagawea dollar
coins substitute for $1 notes, more Federal

Reserve facilities may be operating with
increasing returns to scale in the future unless
some smaller facilities are consolidated. Of
course, the higher unit costs of operating with
increasing returns to scale must be weighed
against the costs of being unable to meet
demand shocks for currency processing and
handling. Also, the cost savings of consolidating
facilities would have to be balanced against the
higher costs of transporting currency deposits
and withdrawals over longer distances.36
Alternatively, the overall level of scale efficiency

■ 36 Bauer et al. (2000) considered whether Federal Reserve
System currency processing costs could be lowered by reallocating
currency volume across facilities or by consolidating facilities. They
found most cost savings could be achieved without closing any existing
currency processing facilities. In addition, they argued that a new facility
located in Phoenix, Arizona, would help to lower System currency
processing costs.

13

T A B L E

achieved by Federal Reserve currency operations could potentially be improved over time
by adjusting standards for normal service levels
or by changing fees for nonstandard access or
nonstandard cash services.37

4

Marginal Costs of Production for Fit Currency,
Destroyed Currency, and Cash Shipments
Functional form/estimation technique
GLS/translog

GLS/hybrid-translog

MLE/translog

Fit currency
(Cents per note)

0.358

0.344

0.398

Destroyed currency
(Cents per note)

0.161

0.145

0.175

Cash shipments
(Dollars per shipment)

22.63

30.43

14.09

Fit currency
(Cents per note)

0.428

0.369

0.521

Destroyed currency
(Cents per note)

0.234

0.200

0.235

Cash shipments
(Dollars per shipment)

38.82

58.15

21.11

Marginal Costs

Arithmetic mean

Weighted mean

SOURCE: Authors’ calculations. Weighted means use the number of notes
processed as the weighting measure.

T A B L E

5

The Effects of Environmental Variables
on the Cost of Currency Operations
Environmental variable

Sample mean

Estimate

The number of depositor errors per
$100,000 notes deposited

13.152

0.0317**

The proportion of incoming notes that
were of $1 denomination

0.370

–0.024

The proportion of notes rejected by highspeed equipment

0.022

0.156***

The mean shipment size (in notes)

70,757

0.345***

The proportion of scheduled operating time
that processing equipment was “down”

0.041

0.083***

The proportion of notes processed on
BPS 3000 machines

0.223

0.068***

The mean throughput of processing
equipment (in notes per hour)

66,030

0.234*

Quality of incoming shipments

Usage intensity of equipment

Recently, some observers have argued that the
Federal Reserve should explicitly charge for
currency services because outgoing currency
from the Reserve Banks is of higher quality
than the incoming currency from depository
institutions.38 Such observers typically argue
that some currency users would be willing to
pay for a higher-quality, “superfit” currency
because they manage newer technologies that
would perform better with higher-quality
currency (such as ATMs or vending machines).
With the view to improve economic efficiency, potential prices for currency services
would likely depend on the marginal costs of
production for fit currency, for destroying
currency, and for currency shipments. We
estimate such marginal costs at each facility by
using cost frontiers that assume all inputs are
variable in the short run. In table 4, both the
arithmetic mean and a weighted mean of
Federal Reserve facility-level marginal cost
estimates for fit currency, destroyed currency,
and cash shipments are presented.
As with the weighted RTS measures presented in table 2, the weights used in table 4
each quarter vary across Federal Reserve
facilities and are equal to the volume of notes
processed at each facility. The average Federal
Reserve System marginal cost for fit currency,
destroyed currency, and cash shipments is
higher when each facility’s marginal cost is
weighted by its output volume. This happens
because the marginal cost is less than the
average cost when a facility operates with
increasing returns to scale, as did most facilities. The average marginal costs for fit and
destroyed currency, unweighted or weighted,
are higher when the MLE technique is used

SOURCE: Authors’ calculations. Estimates significantly different from zero at
the 10, 5, and 1 percent significance level are indicated with *, **, and ***,
respectively. All significance tests used a two-tailed t-test.
■ 37 Although the uniform cash access policy established normal
access for a depository institution as once per week, Reserve Banks can
charge for more frequent nonstandard access. In addition, Reserve Banks
can and do charge for services such as nonstandard packages and nonstandard packaging of same-day express cash orders.
■ 38 See Supel and Todd (1984) and Lacker (1993).

14

than when the GLS technique is used, but of
similar magnitude.
The average Federal Reserve marginal cost
for fit currency is typically less than one-half of
a cent per note during the period studied. The
average marginal cost for destroyed currency is
about half that. Marginal cost estimates for cash
shipments, which are reported at the bottom
lines of the top and bottom panels of table 4,
are very sensitive to both the functional form
of the cost function and the econometric
technique used to estimate the cost frontier:
Depending on which forms and techniques are
used, the arithmetic mean marginal cost for
Federal Reserve cash shipments ranges from
$14 to $30, and the weighted mean ranges
from $21 to $58.

F I G U R E 2
Year-Indicator Effects for Various
Estimation Techniques
Index
1.200
GLS/translog
GLS/hybrid-translog
MLE/translog
1.150

1.100

1.050

Environmental
Factors

1.000

0.950
1990

1991

1992

1994

1993

1995

1996

SOURCE: Authors’ calculations.

F I G U R E 3
Year-Indicator Effects for Various
Model Specifications
Index
1.250

1.200

GLS/translog
GLS/translog, with fixed equipment
GLS/translog, with fixed equipment and buildings

1.150

1.100

1.050

In every specification for the cost frontier we
estimated, environmental factors are important
cost determinants. Table 5 presents the effects
of environmental variables on the cost of
production when the cost frontier estimated is
the GLS/translog and all inputs are considered
variable.39
The quality of incoming currency deposits
significantly influences Federal Reserve costs.
The number of depositor errors per 100,000
notes deposited, the proportion of notes that
are rejected by high-speed equipment, and the
mean shipment size each significantly raises
Federal Reserve currency handling and
processing costs at the 5 percent level of significance. A 1 percent increase in the number of
depositor errors increases costs by 0.03
percent, a 1 percent increase in the proportion of notes rejected by high-speed equipment increases costs by 0.16 percent, and
a 1 percent increase in mean shipment size
increases costs by 0.35 percent. Only the
proportion of incoming notes that are of
$1 denomination is not significant among the
variables that measure incoming currency quality.

1.000

0.950
1991

1992

1993

1994

1995

1996

SOURCE: Authors’ calculations.

■ 39 The environmental effects we present in table 5 are representative of the others we derive using alternative cost frontiers.

15

Operating environment variables also affect
the cost of providing currency services. The
proportion of scheduled operating time that
processing equipment is down and the proportion of notes processed on BPS 3000 machines
each significantly raises costs at the 1 percent
level of significance. In addition, the meanthroughput-of-processing-equipment variable
is significant at the 10 percent level.

F I G U R E 4
Learning-by-Doing Effect for Various
Estimation Techniques
Index
1.300

1.200

Time Adjustment
Factors

1.100

1.000
GLS/translog
GLS/hybrid-translog
MLE/translog
0.900

0.800
0

4
Quarter

2

8

6

SOURCE: Authors’ calculations.

F I G U R E 5
Learning-by-Doing Effect for Various
Model Specifications
Index
1.200

1.100

1.000

0.900
GLS/translog
GLS/translog, with fixed equipment
GLS/translog, with fixed equipment and buildings
0.800
0

2

SOURCE: Authors’ calculations.

4
Quarter

6

8

The last environmental coefficients we
consider are the variables included to control
for technical change in general and the introduction of the new machines in particular.
The estimated coefficients for these variables
are fairly consistent across the various specifications. Figure 2 plots the effects of the yearindicator variable for the GLS/translog,
GLS/hybrid-translog, and MLE/translog models.
Figure 3 plots the GLS/translog model with all
inputs variable, with equipment fixed, and
with equipment and building inputs fixed.
Although it appears that the cost function has
shifted up 10 to 15 percent from 1991 to 1996,
it should be noted that because we include
variables that control for the transition to new
machines, it is not possible to interpret these
coefficients as a technical change index, as is
often done. The reason is that the full effect of
technical change must include the effects of
these other variables.
Recall that we include an indicator variable
for the first quarter during which a processing
site employed the new machines, another
indicator variable for the second quarter of
use, and a time trend to capture the effects in
later quarters. The combined effect of these
three variables, which we call the learning-bydoing effect, is plotted in figure 4 for the GLS/
translog, GLS/hybrid-translog, and MLE /
translog models. As one might suspect, new
machines disrupt the previous work environment, and all the models report higher costs
in the quarter in which the new machines are
introduced. The models do differ as to what
happens after that. Our preferred estimation
technique, GLS, estimates that, after spiking
sharply in the quarter when the new machines
are introduced, costs fairly rapidly decline.
After just two years, the learning-by-doing
effect of the new machines would be more
than enough to offset the year-indicator variables. When the model is estimated using MLE,

16

T A B L E

6

Facility-Level Efficiency Estimate Statistics for
Federal Reserve Currency Operations
Functional form/estimation technique
Type of specification

GLS/translog

GLS/hybrid-translog

MLE/translog

Arithmetic mean

0.815

0.844

0.893

Weighted mean

0.810

0.838

0.888

Minimum

0.596

0.633

0.770

Maximum

1.00

1.00

0.952

Median

0.822

0.848

0.897

Arithmetic mean

0.825

0.863

0.902

Weighted mean

0.819

0.850

0.896

Minimum

0.623

0.662

0.800

Maximum

1.00

1.00

0.946

Median

0.834

0.861

0.903

Arithmetic mean

0.844

0.861

0.917

Weighted mean

0.836

0.866

0.911

Minimum

0.716

0.722

0.871

Maximum

1.00

1.00

0.958

Median

0.839

0.855

0.918

All inputs variable

Equipment inputs fixed

Equipment and building
inputs fixed

SOURCE: Authors’ calculations. The weighted mean level of cost efficiency
weights each facility-level cost-efficiency estimate by the mean number
of notes processed per quarter at that facility during the 1991:IQ to
1996:IVQ period.

the learning-by-doing effect is much smaller,
and consequently, the decline in costs is much
slower. As figure 5 reveals, the GLS estimates
of the learning-by-doing effect demonstrate the
same pattern whether capital or buildings are
treated as variable or fixed inputs.

Cost Efficiency
The estimates for cost efficiency using the
alternative functional forms, estimation techniques, and specifications are presented in
table 6. The GLS/translog model yields the
lowest average cost-efficiency estimates for
Federal Reserve System currency operations.
Depending on whether equipment and building inputs are fixed, the arithmetic average of
facility-level cost-efficiency estimates ranges
between 0.81 and 0.84, and the minimum cost-

efficiency estimates range between 0.60 and
0.72. These statistics suggest that during the
period studied, the average facility operated at
more than 80 percent of the efficiency of the
“best-practice” facility, and the worst performer
in the system could have substantially improved
its efficiency.
Estimates of average cost-efficiency levels
are higher when a hybrid-translog functional
form is used for the cost function. This is not
surprising because the hybrid-translog attributes a larger proportion of the variance in cost
to the structural component of the model. This
attribution results in higher average costefficiency estimates. In addition, estimates of
average cost-efficiency levels are higher when the
MLE technique is used to estimate the frontier.
The weighted means of the cost-efficiency
estimates are lower than the corresponding
arithmetic means. The weighted means are
lower because larger facilities in the System
typically have lower cost-efficiency estimates
than do smaller facilities.40
In general, cost-function specifications with
fixed inputs yield higher average levels of cost
efficiency, regardless of the functional form for
the cost function or the technique used to
estimate the cost frontier. This consistency
across the top, middle, and bottom panels of
table 6 suggests that managers view equipment
and building inputs as largely fixed. Because
changes to such inputs require substantial lead
times, this view seems quite plausible.
If facility-specific efficiency measures are
to be useful for managerial or policymaking
purposes, they should rank facilities consistently and be robust to different specifications
of the cost function or to different econometric
techniques. In tables 7, 8, and 9, we present
facility-specific cost-efficiency estimates for all
of our models. Table 7 reports estimates when
all inputs are variable, table 8 when equipment
inputs are fixed, and table 9 when equipment
and buildings are fixed inputs. Each table
shows the results of all combinations of the
functional form of the cost function and the
two estimation techniques we consider. In
addition, for each model, cost-efficiency
estimates are ranked from most efficient (costefficiency estimate equals one) to least efficient

■ 40 Currency processing and handling facilities with the largest
output levels tend to be Reserve Bank head offices. Such facilities have
additional administrative staff for the District. Even after incorporating a
head office indicator variable into the cost-function models, facilities with
the largest output volume tend to have lower levels of cost efficiency than
smaller facilities.

17

T A B L E

7

Cost-Efficiency Estimates and Relative
Rankings, Cost Model with All Inputs Variable

Federal
Reserve facility

GLS/translog costefficiency estimate

Rank

GLS/hybrid-translog
cost-efficiency estimate

FR 1

1.000

1

0.989

3

0.943

2

FR 2

0.991

2

1.000

1

0.941

3

FR 3

0.982

3

0.992

2

0.952

1

FR 4

0.958

4

0.971

4

0.931

4

FR 5

0.900

5

0.967

5

0.916

8

FR 6

0.872

6

0.910

6

0.899

15

FR 7

0.871

7

0.863

13

0.898

17

FR 8

0.871

8

0.849

17

0.926

5

FR 9

0.868

9

0.900

8

0.924

6

FR 10

0.859

10

0.890

9

0.920

7

FR 11

0.859

11

0.874

11

0.896

21

FR 12

0.855

12

0.848

18

0.916

9

FR 13

0.854

13

0.849

16

0.893

24

FR 14

0.854

14

0.874

10

0.912

10

FR 15

0.845

15

0.901

7

0.899

16

FR 16

0.837

16

0.854

14

0.882

26

FR 17

0.833

17

0.838

20

0.895

22

FR 18

0.831

18

0.821

24

0.893

23

FR 19

0.822

19

0.874

12

0.910

11

FR 20

0.810

20

0.853

15

0.896

20

FR 21

0.809

21

0.826

23

0.881

28

FR 22

0.808

22

0.848

19

0.900

14

FR 23

0.799

23

0.821

25

0.879

29

FR 24

0.786

24

0.817

26

0.897

19

FR 25

0.770

25

0.827

22

0.901

13

FR 26

0.764

26

0.802

29

0.891

25

FR 27

0.760

27

0.808

27

0.872

30

FR 28

0.756

28

0.801

30

0.882

27

FR 29

0.744

29

0.835

21

0.897

18

FR 30

0.728

30

0.804

28

0.906

12

FR 31

0.726

31

0.780

32

0.867

31

FR 32

0.722

32

0.712

36

0.839

35

FR 33

0.714

33

0.726

34

0.866

32

FR 34

0.706

34

0.759

33

0.838

36

FR 35

0.706

35

0.784

31

0.865

33

FR 36

0.693

36

0.723

35

0.852

34

FR 37

0.596

37

0.633

37

0.770

37

Rank

MLE/translog
cost-efficiency estimate

Rank

SOURCE: Authors’ calculations. Facility with rank 1 in each column is the “best-practice” facility using the specification/econometric technique specified. Cost-efficiency estimates are for the full period from 1991:IQ to 1996:IVQ.

18

T A B L E

8

Cost-Efficiency Estimates and Relative
Rankings, Cost Model with Equipment Fixed

Federal
Reserve facility

GLS/translog costefficiency estimate

Rank

GLS/hybrid-translog
cost-efficiency estimate

FR 1

0.921

4

0.957

5

0.931

4

FR 2

0.974

2

0.975

3

0.941

3

FR 3

0.943

3

0.969

4

0.946

1

FR 4

1.000

1

1.000

1

0.943

2

FR 5

0.911

5

0.983

2

0.927

5

FR 6

0.890

6

0.918

7

0.917

13

FR 7

0.862

12

0.891

11

0.908

17

FR 8

0.863

11

0.861

19

0.925

7

FR 9

0.861

14

0.918

6

0.924

8

FR 10

0.867

9

0.904

9

0.920

11

FR 11

0.865

10

0.882

13

0.902

21

FR 12

0.840

18

0.868

16

0.918

12

FR 13

0.815

21

0.851

23

0.883

31

FR 14

0.862

13

0.876

14

0.915

15

FR 15

0.846

16

0.905

8

0.908

16

FR 16

0.875

7

0.868

17

0.899

23

FR 17

0.809

24

0.856

22

0.903

20

FR 18

0.814

22

0.829

28

0.899

24

FR 19

0.844

17

0.885

12

0.916

14

FR 20

0.809

25

0.867

18

0.899

22

FR 21

0.828

20

0.858

21

0.894

28

FR 22

0.856

15

0.897

10

0.921

10

FR 23

0.812

23

0.807

31

0.896

26

FR 24

0.834

19

0.861

20

0.923

9

FR 25

0.788

26

0.842

25

0.894

27

FR 26

0.783

27

0.833

27

0.897

25

FR 27

0.732

33

0.801

32

0.864

33

FR 28

0.755

29

0.823

29

0.892

29

FR 29

0.781

28

0.851

24

0.903

19

FR 30

0.870

8

0.875

15

0.925

6

FR 31

0.750

31

0.810

30

0.891

30

FR 32

0.725

34

0.744

36

0.853

36

FR 33

0.712

35

0.769

35

0.874

32

FR 34

0.709

36

0.788

33

0.862

34

FR 35

0.754

30

0.842

26

0.905

18

FR 36

0.734

32

0.786

34

0.861

35

FR 37

0.623

37

0.662

37

0.800

37

Rank

MLE/translog
cost-efficiency estimate

SOURCE: Authors’ calculations. Facility with rank 1 in each column is the “best practice” facility using the specification/econometric
technique specified. Cost-efficiency estimates are for the full period from 1991:IQ to 1996:IVQ.

Rank

19

T A B L E

9

Cost-Efficiency Estimates and Relative Rankings,
Cost Model with Fixed Equipment and Buildings

Federal
Reserve facility

GLS/translog costefficiency estimate

Rank

GLS/hybrid-translog
cost-efficiency estimate

FR 1

0.963

4

0.961

3

0.943

4

FR 2

0.982

3

0.958

4

0.958

1

FR 3

0.994

2

0.977

2

0.944

3

FR 4

1.000

1

1.000

1

0.946

2

FR 5

0.858

14

0.911

6

0.931

11

FR 6

0.875

11

0.902

7

0.932

9

FR 7

0.844

17

0.854

20

0.919

17

FR 8

0.897

8

0.878

12

0.931

10

FR 9

0.844

18

0.864

16

0.927

14

FR 10

0.847

16

0.864

17

0.925

15

FR 11

0.823

21

0.845

25

0.919

18

FR 12

0.890

9

0.886

11

0.930

12

FR 13

0.817

24

0.845

26

0.905

29

FR 14

0.820

23

0.849

23

0.918

19

FR 15

0.801

28

0.831

27

0.918

21

FR 16

0.927

5

0.891

8

0.933

6

FR 17

0.863

12

0.861

18

0.918

20

FR 18

0.902

7

0.890

9

0.938

5

FR 19

0.837

20

0.853

21

0.914

23

FR 20

0.807

26

0.825

28

0.913

24

FR 21

0.800

29

0.814

32

0.910

28

FR 22

0.909

6

0.922

5

0.927

13

FR 23

0.803

27

0.803

34

0.902

31

FR 24

0.851

15

0.849

22

0.932

7

FR 25

0.839

19

0.875

13

0.910

27

FR 26

0.821

22

0.855

19

0.911

26

FR 27

0.796

31

0.846

24

0.896

32

FR 28

0.747

35

0.807

33

0.886

33

FR 29

0.811

25

0.874

15

0.912

25

FR 30

0.881

10

0.874

14

0.932

8

FR 31

0.753

33

0.819

29

0.903

30

FR 32

0.716

37

0.722

37

0.877

36

FR 33

0.796

30

0.816

30

0.916

22

FR 34

0.766

32

0.815

31

0.884

34

FR 35

0.860

13

0.888

10

0.924

16

FR 36

0.744

36

0.770

35

0.871

37

FR 37

0.747

34

0.760

36

0.882

35

Rank

MLE/translog
cost-efficiency estimate

SOURCE: Authors’ calculations. Facility with rank 1 in each column is the “best practice” facility using the specification/econometric
technique specified. Cost-efficiency estimates are for the full period from 1991:IQ to 1996:IVQ.

Rank

20

F I G U R E 6
Cost-Efficiency Rankings,
GLS/Translog
Rank
40
All inputs variable
Equipment fixed
Equipment and buildings fixed

35
30
25
20
15
10
5
0
0

5

10

15

20
25
Efficiency rank

30

35

40

30

35

40

SOURCE: Authors’ calculations.

F I G U R E 7
Cost-Efficiency Rankings,
GLS/Hybrid-translog
Rank
40
All inputs variable
Equipment fixed
Equipment and buildings fixed

35
30
25
20
15
10
5
0
0

5

10

20

15

Efficiency rank

SOURCE: Authors’ calculations.

25

(cost-efficiency equals the minimum costefficiency estimate). Facilities are identified
using their cost-efficiency estimate rank for the
GLS/translog model that considered all inputs
variable (table 7, second column).
Although the relative ranks of cost-efficiency
estimates for individual facilities vary considerably across our different models, a number of
facilities are consistently ranked among the
most and least efficient in the Federal Reserve
System. Facilities denoted by FR1 through FR4
are consistently ranked among the five most
efficient facilities. FR1 through FR6 are consistently among the top 10 efficient facilities. Consistently ranked among the least efficient facilities are facilities denoted by FR32, FR34, FR36,
and FR37. Interestingly, both the most efficient
and least efficient groups contain head offices
and branches. Rankings in the middle range of
efficiency rankings are more volatile because,
for facilities in that range, a small change in
efficiency leads to a large change in ranking,
and efficiency estimates for facilities in the
middle range are very close. The consistency
of the relative rankings of the most and least
efficient offices and the considerable disagreement about the rankings of facilities in
the middle of the range of rankings is illustrated
in figures 6 through 8.
We calculate Spearman and Kendall Tau-b
rank-order correlations, reported in tables 7, 8,
and 9, to determine the association between
cost-efficiency estimates. The Spearman rankorder correlation measure is concerned with
differences in absolute rankings, putting the
highest weight on facilities at the extremes,
whereas the Kendall Tau-b rank-order correlation measure captures the differences in
relative rankings. These rank-order correlations
are presented in tables 10 and 11 along with
rank-order correlations between cost-efficiency estimates and two accounting-based
performance measures—the number of notes
processed on high-speed equipment per
man-hour (a labor productivity measure) and
the number of notes processed per dollar of
cost (a unit cost measure). These accountingbased measures of performance do not take
into account differences in environmental
variables, nor do they account for the effect of
differences in the scale of operations between
facilities, but they are commonly used as
performance measures.

21

F I G U R E 8
Cost-Efficiency Rankings,
MLE/Translog
Rank
40
All inputs variable
Equipment fixed
Equipment and buildings fixed

35
30
25
20
15
10
5
0
0

5

10

15

20
Efficiency rank

25

30

35

40

SOURCE: Authors’ calculations.

The Spearman correlation coefficients
presented in table 10 are generally greater than
0.70, and each is statistically significant at the
5 percent level. Overall, such correlation
coefficients are larger when the models’
assumptions are more similar or when the
same econometric technique is used to estimate
the cost frontier. Not surprisingly, the weakest
relationships are between cost-frontier-based
and accounting-based measures of facility-level
efficiency.
The Kendall Tau-b correlation coefficients
presented in table 11 are smaller than the corresponding Spearman correlation coefficients
presented in table 10. This implies that the
efficiency rankings do not perform as well for
facilities in the middle cost-efficiency range.
As with the Spearman correlation coefficients,
every Kendall Tau-b correlation coefficient is
statistically significant at the 5 percent level,
which suggests an extremely high degree of
concordance between our cost-efficiency
measures. Again, consistency is greatest
between cost-efficiency measures when the
models’ assumptions are most similar or where
the same econometric technique is used to
estimate the cost frontier. Consistency is weakest
between cost-frontier-based and accounting-

based measures of facility-level efficiency.
Indeed, the Kendall correlation coefficients
between cost-frontier-based and accountingbased measures of facility-level efficiency are
all less than 0.50. This suggests that econometric measures of site-level efficiency may
provide useful insights for the identification
of “best-practice” currency processing and
handling facilities within the Federal Reserve
System, which accounting measures of performance may overlook.
Some Federal Reserve districts may focus on
cost efficiency to a greater degree than others.
Because five of the six consistently most efficient facilities are located in just two of the
twelve Federal Reserve districts, we test for
differences in the estimates of the mean level
of cost efficiency of facilities in those two
Federal Reserve districts against the remaining
ten districts. We find that the average efficiency
of the facilities in the two districts exceeds that
of the remainder of the Federal Reserve System
at the 1 percent significance level using a onesided t-test.
We were curious to see whether costefficiency estimates show any pattern across
services. To do so, we compare our baseline
estimates of currency cost efficiency with those
for check as reported by Bauer and Ferrier
(1996). The Spearman correlation coefficient is
0.309 and is statistically significant only at the
10 percent confidence level. The Kendall
Tau-b correlation coefficient is lower, 0.232,
but is statistically significant at the 5 percent
confidence level. These results suggest that
senior officers who oversee currency operations
well also tend to oversee check operations
well. These correlations also suggest that high
cost efficiency in one area is not achieved by
shifting costs to other operations.

22

T A B L E

10

Spearman Correlation Coefficients Between
Site-Level Measures of Efficiency

Cost-frontier-based cost-efficiency measures
All inputs variable
GLS/
translog

All
inputs
variable

GLS/translog

1.0

GLS/hybridtranslog
MLE/translog

Equipment
fixed

GLS/translog
GLS/hybridtranslog
MLE/translog

Equipment
and
buildings
fixed

GLS/translog
GLS/hybridtranslog
MLE/translog

Other

Labor
productivity
Unit cost

SOURCE: Authors’ calculations.

GLS/
MLE/
hybrid- translog
translog

Equipment fixed
GLS/
GLS/
MLE/
translog hybrid- translog
translog

Equipment and buildings fixed

Accounting-based
performance measures

GLS/
translog

GLS/
hybridtranslog

MLE/
translog

Labor
Productivity

Unit Cost

0.938

0.825

0.881

0.865

0.743

0.690

0.645

0.712

0.598

0.639

1.0

0.848

0.881

0.934

0.772

0.649

0.653

0.684

0.612

0.667

1.0

0.845

0.881

0.910

0.737

0.743

0.750

0.630

0.645

1.0

0.925

0.864

0.794

0.730

0.836

0.578

0.554

1.0

0.892

0.717

0.716

0.766

0.574

0.598

1.0

0.825

0.798

0.867

0.595

0.564

1.0

0.936

0.936

0.575

0.463

1.0

0.871

0.561

0.483

1.0

0.535

0.465

1.0

0.810
1.0

23

T A B L E

11

Kendall tau-b Correlation Coefficients Between
Site-Level Measures of Efficiency

Cost-frontier-based cost-efficiency measures
All inputs variable
GLS/
translog

All
inputs
variable

GLS/translog

1.0

GLS/hybridtranslog
MLE/translog

Equipment
fixed

GLS/translog
GLS/hybridtranslog
MLE/translog

Equipment
and
buildings
fixed

GLS/translog
GLS/hybridtranslog
MLE/translog

Other

Labor
productivity
Unit cost

SOURCE: Authors’ calculations.

GLS/
MLE/
hybrid- translog
translog

Equipment fixed
GLS/
GLS/
MLE/
translog hybrid- translog
translog

Equipment and buildings fixed

Accounting-based
performance measures

GLS/
translog

GLS/
hybridtranslog

MLE/
translog

Labor
Productivity

Unit Cost

0.804

0.670

0.742

0.691

0.592

0.526

0.496

0.565

0.411

0.456

1.0

0.697

0.727

0.802

0.625

0.486

0.480

0.544

0.426

0.471

1.0

0.670

0.703

0.778

0.562

0.574

0.571

0.465

0.486

1.0

0.781

0.689

0.616

0.568

0.661

0.429

0.414

1.0

0.727

0.541

0.534

0.598

0.396

0.423

1.0

0.664

0.662

0.715

0.429

0.429

1.0

0.790

0.787

0.399

0.324

1.0

0.697

0.399

0.354

1.0

0.396

0.345

1.0

0.637
1.0

24

V. Conclusion
I t costs more than half a billion dollars each
year to meet the demands for currency by
depository institutions, businesses, and consumers.41 Yet very little research has been
devoted to understanding the factors that affect
such costs. This paper has attempted to fill a
portion of this gap by considering the scale
and cost efficiency of Federal Reserve currency
operations.
Our finding that there are limited scale
economies for Federal Reserve currency operations suggests that currency services are not a
natural monopoly. As with other paper-based
payments technologies, the average cost curve
implied by the cost frontier for currency
operations is U-shaped with a fairly wide, flat
portion at the bottom of the U. Indeed, the
Federal Reserve System processes the vast
majority of notes at facilities that have nearconstant returns to scale. By the end of
1996, 10 Federal Reserve facilities were already
operating at constant returns to scale, and the
volume of notes processed (figure 1) has continued to increase since that time. Going forward, information on facility-specific marginal
costs and returns-to-scale measures could
potentially be used to improve resource allocations. For example, such information could
be used to set fees for some currency services
or to tailor cash service standards that define
normal service levels to each depository institution. From a policymaking perspective, the
technology for currency handling and processing does not appear to have the declining unit
costs that would give rise to a market failure
that is sufficient to justify or to sustain a
monopoly currency provider. Of course, there
may be other reasons for preserving the current arrangement given the Federal Reserve’s
critical role in supplying and maintaining the
integrity of currency in the United States.
Our finding that the average facility operates
at more than 80 percent of the efficiency of
the “best-practice” facility is comparable to
cost-efficiency estimates reported elsewhere
for private-sector financial institutions.42 Just
like its for-profit services,43 the Federal Reserve
could potentially reduce costs by having
the worst-performing facilities adopt the
procedures and operations of the bestperforming facilities.

Our finding of significant concordance
between cost-efficiency estimates for currency
processing and handling services, which are
not required by the MCA to recover economic
costs in the marketplace, and check processing
services, which are, suggests that the Federal
Reserve may have realized some of the benefits that a more competitive currency market
might have delivered. By encouraging the
formation of a competitive market for check
services, the MCA also gave Reserve Banks a
greater incentive to control costs and improve
resource allocation. Thus, the MCA appears to
have generated spillover benefits by creating a
management culture that increased operational
efficiency even for currency services in which
the Federal Reserve maintained its monopoly.
This spillover benefit from the Federal Reserve’s
participation in competitive payments markets
to services of a more purely governmental
nature has previously been ignored in studies
that have examined the Federal Reserve’s role
in the payments system.44

■ 41 In 2000, for example, the Federal Reserve spent about
$133 million on high-speed currency operations; $4.6 million on currency
cancellation, verification, and destruction; $65 million on paying and
receiving activities; and $456 million for printing of new Federal Reserve
notes, shipment of new notes by the Bureau of Printing and Engraving
(BPE), intra-System shipments of fit notes, counterfeit deterrence research,
the return of currency pallets to the BPE, and reimbursement to the U.S.
Treasury Office of Currency Standards. Federal Reserve outlays do not
include the costs of currency operations borne by depository institutions.
■ 42 See, for example, Bauer, Berger, and Humphrey (1993) and
Berger, Hancock, and Humphrey (1993).
■ 43 See Bauer and Hancock (1993) and Bauer and Ferrier (1996).
■ 44 See, for example, Green and Todd (2001).

25

References
Bauer, Paul W., and Gary D. Ferrier. “Scale
Economies, Cost Efficiencies, and Technological Change in Federal Reserve
Payments Processing,” Journal of Money,
Credit, and Banking, Part 2, vol. 28, no. 4
(November 1996), pp. 1004–39.
Bauer, Paul W., and Diana Hancock.
“The Efficiency of the Federal Reserve in
Providing Check Processing Services,”
Journal of Banking and Finance, vol. 17,
no. 2–3 (April 1993), pp. 287–311.
Bauer, Paul W., Allen N. Berger, and David
B. Humphrey. “Efficiency and Productivity
Growth in U.S. Banking,” in Harold O. Fried,
C.A. Knox Lovell, and Shelton S. Schmidt
(eds.), The Measurement of Productive
Efficiency: Techniques and Applications.
New York: Oxford University Press, 1993,
pp. 386–413.
Bauer, Paul W., Gary D. Ferrier, and
C. Knox Lovell. “A Technique for
Estimating a Cost System that Allows for
Inefficiency,” Federal Reserve Bank of
Cleveland, Working Paper no. 87–04, 1987.
Bauer, Paul W., Apostolos Burnetas,
Viswanath CVSA, and Gregory Reynolds.
“Optimal Use of Scale Economies in the
Federal Reserve’s Currency Infrastructure,”
Federal Reserve Bank of Cleveland,
Economic Review, vol. 36, no. 3 (Quarter 3,
2000), pp. 13–27.
Berger, Allen N. “‘Distribution-free’ Estimates
of Efficiency in the U.S. Banking Industry
and Tests of the Standard Distributional
Assumptions,” Journal of Productivity
Analysis, vol. 4, no. 3 (September 1993),
pp. 261–92.
Berger, Allen N., Diana Hancock, and
David B. Humphrey. “Bank Efficiency
Derived from the Profit Function,” Journal
of Banking and Finance, vol. 17, no. 2–3
(April 1993), pp. 317–48.
Berger, Allen N., John H. Leusner, and
John J. Mingo. “The Efficiency of Bank
Branches,” Journal of Monetary Economics,
vol. 40, no. 1 (September 1997), pp. 141–62.

Booth, George. Currency and Coin Responsibilities of the Federal Reserve: A Historical
Perspective. Cleveland, Oh.: Federal Reserve
Bank of Cleveland, 1989.
Board of Governors of the Federal Reserve
System. Annual Report: Budget Review.
Washington, D.C.: Board of Governors,
2000.
Board of Governors of the Federal Reserve
System. A Comprehensive Assessment of
U.S. Currency Quality, Age, and Cost
Relationships. Washington, D.C.: Board
of Governors, September 1991.
Board of Governors of the Federal Reserve
System. Eight Annual Report of the Federal
Reserve Board, 1921. Washington, D.C.:
Board of Governors, 1922.
Dotsey, Michael. “Cost Function for Federal
Reserve Currency Handling,” in A Comprehensive Assessment of U.S. Currency Quality,
Age, and Cost Relationships. Washington,
D.C.: Board of Governors of the Federal
Reserve System, September 1991.
Fama, Eugene F. “Commodity and Fiduciary
Standards,” unpublished manuscript,
December 1981.
Fama, Eugene F. “Financial Intermediation
and Price Level Control,” Journal of Monetary Economics, vol. 12, no. 1 (July 1983),
pp. 7–28.
Ferrier, Gary D., and C.A. Knox Lovell.
“Measuring Cost Efficiency in Banking:
Econometric and Linear Programming
Evidence,” Journal of Econometrics, vol. 46,
no. 1–2 (October–November 1990),
pp. 229–45.
Gallant, A. Ronald. “Unbiased Determination
of Production Technologies,” Journal of
Econometrics, vol. 20, no. 2, (November
1982), pp. 285–323.
Green, Edward J., and Richard B. Todd.
“Thoughts on the Fed’s Role in the
Payments System,” Federal Reserve Bank
of Minneapolis, 2000 Annual Report,
April 2001, pp. 1–25.

26

Hancock, Diana, David B. Humphrey, and
James A. Wilcox. “Cost Reductions in
Electronic Payments: The Roles of Consolidation, Economies of Scale, and Technical
Change,” Journal of Banking and Finance,
vol. 23, no. 2–4 (February 1999),
pp. 391–421.
Humphrey, David B. “Economies to Scale in
Federal Reserve Check Processing Operation,”
Journal of Econometrics, vol. 15, no. 1
(January 1981), pp. 155–73.
Jondrow, James, C.A. Knox Lovell, Ivan
Materov, and Peter Schmidt. “On the
Estimation of Technical Inefficiency in the
Stochastic Frontier Production Function
Model,” Journal of Econometrics, vol. 19,
no. 2–3 (August 1982) pp. 233–38.
King, Robert G. “On the Economics of Private
Money,” Journal of Monetary Economics,
vol. 12, no. 1 (July 1983), 127–58.

Sumner, Scott. “Privatizing the Mint,” Journal
of Money, Credit, and Banking, vol. 25,
no. 1 (February 1993), 13–29.
Sumner, Scott. “Is Nonprice Competition in
Currency Inefficient? A Comment,” Journal
of Money, Credit, and Banking, vol. 32,
no. 1 (February 2000), 146– 49.
Supel, Thomas M., and Richard M. Todd.
“Should Currency be Priced Like Cars?”
Federal Reserve Bank of Minneapolis,
Quarterly Review, vol. 8, no. 2, (Spring
1984), pp. 3–14.
U.S. Department of Commerce, Office of
Business Economics. Survey of Current
Business, Washington, D.C.: U.S. Government Printing Office, 1990–97.
U.S. Department of Labor, Bureau of Labor
Statistics. Producer Price Index Detailed
Report, Washington, D.C.: Government
Printing Office, 1990–97.

Lacker, Jeffrey M. “Stored Value Cards:
Costly Substitutes for Government Currency,”
Federal Reserve Bank of Richmond,
Economic Quarterly, vol. 82, no. 3
(Summer 1996), 1–25.

U.S. Department of the Treasury. Circular
No.55. Washington, D.C.: Government
Printing Office, 1920.

Means, R.S. Means Square Foot Costs.
17th edition. Kingston, Mass.: R.S. Means
Company, 1996.

White, Lawrence H. “Accounting for Noninterest Bearing Currency: A Critique of
the Legal Restrictions Theory of Money,”
Journal of Money, Credit and Banking,
vol. 19, no. 4 (November 1987), pp. 448–56.

Miller, Merton H., and Daniel Orr. “A Model
of the Demand for Money by Firms,”
Quarterly Journal of Economics, vol. 80,
no. 3 (August 1966), 413–35.
Proxmire, William. “Remarks on the
proposed Monetary Control Act,” Congressional Record (daily edition), March 27,
1980, p. S3168.
Schmidt, Peter, and Robin C. Sickels.
“Production Frontiers and Panel Data,”
Journal of Business and Economic Statistics,
vol. 2 (1984), pp. 367–74.

White, Lawrence H., and Donald J.
Boudreaux. “Is Nonprice Competition in
Currency Inefficient?” Journal of Money,
Credit, and Banking, vol. 30, no. 2
(May 1998), 252–60.
White, Lawrence H., and Donald J.
Boudreaux. “Is Nonprice Competition in
Currency Inefficient? A Reply,” Journal of
Money, Credit, and Banking, vol. 32, no. 1
(February 2000), 150–3.
Zimmerman, Gary C. “The Pricing of Federal
Reserve Services under MCA,” Federal
Reserve Bank of San Francisco, Economic
Review (Winter 1981), pp. 22–40.

27

The Employment of
Nations—A Primer
by Richard Rogerson
Richard Rogerson is the Rondthaler
Professor of Economics at Arizona State
University and a research associate at the
Federal Reserve Bank of Cleveland. He
thanks Nathan Porter and Yasuo Terajima
for research assistance and the National
Science Foundation for financial support.

Introduction
One of the striking macroeconomic developments of the last 30 years has been the marked
rise in European unemployment in comparison
to that in the United States. As of 2000, U.S.
unemployment was basically unchanged from
1970 at around 4 percent, whereas over the
same period, the unemployment rate in the
European Union almost tripled, increasing
from around 3 percent to almost 9 percent.
Not surprisingly, a relatively large literature has
emerged that both documents various aspects
of this differential evolution of unemployment
rates and tries to account for it.1
The premise of this paper is that our understanding of this and other related phenomena
will likely be facilitated by placing them in a
broader context. In particular, since economics
is often defined as the study of how scarce
resources are allocated, I follow the standard
economic approach of approaching the labor
market from the perspective of documenting
differences in resource allocations. Because
labor is a key input in the market production
of goods and services, it follows that how
much time is allocated to the market production of goods and services is likely to be an

important feature of the resource allocation
achieved by an economy. While economists
have long recognized variation in labor input
as a (if not the) central element in business
cycle fluctuations, in contexts other than
business cycles, variation in labor input has
received much less attention. The objective of
this paper, then, is to document the empirical
properties of the low-frequency component of
labor input in a cross section of industrialized
countries over the period 1960–95. This is
done using both aggregate data as well as data
that are disaggregated by age and sex.
While the larger issue of interest is that of
the allocation of time, this study will concern
itself almost exclusively with employment
patterns. As such, it abstracts from differences
in such things as workweeks and vacation
days. Obviously, it would be of interest to

■ 1 The literature is too large to reference exhaustively. See,
however, Bertola and Ichino (1995), Millard and Mortensen (1997),
Ljungqvist and Sargent (1998), Blanchard and Wolfers (2000), and
Blanchard (2000).

28

supplement the analysis provided here with
information on these additional aspects of time
allocation, but due to data limitations, I do not
undertake it in any great detail here. I also
restrict attention to a study of industrialized
countries, since there are reasons to believe
that countries at very different stages of development face considerations which make their
time-allocation problems less comparable.
Specifically, countries with large rural or agricultural populations may behave quite differently.
The paper presents its findings in the form
of nine stylized facts. I will not present them
all in this introduction, but a few are worth
emphasizing. First, large and persistent differences in employment-to-population ratios are
common throughout the period. While differences have become larger since 1970, persistent differences exist throughout the entire
sample and are in no way a novel feature of
the post-1970 world. Second, countries move
about considerably in the distribution of
employment-to-population ratios. Some countries move up, and some countries move
down. Third, and perhaps most important, a
comparison of aggregate and disaggregate data
shows that there is substantial variation in disaggregated labor input that is obscured when
examining aggregate data. When employmentto-population ratios change in a persistent
fashion, the changes are distributed across
groups disaggregated by sex and age in a very
disproportionate fashion. Because changes
across demographic groups display so much
variation, it seems natural to think that they
contain a great deal of information that will be
helpful in sorting out the causes of the aggregate changes.
The paper is unapologetically atheoretical.
There is no attempt to discuss what the facts
uncovered have to say about various models
of the labor market or various explanations for
differences in labor market outcomes. Rather,
the objective here is to simply lay out the facts
that a general theory of low-frequency movements in employment should be able to
account for.
An outline of the paper follows. The next
section gives a brief description of data
sources. Section II examines the aggregate
data and presents the main stylized facts that
follow from them, while section III does the
same for the data disaggregated by age and
sex. Section IV discusses some additional measures of labor input, and section V concludes.

I. Data
The measure of labor input that I use in this
study is employment to population. For the
aggregate analysis, I use the ratio of total
employment to the population of individuals
between the ages of 15 and 64.2
All of the data used in this analysis come
from Organisation for Economic Co-operation
and Development (OECD) sources, which in
turn are based on surveys carried out by individual member countries. Aggregate data on
employment relative to the population of individuals aged 15–64 are taken from various
issues of the OECD publication Historical
Statistics. Data on employment rates disaggregated by sex and age are taken from various
issues of the OECD publication Labor Force
Statistics. This publication actually provides
information on participation rates and unemployment rates, from which I have computed
the employment-to-population ratios. The
disaggregated data are not available for as
many countries or for as long a time, so the
time period and countries analyzed differ for
the two exercises. Section IV of the paper,
which considers some additional measures of
labor input, relies on data that appear in
various issues of the OECD publication
Employment Outlook.
As is true for any study that relies on crosscountry data from country-level sources, an
important caveat that must be mentioned is the
possibility that the data are not strictly comparable. Survey procedures may differ from
country to country, as may classification procedures. Additionally, there are occasional
changes in the accounting methods for some
countries.3 For examining cross-country differences in employment, these accounting issues
do not appear to be very significant. The
OECD publications listed above provide documentation of the various country surveys and
definitions used to determine employment,

■ 2 Note that this measure includes employed individuals above 65
in the numerator but not in the denominator. It is obviously debatable
whether this measure is preferable to the employment-to-population ratio
for all individuals or the employment-to-population ratio for all individuals
above the age of 15. Since much of the analysis is also carried out with
data that are disaggregated by age and the basic findings there are similar,
this is probably not an important issue for the purposes of this paper.
■ 3 For the most part, these did not appear to be too serious.

29

and there is a high degree of uniformity of the
criteria. Basically, in order to be counted as
employed, an individual has to have either
worked at least one hour for pay during the
reference period, had a job from which he or
she was absent (due to sickness, vacation,
strike, weather, and so on), been self-employed,
or been an unpaid employee in a family
business.4 Having said this, I will take the
employment data at face value, and no additional space will be devoted to this potentially
important measurement issue. However, in
section IV, when data are presented on hours
of work and part-time versus full-time employment, measurement issues are likely to be
substantial, and the appropriate caveats will
again be raised.

II. Facts about
Employment Rates I:
The Aggregate Data
This section focuses on patterns found in the
aggregate data. I examine 18 countries over
the period 1960–95.5 Because we are interested in low-frequency rather than highfrequency movements, I present five-year
averages at five-year intervals.6 In each case,
the indicated year is the center of the five-year
period used to construct the average; that is,
data for 1985 represent the average for the
period 1983–87.7 Table 1 presents the data on
aggregate employment relative to the total
population for individuals between the ages of
15 and 64 for 18 countries plus some summary
statistics: the mean, standard deviation, and the
85:15 ratio for the cross-sectional distribution
in each year. The 85:15 ratio is the ratio of the
highest to the lowest value for the employment ratio after having excluded the top two
and bottom two values.8 I report this rather
than the ratio of maximum to minimum values
in order to downplay the possible role of
extreme values. Note also that the mean value
reported is the simple mean of the cross-country
observations and does not weight countries by
their size.
Some of the important features of these data
are described next.

Fact 1. The average employment rate
remained roughly constant over this
period.
As table 1 indicates, there is no evidence of
a secular trend in the average employment rate
across countries. The average values from 1980
onward are slightly lower than the average
values before 1975, but by less than one-half
of a percentage point. The average value does
fluctuate somewhat over time, and the fluctuations seem larger in the post-1975 period.
Fact 2. Differences in employment rates
across countries are large.
No matter which of the cross sections we
look at, the variation in employment rates is
strikingly large. For the sample as a whole, the
maximum values exceed 80 percent, whereas
the lowest values are less than 50 percent. The
standard deviations are also large, though it
can be difficult to gauge what constitutes large
in this context. The 85:15 ratio is perhaps more
informative in this regard. Note that this ratio
exceeds 1.20 for each cross section. To understand the significance of this value, note that
in the largest postwar recession in the United
States (1982), the ratio of employment to population at the peak was only about 1.03 times
larger than it was at the trough. It is particularly significant that the large difference in
employment rates across countries is not a
phenomenon that emerged after the early
1970s. Cross-country differences were large
even in the 1960–70 period.

■ 4 We note that cross-country studies of unemployment are
potentially more problematic because of differences in criteria across
countries, particularly in what constitutes searching for employment.
However, the OECD does publish a series of “standardized” unemployment
rates that attempts to correct for cross-country differences in measurement.
■ 5 See Blanchard and Wolfers (2000) for a presentation of the
low-frequency movements in unemployment rates over this period.
■ 6 An alternative procedure to isolate the low-frequency component
would be to use the trend component generated by applying the Hodrick–
Prescott filter to the data. The results of using this alternative procedure are
very similar and hence are not reported.
■ 7 The exception to this is 1960. Due to data limitations, 1960
simply refers to the average for 1960–62.
■ 8 More accurately, this is really the 83:17 ratio.

30

T A B L E

1

Aggregate Employment/Total
Population, Ages 15–64

1960

1965

1970

1975

1980

1985

1990

1995

Canada
United States
Japan
Australia
New Zealand
Belgium
Denmark
France
Germany
Ireland
Italy
Netherlands
Norway
Portugal
Spain
Sweden
Switzerland
United Kingdom

58.9
62.2
74.3
64.6
62.9
59.5
70.0
68.6
70.1
64.4
62.6
61.1
63.6
59.2
59.9
73.0
78.2
70.8

60.9
62.9
71.5
66.6
63.7
60.5
72.2
66.4
69.6
64.3
58.8
59.9
63.0
59.8
59.8
71.9
78.7
71.4

61.3
64.5
71.1
68.8
64.1
61.3
74.3
65.8
68.6
62.0
56.3
57.3
64.1
62.2
60.1
72.7
77.5
70.4

63.4
64.4
69.7
67.5
64.8
61.1
73.3
65.4
65.7
58.0
55.6
54.5
68.9
64.5
58.1
76.3
75.3
70.8

65.5
66.8
70.3
65.5
64.2
58.4
73.1
63.5
63.6
57.0
55.7
53.6
73.8
63.8
50.2
78.7
73.7
68.7

66.1
68.5
70.6
64.7
62.3
54.5
75.3
59.4
62.5
51.7
54.0
53.3
76.0
64.6
45.4
79.3
74.6
65.7

68.9
72.4
72.5
67.6
66.8
56.6
76.8
60.1
65.8
52.7
55.0
61.5
74.3
68.9
48.9
79.8
81.7
70.7

67.3
73.4
74.4
67.4
69.2
56.3
74.1
58.9
65.2
55.2
52.5
65.3
74.2
67.3
46.4
70.4
79.9
68.6

Mean
Standard
deviation
85:15 ratio

65.7

65.7

65.7

65.4

64.8

63.8

66.7

65.9

5.81
1.23

5.71
1.20

5.91
1.21

6.38
1.26

7.67
1.32

9.45
1.41

9.30
1.40

8.82
1.34

SOURCES: Organisation for Economic Co-operation and Development, Historical Statistics, various issues; and
author’s calculations.

Fact 3. The spread of employment rates
across countries changes substantially
across time.
Whether one looks at the standard deviation
or the 85:15 ratio, the spread of employment
rates across countries has exhibited substantial change over time. The distribution
became significantly more spread out during
the first part of the sample period, and though

it subsequently became more compressed,
there has still been a significant spreading out
over the period as a whole.
Fact 4. Differences in employment rates
across countries are persistent.
To see this, I report the correlation matrix
for the country-level observations across time
in table 2.

31

T A B L E

2

Correlation Matrix for County-Level
Observations across Time

1960
1960
1965
1970
1975
1980
1985
1990
1995

1965

1970

1975

1980

1985

1990

1995

1
.95

1

1
.95
.88
.76
.66
.59
.60
.53

1
.96
.84
.72
.65
.68
.62

1
.93
.81
.75
.77
.69

1
.94
.89
.87
.76

1
.98
.93
.84

1
.97
.89

SOURCES: Organisation for Economic Co-operation and Development, Historical Statistics, various issues;
and author’s calculations.

Note that at one lag (five years), the average
correlation coefficient exceeds 0.95. Obviously,
this indicates a great deal of persistence.9 Note
also, however, that the average persistence
decreases as the number of lags increases. At
four periods (20 years) the average correlation
drops to 0.70. It is also interesting to note that
these correlations are not driven by changes
that occurred before and after the oil shock of
the mid-1970s. For example, the correlation
between 1980 and subsequent years looks
very similar to the pattern of correlations
between 1960 and subsequent years. The fact
that at longer horizons the correlations
decrease considerably suggests our next fact.
Fact 5. There is considerable mobility
within the distribution.
There are several ways to motivate this fact.
One is to simply look at the changes in individual countries over the entire sample. Table 3
reports changes over three horizons—1960 to
1995, 1960 to 1970, and 1985 to 1995. This
split is of interest because the latter two lie on
either side of the 1970s’ oil shocks.

From the perspective of mobility, the key
is dispersion in changes. The dispersion of
changes over each period in table 3 is large.
Over the 1960–95 period, some countries see
their employment rates climb by more than
10 percentage points, while others see their
employment rates fall by more than the same
amount. Moreover, there is very little correlation between a country’s starting value in
1960 and the subsequent change; the
correlation is only –0.18. It is important to
emphasize that the situation is not characterized as one in which countries simply differ in
the extent to which their employment rates
decrease—almost half of the countries experience an increase in their employment rates
over this period, and as we know from above,
there is virtually no secular trend in the average employment rate across these countries.

■ 9 Keep in mind that our observations are themselves five-year
averages, which obviously induce some persistence relative to what
would be found using annual data.

32

T A B L E

3

Changes in Employment Rates

∆ (1960 –95)
Canada
United States
Japan
Australia
New Zealand
Belgium
Denmark
France
Germany
Ireland
Italy
Netherlands
Norway
Portugal
Spain
Sweden
Switzerland
United Kingdom

+8.4
+11.2
+0.1
+2.8
+6.3
–3.2
+4.1
–9.7
–4.9
–9.2
–10.1
+4.2
+10.6
+8.1
–13.5
–2.6
+1.7
–2.2

∆ (1960 –70) ∆ (1985–95)
+2.4
+2.3
–3.2
+4.2
+1.2
+1.8
+4.3
–2.8
–1.5
–2.4
–6.3
–3.8
+0.5
+3.0
+0.2
–0.3
–0.7
–0.4

+1.2
+4.9
+3.8
+2.7
+6.9
+1.8
–1.2
–0.5
+2.7
+3.5
–1.5
+12.0
–1.8
+2.7
+1.0
–8.9
+5.3
+2.9

SOURCES: Organisation for Economic Co-operation and Development, Historical Statistics, various issues;
and author’s calculations.

As suggested earlier, it is significant that
there is substantial mobility even abstracting
from the 1970–85 period. In the period
1960–70, the range of changes exceeds
10 percentage points, and the same holds true
for the period 1985–95 as well. It is also of
interest to note the heterogeneity of experiences
for countries that experience similar changes
over the full sample. Specifically, France, Italy,
and Spain all experience drops in their employment rates of around 10 percent. However,
Italy experiences over half of this drop in the
first 10 years, while France experiences
roughly a third in the first decade, and Spain
experiences virtually no decline in this period.
It should also be noted that there is virtually
zero correlation between changes in the first
decade and changes in the final decade.

Lastly, to highlight the range of mobility
experiences found in the data, consider the
evolutions of the Netherlands and Sweden.
Consider first the Netherlands. Beginning in
1965, the employment rate in the Netherlands
begins a steady decline, losing almost 8 percentage points and bottoming out in 1985.
Subsequently, however, the rate increases
by more than 10 percentage points, and the
Netherlands climbs significantly in the distribution. Next consider Sweden. Between 1960 and
1970, Sweden’s employment rate is relatively
constant, after which it increases substantially,
raising Sweden to the top of the distribution
between 1980 and 1990. Subsequently, however, the employment rate falls back to a value
very near its level in the 1960–70 period.

33

T A B L E

4

Correlation between Employment
and Unemployment Rates

Trend components

Cyclical components

.96
.53
–.30
–.42
–.97
–.89
.73
–.99
–.87
–.98
–.88
–.32
.79
.86
–.98
.10
–.36
–.94

–.92
–.92
–.60
–.90
–.90
–.76
–.88
–.77
–.77
–.85
–.72
–.52
–.81
–.77
–.89
–.93
–.57
–.97

Canada
United States
Japan
Australia
New Zealand
Belgium
Denmark
France
Germany
Ireland
Italy
Netherlands
Norway
Portugal
Spain
Sweden
Switzerland
United Kingdom

SOURCES: Organisation for Economic Co-operation and Development, Historical Statistics, various issues;
and author’s calculations.

A Digression on
Unemployment
versus Employment
The previous analysis has focused on employment-to-population ratios as the measure of
labor input. As noted earlier, many studies
focus instead on unemployment rates. I have
argued that from the perspective of understanding differences in resource allocations, it
is differences in labor input that are of primary
interest. But is the choice really substantive, or
is it the case that movements in employment
and unemployment are close to mirror images
of each other, in which case the issue is largely
irrelevant? In this section I present some
evidence on the matter. It turns out that while
movements in employment and unemployment
are highly negatively correlated at cyclical
frequencies, this is not necessarily true at lower

frequencies. To investigate this I examined the
series for employment-to-population ratios
used earlier and the standardized unemployment rate series produced by the OECD for the
period 1960–95. For each country I apply the
Hodrick–Prescott filter to each series and compute the trend and cyclical components. For
each country I compare the behavior of the
two trend series and the two cyclical series,
in each case computing the correlation
between the two series. Table 4 displays the
correlations.
The two columns of table 4 tell quite a
different story. The first column shows that
there is really no tendency for trend increases
in the employment rate to be associated with
either trend increases or decreases in the
unemployment rate; while there are many
values that are close to negative one, there are
also several that are close to positive one, as

34

well as several that are not close to either of the
extreme values. Reflecting this, the average of
the values in the first column is –0.27. On the
other hand, the values in the second column
are all negative, are all greater than 0.5 in
absolute value, and all have an average value
of –0.80.
The message from this is that when studying
low-frequency movements in the labor market,
one may get a very different picture depending
upon the measure one uses. Having said this, it
would appear that in at least one context—that
of the “European unemployment problem”
mentioned in the introduction—the differences
are likely to be quantitative rather than qualitative in nature, at least in terms of changes over
this period. The main reason for this is that
many of the European countries that experienced large decreases in employment over the
sample period, including Belgium, France,
Germany, Italy, Ireland, and Spain, are countries
for which the correlation between employment
and unemployment is close to negative one.

highest to lowest values after removing the
two highest and lowest values for each category.
Several features of the data in table 5 are
worth remarking on. First, the basic shape of
the life-cycle employment rate profile is the
same across all countries for both men and
women. The basic shape is that of an inverted
U. The peak in most cases occurs for the
35–44 age group, though in some cases it
occurs for the 45–54 age group. However,
although the basic shape is the same across
countries, there are important quantitative
differences in the profiles across countries. In
particular, the disaggregated data show that the
large differences in aggregate data do not
reflect an aggregation phenomenon; that is, it
is not the case that aggregate differences are
accounted for by different age distributions
across countries that have very similar lifecycle profiles. Examination of tables 5(a) and
5(b) lead to the following conclusions.

III. Facts about
Employment Rates II:
The Disaggregated
Data

Fact 6.2. Differences in employment rates
are much larger for young and old individuals than for prime-aged individuals.
These are pretty much self-evident from
table 5. For example, the 80:20 ratios are
U-shaped for both men and women, and the
values are much higher for women than for
men with the lone exception of the over-65
group. It follows that many of the differences
in aggregate data are driven by differences
among women and men who are not of prime
age. Specifically, the tables show that the
highest correlation between disaggregate and
aggregate values occurs for prime-aged
women—the correlation of employment rates
for women in the two age groups 35–44 and
45–54 is 0.86. In contrast, the correlation coefficient between aggregate employment rates and
those for men aged 35–44 is only 0.33 and
0.41 for those aged 45–54. In fact, excluding
workers over the age of 65, the lowest value
of this correlation occurs for males aged
35–44. However, even though differences are
smallest for prime-aged males, it is worth noting that the data still support the following
conclusion.

This section examines employment rates from
the perspective of data that are disaggregated
by age and sex. As mentioned earlier, data
availability limits the scope of the analysis in
terms of time period and the set of countries
that can be examined. I note first that using the
available data, one can establish the equivalent
to facts 2, 4, and 5 for the disaggregated data
as well: there are large differences in the rates
across countries, and they are persistent, but
there is also substantial mobility. Hence, in this
section I focus on additional findings that the
disaggregated data present us with. A basic
issue to explore here is the extent to which
the aggregate data capture the differences that
exist across countries at a point in time and
the changes that take place across countries
over time.
We begin by examining the cross sections
for 1995 for a set of 15 countries. For the timeseries analysis we are restricted to a much
smaller set of countries, but it is of interest to
examine one cross section for a much larger
set. In this case, the data represent a three-year
average, taken over the period 1994 –96.
Table 5 presents the data and several summary
statistics. The 80:20 ratio represents the ratio of

Fact 6.1. Differences in employment rates
are much larger for women than for men.

35

T A B L E

5

Employment Rates over the
Life Cycle, 1995

A. Men
15–24

25–34

35–44

45–54

55–64

65+

Canada
United States
Japan
Korea
New Zealand
Belgium
Denmark
France
Germany
Ireland
Italy
Portugal
Spain
Sweden
United Kingdom

52.2
58.0
44.9
27.6
61.3
29.3
68.5
22.0
51.4
37.7
26.4
42.7
26.6
41.3
61.2

81.7
88.0
94.3
90.5
86.2
86.9
87.0
83.7
83.6
81.1
71.1
86.4
72.2
79.0
84.2

84.6
88.6
96.1
95.2
87.7
89.0
88.5
88.7
90.4
81.7
88.6
91.9
83.2
85.0
86.0

82.8
85.7
95.7
92.3
87.1
82.5
86.1
86.3
87.4
77.2
90.2
86.6
80.3
86.4
85.8

54.4
63.7
80.9
78.8
62.9
34.8
60.5
38.7
47.4
59.0
55.5
58.9
48.8
65.0
56.6

10.1
16.2
36.4
41.6
9.9
2.3
4.3
2.6
4.2
15.0
6.0
21.8
2.8
13.2
7.5

Mean
Standard deviation
Coefficient of variation
80:20 ratio
Correlation with
aggregate aged 15–64
Correlation with
men aged 35–44

43.4
15.0
.35
2.30

83.7
6.2
.07
1.11

88.3
4.0
.05
1.09

86.2
4.5
.05
1.09

57.7
12.5
.22
1.37

12.9
12.1
.93
7.79

.66

.59

.33

.41

.45

.30

–.06

.68

1

.86

.44

.65

36

B. Women
15–24

25–34

35–44

45–54

55–64

65+

Canada
United States
Japan
Korea
New Zealand
Belgium
Denmark
France
Germany
Ireland
Italy
Portugal
Spain
Sweden
United Kingdom

51.6
55.2
44.5
39.3
55.4
24.0
60.7
18.3
48.6
34.0
21.6
33.5
21.4
41.9
57.5

69.4
70.4
57.6
47.6
62.0
69.9
72.6
65.2
65.8
62.1
46.9
73.0
44.9
73.9
66.2

71.8
73.8
63.9
61.6
70.0
63.1
80.9
69.3
67.7
45.7
51.8
74.5
43.1
83.0
71.9

66.4
72.2
68.1
59.6
73.6
45.5
73.7
67.3
62.2
36.2
47.4
61.0
32.0
85.3
71.7

33.9
47.5
47.5
49.9
38.3
12.5
37.8
28.6
24.5
20.3
20.2
34.1
17.6
59.9
38.9

3.3
8.5
15.5
20.5
3.1
0.9
1.2
1.3
1.5
2.7
1.7
9.7
1.3
4.6
3.1

Mean
Standard deviation
Coefficient of variation
80:20 ratio
Correlation with
aggregate aged 15–64
Correlation with
females aged 35–44

40.5
14.4
.36
2.56

63.2
9.8
.15
1.53

66.1
11.7
.18
1.44

61.5
15.0
.24
1.62

34.1
13.6
.40
2.35

5.3
5.9
1.11
7.46

.75

.63

.86

.86

.78

.32

.60

.79

1

.92

.64

.06

SOURCES: Organisation for Economic Co-operation and Development, Labor Force Statistics, various issues; and
author’s calculations.

Fact 7. Differences in employment rates
across countries, even for prime-aged
males, are large.
To see this, simply recall the comment made
earlier to put the 80:20 ratios in perspective.
In postwar time series data for the United States,
a value of 1.03 is big. Hence, values of 1.1
must also be viewed as large.
Fact 8. Differences in disaggregated
employment rates are not proportional
to differences in aggregate data.
Another way to phrase this is to say that
when two countries have very different
aggregate employment rates, the corresponding
life-cycle profiles of employment rates are not
simply shifted up or down in a parallel fashion,
even controlling for sex. The relatively low
correlations between prime-aged employment
rates and the rates for other age groups
(controlling for sex) indicate this. A few examples serve to illustrate the significance of this

point. The three largest economies of continental Europe, Germany, Italy, and France, all
have employment rates at the aggregate level
(or even for males) that are substantially lower
than that of the United States in 1995. However, each of these countries has prime-aged
employment rates for men that actually exceed
the corresponding value for the United States!
Yet, in sharp contrast, for the 55–64 age group
the employment rate in the United States is
roughly double that in these other countries.
Spain has the lowest employment rates in the
aggregate level, and this continues to hold
when the data are disaggregated by sex.
Canada, on the other hand, has a relatively
high aggregate employment rate. Yet the
employment rates for prime-aged males are
roughly similar in these two countries. On the
other hand, male youths in Canada are almost
twice as likely to be employed as their Spanish
counterparts, as is roughly true for females of
all age groups in the two countries.

37

T A B L E

6

Changes in Employment Rates:
1972 to 1995

A. Men
Canada
United States
Japan
France
Germany
Italy
Spain
Sweden

15–24

25–34

35–44

45–54

55–64

65+

–4.7
–8.6
–10.3
–33.7
–18.5
–16.0
–43.2
–22.8

–9.9
–3.1
–2.4
–12.4
–10.8
–23.6
–21.1
–12.5

–8.2
–4.5
–0.9
–9.6
–7.1
—
–13.0
–8.6

–7.7
–3.6
–0.1
–7.7
–6.8
–5.0
–14.0
–6.6

–23.3
–12.1
–4.1
–32.4
–26.6
–13.2
–32.3
–16.3

15–24

25–34

35–44

45–54

55–64

65+

6.2
7.5
–4.8
–25.0
–13.3
–7.6
–29.4
–14.9

27.2
24.4
13.4
10.0
15.0
15.6
16.9
11.3

30.4
23.8
5.6
19.6
17.9
—
18.5
12.9

25.7
18.7
7.9
16.6
13.3
17.5
5.5
16.5

4.4
7.3
3.6
–8.1
–1.3
11.3
–5.7
15.4

1.0
0.2
–0.6
–5.6
–4.3
0.3
–6.5
–2.9

–7.4
–5.5
–9.7
–16.3
–10.2
–1.4
–22.5
–10.8

A. Women
Canada
United States
Japan
France
Germany
Italy
Spain
Sweden

SOURCES: Organisation for Economic Co-operation and Development, Labor Force Statistics, various issues; and
author’s calculations.

Having examined the differences in lifecycle employment rates from a single cross
section, I next turn to a look at the evolution
of life-cycle profiles over time and across
countries. For this exercise, data availability
limits us to a set of eight countries—Canada,
the United States, Japan, France, Germany,
Italy, Spain, and Sweden, and the period
1972–95. The appendix contains a complete
set of tables for the evolution of these life
cycles over time. For the sake of illustration I
focus on changes between 1972 and 1995.
The data for 1995 are a three-year average
centered on 1995, whereas the data for 1972
are simply an average of 1972 and 1973.10
Table 6 presents changes in life-cycle employment profiles, disaggregated by age and sex
for eight countries.11

A look at table 6 indicates a striking pattern
for this period—in all countries there was a
huge reallocation of employment away from
males and toward females. In the table for
males, every entry is negative. In the table for
women, the entries are all positive for ages
25–54. A closer look reveals the following fact.

■ 10 Data for 1971 are not available for all countries, so the two-year
average is used rather than further reducing the number of countries in the
sample. It also seemed preferable to not use 1974 since this marks the
beginning of the oil price shocks.
■ 11 Note that there are two missing values for Italy.

38

T A B L E

7

Changes in Normalized Life-Cycle
Profiles, 1972–1973 to 1994–1996

A. Men
Canada
United States
Japan
France
Germany
Italy
Spain
Sweden
Standard deviation

15–24

25–34

45–54

55–64

Scale

1.00
.91
.82
.43
.79
.69
.44
.71
.21

.97
1.01
.98
.95
.96
.83
.89
.95
.06

1.00
1.01
1.01
1.01
1.00
1.05
.98
1.02
.02

.77
.88
.96
.60
.69
.82
.69
.88
.12

.91
.95
.99
.91
.93
.91
.86
.91
.04

15–24

25–34

45–54

55–64

Scale

.95
1.04
1.19
.85
.95
.90
.92
1.00
.11

.94
.95
1.03
.95
.94
.95
.69
1.05
.11

.66
.80
.99
.56
.70
1.36
.43
1.14
.31

1.73
1.48
1.10
1.39
1.36
1.67
1.75
1.18
.24

B. Women
Canada
United States
Japan
France
Germany
Italy
Spain
Sweden
Standard deviation

.66
.78
.82
.30
.58
.44
.24
.62
.21

Absolute
deviations
.26
.23
.26
1.03
.56
.71
1.00
.48

Absolute
deviations
.79
.51
.41
1.34
.83
1.07
1.72
.57

SOURCES: Organisation for Economic Co-operation and Development, Labor Force Statistics, various issues; and author’s
calculations.

Fact 9. Changes in aggregate employment
rates are associated with large changes in
the shape of life-cycle employment-rate
profiles.
This fact is really a time-series equivalent to
fact 8. There we saw that if two countries have
very different aggregate employment rates, the
life-cycle profiles of employment rates are not
simply parallel transformations of each other.
What table 6 shows is that when the employment rate changes in a country, the life-cycle
employment-rate profiles do not shift in a parallel fashion. Consider a few examples. Aggregate employment and male employment rates
fall significantly over this period in France,
Germany, Italy, and Spain. Yet the change is
disproportionately accounted for by changes
in the employment rates of young and old

workers. The changes would look even more
skewed if we considered them relative to the
starting values—in some countries, youth
employment rates are falling by more than
half. Similarly, female employment rates are
actually increasing over this period in France,
Germany, and Spain, yet despite this, there are
massive decreases in the employment rates for
female youths.
I next present this same information in a
manner that can better highlight the relative
contribution of parallel movements of the lifecycle profile against changes in the shape of
the profile. I adopt the following procedure to
normalize the shape of the life-cycle profile.
Normalize the employment rate for the 35–44
age group to one, and then express all other
values relative to this value. This profile of

39

relative values plus the actual employment rate
for the 35–44 group completely describes the
whole profile. This procedure is carried out for
both the 1972 and 1995 data. Then, to summarize the changes in life-cycle profiles over time,
I compute the ratio of the relative values and
the ratio of the actual values for the 35–44 age
group. I do this for both sexes for each of the
countries in our sample, and table 7 presents
the results. The column labeled scale indicates
the ratio of the employment rate for the 35–44
group in 1995 relative to its value in 1972–73.
The columns with age ranges show the ratios
of the relative life-cycle employment profiles.
The column labeled absolute deviations gives
the sum of the absolute deviations of the four
life-cycle points from one. Note that if the
shape of the profile stayed the same but was
shifted up or down proportionately, each of
the first four columns would have a value of
one, and the sum of absolute deviations would
be zero. Hence, the last column is a measure
of how much the shape of the profile is
changing. Table 7 does not report data for the
over-65 age group since its relatively low
employment rates make it relatively unimportant from our perspective.
Consider the results for men first. It is
striking that the values in the third column
(ages 45–54) are so close to one. This implies
that the shape of the life-cycle employment
profile changes very little for prime-aged
males. For the other age groups, the changes
are much larger. Note that the scale factor is
less than one for each country. Note also that
the vast majority of the values in the table are
less than one as well. Comparing the standard
deviations of the various columns gives us a
way to ascertain the extent to which the
changes in scale dominate the changes in
shape. For three of the four age groups, the
standard deviation of the shape ratios exceeds
the standard deviation of the scale factors.
Now consider the case of women. First note
the magnitude of the scale factors—they are
all much larger than one, indicating that in all
countries the employment rate for women
aged 35–44 increased. However, the rest of the
entries are typically less than one, indicating
that although the employment rates were
increasing for most age groups in most countries, the increase was less than proportionate
to the increase for prime-aged women. Once
again, however, with the exception of Spain,
the entries for the 45–54 age group are fairly
close to one. Interestingly, however, for
women the standard deviation in scale factors
exceeds that of three of the four changes in

shape. This is the opposite of what was found
for men.
Note the asymmetry of the changes for men
and women. For men, the changes for primeaged workers are much smaller proportionately
than are the changes for other age groups.
For women, the opposite is true. Changes are
proportionately largest for prime-aged women.
An important development is that in 1994–96,
women’s life-cycle profiles increasingly resemble those of their male counterparts, whereas
in the 1972–73 period, women’s profiles
tended to be relatively flat.

A Closer Look at
Sweden and the
Netherlands
To illustrate the range of experiences that
exists across countries, we now take a closer
look at Sweden and the Netherlands for the
period 1972–96. Recall from the analysis of
aggregate data that these two countries follow
quite different paths over this period. Sweden
experiences a significant increase in its
employment rate over the first part of the
period and then witnesses a decline at the end
that brings it back roughly to where it began.
In contrast, the Netherlands experiences a
large decrease in its employment rate over the
first part but subsequently experiences a large
increase and ultimately ends the period with a
higher employment rate. What happened to
the life-cycle employment profiles for these
countries over this period? That is the issue I
turn to next. I examine the profiles at the dates
1972, 1980, 1985, 1990, and 1995. In each case,
the data represent a three-year average with
the given year as the midpoint of the period,
except for the initial point, which is a two-year
average based on 1972–73. Data limitations
require that we focus on three age groups:
15–24, 25–54, and 55–64. For each year, I
present normalized profiles that give the
relative employment rates for the two extreme
age groups relative to the 25–54 age group.
I also report a scale factor that gives the employment rate of the 25–54 group relative to its
value in 1972. Table 8 gives the results. These
data are aggregated for men and women.

40

T A B L E

8

Changing Profiles in Sweden
and the Netherlands

A. Sweden

1972
1980
1985
1990
1995

ER15–24
ER25–54

ER55–64
ER25–54

Scale

.76
.75
.68
.71
.50

.79
.75
.72
.76
.76

1.00
1.10
1.13
1.14
1.03

ER15–24
ER25–54

ER55–64
ER25–54

Scale

.93
.71
.64
.76
.75

.77
.56
.44
.41
.40

1.00
1.06
1.04
1.21
1.28

B. Netherlands

1972
1980
1985
1990
1995

SOURCES: Organisation for Economic Co-operation and Development, Labor Force Statistics, various issues;
and author’s calculations.

The differences across the two countries are
rather striking. First consider the case of Sweden.
As the scale column indicates, the employment
rate for prime-aged workers mimics the behavior
of the aggregate employment rate: a substantial
increase between 1972 and 1980, relative
constancy over the 1980s, and then a decline
back to its earlier value. Moreover, until the
final year, the relative life-cycle profile of
employment rates changes little. However,
when the aggregate employment rate falls in
1995, the shape of the profile changes a good
deal, as the employment rate for young workers
falls disproportionately. Next, consider the case
of the Netherlands. What is most striking in the
Netherlands is that the behavior of the primeaged employment rate does not really mimic at
all the behavior of the aggregate employment
rate. Whereas the aggregate employment rate
drops continuously between 1970 and 1985,
the prime-aged-employment rate experiences
a mild increase between 1972 and 1985. Subsequently, it does mirror the large increase
between 1985 and 1995 found in the aggregate
rate. In contrast to the case of Sweden, however, the shape of the life-cycle employment

profile changes substantially. Over the period
1972–85, there are massive relative downward
shifts for both young and old workers. Employment rates for these groups fall relative to those
for prime-aged individuals by roughly a third.
Moreover, over the period 1985–95, although
the employment rate for young individuals
partially recovers, the relative rate for older
individuals falls slightly.

A Closer Look at
Prime-Aged Males
Prime-aged males are a group that attracts
considerable attention in cross-country comparisons of labor market outcomes. One reason
for this interest is that they correspond to a
group whose main activity is presumed to be
market work. They are too old to be in school,
too young to be retired, and, even though
social norms are changing, they typically do
not have primary responsibility for child care
or other family situations. Although fact 7
explicitly deals with differences among primeaged males in the 1995 cross section (noting

41

T A B L E

9

Changes in Employment Rates:
1972 to 1995

Males 35–44
1972

1980

1985

1990

1995

Canada
United States
Japan
France
Germany
Italy
Spain
Sweden

93.0
93.1
97.0
97.3
97.5
—
96.2
93.6

91.9
91.8
96.5
95.2
95.9
97.1
90.6
95.6

88.2
90.3
95.7
92.9
92.6
95.0
85.0
95.1

88.0
90.2
96.6
92.1
91.4
93.1
88.9
94.7

84.6
88.6
96.1
88.7
90.4
88.6
83.2
85.0

Mean
90:10 ratio

95.4
1.05

94.3
1.05

91.9
1.08

91.9
1.07

88.2
1.07

SOURCES: Organisation for Economic Co-operation and Development, Labor Force Statistics, various issues; and
author’s calculations.

that differences across countries are large), it
is of interest to look at the evolution of this
group in our panel of eight countries. Table 9,
which also appears in the appendix, shows the
employment-to-population ratios for males
aged 35–44 over the period 1972–95.
Three points are worth making about the
data in table 9. First, as already mentioned,
one of the striking facts for this age group is
the negative secular trend in all countries, even
for those in which aggregate employment rose
during the period. Second, while it is true that
cross-country differences among other demographic groups tend to be much larger, the
differences for this group are still substantial.
Third, the patterns found for this demographic
group do not reflect the patterns in aggregate
data. For example, in table 9, Germany appears
to be a high-employment country and Canada
appears to be a country with low employment,
but aggregate data suggest just the opposite.

IV. Other Measures
of Labor Input
This paper began by arguing that it is of interest
to understand differences in labor input, both
across time within a given economy and at a
given point in time across different economies.
Until now, however, all of the data analysis
has pertained to employment ratios. In fact,
employment is but one, albeit important,
component of labor input. Other components
include hours of actual work per employed
person and work effort. Differences in actual
hours of work per employed person can be
further subdivided by differences in normal
weekly hours of work, overtime hours, the
extent of multiple job holding, or paid vacation
and sick days. While there seem to be no
attempts to officially document work effort by
national statistical agencies, most countries do
attempt to measure hours of work.
Unfortunately, strictly comparable crosscountry time-series data on hours of work per
employed person do not exist. Differences in
procedures across countries with regard to
such matters as whether the category “hours of
work” refers to hours paid versus hours actually
worked limit the appropriateness of crosscountry comparisons. Having made this qualification, I present in table 10 data reported by

42

T A B L E

10

Annual Hours of Work per
Employed Worker

h1970
1970

1975

1979

1983

1990

1996

h1996

Canada
United States
Japan
France
Germany
Italy
Norway
Spain
Sweden

1890
1889
2201
1962
1949
1969
1766
—
1641

1837
1832
2112
1865
1801
1841
1653
—
1516

1832
1845
2126
1806
1696
1722
1514
2022
1516

1780
1808
2095
1712
1657
1699
1485
1912
1518

1788
1819
2031
1657
1598
1674
1432
1824
1546

1784
1839
1892
1608
1511
1636
1407
1810
1623

1.06
1.03
1.16
1.22
1.29
1.20
1.25
—
1.01

Mean
Standard deviation
90:10 ratio

1908
163.1
1.11

1807
172.2
1.13

1757
197.8
1.22

1719
189.9
1.19

1693
185.0
1.18

1663
165.6
1.22

1.15
.108
1.22

SOURCES: Organisation for Co-operation and Economic Development; Employment Outlook, various issues; and
author’s calculations.

the OECD for annual hours of work for a set of
nine countries over the period 1970–96.12
All of the summary statistics are for the set of
eight countries not including Spain, since
values for Spain are not available before 1979.
I chose to include the available data for Spain
because in our earlier comparisons Spain was
typically the country with the lowest labor
input as measured by employment. However,
in table 10 we see that as of 1996, annual
hours per employed worker in Spain were
some 10–15 percent higher than their corresponding values in France, Germany, and Italy.
Assuming that this comparison is appropriate,
it indicates the care that one must take in
extrapolating from cross-country differences in
employment (and hence unemployment) to
differences in labor input.
There are several patterns worth noting in
table 10. First, in all countries the tendency
since 1970 has been for annual hours of work
per employed person to decrease, though the
range of decreases is very large. Second, as
already discussed, comparisons at a point in
time across countries may be misleading due
to differences in how the data are collected.
However, assuming that the effects of these
differences are roughly constant over time, the

data suggest very large relative movements in
labor input across countries. As the last column
in the table indicates, annual hours per worker
fell in Germany by more than 25 percent relative to the United States.13 One issue to keep in
mind when interpreting these differences is the
fact that in all countries there is a tendency for
the workweek in manufacturing to decrease as
a country becomes richer. While this decrease
for the United States occurred prior to the 1970s,
in many other countries it occurred after 1970.

■ 12 These data are taken from various issues of the OECD
Employment Outlook. The OECD reports these measures with a strong
warning that they should not be used for cross-country comparisons at a
point in time. The value for Sweden in 1996 is not directly comparable to
the earlier values because of survey changes. It seems that the effect of the
change is to increase the value of hours worked in 1996 by around 3 percent.
■ 13 Although cross-country, point-in-time comparisons are not
recommended with these data, there are some indications that such
comparisons may be meaningful in some cases. In their study of the
auto industry, Fuss and Waverman (1992) document differences in
annual hours of work per auto-industry employee in Canada, the
United States, Japan, and Germany between 1961 and 1981. In 1961,
the values were Canada 1970, United States 2042, Japan 2495, and
Germany 2007. In 1981 the values were Canada 1857, United States
1923, Japan 2200, and Germany 1602.

43

T A B L E

11

Full- and Part-Time Employment
Ratios (1970 and 1995)

1970

Full-time
1995

Canada
United States
Japan
Australia
New Zealand
Belgium
France
Germany
Italy
United Kingdom

55.4
54.4
61.2
60.6
56.9
59.0
61.9
61.7
52.7
59.1

54.4
59.4
53.3
49.0
45.3
47.2
50.1
54.4
45.1
53.5

Mean
90:10 ratio

58.3
1.13

51.2
1.20

Part-time
1970
1995
6.0
10.1
9.9
8.2
7.2
2.3
3.9
6.9
3.6
11.3
6.9
2.81

12.9
13.0
16.2
17.2
13.1
9.1
8.8
10.2
5.8
15.9
12.2
1.84

SOURCES: Organisation for Co-operation and Economic Development; Employment Outlook, various issues; and
author’s calculations.

How do these changes in hours correlate
with the changes in employment ratios documented earlier? The correlation between the
change in hours between 1970 and 1996 and
the change in aggregate employment-topopulation ratios between 1970 and 1995 for
this set of eight countries is 0.36. Norway is
somewhat of an outlier in this regard, and if
Norway is excluded, the correlation increases
to 0.77. This suggests that the relative changes
in aggregate labor input are probably substantially larger than are the relative changes in
employment-to-population ratios.
A closely related issue that often comes up
in this context is cross-country differences in
the extent of part-time employment.14 The
measures of hours presented above do include
part-time employment, and hence do control
for these differences. However, it may also be
of interest to directly examine data on fullversus part-time employment. Here again,
however, a major caveat is necessary since
cross-country measures are not directly comparable due to differences in definitions of
part-time employment.

As a crude attempt to decompose the
previously examined employment-to-population ratios into full-time and part-time components so that I can compare relative changes
in the proportion of the population employed
full- and part-time, I compute full- and parttime employment-to-population ratios for 1970
and 1995. I use the data on the fraction of
employment that is part-time in 1973 and 1997
in conjunction with the earlier data on employment-to-population ratios for individuals aged
15–64 for the years 1970 and 1995. Although
the years do not match exactly, to the extent
that changes in part-time versus full-time
employment patterns have been occurring
gradually through time, this comparison should
give a good idea of the changes over the
1970–1995 period.
The values are in table 11.15

■ 14 In particular, some researchers have argued that the dramatic
improvement in employment in the Netherlands is entirely due to increases
in part-time employment. See, for example, Nickel and van Ours (2000) and
its discussion.
■ 15 The 90:10 ratio in this table refers to the ratio of the secondhighest value to the second-lowest value.

44

A few patterns emerge. First, note that in
all countries the employment-to-population
ratio has increased for part-time workers, and
in every country except the United States the
employment-to-population ratio for full-time
employment has decreased. Second, it is in no
way true that differences across countries in
aggregate employment-to-population ratios are
dominated by differences in the extent of
part-time work. The 90:10 ratio for full-time
employment exceeds 1.10 for both years.
Moreover, the value increases over time, suggesting that differences in full-time-employment
ratios have become larger. In contrast, differences in part-time-employment ratios have
actually narrowed slightly, though based on
the 90:10 measure, differences in part-timeemployment ratios still exceed those for
full-time-employment ratios. Third, from the
perspective of the “European unemployment
problem,” the full-time-employment ratios
suggest even larger relative changes than do
the aggregate numbers, since in 1970, part-time
work was much less prevalent in Europe but
has since become increasingly common there
relative to the United States.

V. Conclusion
This paper has studied the empirical properties
of aggregate and disaggregate employment in
a cross section of developed countries over the
period 1960–95. It has documented several
facts that a successful theory of employment
should be able to account for. Though much
information is presented here, there is much
additional information that would give a richer
picture of differences in time allocations across
countries. First, we need better cross-country
measures of hours of actual work, especially
at a disaggregated level. Second, additional
disaggregations would be useful, especially by
family structure. Third, it is important to understand how people spend their time when not
working in the market—are they taking care of
other family members, in school, or what?
Three conclusions seem to bear repeating.
First, large and persistent differences in
employment ratios across countries seem to be
pervasive. In particular, although the relative
changes in unemployment between the
United States and Europe over the last 30 years
have been dramatic, one should not be misled
into thinking that in this regard the world is
dramatically different after 1970 than before.

From the perspective of employment-topopulation ratios, there are large cross-country
differences before 1970 as well as after, and
there were substantial changes in cross-country
relative employment ratios in the period before
1970 as well as after. Second, the changes
found in data disaggregated by sex and age
do not at all mirror the changes found in
aggregate data. Changes tend to be concentrated among the young and the old and
among women. Any successful theory of crosscountry changes in employment must successfully account for this concentration. Third, the
patterns found in aggregate data for a cross
section of countries do not carry over to all
demographic groups within those countries.

45

Appendix—
Disaggregated
Changes in
Employment Rates
This appendix presents the complete set of
tables for disaggregated changes in employment rates between 1972 and 1995 for a set of

T A B L E

eight countries. The data for 1972 represent an
average for 1972 and 1973, whereas for each
other year, the data represent a three-year
average centered on the given year; for example,
for 1980 the data are an average of 1979– 81.
The 90:10 ratio is the ratio of the secondhighest to the second-lowest value.

12

Disaggregated Changes in
Employment Rates

A. Males 15–24
1972

1980

1985

1990

1995

Canada
United States
Japan
France
Germany
Italy
Spain
Sweden

56.9
66.8
55.2
55.7
69.9
42.2
69.8
64.1

62.5
64.0
41.6
46.3
59.4
38.7
54.0
66.8

58.3
62.7
40.7
38.0
57.0
34.3
40.4
61.4

60.7
62.8
41.6
33.6
59.5
33.8
42.5
64.4

52.9
60.9
45.3
25.5
54.0
31.3
31.4
40.8

Mean
90:10 ratio

60.1
1.26

54.2
1.54

49.1
1.62

49.9
1.86

42.8
1.73

B. Males 25 –34
1972

1980

1985

1990

1995

Canada
United States
Japan
France
Germany
Italy
Spain
Sweden

91.6
91.6
96.7
96.1
94.4
94.7
94.3
91.5

89.7
89.5
95.2
93.2
88.8
87.6
85.8
92.8

83.9
88.1
94.4
88.6
82.0
82.5
76.3
91.1

84.8
88.6
95.4
88.4
85.0
79.2
81.4
90.5

81.6
88.0
94.3
83.7
83.6
71.1
72.2
79.0

Mean
90:10 ratio

93.9
1.05

90.3
1.06

85.7
1.11

86.7
1.11

81.7
1.22

46

C. Males 35–44
1972

1980

1985

1990

1995

Canada
United States
Japan
France
Germany
Italy
Spain
Sweden

93.0
93.1
97.0
97.3
97.5
—
96.2
93.6

91.9
91.8
96.5
95.2
95.9
97.1
90.6
95.6

88.2
90.3
95.7
92.9
92.6
95.0
85.0
95.1

88.0
90.2
96.6
92.1
91.4
93.1
88.9
94.7

84.6
88.6
96.1
88.7
90.4
88.6
83.2
85.0

Mean
90:10 ratio

95.4
1.05

94.3
1.05

91.9
1.08

91.9
1.07

88.2
1.07

D. Males 45–54
1972

1980

1985

1990

1995

Canada
United States
Japan
France
Germany
Italy
Spain
Sweden

90.5
89.3
95.9
94.0
94.2
95.2
94.3
93.0

88.8
88.2
95.0
91.6
92.4
96.3
87.3
93.5

85.2
86.9
94.5
88.7
90.7
94.9
80.6
93.4

85.8
87.2
95.7
88.2
88.7
94.2
84.9
93.3

82.8
85.7
95.7
86.3
87.4
90.2
80.3
86.4

Mean
90:10 ratio

93.3
1.05

91.6
1.07

89.4
1.11

89.8
1.10

86.9
1.09

E. Males 54–64
1972

1980

1985

1990

1995

Canada
United States
Japan
France
Germany
Italy
Spain
Sweden

77.7
75.8
85.0
71.1
74.0
—
82.1
81.3

72.5
69.5
81.7
64.5
63.4
72.0
71.6
77.2

64.5
64.9
79.0
46.5
54.4
65.6
59.2
73.2

60.0
64.7
80.5
42.9
51.3
63.1
56.7
73.8

54.4
63.7
80.9
38.7
47.4
55.5
48.8
65.0

Mean
90:10 ratio

78.2
1.11

71.6
1.20

63.4
1.35

61.6
1.44

56.8
1.37

47

F. Males 65+
1972

1980

1985

1990

1995

Canada
United States
Japan
France
Germany
Italy
Spain
Sweden

17.5
21.7
46.1
16.3
14.4
7.4
25.3
24.0

14.5
18.5
40.1
8.2
6.8
7.9
12.3
13.7

12.1
15.5
36.2
5.3
5.0
5.3
6.1
11.6

11.0
15.7
36.3
3.8
4.4
5.2
3.9
13.5

10.1
16.2
36.4
2.6
4.2
6.0
2.8
13.2

Mean
90:10 ratio

21.6
1.76

15.3
2.32

12.1
2.92

11.7
4.03

11.4
5.79

G. Females 15–24
1972

1980

1985

1990

1995

Canada
United States
Japan
France
Germany
Italy
Spain
Sweden

45.4
47.7
49.3
43.3
61.9
29.3
50.8
56.8

55.2
54.0
42.5
33.9
53.2
28.1
34.2
65.2

55.9
55.3
41.6
28.0
50.6
24.1
22.4
61.8

59.1
56.0
43.0
25.1
54.2
25.1
28.3
64.8

51.2
55.2
44.5
18.3
48.6
21.6
21.4
41.9

Mean
90:10 ratio

48.1
1.31

45.8
1.63

42.5
2.32

44.5
2.35

37.8
2.39

H. Females 25–34
1972

1980

1985

1990

1995

Canada
United States
Japan
France
Germany
Italy
Spain
Sweden

42.2
46.0
44.2
55.2
50.8
31.3
28.0
62.3

58.8
60.7
47.2
63.5
57.5
47.1
32.2
79.4

64.1
65.6
50.2
64.4
56.3
47.3
36.3
85.4

70.2
69.0
54.8
65.7
62.5
50.0
44.3
86.3

69.4
70.4
57.6
65.2
65.8
46.9
44.9
73.9

Mean
90:10 ratio

45.0
1.76

55.8
1.34

58.7
1.39

62.9
1.40

61.8
1.50

48

I. Females 35–44
1972

1980

1985

1990

1995

Canada
United States
Japan
France
Germany
Italy
Spain
Sweden

41.4
50.0
58.3
49.7
49.8
—
24.6
70.1

58.1
61.9
60.1
60.4
55.3
43.1
27.5
83.5

64.5
67.8
62.5
64.7
57.6
48.8
28.3
88.9

72.2
73.0
65.1
67.4
64.3
52.8
37.1
91.4

71.8
73.8
63.9
69.3
67.7
51.8
43.1
83.0

Mean
90:10 ratio

49.1
1.41

56.2
1.44

60.4
1.39

65.4
1.38

65.6
1.42

J. Females 45–54
1972

1980

1985

1990

1995

Canada
United States
Japan
France
Germany
Italy
Spain
Sweden

40.7
51.5
60.2
50.7
48.9
29.9
26.5
68.8

50.7
57.2
61.0
54.5
49.6
36.8
26.5
82.2

55.6
61.3
63.4
58.0
50.3
40.5
24.4
87.0

64.1
68.6
67.6
60.9
57.6
45.6
28.1
89.4

66.4
72.2
68.1
67.3
62.2
47.4
32.0
85.3

Mean
90:10 ratio

47.2
2.01

52.3
1.66

55.1
1.57

60.2
1.50

62.6
1.52

K. Females 55–64
1972

1980

1985

1990

1995

Canada
United States
Japan
France
Germany
Italy
Spain
Sweden

29.5
40.2
43.9
36.7
25.8
8.9
23.3
44.5

32.2
40.0
44.6
36.8
26.9
22.0
20.9
54.7

31.1
40.3
43.5
28.6
21.7
20.7
18.5
57.6

32.9
43.8
46.5
28.5
22.2
21.3
18.1
64.3

33.9
47.5
47.5
28.6
24.5
20.2
17.6
60.0

Mean
90:10 ratio

34.8
1.70

34.8
2.03

32.8
2.10

34.7
2.18

35.0
2.35

49

L. Females 65+
Canada
United States
Japan
France
Germany
Italy
Spain
Sweden
Mean
90:10 ratio

1972

1980

1985

1990

1995

4.3
8.3
16.1
6.9
5.8
1.5
7.7
7.5

4.3
7.9
15.5
3.6
3.2
2.7
4.1
3.9

4.0
7.2
15.4
2.2
2.3
2.2
2.4
3.3

3.7
8.2
16.2
1.6
1.9
2.1
1.6
5.3

3.3
8.5
15.5
1.3
1.5
1.7
1.3
4.6

7.3
1.93

5.7
2.47

4.9
3.27

5.1
5.12

SOURCES: Organisation for Economic Co-operation and Development, Labor Force Statistics, various issues; and
author’s calculations.

4.7
6.53

50

References
Bertola, Giuseppe, and Andrea Ichino.
“Wage Inequality and Unemployment:
The United States versus Europe,” NBER
Macroeconomics Annual, Cambridge:
MIT Press, 1995, pp. 13–54.
Blanchard, Olivier. The Lionel Robbins
Lectures. Presented at the London School
of Economics, 2000.
Blanchard, Olivier, and Justin Wolfers.
“The Role of Shocks and Institutions in
the Rise of European Unemployment,”
Economic Journal, vol. 110, no. 462
(March 2000), pp. 1–33.
Ljungqvist, Lars, and Thomas J. Sargent.
“The European Unemployment Dilemma,”
Journal of Political Economy, vol.106, no. 3
(June 1998), pp. 514–550.
Millard, Stephen P., and Dale T. Mortensen.
“The Unemployment and Welfare Effects of
Labour Market Policy: A Comparison of the
USA and the UK,” in Unemployment Policy:
How Should Governments Respond to
Unemployment, edited by Dennis Snower
and Guillermo de la Dehesa, Oxford:
Oxford University Press, 1997.

Mortensen, Dale T., and Chris A. Pissarides.
“Unemployment Responses to ‘Skill Biased’
Technology Shocks: The Role of Labour
Market Policy,” Economic Journal, vol. 109,
no. 455 (April 1999), pp. 242–65.
Nickell, Steve, and Jan van Ours. “The
Netherlands and the United Kingdom:
a European Unemployment Miracle?”
Economic Policy, vol. 0, no. 30 (April 2000),
pp. 135–80.
OECD Employment Outlook, various issues,
Organisation for Economic Co-operation
and Development.
OECD Historical Statistics, various issues,
Organisation for Economic Co-operation
and Development.
OECD Labor Force Statistics, various issues,
Organisation for Economic Co-operation
and Development.

51

Monetary Policy and Asset Prices
with Imperfect Credit Markets
by Charles T. Carlstrom and Timothy S. Fuerst
Charles T. Carlstrom is an economic
advisor at the Federal Reserve Bank
of Cleveland. Timothy S. Fuerst is
the Owens-Illinois Professor in the
Department of Economics at Bowling
Green State University and a
research associate at the Federal
Reserve Bank of Cleveland.

Introduction
In a world with perfect capital markets (the
world of the Modigliani–Miller theorem), a
firm’s financial position—that is, its debt versus
its equity level—is irrelevant to its decisions on
production and investment. The reason is that
perfect capital markets let information flow
freely. If an entrepreneur has a good idea for
a new product, she will be able to produce it
regardless of her personal financial position
because outside investors, well informed and
readily perceiving an attractive profit opportunity,
will provide whatever financing is needed.
The supposition that an entrepreneur’s
financial position is irrelevant has important
implications for monetary policy. A worthy
production activity will be funded whatever
the entrepreneur’s finances may be. Therefore,
monetary policy need not respond to asset
prices.1 The Modigliani–Miller theorem, however, is not necessarily meant to be a description of reality. On the contrary, a voluminous
empirical literature supplies evidence that a
firm’s financial position does affect its ability
to operate. But does this departure from the
Modigliani–Miller theorem provide a rationale
for basing monetary policy partly on equity

prices? As it stands, the theorem provides an
important benchmark and enforces careful
thinking about financial markets’ workings and
the imperfections that would create a world
where a firm’s financial position (hence equity
prices) affects its ability to engage in production.
Many possible imperfections could generate
such a world. This article focuses on failures of
information. Suppose that only the entrepreneur
knows every detail of the proposed project,
while outside investors financing the project
have no way of knowing exactly what the
entrepreneur would do with their funds. Suppose further that outside investors have limited
ability to punish an entrepreneur who runs
off with their money or squanders it on a misguided production activity. In this scenario,
external investors are likely to provide financing only if they know they will be able to

■ 1 A response might be warranted if equity prices help forecast
macro variables of interest such as output and inflation.

52

recoup their investment if the project turns
sour. One way to ensure this is to restrict the
amount of funds they provide to the size of
the entrepreneur’s financial position. That is,
external financing will be no greater than the
collateral that investors can seize after the fact.
We have just summarized a story in which a
firm’s financial position, which we will henceforth call “collateral” or “net worth,” has a
powerful effect on a its ability to produce.
Increases in its equity price will increase this
collateral—and with it a firm’s ability to produce. Clearly, this is no Modigliani–Miller
world, but what role does monetary policy
have in it? Can policy help the economy
respond to the fundamental shocks that buffet
it? Should policy respond to asset prices in
such a world?
This article uses a theoretical model to
address these questions. The model is highly
stylized to keep the analysis tractable, but its
essential point will survive more complicated
modeling environments.2 A key conclusion is
that there is a role for activist monetary policy.
Imperfect information imposes a collateral
constraint on this economy, and monetary
policy can be useful in alleviating this constraint by responding to productivity shocks
or exogenous changes in equity prices.

I. The Model
The theoretical model consists of households
and entrepreneurs. We will discuss the decision problems of each in turn.

(2)

τ

 
Rt

.

Notice that labor supply responds positively to
the real wage with elasticity τ. R t denotes the
gross nominal interest rate. Labor supply is
negatively related to the nominal rate because
we assume that households must use cash to
facilitate their consumption purchases (a “cashin-advance constraint”).3 Because the opportunity cost of holding cash is given by the gross
nominal rate, higher nominal rates make it
more difficult to turn labor income into consumption, thus discouraging labor supply. To
put it another way, the gross nominal interest
rate acts like a wage tax where 1 = (1 – tw ). The
Rt
celebrated “Friedman rule”—that the net nominal interest rate should be zero (or R =1)—is
based directly on the observation that a zero
interest rate eliminates this implicit wage tax.4
A household must also make a decision
about consumption versus saving. Households
can save only by acquiring shares to a real
asset that pays out (real) dividends of Dt consumption goods at the end of time t. It is helpful to think of this as an apple tree that produces Dt apples in time t. The tree trades at
share price q t at the beginning of the period
(before the time-t dividend is paid). Under our
assumption on household preferences, the
equilibrium real share price is given by dividends’ present discounted value (the assumption of linear utility implies that the discount
rate on dividends is the constant β )
(3)

Households

Lt =

wt

q–t =E t  β j Dt+j .
j =0

Households are infinitely lived, discounting the
future at rate β. Their period-by-period utility
function is given by
1+

(1)

U (ct , L t )  ct –

Lt

1
τ

1
1+ τ

,

where ct denotes consumption and Lt denotes
work effort. We choose this form for convenience. Each period, the household decides
how much to work at a real wage of wt . The
resulting labor supply relationship is given by

■ 2 For example, empirical evidence suggests that collateral
constraints have a stronger effect on small firms than on large ones
(see Gertler and Gilchrist [1999]). We abstract from this heterogeneity
and posit a single representative firm. Future efforts to quantify collateral
effects should model heterogeneity more explicitly.
■ 3 See the appendix for a precise statement of the household’s
problem and the resulting first-order conditions.
■ 4 In our model, the first-best policy will be the Friedman rule.
Our policy section includes analysis of a second-best problem where, for
some unspecified reason, the monetary authority desires to keep the
long-run average interest rate above zero, R >1.

53

If the share price were below this level (qt < q–t ),
then household demand for shares would be
infinite; if the share price were above this level
(qt >q–t ), then household demand for shares
would be negative infinity, that is, a desire to
sell short. Thus, households will hold a finite
and positive level of tree shares only if q–t is
the equilibrium price. The dividend process
is given by
Dt +1 = (1 – ρD ) Dss + ρD Dt + ε Dt+1 .
The symbol Et denotes the rational forecast of
future dividends; recall also that β is the rate of
household time preference, which is also the
real interest rate in this environment. Notice
that the asset price depends only on the
exogenous dividend process and that the
share price is increasing in the current and
future dividend levels. The exogenous discount process is an AR1, which means that
next period’s dividend is a weighted average
of today’s dividend (Dt ) and the long-run
average of dividends (Dss ) plus a random i.i.d.
shock (ε Dt+1 ).

Entrepreneurs
Entrepreneurs too are infinitely lived and have
linear preferences over consumption. They are
distinct from households in that they use a
constant-returns-to-scale production technology
in which labor produces consumption goods
(4)

yt = At Ht ,

where At is the current level of productivity,
and Ht denotes the number of workers
employed at real wage wt . Like dividends
productivity, At is an exogenous AR1 random
process given by
At +1 = (1 – ρA ) Ass + ρA At + ε At+1 .
The entrepreneur is constrained by a
borrowing limit. In particular, she must be able
to cover her entire wage bill with collateral
accumulated in advance. We will denote this
collateral as nt (net worth). The loan constraint
is thus
(5)

first supply their labor input but that output is
subsequently produced if and only if the entrepreneur contributes her unique human capital
to the process. This production sequence
implies that the entrepreneur could force
workers to accept lower wages ex post; otherwise, nothing would be produced. Workers,
anticipating this hold-up possibility, will take
steps to prevent it. This is harder than it
sounds. For example, an equity-type arrangement in which worker and entrepreneur agree
ex ante to split the production ex post will not
work. After the worker has supplied his labor,
the entrepreneur can refuse to provide her
unique human capital unless the worker’s
share is made arbitrarily small. In that case, the
worker’s only choices are to accept this small
share or to take nothing. The worker could
seize the entrepreneur’s existing assets, but
then we are back to our collateral constraint.
In fact, as Hart and Moore (1994) and Kiyotaki
and Moore (1997) demonstrate, these hold-up
problems can only be avoided completely if
the entire wage bill is covered by existing
collateral that workers could seize in case
of default.5
We can easily enrich this story by assuming
the existence of financial institutions that intermediate between workers and entrepreneurs.
For example, suppose that such intermediaries
provide within-period financing to entrepreneurs, who use it to pay workers. An intermediary, however, is concerned about the hold-up
problem, so it limits its lending to the firm’s
net worth. This returns us to the collateral
constraint described in equation (5).6
We assume in what follows that the loan
constraint is binding, so that labor demand is
given by
(6)

Ht =


nt

wt

.

Notice that labor demand varies inversely
(with a unit elasticity) to the real wage but is
positively affected by the level of net worth.
Firms that have more collateral can employ

wt Ht ≤ nt .

Notice that all variables are in real terms.
Why is the firm so constrained? Many
possible information stories would motivate
such a constraint. We will assume the classic
hold-up problem: Suppose that hired workers

■ 5 This implicitly assumes a one-period problem so that an
entrepreneur who withholds her labor cannot be punished by being
deprived of future income.
■ 6 Kiyotaki and Moore (1997) use a similar constraint. See Hart
and Moore (1994) for more discussion of the hold-up problem.

54

more workers because hold-up problems are
less severe. The binding collateral constraint
implies that At >wt , that is, the firm would like
to hire more workers but is collateralconstrained.
An entrepreneur’s only source of net worth
is previously acquired ownership of apple
trees. If we let et –1 denote the number of tree
shares acquired at the beginning of time t –1,
then net worth at time t is given by
(7)

nt = et –1 qt ,

so that the loan constraint is given by
(8)

wt Ht ≤ et –1 qt .

As noted above, the assumption of a binding
loan constraint implies that the firm’s marginal
profits per worker employed are (At – wt ).
These profits motivate the entrepreneur to
acquire more net worth. We will need to limit
this accumulation tendency so that collateral
remains relevant. The entrepreneur’s budget
constraint is given by
(9)

c et + et qt = et –1 qt +et Dt + Ht (At – wt ).

The right side of the budget constraint
equation is the entrepreneur’s income in
period t, which consists of her revenue from
the sale of existing trees (et –1 qt ), dividends
from new tree purchases (et Dt ), and profits
(Ht [At – wt ] ). The left side represents her
potential purchases in period t. With her revenue, she purchases either consumption (c et )
or new tree shares (et qt ). Using the binding
loan constraint, we can rewrite this as
At
(10) c te + et (qt – Dt ) = et –1 qt w ,
t

Because of the profit opportunities from net
worth (At >wt ), the entrepreneur would like to
accumulate trees until the constraint no longer
binds (trees are more valuable to collateralconstrained entrepreneurs than they are to
households). To prevent this, we will assume
that entrepreneurs must consume a fraction of
their net income each period
A
(11) cte =(1– )et –1 qt wt ,
t

so that entrepreneurial tree holdings evolve as
A
(12) et (qt – Dt ) = et –1 qt w t .
t

Below we will choose  <1 to offset the
high return to internal funds, thus keeping the
entrepreneur’s collateral constrained in equilibrium. This forced-consumption-savings decision
implies that households will price trees so that
in equilibrium qt = q–t .

Equilibrium
In this theoretical model, there are two active
markets, the market for apple trees and the
labor market (the money market and bond
market are discussed in the appendix). We
normalize the supply of tree shares to unity so
that the asset market clears with et +st =1. The
equilibrium tree price is given by (3). As for
the labor market, equating labor supply with
labor demand (Lt =Ht ) and solving for the real
wage yields
τ

1

(13) wt =nt 1+τ Rtτ +1 .
The equilibrium real wage is increasing in
net worth because higher net worth increases
labor demand. The wage is also increasing in
the nominal interest rate because a higher
nominal rate decreases labor supply. Equilibrium employment is given by

(14) Lt =


nt

τ
1+τ

Rt

.

For the reasons already noted, employment
responds positively to net worth and negatively to the nominal rate.

Log-Linearizing the
Model
Because the model is relatively simple, it is
convenient to express the equilibrium in terms
of log deviations. In what follows, the ~
represents a percent deviation from the
steady state.

τ ~ ~
(15) L~t =
(n –R )
1+ τ t t
~ =q~ +e~
(16) n
t
t
t –1
(17) Et n~t +1 =

τ ~ ~
(n –R ) +A~t ,
1+ τ t t

55

where (17) comes from (12) and the asset
price (3). Using (16) to eliminate nt , we can
rewrite (15) and (17) in terms of et as

τ ~ ~
~
(18) L~t =
(q +e –R t
1+ τ t t –1

)





τ ~
τ
(19) ~
et =
(e –R~ ) +A~t +
– ρ q~ .
1+ τ t –1 t
1+ τ D t
To calculate (19), we have also used the ability
to express the share price (3) as
~
(20) q~t =Dt

1–β
,
1– βρD





where ρD is the autocorrelation in the dividend
process. To sum up, the model consists of equations (18)–(20). There is one predetermined
variable, et –1, and there are three exogenous
shocks: At , Dt , and Rt .

with current collateral. Thus, the collateral
constraint limits the firm’s ability to respond
to shocks.
There is, however, a delayed response.
A positive shock to At has no effect on current
employment, but it increases et and, through
it, tomorrow’s net worth (see [22]). Hence,
employment responds with a lag to productivity shocks.
This lagged response generates persistence
to a temporary shock. That is, even if the
shock to At lasts only one period, the effect
on employment, Lt , and thus on output, lasts
much longer and only dies out at the rate
given by τ /(1+τ). If the shock to productivity is
serially correlated, this effect remains, so that
the collateral constraint prolongs the effect of
the productivity shock.

Second Experiment:
A Shock to
Dividends

II. The Experiments

Proceeding as before, we have:

Before turning to the question of monetary

1–β D
~
(24) L~t = τ e~t –1 + τ
1+τ
1+τ 1– βρD t

policy, it is useful to sharpen one’s economic
intuition about the model by considering
several experiments.

First Experiment:
A Shock to
Productivity (At )
Suppose that we hold all other variables constant and consider only shocks to productivity.
Then we have

τ ~
(21) L~t =
e
1+ τ t –1
τ ~
(22) e~t =
e +A~
1+ τ t –1 t .
By combining, we obtain

τ ~ ~
(23) L~t +1 =
(L +A ) .
1+ τ t t
Notice that contemporaneous employment
does not respond to shocks to productivity, At
(see [21]). This is a manifestation of the collateral constraint. When productivity is high, the
firm would like to expand employment but it
cannot because it must finance current activity

  
  

~.
(25) e~t = τ e~t –1 + τ –ρ 1 – β D
1+τ
1+τ
1– βρD t
By combining, we obtain

τ ~
L~t =
(L +ε~D ).
1+ τ t –1 t
Recall that εDt is the innovation in the dividend
process. The most remarkable observation is
that employment responds positively to dividend
shocks, even though these shocks have no
effect on either worker productivity or labor
supply. Instead, dividends affect employment
solely through the collateral constraint. Because
trees are used as collateral, and a dividend
shock drives up their price, the collateral
constraint is relaxed and the firm can expand
employment. Once again, these effects are
highly persistent.

56

Maximizing Vt with respect to Lt yields the
optimality condition

Third Experiment:
A Monetary
Policy Shock

(30) Lt = Atτ .

We will assume that monetary policy is given
by directives for the gross nominal interest
rate, Rt . The implied path for the money supply
can be backed out of the money demand
relationship (see the appendix).
Proceeding as before, we have

τ ~
(26) L~t =
(e – R~ )
1+ τ t –1 t
τ ~
(27) e~t =
(e – R~ ).
1+ τ t –1 t
By combining, we obtain

τ ~
~
(28) Lt =
(L – R~ ).
1+ τ t –1 t
There are two differences between the interest
rate shock and the productivity shock. First,
the interest rate shock has an immediate effect
on employment because it alters labor supply
contemporaneously. Second, its effect is negative because the higher interest rate lowers the
households’ desire to work. As in the previous
cases, the shock has a persistent effect through
the collateral constraint.

III. Monetary Policy
Optimal Monetary
Policy
What is the nominal interest rate’s optimal
response to productivity and dividend shocks?
To answer such a question, we need a welfare
criterion. The most natural choice in the present context is the sum of household and
entrepreneurial utility, which is given by
(29) Vt ct + cte –

1
1+–
Lt τ

1
1+ τ

= At Lt +Dt –

1
1+–
Lt τ

1
1+ τ

,

where the equality follows from the fact that
total time-t consumption must equal the total
supply of time-t consumption goods, which
comes from the goods produced using the
entrepreneur’s production technology, and
dividends produced by the apple tree. The
only choice variable in Vt is employment.

We will call this solution the “first-best” outcome because the welfare criterion can go no
higher. The first-best has two natural features.
First, employment responds positively to productivity shocks. When productivity is high, it
is efficient for employment to respond positively. Second, the first-best employment does
not respond to dividend or share prices. The
welfare criterion Vt is increasing in Dt , but
these shocks have no effect on labor productivity; thus, it is efficient for employment not to
respond to these shocks.
Is the first-best achievable? If there were no
collateral constraint, we would have wt = At ,
and the first-best could be achieved by setting
Rt =1, that is, by setting the net nominal rate to
zero. This is the celebrated Friedman rule. It
is optimal in this model because the cash-inadvance constraint on consumption distorts the
labor margin.
But in a world with agency costs, the firstbest is impossible because employment is
given by (14), which, as noted above, is
rendered too low (At >wt ) by the collateral
constraint, Furthermore, according to (14),
employment fluctuates with net worth and not
with the level of productivity. Compared to the
first-best outcome, these employment responses
are dreadful. Contemporaneous employment
does not respond to productivity, even though
it is efficient to do so; employment, however,
does respond to share prices which, in an
efficient world, should not affect it. In short, the
collateral constraint causes the economy to
under-respond to productivity shocks and to
over-respond to dividend shocks.
The advantage of the Friedman rule is that
it minimizes the distortion on labor from the
cash-in-advance-constraint.7 The disadvantage
is that a pegged zero nominal interest rate precludes the monetary authority’s responding to
shocks to make employment respond efficiently. It turns out that the benefit of a lower
nominal interest rate always wins out in this

■ 7 Recall that because cash must be held to facilitate transactions,
higher nominal rates discourage labor supply in (2).

57

environment—the first-best policy is simply to
set the nominal interest rate to zero (that is,
R =1) and leave it there. But what happens if
the monetary authority does not set the long-run
interest rate to zero but keeps it positive for
some unspecified reason? 8 Can monetary
policy improve on this economy’s ability to
respond to shocks in this world? Yes. To illustrate, let us consider a second-best exercise.

Optimal Policy in
Log-Deviations
We take the steady state of the economy as
given and use monetary policy so that the
economy responds to shocks efficiently.
Optimal employment (in log deviations) is
given by
(35) L~t = τ A~t .
To find the optimal (second-best) interest rate
policy, we can impose equation (35) in the
system (18) –(19), and back out the implied
interest rate. This exercise yields
(36) R~t =q~t + e~t –1 – (1+ τ )A~t
(37) e~t = (1+ τ )A~t – ρD q~t .
By combining, we obtain
R~t = εtD – (1+ τ) [ εtA + ( ρA –1)A~t –1 ] .
What are the properties of this (secondbest) optimal monetary policy? 9 When there is
a positive shock to productivity At , the central
bank should lower the nominal interest rate
so that employment can expand efficiently.
A constant-interest-rate policy does not allow
this because of the collateral constraint, but a
procyclical interest rate policy overcomes the
collateral constraint and allows the economy to
respond appropriately.
Suppose that productivity shocks are
autocorrelated with coefficient ρA . A positive
technology shock of 1 percent calls for an
immediate interest rate decline of (1+ τ ) percent,
but then an increase to (1+ τ )(1 – ρA ). The
increase is needed to prevent over-expansion
of employment, because net worth rises with
the initial interest rate decline.

In contrast, if there is a shock to share
prices that drives up net worth, nt , the central
bank should increase the interest rate enough
to keep employment constant. It is inefficient
for employment to respond to these dividend
shocks, and the central bank can ensure no
response by raising the nominal rate in response.
Notice, however, that even if a shock to share
prices (dividends) is autocorrelated ( ρD > 0),
the optimal interest rate response is iid.

IV. Conclusion
This article addresses the question of how
monetary policy should be conducted in a
world where asset prices affect real activity
directly because of binding collateral
constraints, that is, a world in which the
Modigliani –Miller theorem does not hold.
How should monetary policy be conducted
in such a world? Should it respond to asset
prices? How should it respond to productivity
movements? In this environment, there is a
welfare-improving role for a monetary policy
that responds actively to asset price and
productivity shocks. This activist interest rate
policy allows the economy to respond to
shocks in a Pareto efficient manner. By
assumption, monetary policy cannot eliminate
the long-run impact of the information constraint, but it can improve welfare by smoothing
the fluctuations in this constraint.
Our results are stark because all firms in the
economy are subject to this hold-up problem.
One can imagine an environment in which
small firms are the ones most subject to agency
costs. This will change the quantitative—but
not the qualitative—predictions of the model.
This article uses a monetary model with
flexible nominal prices. In contrast, Bernanke
and Gertler (1999) analyze a similar question
in a model with sticky prices. They conclude
that as long as monetary policy responds
aggressively to inflation, there is no rationale

■ 8 For example, a positive nominal interest rate may be set to give
the government inflation-tax revenues.
■ 9 Optimal monetary policy refers to how the central bank should
change the interest rate in response to technology shocks and share
prices. Money growth is endogenous and, as discussed in the appendix,
can be backed out of the money demand relationship, that is, the cash-inadvance constraint.

58

for a direct response to asset prices. They
reach this conclusion because, in their model,
asset price shocks directly increase aggregate
demand and thus the price level. Hence, a
policy that responds aggressively to inflation
is automatically responding to asset prices. The
model described in this article creates no direct
link between inflation and asset prices, so
the central bank must respond directly to the
latter. It suggests that, to the extent that asset
prices do not immediately lead to price
inflation, there may be a role for a monetary
policy response to asset price movements.

Appendix 1

s.t.

 

E  β t ct –
t =0

1
1+ τ

Lt

1
1+ τ

Mt –1+Xt
+st –1 qt +st Dt
Pt
+

Rt –1 Bt –1 –Bt
– st qt –ct ≥ 0
Pt

Mt –1 +Xt
+ st –1 qt +st Dt +wt Lt
Pt
+

st =

Rt –1 Bt –1 –Bt
M
–st qt – ct – t ≥ 0 ,
Pt
Pt

where Bt denotes bond holdings (in zero net
supply), and households are assumed to receive
s
lump-sum monetary injections, Xt = Mt –1, at
s
Mt –1
the beginning of the period (Mts denotes the
per capita money supply at time t). Notice that
the bond and tree markets open either simultaneous to or before the consumption market.
The first constraint is the cash-in-advance constraint: The cash remaining after leaving the
bond and tree markets is the cash that can be
used to purchase consumption. The second is
the intertemporal budget constraint.
After minor simplification, household
optimization is defined by the binding cash
constraint and the following Euler equations:
(A1) 1 = β Rt Et (Pt /Pt +1)

if qt < q–t

st indeterminate if qt = q–t
st = 0 if qt > q–t , where
q–t =Dt +Et β q–t +1 .
Substituting (A1) into (A2), we have
Lt =

The household’s maximization problem is
given by
Max

P
(A2) L 1 = wt β Et ( t )
τ
Pt +1

wt τ
,
Rt



which is equation (2) in the text. Along with
the equilibrium conditions given in the text,
we also have Bt = 0 and Mts = Mt . Since we
are following an interest rate policy, the
implied inflation behavior is given by (A1).
The supporting money growth process can
then be backed out of the binding cashin-advance constraint.

59

References
Bernanke, Ben, and Mark Gertler. “Agency
Costs, Net Worth, and Business Fluctuations,” American Economic Review, vol. 79,
no. 1 (1989), pp. 14–31.
Bernanke, Ben, and Mark Gertler. “Monetary
Policy and Asset Price Volatility,” Federal
Reserve Bank of Kansas City Economic
Review, vol. 84 (1999), pp. 17–52.
Bernanke, Ben, Mark Gertler, and Simon
Gilchrist. “The Financial Accelerator in a
Quantitative Business Cycle Framework,”
in John Taylor and Michael Woodford, eds.,
Handbook of Macroeconomics. Volume 1C.
New York: Elsevier Science, 2000,
pp. 1341–93.
Carlstrom, Charles T., and Timothy S.
Fuerst. “Agency Costs, Net Worth, and
Business Fluctuations: A Computable General Equilibrium Analysis,” Federal Reserve
Bank of Cleveland, Working Paper no. 9602,
1996.
Carlstrom, Charles T., and Timothy S.
Fuerst. “Agency Costs, Net Worth, and
Business Fluctuations: A Computable
General Equilibrium Analysis,” American
Economic Review, vol. 87, no. 5 (1997),
pp. 893–910.
Carlstrom, Charles T., and Timothy S.
Fuerst. “Agency Costs and Business
Cycles,” Economic Theory, vol. 12, no. 3
(1998), pp. 583–97.
Carlstrom, Charles T., and Timothy S.
Fuerst. “Monetary Shocks, Agency Costs
and Business Cycles,” Carnegie-Rochester
Series on Public Policy, vol. 54 (2001),
pp. 1–27.
Cooley, Thomas F., and Kwanghee Nam.
“Asymmetric Information, Financial Intermediation, and Business Cycles,” Economic
Theory, vol. 12, no. 3 (1998), pp. 599–620.
Cooley, Thomas F., and Vincenzo
Quadrini. “Monetary Policy and the
Financial Decisions of Firms,” New York
University, unpublished manuscript, 1997.

Fisher, Jonas D.M. “Credit Market Imperfections and the Heterogeneous Response of
Firms to Monetary Shocks,” Journal of
Money, Credit, and Banking, vol. 31,
no. 2 (1999), pp. 187–211.
Fuerst, Timothy S. “Monetary and Financial
Interactions in the Business Cycle,” Journal
of Money, Credit, and Banking, vol. 27,
no. 4 (1995), pp. 1321–38.
Gertler, Mark, and Simon Gilchrist.
“Monetary Policy, Business Cycles, and the
Behavior of Small Manufacturing Firms,”
Quarterly Journal of Economics, vol. 109,
no. 2 (1994), pp. 309–40.
Hart, Oliver, and John Moore. “A Theory of
Debt Based on the Inalienability of
Human Capital,” Quarterly Journal of Economics, vol. 109, no. 4 (1994), pp. 841–79.
Kiyotaki, Nobuhiro, and John Moore.
“Credit Cycles,” Journal of Political Economy,
vol. 105, no. 2 (1997), pp. 211–48.