The full text on this page is automatically extracted from the file linked above and may contain errors and inconsistencies.
Rajashri Chakrabarti Program Design, Incentives, and Response: Evidence from Educational Interventions • In an effort to reform K-12 education, policymakers have introduced vouchers in some U.S. school districts, enabling students to transfer from public to private schools. • The different designs of two school voucher programs—the Milwaukee and Florida programs—have had different effects on public school incentives and performance. • In Milwaukee, vouchers were imposed from the outset; in Florida, schools were first threatened with vouchers and thus had an incentive to avoid them. • The Florida public schools’ efforts to avoid vouchers resulted in performance effects that far exceeded those of Milwaukee’s program. • Program design is critical: Policies that present failing public schools with functional and credible sanctions are best suited to provide the results intended by policymakers. Rajashri Chakrabarti is an economist at the Federal Reserve Bank of New York. rajashri.chakrabarti@ny.frb.org 1. Introduction C oncerns that U.S. students are not performing as well as their counterparts in other developed countries on international math and science tests have led to widespread demands for the reform of K-12 education in the United States. Of the various reforms under consideration, school voucher reform is at the forefront. Vouchers are scholarships that make students eligible to transfer from public to private schools. A basic feature of all publicly funded voucher programs in the United States is the funding of vouchers by public school revenue, so that money always “follows” students. In other words, schools that lose students lose their corresponding funding. Schools therefore recognize the financial implications of vouchers and have an incentive to avoid being subject to voucher programs. This article investigates the role of program design in the context of two such educational interventions in the United States—the Milwaukee and Florida school voucher programs—and analyzes the effects of design on public school incentives and performance.1 We demonstrate that variations in program design have markedly different outcomes for public schools affected by vouchers. The Milwaukee program, introduced in 1990, was the first voucher program in the country. Implemented in 1999, the The author thanks Andrew Haughwout for helpful suggestions. Noah Schwartz provided excellent research assistance. All errors are the author’s. The views expressed are those of the author and do not necessarily reflect the position of the Federal Reserve Bank of New York or the Federal Reserve System. FRBNY Economic Policy Review / October 2010 1 Florida program was the nation’s third, following Cleveland’s. The Milwaukee and Florida voucher programs share the basic feature of funding by public school revenue. But there are crucial differences. Milwaukee’s is a means-tested program targeting low-income students while Florida’s embeds a voucher program in a full-fledged accountability system. Using test-score data from Milwaukee and Florida and implementing a difference-in-differences estimation strategy, our study estimates the impact of each program by comparing the post-program results of the affected schools with a comparable set of control schools. Controlling for potentially confounding pre-program time trends and post-program common shocks, we find that the performance effects of the Florida program far exceed those of Milwaukee’s program. These results are quite robust in that they hold after controlling for other confounding factors, such as mean reversion and a possible stigma effect; they also withstand several sensitivity tests. Our findings have important policy implications, which we consider in the context of New York State’s federal, state, and city accountability programs. These programs include New York City’s accountability policy, known as the “Progress Report” policy, and the federal No Child Left Behind (NCLB) law, as implemented by New York State. Our study is organized as follows. Section 2 describes the Milwaukee and Florida voucher programs. In Section 3, we discuss the incentives created by the programs and the corresponding responses that might be expected from the affected public schools. Our data and empirical strategy are reviewed in Sections 4 and 5, respectively. Section 6 presents our results, and Section 7 considers policy implications. 2. Institutional Details The first publicly funded school voucher program in the United States was established in Milwaukee, Wisconsin, in 1990. The Milwaukee Parental Choice Program made the city’s entire low-income public school population eligible for 1 Our study focuses on the impact of alternative voucher designs on public school performance. A growing body of literature analyzes the many issues associated with school vouchers. Nechyba (1996, 1999, 2000, 2003) analyzes distributional effects of alternative voucher policies in a general equilibrium framework; Epple and Romano (1998, 2002) and Chakrabarti (2009) investigate sorting attributable to vouchers; Manski (1992) considers the impact of vouchers on public school expenditure and social mobility; and McMillan (2004) and Chakrabarti (2008b) model the quality of public schools facing vouchers. Hoxby (2003a, b) and Chakrabarti (2008a) study the effects of the Milwaukee voucher program, while Greene (2001), Greene and Winters (2003), Figlio and Rouse (2006), West and Peterson (2005), and Chakrabarti (2007, 2008a) study the effects of the Florida program. 2 Program Design, Incentives, and Response vouchers. Specifically, starting in the 1990-91 school year, the program made all Milwaukee public school students with family income at or below 175 percent of the poverty line eligible for vouchers to attend nonsectarian private schools. In contrast, the Florida Opportunity Scholarship Program, introduced in 1999, can be looked upon as a “threat-ofvoucher” program. Here, failing public schools were threatened with the imposition of vouchers, with vouchers implemented only if schools failed to meet a governmentdesignated cutoff quality level. The institutional details of the Milwaukee and Florida programs are summarized in Table 1. The Florida Department of Education classified schools according to five grades: A, B, C, D, or F. The state assigned school grades based on Florida Comprehensive Assessment The major difference in program design between the Milwaukee and Florida programs is that in Milwaukee vouchers were imposed at the outset, whereas in Florida failing schools were first threatened with vouchers, with vouchers introduced only if the schools failed to show adequate improvement in performance. Test (FCAT) reading, math, and writing scores. For FCAT reading and math, it categorized students into five achievement levels—1 lowest, 5 highest—that correspond to specific ranges on the raw-score scale. Using current-year data, the Department of Education assigned an “F” grade to a school if it was below the minimum criteria in reading, math, and writing; a “D” if it was below the minimum criteria in one or two of the three subject areas; and a “C” if it was above the minimum criteria in all three subjects, but below the higher performing criteria in all three. In reading and math, at least 60 percent (50 percent) of students had to score level 2 (3) and above; in writing, at least 50 percent (67 percent) had to score 3 and above to meet the minimum (high-performing) criteria in that respective subject.2 Under the Florida Opportunity Scholarship Program, all public school students became eligible for vouchers, or 2 In 1999, seventy-eight schools received an “F” grade. Students in two of those schools became eligible for vouchers. In 2000, four elementary schools received an “F,” although none became eligible for vouchers. In 2001, no schools received an “F” grade. In 2002, sixty-four schools received an “F.” Students in ten of those schools became eligible for vouchers. In 2003, students in nine schools became eligible for vouchers; in 2004, the figure was twenty-one schools. Table 1 Comparison of Milwaukee and Florida Voucher Programs Milwaukee Program Florida Program First U.S. voucher program Third U.S. voucher program Started in 1990-91 school year Started in 1998-99 school year Public school students with family income at or below 175 percent of the poverty line eligible for vouchers to attend nonsectarian private schools Vouchers contingent on school performance Private schools were not permitted, by law, to discriminate against students who apply with vouchers: – Had to accept all students unless oversubscribed – If oversubscribed, had to choose students randomly Average voucher amount equaled the state aid per pupil, and vouchers were financed by an equivalent reduction of state aid to the school district 1990-91 and 1996-97: – Average voucher amounts were $3,346 – Vouchers as a percentage of total revenue per pupil were 45.23 percent Schools classified according to five grades: A, B, C, D, F (A-highest, F-lowest) – Grades based on the Florida Comprehensive Assessment Test (FCAT) reading, math, and writing scores – F, if below the minimum criteria in reading, math, and writing – D, if below the criteria in one or two of the three subjects – C, if above the minimum criteria in all three subjects, but below the higher performing criteria in all three Students categorized into five achievement levels in FCAT reading and math (1-lowest, 5-highest) Minimum criteria: – Reading and math: at least 60 percent must score level 2 and above – Writing: at least 50 percent must score level 3 and above High-performing criteria: – Reading and math: at least 50 percent must score level 3 and above – Writing: at least 67 percent must score level 3 and above All students of a public school became eligible for vouchers if the school received two “F” grades in a period of four years Private schools were not permitted, by law, to discriminate against students who apply with vouchers – Had to accept all students unless oversubscribed – If oversubscribed, had to choose students randomly Average voucher amount equaled the state aid per pupil, and vouchers were financed by an equivalent reduction of state aid to the school district 1999-2000 and 2001-02: – Average voucher amounts were $3,330 – Vouchers as a percentage of total revenue per pupil were 41.55 percent Source: Information and data provided in various Florida Department of Education and Milwaukee Department of Public Instruction reports. “opportunity scholarships,” if the school received two “F” grades in a period of four years. Therefore, a school that received an “F” for the first time was exposed to the threat of vouchers, but did not face them unless and until it got a second “F” within the next three years. Thus, the major difference in program design between the Milwaukee and Florida programs is that in Milwaukee vouchers were imposed at the outset, whereas in Florida failing schools were first threatened with vouchers, with vouchers introduced only if the schools failed to show adequate improvement in performance. Apart from the above differences, the design of the two programs was strikingly similar. In both programs, private schools could not, by law, discriminate against students who applied with vouchers—the schools had to accept all students unless they were oversubscribed, in which case they had to choose students randomly. Indeed, the application form did not ask questions about the student’s race, sex, parents’ education, past scores, or prior records (for example, truancy, violence). The questions were specifically worded only to ascertain whether the student was eligible for the program.3 The system of funding for the Milwaukee and Florida voucher programs was also very similar. Under each program, the average voucher amount was equal to the state aid per pupil, and vouchers were financed by an equivalent reduction of state aid to the school district. Thus, state funding was directly tied to student enrollment, and enrollment losses due to vouchers were reflected in a revenue loss for the public school.4 The average voucher amounts under the Milwaukee (1990-91 through 1996-97) and Florida (1999-2000 through 2001-02) programs were $3,346 and $3,330, respectively. During these periods, vouchers as a percentage of total revenue per pupil were 45.23 percent in Milwaukee and 41.55 percent in Florida. 3 While the schools could not employ any selection criteria for the voucher students, this was not the case for nonvoucher students in the same school. Also note that the private schools had the choice of whether to participate in the program. However, if they decided to participate, they were required by law to accept all students or to choose students randomly, if oversubscribed. FRBNY Economic Policy Review / October 2010 3 3. Discussion: Effects of the Programs on Public School Incentives and Response Chart 1 Analyzing the Effect of “Voucher Threat” versus Vouchers Panel A What incentives would be created by the aforementioned program rules, and how would one expect the affected public schools to respond? Consider a public school subject to the Florida program, a school that has just received its first “F” grade (“F-school” hereafter). The school realizes that if it can avoid another “F” grade in the next three years, it can escape vouchers and the monetary loss and embarrassment associated with them.5 Therefore, it would have an incentive to improve its scores so as not to receive a second “F” grade. In contrast, if the same school were subject to a Milwaukee-type voucher program—in which vouchers have already been introduced— the school could not avoid vouchers (and the revenue loss) Net revenue R* RT Rv v=0 v>0 R1 R2 e v e* e T e Effort Panel B In a Florida-type [voucher] program, the threatened public schools . . . have more of an incentive to respond in order to improve their scores and escape vouchers. Net revenue R* RT Rv v=0 v>0 R1 altogether. In this case, improvements would serve to retain or attract a few students, but the effect would be marginal compared with that of a Florida-type program. In a Floridatype program, the threatened public schools (schools that have received their first “F” grade) have more of an incentive to respond in order to improve their scores and escape vouchers.6 Thus, the key difference between the two programs is that in the Milwaukee program, vouchers have already been implemented, whereas the Florida program first threatens the schools and gives them a window to respond, and an adequate response can preclude sanctions. Sanctions (vouchers) are implemented only if the schools fail to attain the predesignated standard. 4 We focus on the Milwaukee program up to 1996-97. The reason is that following a 1998 Wisconsin Supreme Court ruling, there was a major shift in the program when religious private schools were allowed to participate. Moreover, the financing of the Milwaukee program underwent some crucial changes, so that the voucher amounts and the revenue loss per student due to vouchers were not comparable between the Florida and second-phase Milwaukee programs. See Chakrabarti (2008a) for an analysis of how the shift in the Milwaukee voucher program affected public school incentives and performance as well as for a comparison of public school responses in the two phases of the program. We focus on the Florida program up to 2001-02. This period is chosen because schools that received an “F” grade in 1999 would face the threat of vouchers only through 2002. 5 The loss of students due to vouchers leads to a decrease in both revenue and costs for the school. But for a school operating at full capacity, the cost savings due to the loss of students are marginal, while the loss in revenue is not. This effect is a major reason why public schools do not support vouchers. 6 For a formal proof, see Chakrabarti (2008b). 4 Program Design, Incentives, and Response R2 e* e v e T e Effort The intuition above is shown in Chart 1. Let R 1 illustrate the initial net revenue function of the public school. The public school chooses the effort to maximize net revenue. Let this equilibrium effort be denoted by e and the corresponding net revenue by R . Now assume that Milwaukee-type vouchers are introduced. This leads to a downward shift of the net revenue function—the new net revenue function is denoted by R 2 and the corresponding optimum effort and net revenue by e V and R V , respectively.7 Panel A of the chart illustrates the case in which e V e , and panel B the case in which e V e . The chart implies that any target effort in the range e V , e under a threat-of-voucher regime will induce an effort strictly greater than e V . For example, assume that the policymaker implements a target effort, e T . Satisfying this target would lead to a net revenue of R T while failing to satisfy it would lead to the introduction of vouchers and corresponding revenue of R V ( R T . Therefore, the school has an incentive to implement an effort of e T ( e V . 7 For formal proofs, see Chakrabarti (2008b). 4. Data The Florida data consist of school-level data on test scores, grades, socioeconomic characteristics of schools, and school finances; they are obtained from the Florida Department of Education. School-level data on test scores are obtained from the Florida Comprehensive Assessment Test. Mean scale scores (on a scale of 100-500) on grade 4 reading and grade 5 math are available for 1998-2002. Mean scale scores (on a scale of 1-6) on the Florida grade 4 writing test are available for 1994-2002. Data on socioeconomic characteristics include sex composition (1994-2002), percentage of students eligible for free or reduced-price lunch (1997-2002), and race composition (1994-2002), and are obtained from the school indicators database of the Florida Department of Education. This study refers to school years by the calendar year of the spring semester. School finance data consist of several measures of school-level and district-level per-pupil expenditures, and are obtained from the school indicators database and the Office of Funding and Financial Reporting of the Florida Department of Education. The Wisconsin data consist of school-level data on test scores, socioeconomic characteristics of schools, and per-pupil expenditures (both at the school and district levels). The data are obtained from the Wisconsin Department of Public Instruction, the Milwaukee Public Schools, and the Common Core of Data of the National Center for Education Statistics. School-level data on test scores are obtained for 1) the Third Grade Reading Test (renamed the Wisconsin Reading Comprehension Test, or WRCT, in 1996) and 2) the grade 5 Iowa Test of Basic Skills (ITBS). School scores for the WRCT, which was first administered in 1989, are reported in three “performance standard categories”: percentage of students below, percentage of students at, and percentage of students above the standard.8 Data for these three categories are available for 1989-97. School-level ITBS reading data (mean scores) are available for 1987-93; ITBS math data (mean scores) are available for 1987-97. 5. Empirical Strategy 5.1 Florida In Florida, the schools that received an “F” grade in 1999 were directly exposed to the threat of vouchers because all their students would be eligible for vouchers if the school received 8 The method of reporting ITBS math and WRCT reading scores changed in 1998. Therefore, we use pre-1998 scores. another “F” in the next three years. These F-schools constitute the group of treated schools. Schools that received a “D” grade in 1999 were closest to the F-schools in terms of grade, but were not directly treated by the program. These “D-schools” constitute the group of control schools. The treatment and control groups consist of 65 and 457 elementary schools, If the F-schools and D-schools have similar trends in scores in the pre-program period, any shift of the F-schools compared with the D-schools in the postprogram period can be attributed to the program. respectively.9 Because the program was announced in June 1999 and the grades were based on tests held in February 1999, we classify schools into treatment and control groups on the basis of their pre-program scores and grades. The identifying assumption here is that if the F-schools and D-schools have similar trends in scores in the pre-program period, any shift of the F-schools compared with the D-schools in the post-program period can be attributed to the program. To test this assumption, we first run the following fixed-effects regression (and its ordinary least squares [OLS] counterpart) using only pre-program data: (1) sit = f i + 0 t + 1 F * t + 2 X it + it , where si t is the mean score of school i in year t, fi are schoolfixed effects, t denotes a time trend, F is a dummy variable taking a value of 1 for F-schools and 0 for D-schools, F*t is an interaction between the F dummy and trend, X it denotes the set of school characteristics, and it is a stochastic error term. Scores considered in the Florida part of the analysis include mean school scores in FCAT reading, FCAT math, and FCAT writing. The pre-program difference in trend of the F-schools is captured in 1 . If F-schools and D-schools have similar pre-program trends, we investigate whether the F-schools demonstrate a higher improvement in test scores in the post-program era using specification 2 below. If the treated F-schools demonstrate a differential pre-program trend, then in addition to estimating this specification, we estimate a modified version in which we control for the pre-program differences in trends. We estimate a completely unrestricted and nonlinear model that includes year dummies to control for common year effects and interactions of post-program year dummies with 9 We restrict our analysis to elementary schools because there were too few middle and high schools that received an “F” grade in 1999 (seven and five, respectively) to justify analysis. FRBNY Economic Policy Review / October 2010 5 the F-school dummy to capture individual post-program year effects: 2002 (2) sit = f i + 0j Dj + j = 1999 2002 1j F * D j + 2 X it + it , j = 1999 where D j , j = 1999 2000 2001 2002 are year dummies for 1999, 2000, 2001, and 2002, respectively. While the above specification includes school-fixed effects, we also estimate an OLS counterpart to it. OLS regressions corresponding to both specifications 1 and 2 include a dummy for the treatment group F. Note that this is absorbed in the fixed-effects regressions because it is a time-invariant school effect. Specification 2 does not constrain the post-program yearto-year gains of the F-schools to be equal and allows the program effect to vary across years. The coefficients 1t i = 2000 2001 2002 represent the effect of one, two, and three years into the program, respectively, for the F-schools. Given the nature of the Florida program, the 1999 threatened schools (that is, the schools that received an “F” grade in 1999) would be exposed to the threat of vouchers for the next three years only. Therefore, we track the performance of the threatened schools (relative to the control schools) for three years after the program—2000, 2001, and 2002—when the threat of vouchers would be in effect. The above specifications assume that the D-schools were not affected by the program. Although the D-schools did not face any direct threat from the program, they might have faced an indirect threat because they were close to receiving an “F” grade.10 Therefore, we next allow the F-schools and D-schools to be different treated groups (with varying intensities of treatment) and compare their post-program improvements, if any, with 1999 “C-schools,” which are the next grade up in the scale using the above specifications after adjusting for another treatment group. It should be noted that since D-schools and C-schools may face the threat to some extent, our estimates may be underestimates (lower bounds), but not overestimates. 5.2 Milwaukee Our strategy is based on and is similar to that of Hoxby (2003b). Since students in the Milwaukee Public Schools eligible for free or reduced-price lunch were also eligible for vouchers, the extent of treatment of the Milwaukee schools depended on the percentage of students eligible for free or reduced-price lunch.11 Using this information, Hoxby 10 In fact, there is some anecdotal evidence that D-schools may have responded to the program. The superintendent of Hillsborough County, which had no F-schools in 1999, announced that he would take a 5 percent pay cut if any of his thirty-seven D-schools received an “F” grade on the next school report card. For more information, see Innerst (2000). 6 Program Design, Incentives, and Response classifies the Milwaukee schools into two treatment groups based on the percentages of students eligible for free or reduced-price lunch—“most treated” (at least two-thirds of students eligible in the pre-program period) and “somewhat treated” (less than two-thirds of students eligible in the preprogram period). We classify the schools into three treatment groups (in contrast to Hoxby’s two) based on their pre-program (1989-90 school year) percentage of students eligible for free or reducedprice lunch. Thus, our treatment groups are more homogenous as well as starker from each other. Additionally, to test the Since students in the Milwaukee Public Schools eligible for free or reduced-price lunch were also eligible for vouchers, the extent of treatment of the Milwaukee schools depended on the percentage of students eligible for free or reducedprice lunch. robustness of our results, we consider alternative samples obtained by varying the cutoffs that separate the different treatment groups, departing from the Hoxby approach. The 60-47 (66-47) sample classifies schools that have at least 60 percent (66 percent) of students eligible for free or reducedprice lunch as “more treated,” schools with such population between 60 percent (66 percent) and 47 percent as “somewhat treated,” and schools with such population less than 47 percent as “less treated.” We also consider alternative classifications, such as “66” and “60” samples, where there are two treatment groups—schools that have at least 66 percent (60 percent) of students eligible for free or reduced-price lunch are designated as more treated schools, and schools with such population below 66 percent (60 percent) as somewhat treated schools. Since there were very few middle and high schools in the Milwaukee Public Schools and student participation in the Milwaukee Parental Choice Program was mostly in the elementary grades, we restrict our analysis to elementary schools. 11 Under the Milwaukee program, all households at or below 175 percent of the poverty line are eligible to apply for vouchers. Households at or below 185 percent of the poverty line are eligible for free or reduced-price lunch. However, the cutoff of 175 percent is not strictly enforced (Hoxby 2003a), and households within this 10 percent margin are often permitted to apply. In addition, there were very few students who fell in the 175 percent185 percent range, while in fact 90 percent of students eligible for free or reduced-price lunch qualified for free lunch (Witte 2000). Students below 135 percent of the poverty line qualified for free lunch. Our control group criteria are also based on Hoxby (2003b). Since all schools in Milwaukee were potentially affected by the program, Hoxby constructs a control group that consists of Wisconsin schools outside Milwaukee that satisfy the following criteria in the pre-program period that: 1) had at least 25 percent of their population eligible for free or reduced-price lunch, 2) had black students who make up at least 15 percent of the population, and 3) were urban. Her control group consists of twelve schools. For our control schools, we designate schools that are located outside Milwaukee but within Wisconsin, satisfy Hoxby’s first two criteria, and have locales as similar as possible to the Milwaukee schools. Note that all of these characteristics pertain to the pre-program school year 1989-90.12 Using each sample, we investigate how the different treatment groups in Milwaukee responded to the “voucher shock” program. Using specification 3 below, we first test whether the pre-program trends of the untreated and the different treated groups were the same. We then estimate OLS and fixed-effects versions of specification 4 below. If we observe differences in pre-existing trends between the different treated groups of schools, then in addition to estimating specification 4, we estimate modified versions of the specification that control for pre-existing differences in trends: (3) 1k Ik * t + 2 Xit + it s it = fi + 0 t + Table 2 Pre-Program Demographic Characteristics of Florida and Wisconsin More Treated and Control Schools Percent Panel A: More Treated Schools Wisconsin Black Hispanic White Male Free or reducedprice lunch Hispanic White Male k (4) sit = f i + 2007 2007 j = 1989 j = 1989x 0j Dj + 1k j I k * D j + 2 X it + it , where sit denotes scores of school i in period t; D j , j = 1989 2007 are year dummies for 1989 through 2007, respectively; k MT ST LT for the WRCT and k MT ST for the ITBS, where MT denotes “more treated,” ST denotes “somewhat treated,” and LT denotes “less treated.” The scores considered are mean scores in ITBS reading and ITBS math as well as percentages of students above the standard in WRCT reading. 6. Results Table 2 presents baseline characteristics of treated and control groups in Florida and Wisconsin. It shows that the more treated schools in Florida were indeed similar to the more 12 The more treated and control group characteristics are presented in Table 2. In the 66-47 sample, the somewhat treated (less treated) group had an average of 55.4 percent (37.17 percent) of students eligible for free or reduced-price lunch, 50.99 percent (45.37 percent) who were black, and 4.09 percent (3.83 percent) who were Hispanic. Florida 66-47 60-47 66-47 60-47 62.79 (28.23) 18.95 (23.40) 17.18 (19.54) 51.38 (4.84) 85.80 (9.95) 66.55 (32.22) 18.07 (24.54) 10.21 (10.68) 52.25 (2.60) 84.5 (6.48) 62.90 (29.58) 14.81 (21.86) 17.38 (16.55) 52.33 (2.58) 82.9 (9.04) -3.76 [0.56] 0.88 [0.87] 6.97 [0.07] -0.87 [0.34] 1.3 [0.50] -0.10 [0.99] 4.14 [0.36] -0.20 [0.96] -0.95 [0.22] 2.9 [0.12] Panel B: Control Schools Florida Black Free or reducedprice lunch Florida–Wisconsin 18.12 (14.17) 15.49 (21.23) 63.59 (22.33) 51.38 (4.84) 50.14 (17.51) Wisconsin 22.37 (12.93) 14.84 (6.02) 60.85 (12.80) 50.63 (2.29) 44.95 (11.66) Florida–Wisconsin -4.25 [0.10] 0.17 [0.86] 2.73 [0.49] 0.76 [0.43] 5.19 [0.10] Source: Author’s calculations. Notes: The group of Florida more treated and control schools is composed of F-schools and C-schools, respectively. Samples 66-47 and 60-47 are described in Section 5.2 of the article. Standard deviations are in parentheses; p-values are in brackets. treated schools in Wisconsin and, except in one case, the differences between them were not statistically significant. Similarly, the control schools in Florida were similar to the control schools in Wisconsin, and the differences between them were not statistically significant. However, the treated schools were somewhat different from the control schools within each state. The reason is that Wisconsin schools outside Milwaukee were considerably more advantaged than schools in Milwaukee. We arrived at this control group despite using the strategy (following Hoxby [2003a, b]) of selecting control schools as similar as possible to Milwaukee’s more treated schools in terms of pre-program characteristics. It is important that both the more treated schools and the control groups be similar across the two programs in terms of pre-program characteristics as well as across the two locations. As a result, for purposes of comparing effects across the two FRBNY Economic Policy Review / October 2010 7 programs, we use the C-schools in Florida as the control group. Noticeably, the control group in Wisconsin was very similar to the C-schools in Florida and was not statistically different from them in terms of any characteristics (Table 2). Still another reason for selecting the C-schools as the control group in Florida was that while the D-schools were more similar to the more treated F-schools in terms of grade and demographics, they were very close to receiving an “F” grade; hence, to some extent they perceived an indirect threat and to some extent were treated by the program. Because of differences between the treated and control schools, one might argue that in the absence of the program, the control group would have evolved differently from the more treated group. However, multiple years of pre-program data allow us to check (and control) for any differences in pre-program trends of these groups. In this way, we can dispose of any level differences between the treated and control groups as well as control for differences in pre-program trends, if any. It seems likely that once we control for differences in trends as well as in levels, any remaining differences between the treated and control groups will be minimal. In other words, our identifying assumption is that if the treated schools followed the same trends as the control schools in the immediate preprogram period, they would have evolved similarly in the immediate post-program period in the absence of the program. We also control for time-varying observable characteristics. School-fixed effects remove any time-invariant unobservable characteristics. Note that while time-varying unobserved characteristics cannot be directly controlled for, they did not drive the results as long as the F-schools did not experience a differential shock in unobserved characteristics that coincided with the timing of the program. 6.1 Florida Considerable anecdotal evidence suggests that F-schools have responded to the voucher program. Just after the program’s inception, Escambia County implemented a 210-day extended school year in its F-schools (the typical duration was 180 days), introduced an extended school day at least twice a week, and added small-group tutoring on afternoons and Saturdays and longer time blocks for writing and math instruction. To curb absenteeism, the county started an automated phone system to contact parents when a child is absent. Miami-Dade County hired 210 additional teachers for its twenty-six F-schools, switched to phonics instruction, and encouraged parents (many of whom were dropouts) to go back to school for a high-school- 8 Program Design, Incentives, and Response equivalency diploma. Broward County reduced its class size to eighteen to twenty students in its low-performing schools and increased services for children whose primary language is not English. Palm Beach County targeted its fourth-grade teachers for coaching and began more frequent and closer observation of teachers in its F-schools (Innerst 2000). Carmen Varela-Russo, Associate Superintendent of Technology, Strategic Planning, Considerable anecdotal evidence suggests that F-schools have responded to [Florida’s] voucher program. and Accountability at Broward County Public Schools, described the situation this way: “People get lulled into complacency . . . the possibility of losing children to private schools or other districts was a strong message to the whole community” (Innerst 2000). The analysis below investigates whether the data in Florida support this behavior. Chart 2, which depicts trends in reading, math, and writing scores in F-schools and D-schools, shows that 1999 was the watershed year. In both reading and math, the F-schools had similar trends before the program. However, the F-schools showed improvement relative to the D-schools after the program, and the gap between F- and D-schools narrowed. In writing, while the F-schools were deteriorating relative to the D-schools before the program, this pattern changed after it. The F-schools showed improvement relative to the D-schools to the extent that they successfully closed the “F” to “D” gap after the program. We now turn to our estimation results. All regressions control for ethnicity (the percentage of students in different racial categories in a school), the percentage of male students, the percentage of students eligible for free or reduced-price lunch, and real per-pupil expenditures. Table 3 presents preprogram trends in reading, math, and writing. It reveals that F-schools have no significant differences in trend compared with D-schools in reading and math, although they exhibit a small, negative differential trend in writing. Compared with C-schools, F-schools exhibit a negative differential trend in reading and writing, but no significant differential trend in math. D-schools exhibit a negative trend in reading and a positive trend in math and writing compared with C-schools. Whenever there is a difference in pre-program trends, our reported regressions control for these differences by including interactions between trend and the respective treated dummies. D-schools are considered as an additional treatment group in Table 4, columns 4-6. Here, we see how the program affects F-schools (more treated) and D-schools (less treated) compared with C-schools. All columns control for differences in pre-existing trends between groups. The results show positive, significant year effects in reading, math, and writing for F-schools in each of the years after the program’s implementation. Although many of the D-school effects are also positive and significant, the F-school shifts are statistically larger in each year.15 The F-school effects are economically meaningful as well. In reading, relative to the base year, F-schools showed a 3.6 percent improvement in the first year after the program, a 4.2 percent improvement after the second year, and a 6.3 percent improvement after the third year. In math, F-schools showed a 3.4 percent, 4.2 percent, and Chart 2 Florida “Threat-of-Voucher” Program Score 290 FCAT Reading 280 D-schools 270 F-schools 260 250 1998 300 99 FCAT Math 00 01 02 D-schools 290 Our results show considerable improvement in the F-schools after the program compared with the control schools. F-schools 280 270 260 1998 3.5 99 00 01 02 FCAT Writing 3.0 D-schools F-schools 2.5 2.0 1.5 1994 95 96 97 98 99 00 01 02 Source: Author’s calculations. Note: FCAT is the Florida Comprehensive Assessment Test. Table 4, columns 1-3 present the effects of the Florida threat-of-voucher program on F-school reading, math, and writing scores compared with those for D-schools. All models reported include school-fixed effects. The results from our OLS estimation are similar to the fixed-effects estimates and hence are not reported. The regressions for writing include interactions of the “F” dummy with trend to control for differences in pre-program trends seen above.13,14 The table shows economically large, positive, and statistically significant effects in each subject area and year. 4.5 percent improvement in the first, second, and third years, respectively, after implementation of the program. In writing, the percentage improvement was around 15 percent. At the end of 2002 (three years after program implementation), the pre-program gap between F-schools and C-schools was closed by 37.08 percent in reading, 30.31 percent in math, and around 75 percent in writing. In summary, based on different samples, different subjects, and both OLS and fixed-effects estimates, our results show considerable improvement in the F-schools after the program compared with the control schools. Although D-schools show non-negligible improvement (at least in reading and writing), their improvement is considerably less than and statistically different from that of the F-schools. 13 Note that the table reports only the coefficients that reflect program effects; therefore, the coefficient corresponding to this interaction term (which captures the differential pre-existing trend) is not reported. Pre-existing trends are reported in Table 3. 14 The regressions for reading and math (columns 1 and 2) do not include this interaction term because there is no evidence of differential pre-program trends in reading and math for F-schools and D-schools (Table 3). Note that the results with inclusion of this term remain very similar. 15 Here, we test whether the F-school effects are statistically different from the D-school effects against the null hypothesis that they are equal. FRBNY Economic Policy Review / October 2010 9 Table 3 Pre-Program Trend of F-, D-, and C-Schools in Florida Sample of F- and D-Schools FCAT Reading Trend F * trend FCAT Math OLS (3) FE (4) Sample of F-, D-, and C-Schools FCAT Writing OLS (5) FE (6) OLS (7) FE (8) FCAT Math OLS (9) FE (10) FCAT Writing OLS (1) FE (2) 0.41 (0.56) -1.78 (2.47) -0.05 (0.47) -2.01 (1.46) 13.20*** (0.55) -0.98 (1.44) 13.02** (0.61) -0.72 (1.48) 0.20** (0.008) -0.05*** (0.011) 0.21** (0.003) -0.04*** (0.007) 2.66** (0.57) -3.80 (2.29) -2.29*** (0.66) 2.70 (0.36) -4.77*** (1.41) -2.69*** (0.57) 9.79*** (0.53) 2.46 (1.51) 3.46*** (0.60) 10.20*** (0.38) 1.96 (1.43) 2.79*** (0.67) 0.18*** (0.01) -0.03*** (0.01) 0.02** (0.007) 0.19*** (0.002) -0.03*** (0.01) 0.02*** (0.003) Y 1,013 0.58 Y 1,013 0.93 Y 1,006 0.59 Y 1,006 0.90 Y 2,948 0.64 Y 2,948 0.80 Y 2,386 0.76 Y 2,386 0.95 Y 2,377 0.74 Y 2,377 0.93 Y 6,982 0.65 Y 6,982 0.82 D * trend Controls Observations R2 FCAT Reading OLS (11) FE (12) Source: Author’s calculations. Notes: FCAT is the Florida Comprehensive Assessment Test; OLS is ordinary least squares regression; FE is fixed-effects regression. Controls include race, sex, percentage of students eligible for free or reduced-price lunch, and real per-pupil expenditure. Huber-White standard errors are in parentheses. All regressions are weighted by the number of students tested. ***Statistically significant at the 1 percent level. ***Statistically significant at the 5 percent level. ***Statistically significant at the 10 percent level. 10 Program Design, Incentives, and Response Table 4 Effect of “Threatened Status” on FCAT Reading, Math, and Writing Scores Sample of Treated F- and Control D-Schools in Florida Reading FE (1) Treated * one year after program Treated * two years after program Treated * three years after program 4.85*** (1.68) 3.30* (1.71) 7.08*** (1.78) Math FE (2) 6.78*** (1.63) 7.25*** (1.82) 5.35*** (2.00) Writing FE (3) Less treated * two years after program Less treated * three years after program More treated * one year after program More treated * two years after program More treated * three years after program Y Y Y 2,550 0.77 0.00 Y Y Y 2,524 0.76 0.00 Math FE (5) Writing FE (6) 0.35*** (0.04) 0.37*** (0.04) 0.43 (0.05) Less treated * one year after program School-fixed effects Year dummies Controls Observations R2 p-valuec Reading FE (4) Y Y Y 4,476 0.85 0.00 3.53*** (0.76) 5.52*** (0.80) 7.94*** 0.97 (0.85) 2.54*** (0.94) 3.47*** 0.05** (0.02) 0.00 (0.02) -0.03 (0.87) 9.32b*** (1.87) 10.75a*** (1.87) 16.03b*** (1.91) (0.92) 8.96b*** (1.59) 11.00b*** (1.77) 11.94b*** (1.95) (0.02) 0.39b*** (0.04) 0.37a*** (0.04) 0.39a*** (0.05) Y Y Y 5,933 0.86 0.00 Y Y Y 5,909 0.83 0.00 Y Y Y 10,587 0.86 0.00 Source: Author’s calculations. Notes: FCAT is the Florida Comprehensive Assessment Test. FCAT scores for reading and math are for the period 1998-2000; FCAT scores for writing are for the period 1994-2002. FE is fixed-effects regression. Huber-White standard errors are in parentheses. Controls include race, sex, percentage of students eligible for free or reduced-price lunch, and real per-pupil expenditure. All regressions are weighted by the number of students tested. a More treated significantly different from less treated at 5 percent level. More treated significantly different from less treated at 1 percent level. c p-value of F-test of the program effect on treated schools. b ***Statistically significant at the 1 percent level. ***Statistically significant at the 5 percent level. ***Statistically significant at the 10 percent level. FRBNY Economic Policy Review / October 2010 11 Table 5 Pre-Program Trend of More Treated, Somewhat Treated, and Less Treated Schools in Milwaukee WRCT (Percentage above) OLS (1) Trend More treated * trend Somewhat treated * trend Less treated * trend Observations R2 ITBS Reading FE (2) ITBS Math OLS (3) FE (4) OLS (5) FE (6) -3.84 (2.33) -3.08 (3.41) -4.41 (3.01) -2.33 (3.61) -4.34** (2.16) -2.03 (3.35) -3.61 (2.67) -3.23 (3.10) -4.09 (4.11) 4.01 (3.69) 3.14 (4.05) -3.45 (3.42) -1.88 (2.73) 2.12 (3.17) -3.04* (1.66) 0.56 (1.97) 0.73 (1.83) 2.52** (0.98) 0.32 (1.40) 0.31 (1.21) 242 0.50 242 0.87 411 0.30 411 0.56 410 0.30 410 0.71 Source: Author’s calculations. Notes: WRCT is the Wisconsin Reading Comprehension Test; ITBS is the Iowa Test of Basic Skills; OLS is ordinary least squares regression; FE is fixed-effects regression. Controls include race, sex, and percentage of students eligible for free or reduced-price lunch. ***Statistically significant at the 1 percent level. ***Statistically significant at the 5 percent level. ***Statistically significant at the 10 percent level. 6.2 Milwaukee These results seem to be robust in that they are replicated in the analysis with other samples.17 Chart 3 presents the trends in The Milwaukee analysis uses the 66-47 sample. Estimation results for pre-program trends are presented in Table 5. The results show no statistical difference in trends between the various treated and control groups in any subject area. Table 6 examines the effect of the Milwaukee “voucher shock” program on the WRCT (the percentage above), ITBS reading, and ITBS math scores of different treated groups. Except for the positive and statistically significant effect in WRCT reading in the test’s second year, there is no statistically significant evidence of any effect of the program. Although the second year’s somewhat treated effect in ITBS math is statistically significant, it is more than the corresponding more treated effect.16 Thus, the results in Milwaukee are mixed. The program seems to have had a positive and significant effect in the second year after the program’s implementation, at least in the WRCT. 16 Since the ITBS was administered in Milwaukee as a district assessment program, we do not have data on non–Milwaukee, Wisconsin, schools for this test. As a result, our comparison group is the less treated group of schools. Since the comparison group is also treated to some extent, we expect our estimates for the ITBS to be lower bounds. 12 Program Design, Incentives, and Response The results show no statistical difference in trends between the various treated and control groups in any subject area. . . . Except for the positive and statistically significant effect in [Wisconsin Reading Comprehension Test] reading in the test’s second year, there is no statistically significant evidence of any effect of the program. . . . Thus, the results in Milwaukee are mixed. ITBS scores for the various groups. As expected, there is no evidence of any program effect. 17 These results are not reported here, but are available from the author. Table 6 Chart 3 Effect of the Milwaukee “Voucher Shock” Program Milwaukee “Voucher-Shock” Program WRCT (1) Somewhat treated * one year after program ITBS Reading (2) ITBS Math (3) More treated * three years after program 2.03 (2.81) 5.38** (2.43) 5.01 (3.03) -0.92 (3.33) 6.06* (3.14) 5.69 (3.98) 4.15 (4.49) 7.83 (5.17) 6.78 (5.31) 1.12 (3.86) 6.59 (5.15) 2.85 (5.18) -1.35 (2.94) 6.14* (3.38) 2.47 (3.31) -4.02 (3.26) 4.36 (3.83) -2.22 (3.54) School-fixed effects Year dummies Controls Observations R2 a p-value Y Y Y 1,195 0.58 0.11 Y Y Y 717 0.55 0.62 Y Y Y 1,127 0.60 0.27 Somewhat treated * two years after program Somewhat treated * three years after program More treated * one year after program More treated * two years after program Score 70 ITBS Reading 60 Control 50 Somewhat treated schools 40 30 More treated schools 20 1987 70 88 89 90 91 93 ITBS Math 60 Control 50 Somewhat treated schools 40 More treated schools Source: Author’s calculations. Notes: WRCT is the Wisconsin Reading Comprehension Test; ITBS is the Iowa Test of Basic Skills. Huber-White standard errors are in parentheses. All regressions include school-fixed effects and control for race, sex, percentage of students eligible for free or reduced-price lunch, and real per-pupil expenditure. 92 30 1987 88 89 90 91 92 93 94 95 96 97 Source: Author’s calculations. Note: ITBS is the Iowa Test of Basic Skills. a p-value of the F-test of joint significance of more treated shift coefficients. ***Statistically significant at the 1 percent level. ***Statistically significant at the 5 percent level. ***Statistically significant at the 10 percent level. FRBNY Economic Policy Review / October 2010 13 7. Robustness Checks Table 7 7.1 Mean Reversion Panel A: 98F- and 98D-Schools Mean Reversion of 98F-Schools Compared with 98D- and 98C-Schools, 1998-99 Dependent Variable: FCAT Score, 1998-99 Several factors might bias the results; we consider each factor and its potential solutions. First is the issue of mean reversion. Mean reversion is the statistical tendency whereby highor low-scoring schools tend to score closer to the mean subsequently. Because the F-schools scored low in 1999, a natural question would be whether the improvement in Florida is driven by mean reversion rather than the voucher program. Since we conduct a difference-in-differences analysis, our estimates will be tainted by mean reversion only if F-schools mean-revert to a greater extent than do the D-schools or the C-schools, or both. To investigate mean reversion, we examine whether and by how much schools that received an “F” grade in 1998 improved during the 1998-99 academic year compared with those that received a “D” (or “C”) grade in 1998. Since these years fall within the pre-program period, the gain can be taken to approximate the mean-reversion effect and can be subtracted from the post-program gain of F-schools compared with D-schools (or C-schools) to get at the mean-reversioncorrected program effect. The accountability system of assigning letter grades to schools began in 1999. The pre-1999 accountability system classified schools into four groups, designated 1 (low) to 4 (high). However, using the state grading criteria and data on the percentage of students in different achievement levels in each FCAT reading, math, and writing, we assigned letter grades to schools in 1998 and implemented the above strategy. Schools receiving “F,” “D,” and “C” grades in 1998 using this procedure are referred to as “98F-schools,” “98D-schools,” and “98C-schools,” respectively. Using Florida data for 1998 and 1999, we demonstrate in Table 7, panel A, that when compared with the 98D-schools, the 98F-schools show no evidence of mean reversion either in reading or math, although there is mean reversion in writing. Compared with the 98C-schools (panel B), there is no evidence of mean reversion in reading; both 98D-schools and 98Fschools show comparable amounts of mean reversion in math; only 98F-schools show mean reversion in writing. 14 Program Design, Incentives, and Response Reading FE (1) Trend 98F * trend Observations R2 2.01*** (0.43) -0.65 (1.14) 1,353 0.93 Math FE (2) 14.02*** (0.49) 1.17 (1.19) 1,354 0.91 Writing FE (3) 0.04*** (0.01) 0.14*** (0.02) 1,355 0.85 Panel B: 98F-, 98D-, and 98C-Schools Dependent Variable: FCAT Score, 1998-99 Reading FE (1) Trend 98F * trend 98D * trend Observations R2 1.76*** (0.35) -0.55 (1.12) 0.16 (0.54) 2,605 0.96 Math FE (2) 9.71*** (0.36) 4.63*** (1.16) 4.22*** (0.58) 2,608 0.94 Writing FE (3) 0.03*** (0.01) 0.14*** (0.02) 0.01 (0.01) 2,608 0.87 Source: Author’s calculations. Notes: FCAT is the Florida Comprehensive Assessment Test; FE is fixedeffects regression. All regressions control for race, sex, percentage of students eligible for free or reduced-price lunch, and real per-pupil expenditure. The ordinary least squares regressions include 98F- and 98D-school dummies. In the sample of 98F- and 98D-schools, the standard deviations of FCAT reading, math, and writing are 18.9, 18.05, and 0.30, respectively. In the sample of 98F-, 98D-, and 98C-schools, the standard deviations of FCAT reading, math, and writing are 21.16, 21.56, and 0.31, respectively. ***Statistically significant at the 1 percent level. ***Statistically significant at the 5 percent level. **Statistically significant at the 10 percent level. Table 8 Is There a Stigma Effect of Getting the Lowest Performing Grade? Effect of Being Categorized in Group 1 on FCAT Writing Scores Using FCAT Writing Scores, 1997-98 Sample: Group 1, 2 Schools OLS (1) Trend Group 1 * trend 0.52*** (0.04) -0.01 (0.08) FE (2) 0.52*** (0.03) -0.02 (0.06) Sample: Group 1, 2, 3 Schools FE (3) 0.48*** (0.04) -0.02 (0.06) Group 2 * trend Controls Observations R2 N 314 0.49 N 314 0.84 Y 314 0.85 OLS (4) FE (5) FE (6) 0.48*** (0.02) 0.03 (0.07) 0.03 (0.04) 0.48*** (0.01) 0.01 (0.05) 0.04 (0.03) 0.46*** (0.02) 0.02 (0.05) 0.04 (0.03) N 1,361 0.52 N 1,361 0.87 Y 1,358 0.87 Source: Author’s calculations. Notes: FCAT is the Florida Comprehensive Assessment Test; OLS is ordinary least squares regression; FE is fixed-effects regression. Huber-White standard errors are in parentheses. All regressions are weighted by the number of students tested; controls include race, sex, percentage of students eligible for free or reduced-price lunch, and real per-pupil expenditure. The OLS regressions include group 1 and group 2 dummies. ***Statistically significant at the 1 percent level. ***Statistically significant at the 5 percent level. ***Statistically significant at the 10 percent level. 7.2 Stigma Effect of Getting the Lowest Performing Grade A second concern in Florida is the potential stigma effect of receiving a performance grade of “F.” If there is such a stigma, the F-schools will try to improve only to avoid this stigma rather than in response to the program. We use several If there is [a low-performance] stigma, the F-schools will try to improve only to avoid this stigma, rather than in response to the program. alternative strategies to investigate this possibility. First, although the system of assigning letter grades to schools started in 1999, Florida had an accountability system in the pre-1999 period when schools were categorized into four groups, designated 1 (low) to 4 (high), based on FCAT writing and reading and math norm-referenced test scores. Using FCAT writing data for two years (1997 and 1998), we investigate whether the schools, which were categorized in group 1 in 1997, improved in relation to the 1997 group 2 and group 3 schools in 1997-98.18 Our rationale is that if a stigma effect is associated with getting the lowest performing grade, the group 1 schools should improve relative to the group 2 and 3 schools, even in the absence of the threat-of-voucher program. Table 8, using pre-program FCAT writing scores, shows that no such stigma effect exists—group 1 schools display no improvement relative to the group 2 or group 3 schools. Second, all the schools that received an “F” grade in 1999 received higher grades in 2000, 2001, and 2002. Therefore, although the stigma effect on F-schools may be operative in 2000, this is not likely to be the case in 2001 or 2002 since none of the F-schools received an “F” grade in the preceding year (2000 or 2001, respectively). However, the F-schools would face the threat of vouchers until 2002, so any improvement in 18 We do not use the pre-1999 reading and math norm-referenced test (NRT) scores because different districts used different NRTs during this period, which varied in content and norms. Also, districts often chose different NRTs in different years. Thus, these NRTs were not comparable across districts and across time. Moreover, since districts could choose the specific NRT to administer each year, the choice was likely related to timevarying (and also time-invariant) district-unobservable characteristics that also affected test scores. FRBNY Economic Policy Review / October 2010 15 2001 and 2002 would provide evidence in favor of the threatof-voucher effect and against the stigma effect. F-schools showed strong gains in both 2001 and 2002—a result that provides further support for the threat-of-voucher effect and against the stigma effect. Table 9 Effect of Milwaukee Program on Demographic Composition of Schools Percent Black (1) Less treated * program 7.3 Sorting Another factor relates to sorting in the context of the Milwaukee voucher program. Vouchers affect public school quality not only through direct public school response but also through changes in student composition and peer quality brought about by sorting. These three factors are then reflected in the public school scores.19 This issue is important in Milwaukee because over the years students have left the city’s public schools with vouchers. In contrast, no Florida school became eligible for vouchers in 2000 or 2001. Therefore, the program effects (for each of the years 2000, 2001, and 2002) are not likely to be tainted by this factor.20 Moreover, as we discuss shortly, the demographic compositions of the different groups of schools remained very similar across the years under consideration. We also examine whether the demographic composition of the different Milwaukee treated groups changed over the years (Table 9). No such evidence is found. Only a few of the coefficients are statistically significant, and they are always very Vouchers affect public school quality not only through direct public school response but also through changes in student composition and peer quality brought about by sorting. small in magnitude. They imply changes of less than 1 percent, more precisely, ranging between 0.22 percent and 0.65 percent. This result suggests that sorting was not an important factor. Note that we conducted the same exercise for Florida as well and found no evidence of any relative shift of the demographic composition of the F-schools compared with the D-schools or C-schools. 19 See Hsieh and Urquiola (2006) for a discussion. This does not mean that the Florida program was not credible. Ten schools received a second “F” grade in 2002, nine schools in 2003, and twenty-one in 2004; all of these students became eligible for vouchers. 20 16 Program Design, Incentives, and Response Hispanic (2) Asian (3) Somewhat treated * program * trend More treated * program * trend 0.90 (1.59) -0.25 (1.35) -1.0 (1.34) 0.22 (0.32) 0.70 (0.25) 0.08 (0.23) 0.40 (0.83) 1.06 (0.63) 1.57 (0.81) 0.16 (0.15) -0.12 (0.13) -0.39*** (0.14) 0.04 (0.37) 0.53 (0.37) 0.65* (0.37) 0.24*** (0.07) 0.29*** (0.07) -0.22*** (0.07) Observations R2 1,228 0.95 1,226 0.97 1,216 0.91 Somewhat treated * program More treated * program Less treated * program * trend Source: Author’s calculations. Notes: Huber-White standard errors are in parentheses. All regressions are weighted by the number of students tested. All columns include a time trend, a program dummy that takes a value of 1 after the program, and an interaction between program dummy and trend. ***Statistically significant at the 1 percent level. ***Statistically significant at the 5 percent level. ***Statistically significant at the 10 percent level. A Comparison of Program Effects in Florida and Milwaukee Since Florida and Milwaukee are in different regions, we argue that our comparison of the effects of the two programs is fair and reasonable. First, as mentioned earlier, apart from the crucial design differences between the two programs, the other features of the programs were very similar. In both programs, private schools could not discriminate against voucher applicants. Also, the method of funding for the two programs, the average voucher amounts, and the per-pupil revenue losses from vouchers were very similar. Second, state and local revenues constituted very similar proportions of total revenue during the relevant periods—the percentages of revenue from state and local sources were 51 percent and 41 percent, respectively, in Florida, and 55 percent and 36 percent, respectively, in Milwaukee. Third, the demographic characteristics of the more treated and control schools in Florida were very similar, both economically and statistically, to those of the more treated and control schools in Milwaukee Table 10 Comparison of Results from Florida “Threat-of-Voucher” and Milwaukee “Voucher-Shock” Programs Using Standardized Reading and Math Scores Corrected for Mean Reversion Reading Wisconsin WRCT (1) More treated * one year after program More treated * two years after program More treated * three years after program Math Florida FCAT (2) -0.06 Wisconsin ITBS (3) Reading Florida FCAT (4) 0.47*** -0.24 0.45*** 0.38* 0.50*** 0.26 0.55*** 0.35 0.80*** -0.13 0.60*** Wisconsin WRCT (5) -0.06 Math Florida FCAT (6) Wisconsin ITBS (7) Florida FCAT (8) 0.47*** -0.24 0.24*** 0.38* 0.50*** 0.26 0.34*** 0.36 0.80*** -0.13 0.39*** Source: Author’s calculations. Notes: Reading test scores are from the Wisconsin Reading Comprehensive Test (WRCT), 1989-97, and the Florida Comprehensive Assessment Test (FCAT) Reading, 1998-2002. Math test scores are from the Iowa Test of Basic Skills (ITBS) Math, 1986-97, and the FCAT Math, 1998-2002. All figures are respective sample standard deviations. All figures are obtained from regressions that contain school-fixed effects, year dummies, interactions of year dummies with the respective treatment dummies, race, sex, percentage of students eligible for free or reduced-price lunch, and real per-pupil expenditure. Standard deviation of FCAT reading scores = 20; standard deviation of FCAT math scores = 20; standard deviation of WRCT (percentage above) reading scores = 16; standard deviation of ITBS reading scores = 18.45; standard deviation of ITBS math scores = 16.71. For standard deviations corresponding to the mean reversion sample, see the notes to Table 4. ***Statistically significant at the 1 percent level. ***Statistically significant at the 5 percent level. ***Statistically significant at the 10 percent level. (Table 2). Fourth, we repeat our analysis by comparing the improvement in Milwaukee with that of a large urban district in Florida: Miami-Dade County (the state’s largest school district). The results are very similar and hence are not reported here. Finally, and perhaps most importantly, since we follow a difference-in-differences strategy in trends, any level or even trend differences between the two regions (that are common to schools in that region) are differenced out. It is unlikely that any remaining difference, which differentially affects the trends in the two regions only in the post-program period, will be large. Table 10 compares the effects of the Florida and Milwaukee programs on their respective more treated schools both before and after correcting for mean reversion. Figures are based on data in Tables 4 and 6, and all numbers are expressed in terms of their respective sample standard deviations. Columns 1-4 present results before correcting for mean reversion; columns 5-8 present results corrected for mean reversion. Precorrection results show positive and significant effect sizes in each of the years and subject areas in Florida, which always exceed the corresponding Milwaukee effect sizes (which are not significant, except in second-year reading). Mean-reversioncorrected effect sizes are obtained by subtracting the effect size attributed to mean reversion (obtained from expressing the relevant coefficients in Table 7, panel B, in terms of respective standard deviations) from the F-school effect sizes (obtained from expressing the more treated coefficients in Table 4, columns 4-6, in terms of respective sample standard deviations) in each of the three years after the program. The estimates in reading are the same as those described earlier. In math, although the effect sizes fall in Florida, they are still positive and considerably larger than those in Milwaukee. In reading (math), relative to the control schools, the F-schools show an improvement of 0.47 (0.24) standard deviations in the first year after the program, 0.5 (0.34) standard deviations after the second year, and 0.8 (0.39) standard deviations after the third year. Mean-reversion-corrected effect sizes in writing are 0.29, 0.25, and 0.29 in the first, second, and third years, respectively, after the program. Note that since none of the F-schools received an “F” grade in either 2000 or 2001, the mean-reversion-corrected effect sizes attributed to the Florida program in the second and third years may be underestimates. FRBNY Economic Policy Review / October 2010 17 8. Lessons for New York City Our analysis of school voucher programs implies that policies that threaten underperforming public schools (or other agents) with functional and credible sanctions can induce them to respond in a way intended or desired by the policymaker. This finding has important implications for some educational policies in New York City. These include New York City’s own accountability policy, also known as the “Progress Report” policy, and the federal No Child Left Behind law, as implemented by New York State. The Progress Report policy was introduced in New York City in 2007. It rates schools on a scale of A to F, with grades based on three components: school environment, student performance, and student progress. A school’s environment As in Florida’s voucher program, public schools in New York face valid sanctions if they fail to perform. Therefore, incentives faced by New York’s low-performing schools are similar to those faced by the F-schools in Florida, and one would expect a similar response from them. score is based on attendance rates and responses from surveys given to teachers, parents, and students. The other two scores are based on student performance in state math and English Language Arts (ELA) examinations. While student performance measures rely on level scores, student progress measures rely on growth or changes in student scores over years. The program attaches consequences to the letter grades. Higher grade (A) schools are eligible for increases in per-pupil funding and bonuses for principals. Schools receiving “F” or “D” grades are required to implement “school improvement measures and target setting.” Low-performing (F- and Dschools) are also threatened with changes in their principal, and possible restructuring and closure if they continue to receive poor grades. The program also makes students in F-schools eligible to transfer to better performing schools. Although the Progress Report program does not have a voucher element, it is in many ways similar to the Florida voucher program; indeed, its design was based on the Florida program. Like the Florida program, it embeds sanctions in an accountability framework with consequences/sanctions imposed on low-performing schools if they fail to improve. Additionally, the criteria of the New York City program that 18 Program Design, Incentives, and Response make students in low-performing schools eligible to transfer to other higher performing schools are similar to those of Florida’s program. The only distinction is that in New York, students can transfer to public schools only—not to private schools, as in the Florida program. The threat of removal of the principal and the possibility of restructuring are sanctions imposed over and above the transfer option. These sanctions are credible and pose a valid threat to administrators. For example, as reported in Rockoff (2008), “Seven schools receiving an F and two schools receiving a D were told in December of 2007 that they would be closed immediately or phased out after the school year 200708. . . . Additionally, 17 percent of the remaining F-school principals (and 12 percent of the D-school principals) did not return in the school year 2008-09, relative to 9 percent of principals receiving a C, B, or A grade.” Thus, as in Florida’s voucher program, public schools in New York face valid sanctions if they fail to perform. Therefore, incentives faced by New York’s low-performing schools are similar to those faced by the F-schools in Florida, and one would expect a similar response from them. Accordingly, the above analysis would indicate that low-performing schools under the Progress Report program would have an incentive to improve. In fact, there is some evidence in favor of such improvement. In 2009, 82 percent of students passed in math and 69 percent in English, up from 42 percent and 38 percent, respectively, in 2002. Earlier, all five boroughs of New York City ranked toward the bottom in the state; now Queens and Staten Island rank toward the top in elementary-school math scores. The racial achievement gap in passing rates has been closed by half in some tests. (Statistics are from Elissa Gootman and Robert Gebeloff, New York Times, August 4, 2009.) Gootman and Gebeloff also report: At Public School 398 in Brownsville, Brooklyn, 77 percent of students passed the math tests this year and 60 percent passed English, up from 56 and 43 percent last year. Gene McCarthy, a fifthgrade teacher, attributed the improvement to “a tremendous amount of test prep,” but said that with a little creativity on his part, “ultimately I think it’s learning.” The principal, Diane Danay-Caban, said at P.S. 398, which had struggled for years with low scores and discipline problems, she has come to feel that the push to raise scores has brought genuine gains in knowledge. Rockoff and Turner (2008) find that schools labeled “F” improved their performance in both ELA and math, with larger effects in math. Winters (2008), analyzing the same program, finds improvement of F-schools in math, although he finds no such effect in ELA. NCLB, a major reform of the Elementary and Secondary Education Act, was signed into law on January 8, 2002. The states, including New York, implemented it soon thereafter. In compliance with the law, New York established Adequate Yearly Progress (AYP) targets, and all schools were graded on the basis of the targets. AYP is determined based on each school’s progress toward meeting the state proficiency level for all students in English language arts, mathematics, science, as well as the high-school graduation rate. Schools are held accountable for the achievement of students of different races and ethnic groups, students with disabilities, students with limited English proficiency, and students of low-income families. Schools must also have an average over two years of 95 percent of their students participating in state tests. If a school does not meet requirements in any one of these categories, it is said to miss AYP. Schools that receive Title I money are subject to NCLB sanctions if they miss AYP in two consecutive years. A school missing AYP for two consecutive years is required to provide public school choice to students. A school missing AYP for three consecutive years is required to provide supplemental educational services (such as tutoring) in addition to the above sanctions. Missing AYP for four consecutive years leads to corrective action in addition to Only a fraction of eligible students took advantage of the transfer option in New York as well as in the nation as a whole. This result is attributable mainly to two factors: the absence of an adequate number of spaces in nearby schools and the lack of adequate information. the above sanctions; for five consecutive years, it results in restructuring in addition to the above sanctions. Thus, sanctions start with two years of missed AYP and escalate from there. While NCLB does not have any voucher component, the accountability-sanctions component is similar in spirit to that of Florida’s voucher program. In fact, the design of NCLB was based on that program. As in the Florida program, NCLB first threatens failing schools with sanctions, and sanctions are introduced only if the schools fail to meet the predesignated targets in the following years.21 Therefore, one would expect similar incentives to be created by NCLB and threatened 21 Note, though, that while under NCLB all low-performing schools face stigma (embarrassment) due to public reporting of scores and grades, only Title I schools (schools that receive Title I money) are subject to sanctions. schools to respond in a way similar to the F-schools under the Florida program. In other words, one would expect schools threatened by the NCLB sanctions to improve their performance in an effort to make AYP. However, it should be emphasized that these incentives and responses would be The challenge to policymakers in [accountability] programs is to establish— and enforce—credible sanctions that function as valid threats to the agents (here, public schools). applicable only if the sanctions are credible and pose a valid threat to the affected schools. Under NCLB, though, implementation of the sanctions has been largely limited. For example, only a fraction of eligible students took advantage of the transfer option in New York as well as in the nation as a whole. This result is attributable mainly to two factors: the absence of an adequate number of spaces in nearby schools and the lack of adequate information. For example, as reported in the New York Daily News, “Some parents of kids in failing schools told the Daily News they weren’t even aware they could transfer out, and some were turned away from better schools that are already overcrowded” (February 3, 2008). In summary, both New York City’s Progress Report program and NCLB have the potential to induce improvement from threatened schools, but the incentives and response ultimately depend on how functional and credible the threats under consideration are. The challenge to policymakers in such programs is to establish—and enforce—credible sanctions that function as valid threats to the agents (here, public schools). Only in such cases would the agents have an incentive to respond in the direction intended or deemed appropriate by the policymakers. 9. Conclusion This article examines the role of program design in the context of two educational interventions in the United States—the Florida and Milwaukee school voucher programs. Even though both programs involve vouchers, their designs are quite different: the Milwaukee program makes low-income Milwaukee public school students eligible for vouchers, while the Florida system ties vouchers to low school performance. Specifically, Florida students become eligible for vouchers if FRBNY Economic Policy Review / October 2010 19 and only if their school receives two “F” grades in a period of four years. This study shows that program design matters; indeed, the design differences have had very different incentive and performance effects on schools subject to the two programs. Specifically, the Florida program led to considerably larger improvements from the threatened schools compared with corresponding schools under the Milwaukee program. These findings are robust to several sensitivity checks. The lessons drawn from our analysis are applicable to some of New York City’s educational policies. These policies include the No Child Left Behind Act, as implemented by the state, and New York City’s “Progress Report” policy. While 20 Program Design, Incentives, and Response neither of these programs has voucher components, both are accountability programs that have consequences for schools that fail to perform. In that sense, one would expect the incentives and responses generated by these programs to be similar to those created by the Florida program. Hence, the threatened schools could be expected to improve in an effort to avoid the sanctions. In fact, there is some evidence of such improvement in the affected schools, especially in schools treated by New York City’s Progress Report program. However, the extent of the responses and the performance effects ultimately depends on the credibility of the sanctions and the validity of the threat posed to the affected schools. References Chakrabarti, R. 2007. “Vouchers, Public School Response, and the Role of Incentives: Evidence from Florida.” Federal Reserve Bank of New York Staff Reports, no. 306, October. ———. 2008a. “Can Increasing Private School Participation and Monetary Loss in a Voucher Program Affect Public School Performance? Evidence from Milwaukee.” Journal of Public Economics 92, no. 5-6 (June): 1371-93. ———. 2008b. “Impact of Voucher Design on Public School Performance: Evidence from Florida and Milwaukee Voucher Programs.” Federal Reserve Bank of New York Staff Reports, no. 315, January. ———. 2009. “Do Vouchers Lead to Sorting under Random PrivateSchool Selection? Evidence from the Milwaukee Voucher Program.” Federal Reserve Bank of New York Staff Reports, no. 379, July. Epple, D., and R. E. Romano. 1998. “Competition between Private and Public Schools, Vouchers, and Peer Group Effects.” American Economic Review 88, no. 1 (March): 33-62. ———. 2002. “Educational Vouchers and Cream Skimming.” International Economics Review 49, no. 4 (November): 1395-1435. Figlio, D. N., and C. E. Rouse. 2006. “Do Accountability and Voucher Threats Improve Low-Performing Schools?” Journal of Public Economics 90, no. 1-2 (January): 239-55. Greene, J. P. 2001. “An Evaluation of the Florida A-Plus Accountability and School Choice Program.” Manhattan Institute for Policy Research Civic Report (with Florida State University and the Program on Education Policy and Governance at Harvard University), February. ———. 2003b. “School Choice and School Productivity: Could School Choice Be a Tide that Lifts All Boats?” In C. M. Hoxby, ed., The Economics of School Choice, 287-342. National Bureau of Economic Research Conference Report. Chicago: University of Chicago Press. Hsieh, C., and M. Urquiola. 2006. “The Effects of Generalized School Choice on Achievement and Stratification: Evidence from Chile’s Voucher Program.” Journal of Public Economics 90, no. 8-9 (September): 1477-1503. Innerst, C. 2000. “Competing to Win: How Florida’s A+ Plan Has Triggered Public School Reform.” Available at http:// www.edreform.com/published_pdf/Competing_To_Win_ How_Floridas_A_Plus_Plan_Has_Triggered_Public_ School_Reform.pdf. Manski, C. F. 1992. “Educational Choice (Vouchers) and Social Mobility.” Economics of Education Review 11, no. 4 (December): 351-69. McMillan, R. 2004. “Competition, Incentives, and Public School Productivity.” Journal of Public Economics 88, no. 9-10 (August): 1871-92. Nechyba, T. J. 1996. “Public School Finance in a General Equilibrium Tiebout World: Equalization Programs, Peer Effects, and Private School Vouchers.” National Bureau of Economic Research Working Paper no. 5642, June. ———. 1999. “School-Finance-Induced Migration and Stratification Patterns: The Impact of Private School Vouchers.” Journal of Public Economic Theory 1, no. 1 (January): 5-50. ———. 2000. “Mobility, Targeting, and Private-School Vouchers.” American Economic Review 90, no. 1 (March): 130-46. Greene, J. P., and M. A. Winters. 2003. “When Schools Compete: The Effects of Vouchers on Florida Public School Achievement.” Manhattan Institute for Policy Research Education Working Paper no. 2, August. ———. 2003. “Introducing School Choice into Multidistrict Public School Systems.” In C. M. Hoxby, ed., The Economics of School Choice, 145-94. National Bureau of Economic Research Conference Report. Chicago: University of Chicago Press. Hoxby, C. M. 2003a. “School Choice and School Competition: Evidence from the United States.” Swedish Economic Policy Review 10, no. 2: 9-65. Rockoff, J. E., and L. J. Turner. 2008. “Short-Run Impacts of Accountability on School Quality.” National Bureau of Economic Research Working Paper no. 14564, December. FRBNY Economic Policy Review / October 2010 21 References (Continued) West, M. R., and P. E. Peterson. 2005. “The Efficacy of Choice Threats within School Accountability Systems: Results from Legislatively Induced Experiments.” Harvard University Program on Education Policy and Governance, PEPG no. 05-01. Witte, J. F. 2000. The Market Approach to Education: An Analysis of America’s First Voucher Program. Princeton, N.J.: Princeton University Press. Winters, M. A. 2008. “Grading New York: An Evaluation of New York City’s Progress Report Program.” Manhattan Institute for Policy Research Civic Report no. 55, November. The views expressed are those of the author and do not necessarily reflect the position of the Federal Reserve Bank of New York or the Federal Reserve System. The Federal Reserve Bank of New York provides no warranty, express or implied, as to the accuracy, timeliness, completeness, merchantability, or fitness for any particular purpose of any information contained in documents produced and provided by the Federal Reserve Bank of New York in any form or manner whatsoever. 22 Program Design, Incentives, and Response Argia M. Sbordone, Andrea Tambalotti, Krishna Rao, and Kieran Walsh Policy Analysis Using DSGE Models: An Introduction • Dynamic stochastic general equilibrium models are playing an important role in the formulation and communication of monetary policy at many of the world’s central banks. • These models, which emphasize the dependence of current choices on expected future outcomes, have moved from academic circles to the policymaking community—but they are not well known to the general public. • This study adds to the understanding of the DSGE framework by using a small-scale model to show how to address specific monetary policy questions; the authors focus on the causes of the sudden pickup in inflation in the first half of 2004. • An important lesson derived from the exercise is that the management of expectations can be a more effective tool for stabilizing inflation than actual movements in the policy rate; this result is consistent with the increasing focus on central bankers’ pronouncements of their future actions. Argia M. Sbordone is an assistant vice president and Andrea Tambalotti a senior economist at the Federal Reserve Bank of New York; Krishna Rao is a graduate student at Stanford University; Kieran Walsh is a graduate student at Yale University. Correspondence: argia.sbordone@ny.frb.org, andrea.tambalotti@ny.frb.org 1. Introduction I n recent years, there has been a significant evolution in the formulation and communication of monetary policy at a number of central banks around the world. Many of these banks now present their economic outlook and policy strategies to the public in a more formal way, a process accompanied by the introduction of modern analytical tools and advanced econometric methods in forecasting and policy simulations. Official publications by central banks that formally adopt a monetary policy strategy of inflation targeting—such as the Inflation Report issued by the Bank of England and the monetary policy reports issued by the Riksbank and Norges Bank—have progressively introduced into the policy process the language and methodologies developed in the modern dynamic macroeconomic literature.1 The development of medium-scale DSGE (dynamic stochastic general equilibrium) models has played a key role in this process.2 These models are built on microeconomic foundations and emphasize agents’ intertemporal choice. The dependence of current choices on future uncertain 1 The Bank of England has published a quarterly Inflation Report since 1993. The report sets out the detailed economic analysis and inflation projections on which the Bank’s Monetary Policy Committee bases its interest rate decisions. The Riksbank and Norges Bank each publish monetary policy reports three times a year. These reports contain forecasts for the economy and an assessment of the interest rate outlook for the medium term. 2 A simple exposition of this class of models can be found in Galí and Gertler (2007). Woodford (2003) provides an exhaustive textbook treatment. The authors thank Ariel Zetlin-Jones for his contribution to the early stages of this work. The views expressed are those of the authors and do not necessarily reflect the position of the Federal Reserve Bank of New York or the Federal Reserve System. FRBNY Economic Policy Review / October 2010 23 outcomes makes the models dynamic and assigns a central role to agents’ expectations in the determination of current macroeconomic outcomes. In addition, the models’ general equilibrium nature captures the interaction between policy actions and agents’ behavior. Furthermore, a more detailed specification of the stochastic shocks that give rise to economic fluctuations allows one to trace more clearly the shocks’ transmission to the economy. The use of DSGE models as a potential tool for policy analysis has contributed to their diffusion from academic to policymaking circles. However, the models remain less wellknown to the general public. To broaden the understanding of these models, this article offers a simple illustration of how an estimated model in this class can be used to answer specific monetary policy questions. To that end, we introduce the structure of DSGE models by presenting a simple model, meant to flesh out their distinctive features. Before proceeding to a formal description of the optimization problems solved by firms and consumers, we use a simple diagram to illustrate the interactions among the main agents in the economy. With the theoretical structure in place, we discuss the features of the estimated model and the extent to which it approximates the volatility and comovement of economic time series. We also discuss important outcomes of the estimation—namely, the This article offers a simple illustration of how an estimated [DSGE] model . . . can be used to answer specific monetary policy questions. possibility of recovering the structural shocks that drive economic fluctuations as well as the historical behavior of variables that are relevant for policy but are not directly observable. We conclude by applying the DSGE tool to study the role of monetary policy in a recent episode of an increase in inflation. The lesson we emphasize is that, while they are a very stylized representation of the real economy, DSGE models provide a disciplined way of thinking about the economic outlook and its interaction with policy.3 We work with a small model in order to make the transmission mechanism of monetary policy, whose basic contours our model shares with most DSGE specifications, as transparent as possible. Therefore, the model focuses on the behavior of only three major macroeconomic variables: inflation, GDP growth, and the short-term interest rate. 3 Adolfson et al. (2007) offer a more extended illustration of how DSGE models can be used to address questions that policymakers confront in practice. Erceg, Guerrieri, and Gust (2006) illustrate policy simulations with an open-economy DSGE model. 24 Policy Analysis Using DSGE Models: An Introduction However, the basic framework that we present could easily be enriched to provide more details on the structure of the economy. In fact, a key advantage of DSGE models is that they share core assumptions on the behavior of households and firms, which makes them easily scalable to include details that are relevant to address the question at hand. Indeed, several A key advantage of DSGE models is that they share core assumptions on the behavior of households and firms, which makes them easily scalable to include details that are relevant to address the question at hand. extensions of the basic framework presented here have been developed in the literature, including the introduction of wage stickiness and frictions in the capital accumulation process (see the popular model of Smets and Wouters [2007]) and a treatment of wage bargaining and labor market search (Gertler, Sala, and Trigari 2008).4 Recently, the 2008 financial crisis has highlighted one key area where DSGE models must develop: the inclusion of a more sophisticated financial intermediation sector. There is a large body of work under way to model financial frictions within the baseline DSGE framework—work that is very promising for the study of financial intermediation as a source and conduit of shocks as well as for its implications for monetary policy. However, this last generation of models has not yet been subjected to extensive empirical analysis. Our study is organized as follows. Section 2 describes the general structure of our model while Section 3 illustrates its construction from microeconomic foundations. Section 4 briefly describes our approach to estimation and presents some of the model’s empirical properties. In Section 5, we use the model to analyze the inflationary episode of the first half of 2004. Section 6 concludes. 2. DSGE Models and Their Basic Structure Dynamic stochastic general equilibrium models used for policy analysis share a fairly simple structure, built around three interrelated blocks: a demand block, a supply block, and a 4 Some of these larger DSGE models inform policy analysis at central banks around the world: Smets and Wouters (2007) of the European Central Bank; Edge, Kiley, and Laforte (2008) of the Federal Reserve System; and Adolfson et al. (2008) of the Riksbank. The Basic Structure of DSGE Models Mark-up shocks Demand shocks Y = f YY(Y e, i - π ee,...) Demand Productivity shocks π = f π(π e,Y,...) Supply Y e,π e Expectations i = f i(π − π *,Y,...) Monetary policy Policy shocks monetary policy equation. Formally, the equations that define these blocks derive from microfoundations: explicit assumptions about the behavior of the main economic actors in the economy—households, firms, and the government. These agents interact in markets that clear every period, which leads to the “general equilibrium” feature of the models. Section 3 presents the microfoundations of a simple DSGE model and derives the equations that define its equilibrium. But first, we begin by introducing the basic components common to most DSGE models with the aid of a diagram. In the diagram, the three interrelated blocks are depicted as rectangles. The demand block determines real activity Y as a function of the ex ante real interest rate—the nominal rate e minus expected inflation i – —and of expectations about e future real activity Y . This block captures the idea that, when real interest rates are temporarily high, people and firms would rather save than consume or invest. At the same time, people are willing to spend more when future prospects are promising e ( Y is high), regardless of the level of interest rates. The line connecting the demand block to the supply block shows that the level of activity Y emerging from the demand block is a key input in the determination of inflation , e together with expectations of future inflation . In prosperous times, when the level of activity is high, firms must increase wages to induce employees to work longer hours. Higher wages increase marginal costs, putting pressure on prices and generating inflation. Moreover, the higher inflation is expected to be in the future, the higher is this increase in prices, thus contributing to a rise in inflation today. The determination of output and inflation from the demand and supply blocks feeds into the monetary policy block, as indicated by the dashed lines. The equation in that block describes how the central bank sets the nominal interest rate, usually as a function of inflation and real activity. This reflects the tendency of central banks to raise the short-term interest rate when the economy is overheating as well as when inflation rises and to lower it in the presence of economic slack. By adjusting the nominal interest rate, monetary policy in turn affects real activity and through it inflation, as represented by the line flowing from the monetary policy block to the demand block and then to the supply block. The policy rule therefore closes the circle, giving us a complete model of the relationship between three key endogenous variables: output Y , inflation , and the nominal interest rate i . While this brief description appears static, one of the fundamental features of DSGE models is the dynamic interaction between the blocks—hence, the “dynamic” aspect of the DSGE label—in the sense that expectations about the future are a crucial determinant of today’s outcomes. These expectations are pinned down by the same mechanism that generates outcomes today. Therefore, output and inflation One of the fundamental features of DSGE models is the dynamic interaction between the [three interrelated] blocks— hence, the “dynamic” aspect of the DSGE label—in the sense that expectations about the future are a crucial determinant of today’s outcomes. tomorrow, and thus their expectations as of today, depend on monetary policy tomorrow in the same way as they do today— of course, taking into account what will happen from then on into the infinite future. The diagram highlights the role of expectations and the dynamic connections between the blocks that they create. The influence of expectations on the economy is represented by the arrows, which flow from monetary policy to the demand and then the supply block, where output and inflation are determined. This is to emphasize that the conduct of monetary policy has a large influence on the formation of expectations. In fact, in DSGE models, expectations are the main channel through which policy affects the economy, a feature that is consistent with the close attention paid by financial markets and the public to the pronouncements of central banks on their likely course of action. The last component of DSGE models captured in the diagram is their stochastic nature. Every period, random exogenous events perturb the equilibrium conditions in each block, injecting uncertainty in the evolution of the economy FRBNY Economic Policy Review / October 2010 25 and thus generating economic fluctuations. Without these shocks, the economy would evolve along a perfectly predictable path, with neither booms nor recessions. We represent these shocks as triangles, with arrows pointing toward the equilibrium conditions on which they directly impinge. Markup and productivity shocks, for example, affect the pricing and production decisions of firms that underlie the supply block, while demand shocks capture changes in the willingness of households to purchase the goods produced by those firms. 3. Microfoundations of a Simple DSGE Model We present the microfoundations of a small DSGE model that is simple enough to fit closely into the stylized structure outlined in our diagram. Our objective is to describe the basic components of DSGE models from a more formal perspective, using the mathematical language of economists, but avoiding unnecessary technical details. Despite its simplicity, our model is rich enough to provide a satisfactory empirical account of the evolution of output, inflation, and the nominal interest rate in the United States in the last twenty years, as we discuss in the next section. Given the constraints we impose on this treatment for the sake of simplicity, our model lacks many features that are standard in the DSGE models that central banks typically use. For example, we ignore the process of capital accumulation, Despite its simplicity, our model is rich enough to provide a satisfactory empirical account of the evolution of output, inflation, and the nominal interest rate in the United States in the last twenty years. which would add another dimension—investment decisions by firms—to the economy’s demand block. Nor do we attempt to model the labor market in detail: for example, we make no distinction between the number of hours worked by each employee and the number of people at work, an issue that is hard to overlook in a period with unemployment close to 10 percent. Finally, we exclude any impediment to the smooth functioning of financial markets and assume that the central bank can perfectly control the short-term interest rate—the only relevant rate of return in the economy. The 2008 financial crisis has proved that this set of assumptions can fail miserably 26 Policy Analysis Using DSGE Models: An Introduction in some circumstances and has highlighted the need for a more nuanced view of financial markets within the current generation of DSGE models, as we observe in our introduction. 3.1. The Model Economy Our model economy is populated by four classes of agents: a representative household, a representative final-goodproducing firm (f-firm), a continuum of intermediate firms (i-firms) indexed by i 0 1, and a monetary authority. The household consumes the final good and works for the i-firms. Each of these firms is a monopolist in the production of a particular intermediate good i , for which it is thus able to Our model economy is populated by four classes of agents: a representative household, a representative final-goodproducing firm, . . . a continuum of intermediate firms, . . . and a monetary authority. set the price. The f-firm packages the differentiated goods produced by the i-firms and sells the product to households in a competitive market. The monetary authority sets the nominal interest rate. The remainder of this section describes the problem faced by each economic agent, shows the corresponding optimization conditions, and interprets the shocks that perturb these conditions. These optimization conditions result in dynamic relationships among macroeconomic variables that define the three model blocks described above. Together with market clearing conditions, these relationships completely characterize the equilibrium behavior of the model economy. 3.2. Households and the Aggregate Demand Block At the core of the demand side of virtually all DSGE models is a negative relationship between the real interest rate and desired spending. In our simple model, the only source of spending is consumption. Therefore, the negative relationship between the interest rate and demand emerges from the consumption decision of households. We model this decision as stemming from the optimal choice of a very large representative household—the entire U.S. population—which maximizes its expected discounted lifetime utility, looking forward from an arbitrary date t 0 s Et bt + s log Ct + s Max 0 0 0 s = 0 s = 0 Bt + s Ct + s Ht + s i To find the solution to the optimal problem above, we form the Lagrangian s bt 0 + s log Ct + s – Ct + s – 1 L = Et 0 0 0 s=0 0 0 0 – Ct i 0 1 1 0+s–1 0 – v Ht 0+s B Pt C t + -----t B t – 1 + Rt –t i di subject to the sequence of budget constraints – Bt 0+s Pt 0+s–1 0+s – Ct 1 0 Wt 0+s 0+s i di + Bt 0 +s –1 0+s Rt , i Ht + s i di 0 +s 0 with first-order conditions 1 0 wt i Ht i di , for t = t 0 t 0 + 1 ,, and given Bt – 1. The members of this 0 household, we call them “Americans,” like consumption, Ct , but dislike the number of hours they spend at work, Ht , to an extent described by the convex function . The utility flow from consumption depends on current as well as past consumption, with a coefficient . As a result of this “habit,” consumers are unhappy if their current consumption is low, but also if it falls much below the level of their consumption in the recent past. To afford consumption, Americans work a certain amount of hours Ht i in each of the i-firms, where they earn an hourly nominal wage Wt i which they take as given when deciding how much to work.5 With the income thus earned, they can purchase the final good at price Pt or save by accumulating one-period discount government bonds Bt , whose gross rate of return between t and t + 1 is Rt . From the perspective of time t , the household discounts utility in period t + 1 by a time-varying factor bt + 1 bt , where bt + 1 bt is an exogenous stochastic process. Changes in bt + 1 bt represent shocks to the household’s impatience. When bt + 1 increases relative to bt , for example, the household cares more about the future and thus wishes to save more and consume less today, everything else equal. In this respect, bt + 1 bt acts as a traditional demand shock, which affects desired consumption and saving exogenously. A persistent increase in bt + 1 bt is one way of interpreting the current macroeconomic situation in the United States, in which households have curtailed their consumption—partly to build their savings. Of course, in reality there are many complex reasons behind this observed change in behavior, and an increase in people’s concern about the future is surely one of them. For simplicity, the model focuses on this one reason exclusively. 5 1 0 – v Ht In equilibrium, the wage—and thus the number of hours worked—will settle at the level at which the supply of labor by the household equals the demand of labor by firms. This labor demand in turn is determined by the need of firms to hire enough workers to satisfy the demand for their products, as we describe in the next section. (3.1a) (3.1b) L------: t = Et t + 1 R t Bt bt + 1 bt L 1 -------- : -----t P = -------------------------- . – Et -------------------------C Ct bt t Ct – Ct – 1 t + 1 – Ct for t = t 0 t 0 + 1 , and (3.2) Ht i L --------------- : --------------------- = Wt i Ht i t bt for t = t 0 t 0 + 1 , and i 0 1 , together with the sequence of budget constraints. These conditions yield a fully state-contingent plan for the household’s choice variables—how much to work, consume, and save in the form of bonds—looking forward from the planning date t 0 and into the foreseeable future. At any point in time, the household is obviously uncertain about the way in which this future will unfold. However, we assume that the household is aware of the kind of random external events, or shocks, that might affect its decisions and, crucially, that it knows the probability with which these shocks might occur. Therefore, the household can form expectations about future outcomes, which are one of the inputs in its current choices. We assume that these expectations are rational, meaning that they are based on the same knowledge of the economy and of the shocks that buffet it as that of the economist constructing the model. We use the notation Et X t + s to denote expectations formed at time t of any future variable X t + s , as in equations 3.1, for example. The optimal plan, then, is a series of instructions on how to behave in response to the realization of each shock, given expectations about the future, rather than a one-time decision on exactly how much to work, consume, and save on each future date. Together, the optimality conditions in equations 3.1 establish the negative relationship between the interest rate and desired consumption that defines the demand side of the model. The nature of this relationship is more transparent in the special case of no habit in consumption ( = 0 ), when we FRBNY Economic Policy Review / October 2010 27 can combine the two equations to obtain (3.3) Rt bt + 1 1 1 - . - ----------- ----------------------- = E t ------------bt C t + 1 Pt + 1 Pt Ct According to this so-called Euler equation, desired consumption decreases when the (gross) real interest rate Rt ------------------- increases, when expected future consumption Pt + 1 Pt decreases, and when households become more patient ( bt + 1 rises). A log-linear approximation of the Euler equation (3.3), after some manipulation, gives (3.4) yt = Et yt + 1 – i t – Et t + 1 – t , where t log Pt Pt – 1 is the quarterly inflation rate, i t log Rt is the continuously compounded nominal interest rate, t Et log bt + 1 bt is a transformation of the demand shock, and y t log Yt is the logarithm of total output. In this expression, we can substitute consumption of the final good Ct with its output Yt because in our model consumption is the only source of demand for the final good. Therefore, market clearing implies Yt = Ct . In this framework, equation 3.4 is similar to a traditional IS equation, since it describes the relationship between aggregate activity y t and the ex ante real interest rate i t – Et t + 1 , which must hold for the final-good market to clear. Unlike a traditional IS relationship, though, this equation is dynamic and forward looking, as it involves current and future expected variables. In particular, it establishes a link between current output and the entire future expected path of real interest rates, as we see by solving the equation forward (3.5) yt = – Et it + s – t + s + 1 – t + s . s=0 Through this channel, expectations of future monetary policy directly affect current economic conditions. In fact, this equation shows that future interest rates are just as important to determine today’s output as the current level of the shortterm rate, as we describe in our discussion of the role of policy expectations. In our full model, the Euler equation is somewhat more complicated than in equation 3.4 because of the consumption habit ( 0), which is a source of richer, and more realistic, output dynamics in response to changes in the interest rate. Nevertheless, these more intricate dynamics do not change the qualitative nature of the relationship between real rates and demand. The third first-order condition of the household optimization problem, equation 3.2, represents the labor supply decision. It says that Americans are willing to work more hours in firms that pay a higher wage and at times when 28 Policy Analysis Using DSGE Models: An Introduction wages are higher, at least for differences in wages modest enough to have no significant effect on their income.6 Large wage changes, in fact, would trigger an income effect and lead the now richer workers to curtail their labor supply. Mathematically, workers with higher income could afford more consumption, which would lead to a drop in the marginal utility t and thus to a decrease in labor supply at any given wage level. We can think of the labor supply schedule (equation 3.2) as a relationship determining the wage that firms must pay to induce Americans to work a certain number of hours. In prosperous times, when demand is high and firms are producing much, firms require their labor force to work long hours and they must correspondingly pay higher hourly wages. This is an important consideration in the production and pricing decisions of firms, as we discuss in the next section. 3.3. Firms and the Aggregate Supply Block The supply block of a DSGE model describes how firms set their prices as a function of the level of demand they face. Recall that in prosperous times, demand is high and firms must pay their workers higher wages. As a result, their costs increase as do their prices. In the aggregate, this generates a positive relationship between inflation and real activity. In terms of microfoundations, establishing this relationship requires some work, since firms must have some monopoly power to set prices. This is why our production structure includes a set of monopolistic i-firms, which set prices, as well as an f-firm, which simply aggregates the output of the i-firm into the final consumption good. Because all the pricing action occurs within the i-firms, we focus on their problem and omit that of the f-firm. Intermediate firm i hires Ht i units of labor of type i on a competitive market to produce Yt i units of intermediate good i with the technology (3.6) Yt i = At Ht i , where At represents the overall efficiency of the production process. We assume that At follows an exogenous stochastic process, whose random fluctuations over time capture the unexpected changes in productivity often experienced by modern economies—for example, the productivity boom in the mid-1990s that followed the mass adoption of information technology. We call this process an aggregate productivity shock, since it is common to all firms. Labor supply is upward sloping because is an increasing function, as is convex. In other words, people dislike working an extra hour more intensely when they are already working a lot rather than when they are working little. 6 The market for intermediate goods is monopolistically competitive, as in Dixit and Stiglitz (1977), so firms set prices subject to the requirement that they satisfy the demand for their good. This demand comes from the f-firm and takes the form Pt i –t (3.7) , Yt i = Yt ----------Pt where Pt i is the price of good i and t is the elasticity of demand. When the relative price of good i increases, its demand falls relative to aggregate demand by an amount that depends on t . Moreover, we assume that firms change their prices only infrequently. The fact that firms do not adjust prices continuously, but leave them unchanged in some cases for long periods of time, should be familiar from everyday experience and is well established in the economic literature (for example, Bils and Klenow [2004]; Nakamura and Steinsson [2008]). To model this fact, we follow Calvo (1983) and assume that in every period only a fraction 1 – of firms is free to reset its price while the remaining fraction maintains its old price.7 The subset of firms that are able to set an optimal price at t , call it t 0 1 , maximize the discounted stream of expected future profits, taking into account that s periods from now s there is a probability that they will be forced to retain the price chosen today. The objective function of each of these firms is therefore s s t + s ---------------- Pt i Yt + s i – Wt + s i Ht + s i Max E t t P Pt i s=0 for all i t , subject to the production function 3.6 and to the additional constraint that they must satisfy the demand for their product at every point in time Pt i –t + s (3.8) , Yt + s i = Yt + s ---------- Pt + s for s = 0 , 1, , . Profits, which are given by total revenue at the price chosen today, Pt i Yt + s i minus total costs s Wt + s i Ht + s i , are discounted by the multiplier t + s t , sometimes called a stochastic discount factor, which translates dollar profits in the future into a current dollar value. The first-order condition of this optimization problem is t + s – 1 Wt + s i s - = 0, t + s Yt + s Pt + s Pt i – t + s -----------------(3.9) E t At + s s=0 for all i t , where Pt i denotes the optimal price chosen by firm i , Wt + s i At + s is the firm’s nominal marginal cost at time t + s –1 t + s , and t + s = ---------------- is its desired mark-up—the mark-up t + s In our estimated model in Section 5, we actually assume that the fraction of firms that cannot choose their price freely can in fact adjust it in part to catch up with recent inflation. This assumption improves the model’s ability to fit the data on inflation, but it would complicate our presentation of the model’s microfoundations without altering its basic message. it would charge if prices were flexible. As rational monopolists, optimizing firms set their price as a mark-up over their marginal cost. However, this relationship holds in expected present discounted value, rather than every period, since a s price chosen at time t will still be in effect with probability in period t + s. We can rewrite the marginal cost of a firm that at time t + s is still forced to retain the price Pt i as (3.10) Wt + s i H t + s i 1 S t + s i ------------------ = ----------------------------- ---------- t + s bt + s At + s At + s P i –t + s Y A t + s Pt + s = --------------------------------------------------, A t + s t + s bt + s t + s t - ----------- --------- where we use the labor supply relation 3.2 to substitute for the wage as well as the production function 3.6 and the demand function 3.8 to give us an expression for the labor demand Ht + s i .8 Inserting this expression into the first-order condition 3.9, we see that the solution to the optimal pricing problem is the same for all firms in the set t , since it depends only on the aggregate variables Yt + s , At + s , Pt + s , t + s s = 0 . We denote this common optimal price as Pt . t + s –1 The equation for the desired mark-up— t + s = ---------------- , also t + s known as Lerner’s formula—says that monopolists facing a more rigid demand optimally charge a higher mark-up, and thus higher prices, since their clients are less sensitive to changes in the latter. We assume that this sensitivity—the elasticity of demand—and thus the desired mark-up, follows an exogenous stochastic process. Positive realizations of this desired mark-up shock, which correspond to a fall in the elasticity of demand, represent an increase in firms’ market power, to which they respond by increasing their price. Equation 3.9, together with the definition of the aggregate price level as a function of newly set prices Pt and of the past price index Pt – 1 1 Pt 1 – Pt 1 – t ------------1 – t 1 – t + Pt –1 yields an approximate New Keynesian Phillips curve—a relationship between current inflation, future expected inflation, and real marginal cost—of the form (3.11) t = st + Et t + 1 + ut , where ut = log ut is a transformation of the mark-up shock and st log St Pt is the logarithm of the real marginal cost.9 The sensitivity of inflation to changes in the marginal cost, , depends on the frequency of price adjustment , as well as on 7 8 These substitutions are equivalent to “solving” for equilibrium in the labor market. 9 Variables are in logarithms, because equation 3.11, like equation 3.4, is obtained by a log-linear approximation. FRBNY Economic Policy Review / October 2010 29 1 – 1 – , other structural parameters, according to ------------------------------------- 1 + H where ----------- is the elasticity of the marginal disutility of work, while is the average value of the elasticity of demand t . This Phillips curve, together with the expression for marginal costs (3.10), provides the relationship between inflation and real activity that defines the supply block of the model. In fact, we see from equation 3.10 that marginal cost depends on the level of aggregate activity, among other factors. Higher economic activity leads to higher wages and marginal costs. Thus, firms increase their prices, boosting aggregate inflation. Another important feature of the Phillips curve is that it is forward looking, just as the Euler equation in the previous section is. As in that case, therefore, we can iterate equation 3.11 forward to obtain 4Q e i + y yt – yt +t , e e where rt , t, and yt are the baselines for the real interest rate, 4Q inflation, and output, respectively, and t log Pt Pt – 4 is the rate of inflation over the previous four quarters. The i monetary policy shock t , a random variable with mean zero, captures any deviation of the observed nominal interest rate from the value suggested by the rule. This rule implies that, if inflation and output rise above their baseline levels, the nominal interest rate is lifted over time above its own baseline, e rt + t, by amounts dictated by the parameters and y and at a speed that depends on the coefficient . The higher policy rate, which is expected to persist even after output and inflation have returned to normal, exerts a restraining force on the t = E t st + s + u t + s , s When output is at its efficient level, inflation is not stable, as policymakers would like it to be, but fluctuates because of the presence of mark-up shocks. This is the essence of the monetary policy trade-offs in the economy. s=0 which highlights how inflation today really depends on the entire future expected path of marginal costs, and through those, of real activity. But this path depends in turn on expectations about interest rates, and thus on the entire future course of monetary policy, as equation 3.5 shows. Hence, we have the crucial role of policy expectations for the determination of current economic outcomes in this model, a feature we discuss in Section 2. Monetary Policy Recall that when the interest rate—current and expected—is low, people demand more consumption goods (equation 3.5). But if demand is high, firms’ marginal costs increase and so do their prices. The end result is inflation. The opposite is true when the interest rate is high. But where does the interest rate come from? In DSGE models, as in the real world, the shortterm interest rate is set by monetary policy. In practice, this is a decision made by a committee (the Federal Reserve’s Federal Open Market Committee, or FOMC) using various inputs: large data sets, projections from several models, and the judgment of policymakers. Despite the apparent complexity of the process, Taylor (1993) famously demonstrated that it can be reasonably well approximated by assuming that the Federal Reserve raises the federal funds rate when inflation and/or output is “high” with respect to some baseline. This behavior is assumed in almost all variants of DSGE models, although the definition of the appropriate baselines is somewhat controversial. In our model, we assume that interest rates are set according to the policy rule 30 e i t = i t – 1 + 1 – r t + t + t – t (3.12) Policy Analysis Using DSGE Models: An Introduction economy—curbing demand, marginal costs, and inflation. In e this respect, t and yt can be regarded as targets of monetary policy—the levels of inflation and output that the central bank considers consistent with its mandate—and therefore do not elicit either a restrictive or a stimulative policy. In equation 3.12, we denote the central bank’s objective in e terms of production as yt , the “efficient” level of output. This unobserved variable can be derived from the microfoundations of the model.10 It represents the level of output that would prevail in the economy if we could eliminate at once all distortions—namely, force the i-firms to behave competitively rather than as monopolists and allow them to change their prices freely. The level of activity that would result from such a situation is ideal from the perspective of the representative household in the model, as its name suggests. This is what makes it a suitable target for monetary policy. However, when output is at its efficient level, inflation is not stable, as policymakers would like it to be, but fluctuates because of the presence of mark-up shocks. This is the essence of the monetary policy trade-offs in the economy. Achieving the 10 The precise mathematical definition of efficient output in the model is irrelevant for our purposes. We present in Section 4 an estimate of the behavior of this variable over the last twenty years. efficient level of output requires undesirable movements in inflation. In contrast, a stable inflation implies deviations from the efficient level of output. The two objectives cannot be reconciled, but must be traded off of each other. Related to the efficient level of output is the efficient real e interest rate, rt , which is the rate of return we would observe in the efficient economy described above. This definition implies that, when the actual real interest rate is at its efficient level and is expected to remain there in the future, output will e also be at its efficient level. This is why we include rt in our definition of the baseline interest rate. The other component of this baseline rate is the inflation target t. We allow this target to vary slowly over time to accommodate the fact that inflation hovered at about 4 percent for a few years around 1990 before declining to nearly 2 percent after the recession that ended in 1991. Nominal interest rates were correspondingly higher in the first period, thus implying a stable average for the real interest rate. We now present our estimate of the evolution of the inflation target. 4. Empirical Approach and Estimation Results We estimate our model using data on the growth rate of real GDP to measure output growth, yt , the growth rate of the personal consumption expenditures chain price index excluding food and energy (core PCE) to measure inflation, t , and the quarterly average of the monthly effective federal funds rate to measure the nominal interest rate, i t . We measure inflation by core PCE, rather than by a more comprehensive measure, because the monetary policy debate in the United States tends to focus on this index. Our data span the period 1984:1 to 2007:4 (Chart 1). This is the longest possible data set over which it is reasonable to argue that U.S. monetary policy can be represented by a stable interest rate rule. It follows the period of extremely high interest rates in the early 1980s that brought inflation under control. However, in the first few years of this sample, inflation and the federal funds rate were still relatively high, with a fairly abrupt reduction taking place around the 1991 recession. We capture this low-frequency movement in inflation and the nominal interest rate by including the slow-moving inflation target t in the policy rule. We use Bayesian methods to characterize the posterior distribution of the parameters of the model. This distribution combines the model’s likelihood function with prior information on the parameters, using techniques surveyed, for Chart 1 Observable Variables Percent, annualized 12 Federal funds rate 10 8 6 4 2 0 Inflation -2 GDP growth -4 1984 86 88 90 92 94 96 98 00 02 04 06 Sources: Bureau of Economic Analysis; Board of Governors of the Federal Reserve System. example, by An and Schorfheide (2007).11 A discussion of these methods is beyond the scope of this article. Instead, we focus on the implications of these estimates for some key properties of the model. Our objective is to show that the estimated model provides a good fit to the data across many dimensions, but also to highlight some of the model’s most notable shortcomings. 4.1. Moment Comparisons We compare the second moments implied by the estimated model with those measured in the data. Table 1 presents the standard deviations of the observable variables, reported as annualized percentages. In the model column, we report the median and 5th and 95th percentiles of the standard deviations across draws from the model’s posterior. This interval reflects the uncertainty on the structural parameters—and thus on the model-implied moments—encoded in the parameters’ posterior distribution. In the data column, we report the median and 5th and 95th percentiles of the empirical standard deviations in the data. This interval represents the uncertainty on the true empirical moments because of the small sample available for their estimation. Our model does a very good job replicating the volatilities in the data. It captures the standard deviation of output growth and replicates quite closely the volatility of the nominal interest rate, although it overestimates the standard deviation of 11 The technical appendix provides information on the priors for the parameters. FRBNY Economic Policy Review / October 2010 31 Table 1 Model-Implied and Empirical Standard Deviations Percent Variable GDP growth Core PCE inflation Federal funds rate Model Data 2.03 [1.79, 2.37] 1.41 [0.98, 2.40] 2.23 [1.61, 3.56] 2.03 [1.74, 2.27] 1.15 [0.67, 1.38] 2.46 [1.55, 2.94] Source: Authors’ calculations. Notes: The table reports the standard deviations of the observable variables. The model-implied standard deviations are medians across draws from the posterior; the 5th and 95th posterior percentiles across those same draws are in brackets. The empirical standard deviations are medians across bootstrap replications of a VAR(4) fit to the data; the 5th and 95th percentiles across those same replications are in brackets. PCE is personal consumption expenditures. inflation. The ability of the model to accurately reproduce the volatility of the observable variables is not a preordained conclusion, even if we freely estimate the standard deviations of the shocks. The reason is that a likelihood-based estimator tries to match the entire autocovariance function of the data, and thus must strike a balance between matching standard deviations and all the other second moments—namely, autocorrelations and cross-correlations. These other moments are displayed in Chart 2. The black line represents the model-implied correlation, with the shaded area representing a 90 percent posterior interval. The solid blue Our model does a very good job replicating the volatilities in the data. It captures the standard deviation of output growth and replicates quite closely the volatility of the nominal interest rate, although it overestimates the standard deviation of inflation. line is instead the correlation measured in the data, with a 90 percent bootstrap interval around this estimate represented by the dashed lines. The serial correlation of output growth in the model is very close to its empirical counterpart and well within the data-uncertainty band. For inflation and the interest rate, the model serial correlations are on the high end of the 32 Policy Analysis Using DSGE Models: An Introduction band. This excessive persistence is a result of the low-frequency component in both variables associated with the inflation target, as we can infer from the variance decomposition in Table 2. According to the model, shocks to the inflation target account for 85 percent of the unconditional variance of inflation and 38 percent of that of the nominal interest rate. Although we do not calculate a variance decomposition by frequency, we know that the contribution of the inflation target shock is concentrated at low frequencies, since this shock is very persistent (the posterior mean of its autocorrelation coefficient is 0.98). This finding suggests that the model faces a trade-off between accommodating the downward drift in inflation in the first part of our sample and providing a more balanced account of the sources of inflation variability. The rest of the variance decomposition accords well with conventional wisdom. The productivity shock plays an According to the model, shocks to the inflation target account for 85 percent of the unconditional variance of inflation and 38 percent of that of the nominal interest rate. This finding suggests that the model faces a trade-off between accommodating the downward drift in inflation in the first part of our sample and providing a more balanced account of the sources of inflation variability. important role in accounting for the volatility of output growth, although the demand shock and the monetary policy shocks (interest rate plus inflation target) are also nonnegligible. Moreover, the demand shock accounts for more than half of the variance of the nominal interest rate. Finally, mark-up shocks play a minor role as sources of volatility. This finding has potentially important policy implications, since in the model mark-up shocks are the only source of the aforementioned policy trade-off between inflation and real activity. Returning now to the cross-correlations in Chart 2, we see that the model is quite successful in capturing the lead-lag relationships in the data. In our sample, there is no statistically significant predictability of future inflation through current output growth, a pattern that is reproduced by the model. The Chart 2 Correlations Corr. (Δ yt , π t+k ) Corr. ( Δyt , Δ yt-k ) 1.0 0.4 0.8 0.3 0.2 0.6 0.1 0.4 0 0.2 -0.1 0 -0.2 -0.2 -0.3 -0.4 -0.4 Corr. ( π t , π t-k ) Corr. ( π t , it+k ) 1.0 0.8 0.8 0.6 0.6 0.4 0.4 0.2 0.2 0 0 -0.2 Corr. ( it , Δ yt+k ) Corr. ( it , i t-k ) 1.0 0.6 0.8 0.4 0.6 0.2 0.4 0 0.2 -0.2 0 -0.2 0 1 2 3 4 k 5 6 7 8 -0.4 -8 -7 -6 -5 -4 -3 -2 -1 0 k 1 2 3 4 5 6 7 8 Source: Authors’ calculations. Notes: The black line represents the median model-implied correlation across draws from the posterior; the shaded area represents the interval between the 5th and 95th percentiles across those same draws. The solid blue line represents the median autocorrelation across bootstrap replications of a VAR(4) fit to the data; the dashed blue lines represent the interval between the 5th and 95th percentiles across those same replications. Each statistic is calculated at horizons k = 0, . . . 8 for autocorrelations and at horizons k = -8, . . . , 0, . . . 8 for cross-correlations. model also reproduces the positive correlation between inflation and the nominal interest rate present in the data both in the leads and in the lags (the middle right panel of the chart). The positive correlation between current interest rates and future inflation might seem puzzling at first. We would expect higher interest rates to bring inflation down over time, which should make the correlation negative. However, over our sample, this negative relationship is confounded by the lowfrequency positive comovement between inflation and the nominal interest rate induced by the Fisher effect. Recall that inflation and the nominal interest rate in fact are persistently above their unconditional sample average over the first third of the sample and are persistently below it after about 1992. The bottom right panel of Chart 2 reports the dynamic FRBNY Economic Policy Review / October 2010 33 Table 2 Variance Decomposition Percent Shocks Variable GDP growth Core PCE inflation Federal funds rate Demand Productivity Mark-Up Interest Rate Inflation Target 0.20 [.13, .28] 0.04 [.01, .06] 0.54 [.32, .76] 0.56 [.45, .67] 0.01 [.00, .02] 0.06 [.00, .13] 0.07 [.04, .10] 0.10 [.04, .17] 0.01 [.00, .02] 0.04 [.02, .06] 0.00 [.00, .01] 0.01 [.01, .02] 0.13 [.07, .19] 0.85 [.76, .94] 0.38 [.16, .61] Source: Authors’ calculations. Notes: The table reports the share of the unconditional variance of each observable variable contributed by each shock. The point estimates are medians across draws from the posterior; the 5th and 95th posterior percentiles across those same draws are in brackets. PCE is personal consumption expenditures. correlation between output growth and the nominal interest rate. In the data, high growth rates of output predict high nominal interest rates one to two years ahead, but this predictability is much less pronounced in the model. Moreover, this discrepancy is statistically significant in the sense that the model-implied median autocorrelation lies outside the 90 percent bootstrap interval computed from the data. This failure to match the data highlights the main empirical weakness of our model: its demand-side specification. As in most of the DSGE literature, our demand block consists of the Euler equation of a representative consumer. Standard specifications of a Euler equation of the type adopted here provide an inaccurate description of the observed relationship between the growth rate of consumption (or output, as in our case) and financial returns, including interest rates, as first documented by Hansen and Singleton (1982, 1983) and subsequently confirmed by many others (see Campbell [2003] for a review). Improving the performance of the current generation of DSGE models in this dimension would be an important priority for future research. We now report our estimates of a few of the variables that play an important role in the model, but that are not directly observable. We focus on the three latent variables that enter the interest rate rule: the inflation target t, the output gap e e yt – yt , and the efficient real interest rate rt (Chart 3). As in Charts 1 and 2, the black line is the median estimate across draws from the model’s posterior and the shaded area represents a 90 percent posterior probability interval. Starting from the top panel, we note that the estimated inflation target captures well the step-down in inflation from a local mean above 4 percent between 1984 and 1991 to an average value of around 2 percent since 1994. This permanent reduction in inflation represents the last stage of the 34 Policy Analysis Using DSGE Models: An Introduction disinflation process initiated by Federal Reserve Chairman Volcker in 1979, which became known as an example of opportunistic disinflation (Orphanides and Wilcox 2002). Needless to say, the estimated target is not completely smooth, but it also displays some higher frequency variation. For example, it reaches a minimum of around 1 percent at the beginning of 2003, but moves closer to 2 percent over 2004. (The next section studies in more detail the implications of these movements.) The middle and bottom panels of Chart 3 report estimates of the output gap—the percentage deviation of output from its efficient level—and of the efficient real interest rate. Several features of the estimated output gap are noteworthy. First, its two deepest negative troughs correspond to the two recessions in our sample. In this respect, our model-based output gap conforms well with more conventional measures of this variable, such as the one produced by the Congressional Budget Office (CBO). However, the shortfall of output from its efficient level is never larger than 0.7 percent, even in these recessionary episodes. By comparison, the CBO output gap is as low as -3.2 percent in 1991. The amplitude of the business cycle fluctuations in our estimated output gap is small because the efficient level of output is a function of all the shocks in the model and therefore it tracks actual output quite closely. The last notable feature of the efficient output gap is that it displays a very pronounced volatility at frequencies higher than the business cycle. During the 1990s expansion, for example, it crosses the zero line about a dozen times. Compared with the output gap, the efficient real rate is significantly smoother. Although some high-frequency variation remains, the behavior of the efficient real rate is dominated by swings at the business cycle frequency. The rate spikes and then plunges for some time before the onset of Chart 3 Kalman Smoother Estimates of Latent Variables π *t 7 6 To show how our model can be used to address specific policy questions, we examine a particular historical episode: the puzzling pickup in inflation in the first half of 2004. This exercise allows us to illustrate how we use the model’s forecasts to construct alternative scenarios for counterfactual policy analysis. Moreover, our analysis offers potentially interesting lessons for the current situation—although inflation has recently been quite low, there has been some concern that it might accelerate in the near future. After approaching levels close to 1 percent between 2002 and 2003, core PCE inflation started moving higher in mid2003. This pickup accelerated significantly in the first half 5 4 3 2 1 0 -1 0.8 0.6 5. The Model at Work: The Pickup in Inflation in the First Half of 2004 yt - yte 0.4 0.2 0 -0.2 To show how our model can be used to address specific policy questions, we examine a particular historical episode: the puzzling pickup in inflation in the first half of 2004. This exercise allows us to illustrate how we use the model’s forecasts to construct alternative scenarios for counterfactual policy analysis. -0.4 -0.6 -0.8 -1.0 e rt 10 8 6 4 2 0 -2 -4 1984 86 88 90 92 94 96 98 00 02 04 06 Source: Authors’ calculations. Notes: The black line represents the Kalman smoother estimate of the relevant latent variable conditional on the posterior mean of the parameters; the shaded area represents the interval between the 5th and 95th percentiles of the Kalman smoother estimates across draws from the posterior. The vertical bands indicate NBER recessions. recessions and recovers a few quarters into the expansions. It is interesting to note that the efficient real rate was negative for the entire period between 2001 and 2004—a time when the FOMC was concerned about the possibility that the U.S. economy would fall into a liquidity trap.12 A negative efficient real interest rate is a necessary condition for the zero bound on nominal interest rates to become binding, and hence for the liquidity trap to become a problem. 12 A liquidity trap describes a situation in which nominal interest rates have reached their zero lower bound, as in Japan in the 1990s, and therefore cannot be lowered any further. of 2004, when (year-over-year) core inflation moved from about 1.5 percent to more than 2 percent, where it remained until the end of 2008. We use our DSGE model to analyze the sources of this unusually rapid and persistent step-up in the level of inflation. We organize our discussion around three questions. First, was the surge in inflation forecastable? As we will see, the answer to this question is no, at least from the perspective of our model. Second, what accounts for the discrepancy between the model’s forecast and the observed paths of inflation, output growth, and the federal funds rate? Third, could monetary policy have achieved a smooth transition to inflation rates below 2 percent and, if so, at what cost in terms of added volatility in output and the interest rate? Chart 4 presents forecasts of quarterly core PCE inflation, real GDP growth, and the federal funds rate from the DSGE model. The forecast starts in 2003:1, when quarterly inflation reached 1.1 percent (at an annual rate)—its lowest level following the 2001 recession—and extends through the beginning of 2005. In each panel, the dashed line represents FRBNY Economic Policy Review / October 2010 35 assessed only about one in ten chances (12.5 percent) of inflation being as high as it was in that period. From an economic perspective, it is interesting to note that these sizable forecast errors for inflation roughly correspond to the “considerable period” era that extended from June 2003 to June 2004. At that time, the FOMC kept the federal funds rate constant at 1 percent to guard against the risk of deflation, while indicating in its statement that “policy accommodation can be maintained for a considerable period.”13 According to the model’s projection, this path for the federal funds rate represents a deviation from the policy stance historically maintained by the Federal Reserve in similar macroeconomic circumstances. Based on the estimated interest rate rule, in fact, Chart 4 Forecasts of Observable Variables 3.5 Core PCE Inflation 2.5 1.5 0.5 -0.5 8 Real GDP Growth 6 The DSGE forecast is just a description of what would happen to the variables of interest if we allowed the model economy to “run” from its initial condition, without introducing any innovations. Any observed deviation from the forecast, therefore, must be attributable to the realization of a particular combination of such innovations. 4 2 0 -2 5 Federal Funds Rate 4 3 2 1 0 -1 2002 2003 2004 Source: Authors’ calculations. Notes: The dashed line represents the forecast of the relevant variable conditional on the posterior mean of the parameters; the solid line represents the observed realization. The shaded areas represent 50 (light blue), 75 (medium blue), and 90 percent (dark blue) symmetric probability intervals for the forecast at each horizon. PCE is personal consumption expenditure. the expected value of the forecast, while the bands show the 50 (light blue), 75 (medium blue), and 90 (dark blue) percent probability intervals. The solid line shows the realized data. The model performs well in its forecast of output and the federal funds rate, especially in the medium term. Inflation, by comparison, is close to the mean forecast in 2003, but is well above it in 2004 and beyond. Moreover, the probability intervals for the forecast suggest that this realization of inflation was quite unusual, as we see from the fact that the solid line borders the 75 percent probability interval in the first half of 2004. This means that in 2003:1, the model would have 36 Policy Analysis Using DSGE Models: An Introduction the DSGE predicts a slow rise in the interest rate over 2003 and 2004. Instead, the FOMC maintained the federal funds rate at 1 percent through the first half of 2004. However, the pickup in inflation over this period is significantly more “unusual” than the deviation of the federal funds rate from the historical norm. Actual inflation in 2004 is mostly outside the 50 percent probability interval of the model forecast (the light blue band), while the actual federal funds rate remains well within it. Moreover, the acceleration in inflation is not accompanied by unexpectedly high real growth, suggesting that it cannot be fully explained by the traditional channel of transmission from an overheated economy to higher inflation. What else, then, accounts for the unexpected and unlikely deviation of inflation from the model’s forecast over 2004? The DSGE framework provides a particularly useful way of addressing this question. As we discuss in Section 2, the economic outcomes predicted by the model—the levels of inflation, output, and the interest rate—are the result of the endogenous responses of the agents in the economy to the 13 This formulation was maintained in the FOMC statement from August 2003 to December 2003, and was later substituted with “policy accommodation can be removed at a pace that is likely to be measured.” Chart 5 Forecasts of Shocks 2 2 Demand Productivity 1 0 0 -2 -1 -4 -2 -3 -6 1.0 2.0 Mark-Up Inflation Target 1.5 0.5 1.0 0 0.5 -0.5 0 -0.5 -1.0 2002 2003 2004 2002 2003 2004 Source: Authors’ calculations. Notes: The dashed line represents the forecast of the relevant shock conditional on the posterior mean of the parameters while the solid line represents an estimate of the realization based on the Kalman smoother. The shaded area represents the 75 percent symmetric probability interval for the forecast at each horizon. realization of a set of exogenous processes, such as productivity or desired mark-ups. The innovations to these driving processes account for the deviations of the data from the model’s forecast. In fact, the DSGE forecast is just a description of what would happen to the variables of interest if we allowed the model economy to “run” from its initial condition, without introducing any innovations. Any observed deviation from the forecast, therefore, must be attributable to the realization of a particular combination of such innovations.14 Chart 5 depicts the combinations of exogenous driving processes that, according to the estimated DSGE model, are responsible for the observed path of inflation, output, and the interest rate over the period we analyze. In each panel, the dashed line represents the evolution of the shock associated with the mean forecast while the solid line represents the sequence of shocks corresponding to the actual realization of the observable variables. As in Chart 4, the medium blue band denotes the 75 percent probability interval for the forecast. 14 In this study, we distinguish between exogenous driving processes—shocks, for short—and innovations. Driving processes can be autocorrelated, and thus forecastable, while their innovations are i.i.d. The contribution of three shocks stands out. First, the demand shock recovers from almost -4 percent to around -1 percent. This movement is particularly pronounced during 2004, when inflation was picking up. However, this profile is broadly consistent with the shock’s expected evolution, represented by the dashed line. The productivity shock is also broadly in line with expectations, with the exception of 2003:3; this spike in productivity accounts for the corresponding spike in output growth in that quarter. However, the most significant and direct contribution to the surge in inflation comes from a sizable upward movement in the inflation target, t. According to our estimates, this target moves by about 1 percentage point, from less than 1 percent to close to 2 percent. Moreover, this movement is at the edge of the 75 percent probability interval for the forecast, suggesting that the realization of this driving process is indeed quite unusual. To quantify more directly the effect on inflation of the unexpected increase in the implicit inflation target, we depict what would have happened to core PCE inflation in the absence of such an increase (Chart 6). Here, the solid line FRBNY Economic Policy Review / October 2010 37 Chart 6 Conditional Forecast of Inflation 3 Core PCE Inflation 2 1 0 2002 2003 2004 Source: Authors’ calculations. Notes: The dashed line represents a forecast of inflation conditional on the Kalman smoother estimates of all shocks except for those to the inflation target; the solid line represents the observed realization. The shaded areas represent 50 (light blue), 75 (medium blue), and 90 percent (dark blue) symmetric probability intervals for this conditional forecast. Therefore, they represent uncertainty stemming from future realizations of the inflation target shock alone. PCE is personal consumption expenditure. is realized inflation. The dashed line represents the counterfactual path of inflation predicted by the model in the absence of shocks to the inflation target. In other words, this is a forecast for inflation, conditional on the estimated path of all but the inflation target shock. The bands therefore represent the usual probability intervals, but in this case they are computed around this conditional forecast. The chart confirms our conclusion on the role of innovations to the inflation target in accounting for the observed evolution of inflation. According to the model, core inflation would not have increased to above 2 percent, as it did for most of 2004, without the steady increase in the inflation target over the same period. In fact, inflation would have remained within the “comfort zone” of 1 to 2 percent. Moreover, note that the solid line of realized inflation is mostly inside the area associated with the 90 percent probability interval for the conditional forecast. This suggests that the share of the forecast error in inflation accounted for by the innovations in the inflation target in this episode is unusually large compared with the historical average. This is just a more formal way of saying that the increase in the inflation target is disproportionately responsible for the observed increase in inflation that we examine. The estimated rise in the implicit inflation target provides the missing link for a unified explanation of the pickup in inflation, the “considerable period” monetary policy, and the absence of a concomitant acceleration in output growth. In the model, the inflation target is the main driver of movements 38 Policy Analysis Using DSGE Models: An Introduction in inflation expectations, which are a key determinant of firms’ pricing behavior together with the amount of slack in the economy. According to the DSGE model, therefore, a significant fraction of the inflation acceleration in 2003-04 can be attributed to a change in inflation expectations, driven by an increase in the Federal Reserve’s implicit inflation target as perceived by the private sector. This increase in the perceived target in turn is consistent with the unusually loose stance of monetary policy maintained by the FOMC during the “considerable period” era. This brings us to the third question: If the DSGE model is correct, and the pickup in inflation in 2004 is attributable to an increase in the implicit inflation target perceived by the public, could the Federal Reserve have prevented this development? Charts 7 and 8 show the results of this counterfactual analysis. Both charts display the data (solid line) along with the counterfactual outcomes for the economy predicted by the model under a policy consistent with the stabilization of core inflation at 1.6 percent through 2004. The way in which this According to the DSGE model . . . a significant fraction of the inflation acceleration in 2003-04 can be attributed to a change in inflation expectations, driven by an increase in the Federal Reserve’s implicit inflation target as perceived by the private sector. policy is implemented, however, is different in the two cases. In Chart 7, we present the outcomes associated with what we call a “no-communication” monetary strategy (dashed line) while in Chart 8 we compare these results with those that would emerge under a “full-communication” strategy (blue line). Under the no-communication strategy, the path for the interest rate compatible with the desired evolution of inflation is achieved each period through “surprise” departures from the historical rule. In contrast, under the full-communication strategy, the Federal Reserve implements the same path for inflation by announcing an inflation target that is consistent with it.15 15 Technically, in both cases we choose shocks to the monetary policy rule that are compatible with the desired evolution of inflation, conditional on the smoothed value of all other disturbances. Under the no-communication i strategy, the shocks we choose are the i.i.d. monetary shocks, t . Under the full-communication strategy, our chosen shocks are to the inflation target, t . 3 Chart 7 Chart 8 “No-Communication” Counterfactual “Full-Communication” Counterfactual 3 Core PCE Inflation 2 2 1 1 0 0 12 12 Real GDP Growth 10 10 8 8 6 6 4 4 2 2 0 0 -2 -2 5 5 Federal Funds Rate 4 4 3 3 2 2 1 1 0 0 -1 -1 Core PCE Inflation Real GDP Growth Federal Funds Rate -2 -2 2002 2003 2004 2002 2003 2004 Source: Authors’ calculations. Source: Authors’ calculations. Notes: The dashed line represents the counterfactual evolution of the economy predicted by our model had monetary policy been set to achieve the path for inflation depicted in the top panel. This counterfactual is conditional on the posterior mean of the parameters. Under the “no-communication” scenario, the desired path for inflation is achieved by the choice of the interest rate shock while all other shocks are set at their Kalman smoother estimate. The shaded area represents the 75 percent symmetric probability interval for the unconditional forecast, which is the same as in Chart 4. The black line represents the observed realization of each series. PCE is personal consumption expenditure. Notes: The blue line represents the counterfactual evolution of the economy predicted by our model had monetary policy been set to achieve the path for inflation depicted in the top panel. This counterfactual is conditional on the posterior mean of the parameters. Under the “full-communication” scenario, the desired path for inflation is achieved by the choice of the inflation target while all other shocks are set at their Kalman smoother estimate. The shaded area represents the 75 percent symmetric probability interval for the unconditional forecast, which is the same as in Chart 4. The black line represents the observed realization of each series. The dashed line is the conditional forecast under the “no-communication” scenario. PCE is personal consumption expenditure. The crucial difference between the results obtained under the two scenarios stems from the key role that expectations play in the DSGE model. Under the full-communication strategy, inflation expectations are immediately affected by the announcement of an inflation target. These expectations in turn have a direct effect on actual inflation without requiring a contraction in real activity to force businesses to contain their price increases. Under the no-communication strategy, inflation expectations remain at their historical level. As a result, inflation can be controlled only by increasing interest rates to contain GDP growth. The way in which we model the full-communication scenario is quite stark. In practice, expectations would be unlikely to adjust instantaneously, even if the Federal Reserve FRBNY Economic Policy Review / October 2010 39 were completely transparent about its inflation target. Nevertheless, the differences between the results of the two policy strategies are striking. In the no-communication case, inflation can be stabilized only through wild movements in the federal funds rate. As a result, GDP growth is also extremely volatile: it falls below zero in 2004:1, but then recovers to a quarterly (annualized) growth rate of 10 percent and ends the period Reserve around 2003 was to minimize the U.S. economy’s likelihood of falling into such a trap.16 From this perspective, our analysis might be interpreted as supportive of the policy stance adopted by the central bank in 2003-04 as part of a successful preemptive strike against a liquidity trap. 6. Conclusion Our analysis might be interpreted as supportive of the policy stance adopted by the central bank in 2003-04 as part of a successful preemptive strike against a liquidity trap. at zero. These movements in output are indeed extreme. They lie well outside the 75 percent forecast probability interval reported in the chart. In fact, the quantitative details of the evolution of output and the interest rate under the counterfactual simulations should not be taken literally, since they depend significantly on the details of the model and on the assumption that the central bank insists on perfectly stabilizing current inflation. However, the qualitative pattern of higher volatility under the no-communication strategy is a robust feature of models in which expectations matter. Under the full-communication strategy, in contrast, the desired path for inflation can be achieved with much less pronounced fluctuations in real growth and an almost unchanged policy relative to the actual path. Interest rates need not rise and output need not fall significantly because a shift in expectations brought about by clear communication of the Federal Reserve’s inflation objective largely brings inflation under control. Note that the results of these counterfactual exercises should be interpreted with caution. Their objective is not to prescribe an alternative to the policy followed in 2004, but rather to investigate how a different path for inflation might have been achieved. In fact, according to Krugman (1998) and Eggertsson and Woodford (2003), an increase in inflation expectations might be the best monetary strategy to escape a liquidity trap. Many have argued that the main objective of the Federal 40 Policy Analysis Using DSGE Models: An Introduction This article provides an introduction to dynamic stochastic general equilibrium models and presents an example of their use as tools for monetary policy analysis. Given the mainly educational nature of our presentation, we simplify by using a small-scale model designed to account for the behavior of three key macroeconomic variables: GDP growth, core PCE inflation, and the federal funds rate. Despite its simplicity, our model is rich enough to reproduce some of the salient features of the series of interest. It also allows us to highlight the components common to more articulated and realistic DSGE specifications. Our model offers insight into the causes of the abrupt pickup in inflation in the first half of 2004, from levels close to 1 percent at the beginning of 2003 to values steadily above 2 percent through the end of 2008. This exercise highlights the central role of expectations in the transmission of shocks and policy impulses in DSGE models. The main lesson that we derive from the exercise is that the most effective approach to controlling inflation is through the management of expectations, rather than through actual movements of the policy instrument. This lesson seems to be well understood by the public, given the amount of attention and speculation that usually surround the pronouncements of central bankers on their likely future actions. DSGE models have the potential to broaden this understanding by adding a quantitative assessment of the link between current policy, expectations, and economic outcomes—and thus to clarify the effect that different systematic approaches to policy have on those outcomes. 16 In its August 2003 statement, the FOMC observed that “on balance, the risk of inflation becoming undesirably low is likely to be the predominant concern for the foreseeable future.” Very low or negative levels of inflation are one of the most likely triggers of a liquidity trap. Technical Appendix The table reports information on the prior distribution for the parameters of the model. Further details on the parameters and the structure of the model are available from the corresponding authors. Parameter Distribution Mean Standard Deviation y Calibrated Gamma Gamma Beta Beta Beta Normal Normal Normal Normal Normal Beta Beta Beta Beta InvGamma InvGamma InvGamma InvGamma InvGamma 0.99 0.1 1.0 0.6 0.6 0.7 1.5 0.5 2.0 2.0 3.0 0.95 0.5 0.5 0.5 0.5 0.5 0.5 0.2 0.5 — 0.05 0.2 0.2 0.2 0.15 0.25 0.2 1.0 1.0 0.35 0.04 0.2 0.2 0.2 2.0 2.0 2.0 1.0 2.0 r u u i FRBNY Economic Policy Review / October 2010 41 References Adolfson, M., M. K. Andersson, J. Lindé, M. Villani, and A. Vredin. 2007. “Modern Forecasting Models in Action: Improving Macroeconomic Analyses at Central Banks.” International Journal of Central Banking 3, no. 4 (December): 111-44. Adolfson, M., S. Laséen, J. Lindé, and M. Villani. 2008. “Evaluating an Estimated New Keynesian Small Open-Economy Model.” Journal of Economic Dynamics and Control 32, no. 8 (August): 2690-2721. An, S., and F. Schorfheide. 2007. “Bayesian Analysis of DSGE Models.” Econometric Reviews 26, nos. 2-4 (April): 113-72. Bernanke, B. S., M. Gertler, and S. Gilchrist. 1999. “The Financial Accelerator in a Quantitative Business Cycle Framework.” In J. B. Taylor and M. Woodford, eds., Handbook of Macroeconomics, vol. 1, 1341-93. Elsevier. Curdia, V., and M. Woodford. 2009. “Conventional and Unconventional Monetary Policy.” Federal Reserve Bank of New York Staff Reports, no. 404, November. Dixit, A. K., and J. E. Stiglitz. 1977. “Monopolistic Competition and Optimum Product Diversity.” American Economic Review 67, no. 3 (June): 297-308. Edge, R. M., M. T. Kiley, and J.-P. Laforte. 2008. “Natural Rate Measures in an Estimated DSGE Model of the U.S. Economy.” Journal of Economic Dynamics and Control 32, no. 8 (August): 2512-35. Eggertsson, G. B., and M. Woodford. 2003. “The Zero Bound on Interest Rates and Optimal Monetary Policy.” Brookings Papers on Economic Activity, no. 1: 139-211. Bils, M., and P. J. Klenow. 2004. “Some Evidence on the Importance of Sticky Prices.” Journal of Political Economy 112, no. 5 (October): 974-85. Eichenbaum, M. S., L. Hansen, and K. J. Singleton. 1988. “A Time Series Analysis of Representative Agent Models of Consumption and Leisure Choice under Uncertainty.” Quarterly Journal of Economics 103, no. 1 (February): 51-78. Calvo, G. A. 1983. “Staggered Prices in a Utility-Maximizing Framework.” Journal of Monetary Economics 12, no. 3 (September): 383-98. Erceg, C. J., L. Guerrieri, and C. Gust. 2006. “SIGMA: A New OpenEconomy Model for Policy Analysis.” International Journal of Central Banking 2, no. 1 (March): 1-50. Campbell, J. Y. 2003. “Consumption-Based Asset Pricing.” In G. M. Constantinides, M. Harris, and R. M. Stulz, eds., Handbook of the Economics of Finance, vol. 1, 803-87. North-Holland. Galí, J., and M. Gertler. 1999. “Inflation Dynamics: A Structural Econometric Analysis.” Journal of Monetary Economics 44, no. 2 (October): 195-222. Campbell, J. Y., and J. H. Cochrane. 2000. “Explaining the Poor Performance of Consumption-Based Asset Pricing Models.” Journal of Finance 55, no. 6 (December): 2863-78. ———. 2007. “Macroeconomic Modeling for Monetary Policy Evaluation.” Journal of Economic Perspectives 21, no. 4 (fall): 25-45. Christiano, L., R. Motto, and M. Rostagno. 2007. “Financial Factors in Business Cycles.” Unpublished paper. Gerali, A., S. Neri, L. Sessa, and F. Signoretti. 2008. “Credit and Banking in a DSGE Model.” Unpublished paper, Bank of Italy, June. Clarida, R. G., J. Galí, and M. Gertler. 1999. “The Science of Monetary Policy: A New Keynesian Perspective.” Journal of Economic Literature 37, no. 4 (December): 1661-1707. Curdia, V., A. Ferrero, and A. Tambalotti. 2009. “Which Output Gap? Evaluating Interest Rate Rules in an Estimated DSGE Model.” Unpublished paper, Federal Reserve Bank of New York. 42 Policy Analysis Using DSGE Models: An Introduction Gertler, M., and P. Karadi. 2009. “A Model of Unconventional Monetary Policy.” Unpublished paper. Gertler, M., L. Sala, and A. Trigari. 2008. “An Estimated Monetary DSGE Model with Unemployment and Staggered Nominal Wage Bargaining.” Journal of Money, Credit, and Banking 40, no. 8 (December): 1713-64. References (Continued) Gilchrist, S., A. Ortiz, and E. Zakrajsek. 2009. “Credit Risk and the Macroeconomy: Evidence from an Estimated DSGE Model.” Unpublished paper. Hansen, L. P., and K. J. Singleton. 1982. “Generalized Instrumental Variables Estimation of Nonlinear Rational Expectations Models.” Econometrica 50, no. 5 (September): 1269-86. ———. 1983. “Stochastic Consumption, Risk Aversion, and the Temporal Behavior of Asset Returns.” Journal of Political Economy 91, no. 2 (April): 249-65. Krugman, P. 1998. “It’s Baaack! Japan’s Slump and the Return of the Liquidity Trap.” Brookings Papers on Economic Activity, no. 2: 137-87. Nakamura, E., and J. Steinsson. 2008. “Five Facts about Prices: A Reevaluation of Menu Cost Models.” Quarterly Journal of Economics 123, no. 4 (November): 1415-64. Orphanides, A., and D. W. Wilcox. 2002. “The Opportunistic Approach to Disinflation.” International Finance 5, no. 1 (spring): 47-71. Smets, F., and R. Wouters. 2003. “An Estimated Dynamic Stochastic General Equilibrium Model of the Euro Area.” Journal of the European Economic Association 1, no. 5 (September): 1123-75. ———. 2007. “Shocks and Frictions in U.S. Business Cycles: A Bayesian DSGE Approach.” American Economic Review 97, no. 3 (June): 586-606. Taylor, J. B. 1993. “Discretion versus Policy Rules in Practice.” Carnegie-Rochester Conference Series on Public Policy 39, December: 195-214. ———. 2008. “The Costs and Benefits of Deviating from the Systematic Component of Monetary Policy.” Remarks delivered at the Federal Reserve Bank of San Francisco Conference on Monetary Policy and Asset Markets, San Francisco, California, February 22. Woodford, M. 2003. Interest and Prices: Foundations of a Theory of Monetary Policy. Princeton, N.J.: Princeton University Press. Sbordone, A. M. 2002. “Prices and Unit Labor Costs: A New Test of Price Stickiness.” Journal of Monetary Economics 49, no. 2 (March): 265-92. The views expressed are those of the authors and do not necessarily reflect the position of the Federal Reserve Bank of New York or the Federal Reserve System. The Federal Reserve Bank of New York provides no warranty, express or implied, as to the accuracy, timeliness, completeness, merchantability, or fitness for any particular purpose of any information contained in documents produced and provided by the Federal Reserve Bank of New York in any form or manner whatsoever. FRBNY Economic Policy Review / October 2010 43 Kenneth D. Garbade, Frank M. Keane, Lorie Logan, Amanda Stokes, and Jennifer Wolgemuth The Introduction of the TMPG Fails Charge for U.S. Treasury Securities • Prior to May 2009, market convention enabled a seller of Treasury securities to postpone— without any explicit penalty and at an unchanged invoice price—its obligation to deliver the securities. • The September 2008 insolvency of Lehman Brothers exposed a flaw in this convention, when a decline in short-term interest rates set the stage for an extraordinary volume of settlement fails that threatened to erode the perception of the market as being free of credit risk. • In response, the Treasury Market Practices Group introduced a “dynamic fails charge” for Treasury securities in May 2009. • The fails charge incentivizes timely settlement by providing that a buyer of Treasury securities can claim monetary compensation from a seller if the seller fails to deliver on a timely basis. • The fails charge mitigated a key dysfunctionality in the market and illustrates the value of public and private sector cooperation in responding to altered market conditions. Kenneth D. Garbade is a senior vice president, Frank M. Keane an assistant vice president, Lorie Logan a vice president, Amanda Stokes a trader/analyst, and Jennifer Wolgemuth a counsel and assistant vice president at the Federal Reserve Bank of New York. Correspondence: kenneth.garbade@ny.frb.org 1. Introduction S ecurities transactions commonly involve a variety of market conventions—widely accepted ways of doing business that persist through time even though not mandated by law or regulation. Commonplace examples include the quotation of prices for Treasury bonds in increments of 32nds (and fractions of a 32nd) of a percent of principal value (rather than in decimal increments) and the quotation of Treasury bills in terms of discount rates (rather than prices or yields). In most cases, market conventions are useful or, at worst, innocuous. In some cases, however, a new use for an old instrument can render a convention in need of revision. One particularly notorious example was the convention—observed prior to 1982—of ignoring accrued interest on Treasury bonds sold on repurchase agreements (also known as repos, or RPs). The convention made sense as long as repos were used primarily to borrow money from creditworthy lenders that held the bonds simply to limit their exposure to credit risk. It made less sense when highly leveraged securities dealers began to use repos to borrow bonds to deliver on short sales. The 1982 Drysdale episode illuminated the risks involved in ignoring accrued interest and prompted the Federal Reserve Bank of New York to orchestrate a change in the convention.1 1 Garbade (2006). The authors are grateful to Murray Pozmanter and Cartier Stennis for assistance in researching this article and to two anonymous referees for helpful suggestions. They are especially grateful to the members of the Treasury Market Practices Group, identified in Box 4, for the time and effort they devoted to making the fails charge a reality, as well as to all of the other individuals, identified in the appendix, who provided invaluable feedback and assistance during the implementation of the fails charge. The views expressed are those of the authors and do not necessarily reflect the position of the Federal Reserve Bank of New York or the Federal Reserve System. FRBNY Economic Policy Review / October 2010 45 A market convention may also require revision following a change in the economic environment. This article discusses a recent example: the convention of postponing—without any explicit penalty and at an unchanged invoice price—a seller’s obligation to deliver Treasury securities if the seller fails to deliver the securities on a scheduled settlement date. As discussed in more detail below, as long as short-term interest rates were above about 3 percent, the time value of money usually sufficed to incentivize timely settlement of transactions in Treasury securities. However, when short-term rates fell to A market convention may . . . require revision following a change in the economic environment. This article discusses a recent example: the convention of postponing—without any explicit penalty and at an unchanged invoice price—a seller’s obligation to deliver Treasury securities if the seller fails to deliver the securities on a scheduled settlement date. near zero following the insolvency of Lehman Brothers Holdings Inc. in September 2008, the time value of money no longer provided adequate incentive and the Treasury market experienced an extraordinary volume of settlement fails. Both the breadth of the fails across a large number of securities and the persistence of the fails were unprecedented and threatened to erode the perception of the Treasury market as a market free of credit risk.2 In response, the Treasury Market Practices Group (TMPG)—a group of market professionals committed to supporting the integrity and efficiency of the U.S. Treasury market—worked over a period of six months to revise the market convention for settlement fails, developing a “dynamic fails charge” that, when short-term interest rates are below 3 percent, produces an economic incentive to settle trades roughly equivalent to the incentive that exists when rates are at 3 percent. Thus, the TMPG fails charge preserves a significant economic incentive for timely settlement even when interest rates are close to zero. 2 See, for example, Wrightson, Federal Reserve Data, October 17, 2008 (“The breakdown in the clearing mechanism for the Treasury market is beginning to emerge as a top-tier policy concern. The safe-haven status of Treasury securities is one of the few advantages the government market has left in a year in which net Treasury borrowing needs . . . are likely to exceed $1 trillion by a large margin. At some point, though, buyers will think twice about buying a ‘safe-haven’ asset for peace of mind if they have doubts about their counterparty’s ability to deliver the security.”). 46 The Introduction of the TMPG Fails Charge This article describes the introduction of the TMPG fails charge. The introduction of the fails charge is important for two reasons. First, it mitigated an important dysfunctionality in a market of critical significance both to the Federal Reserve in its execution of monetary policy and to the country as a whole. Second, it exemplified the value of cooperation between the public and private sectors in responding to altered market conditions in a flexible, timely, and innovative fashion. Our study is divided into three parts. The first part (Sections 2-5) describes settlement processes and settlement fails in the Treasury market, explains why sellers usually try to avoid fails, describes industry and Federal Reserve efforts to mitigate settlement fails prior to 2008, and briefly reviews three episodes of chronic fails in the Treasury market. The second part (Section 6) describes the tsunami of fails that followed Lehman’s insolvency. The balance of the study (Sections 7-10) explains the TMPG’s response. Section 11 concludes. 2. Settlements and Settlement Fails in U.S. Treasury Securities A transaction in Treasury securities is said to “settle” when the seller delivers the securities to, and receives payment from, the buyer. The two most important settlement processes are bilateral settlement and multilateral net settlement. Before describing those processes, we explain how market participants establish and transfer ownership of Treasury securities. 2.1 Establishing and Transferring Ownership of Treasury Securities For more than three decades, investors have established ownership of Treasury securities through Federal Reserve book-entry securities accounts.3 Book-entry account holders that own Treasury securities can house their securities directly in their accounts and can transfer the securities to other bookentry accounts by issuing appropriate instructions to the Federal Reserve. Federal Reserve book-entry accounts are generally available only to depository institutions and certain other organizations, such as government-sponsored enterprises and foreign central banks. All other market participants establish ownership of Treasury securities through commercial book-entry accounts at depository institutions that act as custodians for their customers. Depository institutions that offer commercial 3 See “Book-Entry Securities Account Maintenance and Transfer Services,” Federal Reserve Banks Operating Circular no. 7, August 19, 2005. Garbade (2004) describes the origins of the Federal Reserve book-entry system. book-entry accounts hold their customers’ securities in their Federal Reserve book-entry accounts commingled with their own securities. A market participant with a commercial book-entry account can transfer Treasury securities to another market participant through their respective custodians. For example, participant A can transfer a Treasury security to participant B by instructing its custodian to debit its commercial book-entry account and to transfer the security to B’s custodian for credit to B’s commercial book-entry account. Upon receipt of instructions, A’s custodian will debit A’s account and instruct the Federal Reserve to 1) debit its Federal Reserve book-entry account and 2) credit the Federal Reserve book-entry account of B’s custodian. Following receipt of the security in its Federal Reserve book-entry account, B’s custodian will complete the transfer by crediting B’s commercial book-entry account. (If A and B have a common custodian, the transfer can be completed on the books of that common custodian without involving the Federal Reserve.) 2.2 Bilateral Settlement The simplest type of settlement occurs when a market participant has sold Treasury securities for bilateral settlement on a deliver-versus-payment basis. The sale may be a conventional sale of securities but it may alternatively be The simplest type of settlement occurs when a market participant has sold Treasury securities for bilateral settlement on a deliver-versus-payment basis. the starting leg, or the “off” leg, of a repurchase agreement. (We describe repurchase agreements in more detail below.) Suppose, for example, an investor sells ten Treasury bonds at a price of $100 per bond for settlement on June 2. Following negotiation of the terms of the sale, the seller will instruct its custodian to send ten bonds to the buyer’s custodian on June 2 against payment of $1,000. The buyer will concurrently instruct its custodian to receive, on June 2, ten bonds from the seller’s custodian and to pay $1,000 upon receipt of the bonds. On June 2, the seller’s custodian will instruct the Federal Reserve to 1) debit its Federal Reserve book-entry account for ten bonds, 2) credit the Federal Reserve book-entry account of the buyer’s custodian for ten bonds and simultaneously debit the account of the buyer’s custodian for the $1,000 due upon Exhibit 1 Bilateral Settlement Buyer Seller Delivery instructions Receive instructions Ten bonds Seller’s custodian Buyer’s custodian $1,000 Note: This transfer of bonds and funds is effected through the Federal Reserve if the seller and buyer have different custodians, and is effected on the books of the common custodian if they have the same custodian. delivery, and 3) credit the seller’s custodian’s account for the $1,000. The resulting transfers of securities and funds are shown in Exhibit 1.4 Following notification that ten bonds have come into its Federal Reserve book-entry account and that $1,000 has been withdrawn, the buyer’s custodian will verify that the bonds and money are consistent with the buyer’s instructions. In most cases, they are and the custodian will credit the buyer’s account for the ten bonds and debit that account for $1,000. In some cases, however, the buyer will have provided different instructions—perhaps referencing a different security or a different invoice price—or no instructions. In any of these cases, the buyer’s custodian will reverse the settlement, instructing the Federal Reserve to return the ten bonds and recover the $1,000 payment. The buyer and seller and their respective custodians will then have to communicate and come to a common understanding of the terms of the underlying transaction, following which the seller will reinitiate the settlement process. 2.3 Multilateral Net Settlement Bilateral settlement is a simple process that satisfies the purpose of settlement: moving securities from sellers to buyers and moving funds from buyers to sellers. Alternative settlement structures, however, can sometimes be more efficient. 4 In the event the buyer and seller have a common custodian, settlement can be completed on the books of the common custodian, with cash and securities moving between the accounts of the respective customers, without involving the Federal Reserve. FRBNY Economic Policy Review / October 2010 47 Exhibit 2 Bilateral Settlement of Three Transactions because it will receive more for the bonds sold to C than the original contract price and C loses for the same reason. These gains and losses are exactly offset with further agreements to make small side payments of cash. In particular: A • A agrees to pay $8 to the CCP, reflecting the $8 gain from marking the price of the eight bonds bought from C down from $101 per bond to $100 per bond, Instructions • B agrees to pay $10 to the CCP, reflecting the $10 gain from marking the price of the ten bonds sold to C up from $99 to $100 per bond, and A’s custodian Eight bonds $808 Ten bonds • the CCP agrees to pay $18 to C, in compensation for the $8 loss from marking the price of the eight bonds sold to A down from $101 per bond, and for the $10 loss from marking the price of the ten bonds bought from B up from $99 per bond. $1,000 $990 C’s custodian B’s custodian Ten bonds Instructions Instructions C B On the night before the settlement date, the CCP nets out the deliver and receive obligations of A, B, and C and novates5 their respective contracts, becoming the buyer from every net seller and the seller to every net buyer, all at the common settlement price. After netting and novation: • A is obligated to deliver two bonds to the CCP against payment of $200, Consider, for example, the case where: • B has no deliver or receive obligations, and • participant A sells ten bonds to participant B at a price of $100 per bond for settlement on the following business day, • the CCP is obligated to deliver two bonds to C against payment of $200. • B sells ten of the same bonds to participant C at a price of $99 per bond, also for settlement on the following business day, and • C sells eight of the same bonds to A at a price of $101 per bond, again for settlement on the following business day. As shown in Exhibit 2, bilateral settlement of the three transactions requires the delivery of twenty-eight bonds against payments of $2,798. As an alternative, the participants might agree to settle through a central counterparty (CCP). The CCP first marks all of the deliver and receive obligations to a common price—say, $100 per bond. After marking to the common price, • A is obligated to deliver ten bonds to B against payment of $1,000, • B is obligated to deliver ten bonds to C against payment of $1,000, and • C is obligated to deliver eight bonds to A against payment of $800. Marking to a common price results in gains for some participants and losses for others. In the example, B gains 48 The Introduction of the TMPG Fails Charge On settlement day, the obligations of A to deliver two bonds to the CCP and the CCP to deliver two bonds to C are settled with bilateral deliver-versus-payment settlements. In addition, A, B, and the CCP make the agreed-upon side payments of cash. Exhibit 3 shows that multilateral net settlement requires the delivery of four bonds and payments of $436—about 15 percent of the deliveries and payments shown in Exhibit 2. 2.4 Some Concrete Identities The foregoing description of settlement processes referred to abstract entities like “participant A” and an unnamed “central counterparty.” Before we begin to discuss settlement fails, it may be helpful to identify some of the key participants in the Treasury market. At the center of the market is a group of dealers that provide liquidity to customers, quoting bid prices at which they are willing to buy and offer prices at which they are prepared to sell. A subset of dealers, called “primary dealers,” make markets 5 To novate is to substitute one legal obligation for another. Box 1 Exhibit 3 Multilateral Net Settlement of Three Transactions A Instructions Primary Dealers in Mid-2008a Banc of America Securities LLC Barclays Capital Inc. Bear, Stearns & Co., Inc.b BNP Paribas Securities Corp. Cantor, Fitzgerald & Co. A’s custodian $8 Two bonds B $200 Instructions Central Instructions Central counterparty’s counterparty custodian Two bonds $10 Citigroup Global Markets, Inc. Credit Suisse Securities (USA) LLC Daiwa Securities America Inc. Deutsche Bank Securities Inc. Dresdner Kleinwort Securities LLC B’s custodian $200 Goldman, Sachs & Co. Greenwich Capital Markets, Inc. HSBC Securities (USA) Inc. J. P. Morgan Securities Inc. Lehman Brothers Inc.c $18 C Instructions C’s custodian Note: Security settlements are shown with solid blue and black lines. Side payments, represented by dashed black lines, take place independently of security settlements. to the Federal Reserve Bank of New York when the Bank is conducting open market operations on behalf of the Federal Reserve System.6 Box 1 identifies the primary dealers as of mid2008. Dealers sometimes trade directly with each other, but more commonly through specialized interdealer brokers. A dealer that sells securities to another dealer through an interdealer broker agrees to deliver securities (against payment) to the broker. The broker, in turn, agrees to deliver the same securities (also against payment) to the ultimate buyer. This arrangement allows the dealers to trade on a “blind,” or undisclosed, basis. All of the dealers, and all of the interdealer brokers, maintain commercial book-entry accounts at one of two banks: JPMorgan Chase Bank, N.A., and The Bank of New York Mellon. These two “clearing” banks offer custodial services refined over many years to meet the needs of brokers and dealers that deliver and receive large volumes of securities on a daily basis. The Fixed Income Clearing Corporation (FICC), a subsidiary of the Depository Trust & Clearing Corporation, is the central counterparty in the Treasury market. All of the 6 See, generally, Federal Reserve Bank of New York, “Administration of Relationships with Primary Dealers,” January 11, 2010, available at http://www.newyorkfed.org/markets/pridealers_ policies.html. Merrill Lynch Government Securities Inc.d Mizuho Securities USA Inc. Morgan Stanley & Co. Incorporated UBS Securities LLC a Federal Reserve Bank of New York, “Primary Dealers List,” July 15, 2008, available at http://www.newyorkfed.org/newsevents/news/markets/2008/ an080715.html. b Removed October 1, 2008, following its acquisition by J. P. Morgan Securities Inc. c Removed September 22, 2008. d Removed February 11, 2009, following its acquisition by Bank of America Corporation. primary dealers and all of the interdealer brokers, as well as a number of other market participants, are netting members of FICC. FICC maintains commercial book-entry accounts at both JPMorgan Chase and The Bank of New York Mellon and is prepared to receive securities from, and deliver securities to, any of its netting members in a timely and efficient fashion. Beyond the dealers, the interdealer brokers, and FICC, the Treasury market consists of a large number of other participants, including “real-money” investors such as mutual funds, pension funds, and corporate treasurers, and “leveraged accounts” such as hedge funds. Some of these participants trade directly with dealers, others trade anonymously in electronic markets. All use custodians that offer more or less complex (and more or less costly) services tailored to their needs. FRBNY Economic Policy Review / October 2010 49 2.5 Settlement Fails A settlement fail occurs when the obligation of a seller to deliver securities to a buyer remains outstanding following the close of business on the scheduled settlement date of a transaction. This can occur either because the seller’s custodian failed to tender any securities to the buyer’s custodian, or because the buyer’s custodian rejected whatever securities were tendered by the seller’s custodian. In the event of a settlement fail in Treasury securities, the market convention is to postpone settlement to the following business day without any A settlement fail occurs when the obligation of a seller to deliver securities to a buyer remains outstanding following the close of business on the scheduled settlement date of a transaction. change in the funds due upon delivery and (prior to May 2009) without any explicit penalty or charge.7 The process of failing (to settle) and deferring settlement to the next business day can take place repeatedly, day after day, until settlement occurs or the trade is canceled. Settlement fails can occur for any of several reasons. First, a fail can result from miscommunication. A buyer and seller may not have a common understanding of the terms of a trade, or one or the other may have failed to communicate settlement instructions to its custodian, or may have communicated incorrect instructions, or one of the custodians may have misunderstood the instructions that it received. In any of these cases, the buyer’s custodian will reject whatever securities are tendered by the seller’s custodian. After becoming aware of the failed attempt to settle (or of the absence of any attempt to settle), the buyer and seller and their respective custodians communicate to resolve the problem. This usually results in successful settlement within a day or two. A fail may also stem from operational problems. One well-known instance occurred on Thursday, November 21, 1985, when a computer outage at The Bank of New York 7 This convention was memorialized in Chapter 8, Section C, of the Government Securities Manual of the Public Securities Association: “If securities are not delivered on the agreed upon settlement date, there is a fail. Regardless of the date the securities were actually delivered, the buyer of the securities pays the seller the original settlement date figures.” The Public Securities Association was the forerunner of the Bond Market Association, which joined with the Securities Industry Association in 2006 to form the Securities Industry and Financial Markets Association. 50 The Introduction of the TMPG Fails Charge (a predecessor of The Bank of New York Mellon) prevented that bank from effecting deliveries of Treasury securities. The bank was unable to resolve the problem until the following day, and had to finance overnight (at its own expense) the customer securities that it was unable to deliver. It borrowed in excess of $20 billion from the Federal Reserve Bank of New York and incurred interest expenses of $5 million.8 A settlement fail can also occur because the seller does not have the requisite securities in its commercial book-entry account. This is the most common reason for failing when fails are chronic, but it is usually avoided at other times by borrowing securities and delivering the borrowed securities. 3. Repurchase Agreements and Borrowing Securities to Avoid or Cure Settlement Fails A repurchase agreement is a sale of securities coupled with an agreement to repurchase the same securities at a specified price on a later date.9 Market participants use repos to borrow money when they buy securities but do not have sufficient cash on hand to pay for them, that is, to finance long positions, as well as to borrow securities when they sell securities they do not already own, that is, to finance short positions. A repo is analogous to a loan, where the proceeds of the initial sale correspond to the principal amount of the loan and the excess of the repurchase price over the original sale price corresponds to the interest paid on the loan. A market participant might, for example, sell securities for $10 million and simultaneously agree to repurchase the securities ten days later for $10,008,333. This is analogous to borrowing $10 million for ten days at an interest rate of 3 percent per annum.10 Market participants commonly think of repos as loans, rather than as purchases and sales, and quote repos in terms of interest rates rather than in terms of sale and repurchase prices.11 8 A Computer Snafu Snarls the Handling of Treasury Issues,” Wall Street Journal, November 25, 1985, p. 58; Committee on Banking, Finance, and Urban Affairs (1985); “Fed is Queried on Failure of Bank Computer System,” Wall Street Journal, December 13, 1985, p. 49; “Fed Weighs a Penalty,” New York Times, December 13, 1985, p. D2; Sender (1986). 9 Repurchase agreements are complex financial instruments whose contracting conventions have evolved over the past four decades. See Garbade (2006) and Fleming and Garbade (2003, 2004). 10 $8,333 = (repo term of 10 days / 360 days in a year) × 3 percent per annum × $10 million, where the calculation uses the money market convention of a 360-day year. 11 The quotation convention does not change the nature of a repo— a transaction in which one party sells securities subject to an agreement to repurchase them at a later date. Repos are most commonly arranged on an overnight basis but can run for days or weeks. They can also be arranged on an “open,” or continuing, basis (with a daily adjustment of the interest rate) at the mutual consent of the parties. Industry standard documentation for a repo provides that if the original seller fails to repurchase the securities on the agreed-upon repurchase date, the original buyer has the contractual right to, among other things, sell the securities to a third party and use the proceeds to satisfy the original seller’s repurchase obligation. Conversely, if the original buyer does not deliver the securities back to the original seller on the repurchase date, the original seller has the contractual right to, among other things, use the funds that it otherwise would have used to repurchase the securities to “buy in,” or replace, the securities. Exhibit 4 Lending Treasury Bond B (against Borrowing Money at 2 Percent) on a Special Collateral Repurchase Agreement and Relending the Money on a General Collateral Repurchase Agreement at 3 Percent Investor lending bond B (against borrowing money at 2 percent) and lending money at 3 percent (against general collateral) Bond B General collateral Funds at 2 percent 3.1 Types of Repurchase Agreements Participant borrowing bond B against lending money at 2 percent Repos come in two flavors: general collateral repos (used to borrow money) and special collateral repos (used to borrow securities). Starting leg of special collateral repurchase agreement General collateral repos: A general collateral repo is a repo in which the lender of funds is willing to accept any member of a stated class of securities as collateral. Any of a variety of securities is acceptable because the lender is concerned Repos come in two flavors: general collateral repos (used to borrow money) and special collateral repos (used to borrow securities). primarily with earning interest on its money and having possession of liquid assets that can be sold quickly in the event of a default by the borrower. Interest rates on overnight general collateral repos are usually quite close to rates on overnight loans in the federal funds market. This reflects the essential character of a general collateral repo as a device for borrowing and lending money. Repo rates for the most liquid and creditworthy collateral, Treasury securities, are lowest. Repo rates for other classes of collateral, such as fixed-income securities issued by a federal agency or mortgage-backed securities issued by a government sponsored enterprise, are somewhat higher. Special collateral repos: A special collateral repo is a repo in which the lender of funds designates a particular security as the only acceptable collateral.12 Treasury market participants Funds at 3 percent Participant borrowing money at 3 percent against general collateral Starting leg of general collateral repurchase agreement Note: For simplicity, the separate roles of custodians are not shown explicitly. commonly lend money on special collateral repos in order to borrow specific securities that they need. The interest rate on a special collateral repo is called a “specials rate.” The owner of a Treasury security that other market participants want to borrow may be incentivized to lend the security if that owner is offered an opportunity to borrow money at a specials rate less than the Treasury general collateral repo rate. For example, if the rate on a special collateral repo involving bond B is 2 percent and the general collateral repo rate is 3 percent, an investor who owns bond B can earn a 100 basis point spread by lending the bond and borrowing money on a special collateral repo and then relending the money on a general collateral repo (Exhibit 4). The difference between the general collateral repo rate for Treasury securities and the special collateral repo rate for a particular Treasury security is a measure of the “specialness” of the security and is commonly called the security’s “specialness spread.” We show below that a security’s specialness spread is exactly the opportunity cost of borrowing the security to avoid or cure a settlement fail. 12 See Duffie (1996), Keane (1996), Jordan and Jordan (1997), Fisher (2002), and Fleming and Garbade (2002). FRBNY Economic Policy Review / October 2010 51 3.2 Incentives, prior to May 2009, to Borrow Securities to Avoid or Cure a Settlement Fail Prior to May 2009, sellers of Treasury securities, including short sellers, borrowed securities to avoid or cure settlement fails primarily because they did not get paid until they delivered the securities that they had sold. Prior to May 2009, market participants usually quantified the cost to a seller of a settlement fail in Treasury securities as the overnight Treasury general collateral repo rate—the rate the seller could have Prior to May 2009, a seller had an incentive to avoid failing to deliver a security (by borrowing the security on a special collateral repo and delivering the borrowed security) as long as the cost of borrowing the security was less than the cost of failing. earned on a riskless overnight investment of the sale proceeds that it did not receive. (It should be noted, however, that even prior to May 2009, the cost of a settlement fail was not limited to foregone interest earnings. Settlement fails also expose market participants to the risk of counterparty insolvency and can lead to increased capital charges for some participants. These other costs are discussed in Box 2.) A seller who does not have the securities needed to settle a sale can avoid failing by borrowing (on a special collateral repo) the securities that it needs. However, borrowing securities is not costless because the borrower has to lend money (on the special collateral repo) at a rate lower than the general collateral repo rate that it could have earned on the money. The cost of borrowing securities to avoid a fail in Treasury securities may be quantified as the difference between the overnight Treasury general collateral repo rate (the rate the borrower could have earned on its money) and the overnight special collateral repo rate on the borrowed securities (the rate the borrower actually earns on its money)—that is, the securities’ specialness spread. Prior to May 2009, a seller had an incentive to avoid failing to deliver a security (by borrowing the security on a special collateral repo and delivering the borrowed security) as long as the cost of borrowing the security was less than the cost of failing. This was certainly the case if the specialness spread for the security was less than the general collateral repo rate or, equivalently, if the special collateral repo rate for the security 52 The Introduction of the TMPG Fails Charge Box 2 Other Costs of Settlement Fails Settlement fails impose risks on both buyers and sellers and can lead to increased capital charges for some market participants. A buyer who fails to receive securities faces the risk that the seller might become insolvent before the transaction is settled and that, to replace the securities, it will have to pay more than the price negotiated with the insolvent seller. Conversely, a seller who fails to deliver securities faces the risk that the buyer might become insolvent before the transaction is settled and that the seller will then have to sell its securities to someone else at a price lower than the price negotiated with the insolvent buyer. These risks may be small for fails that are no more than a day or two old (because securities prices usually do not change much from day to day), but they can become significant when fails persist for weeks or months. In view of the greater risks faced by both buyers and sellers during the life of a long-outstanding, or “aged,” fail, some market participants are required to absorb incremental capital charges when a settlement fail has been outstanding for more than a few weeks.a Such capital charges can drain capital from more rewarding activities and limit balance sheet capacity, thereby imposing opportunity costs on market participants—and if sufficiently widespread and chronic can threaten overall market functioning.b a See, for example, the net capital rule of the Securities and Exchange Commission, Code of Federal Regulations, Chapter 17, Section 15c3-1. b See, for example, “Minutes of the Meeting of the Treasury Borrowing Advisory Committee of the Bond Market Association, November 4, 2003,” November 5, 2003, available at http://www.ustreas.gov/press/releases/ js933.htm (“While the situation is much improved since this past summer, members commented that fails were still at an elevated level which does hurt general market liquidity because dealers are forced to reduce their market making activities as the fails take up space on their balance sheets.”). One Treasury official suggested that opportunity costs resulting from higher capital charges might not be all bad. See “Remarks by Jeff Huther, Director of the Office of Debt Management, to the Bond Market Association’s Annual Meeting,” April 22, 2004, available at http:// www.treas.gov/press/releases/js1455.htm (noting that “capital charges resulting from chronic—widespread and persistent—fails soak up dealer capital that might otherwise be used to support profit-making activities, thereby focusing management attention on the underlying fails problem and incentivizing managers to remedy the situation.”). was greater than zero.13 As long as a seller could earn more than a de minimis amount of interest on a special collateral repo, it made economic sense to lend the money, earn the interest, and avoid the fail. Using economic terminology, let Rgc denote the general collateral repo rate and Rsp denote the special collateral repo rate for the security. The specialness spread Rgc – Rsp will be less then Rgc if and only if Rsp is greater than zero. 13 3.3 Equilibrium in the Market for Special Collateral Repos The market for borrowing and lending a particular Treasury security comes into equilibrium as a result of fluctuations in the special collateral repo rate for the security relative to the Treasury general collateral repo rate. If the demand to borrow the security is modest relative to the supply available for lending, a market participant seeking to borrow the security will usually be able to lend its money at a specials rate no lower than about 15 to 25 basis points below the general collateral repo rate. However, if the demand to borrow the security expands, some borrowers (in order to avoid failing) will be willing to accept less interest on the money they lend. Downward pressure on the specials rate for the security relative to the general collateral repo rate makes lending the security more remunerative, thereby attracting additional lenders. It may also ration some borrowers out of the market, particularly short sellers who decide to liquidate their short positions rather than continue to finance those positions on special collateral repos earning lower rates of interest. The collateral market will return to equilibrium, that is, to a state where the quantity of the security sought to be borrowed at the prevailing specials rate equals the quantity of the security available for lending at that rate, when the lower specials rate has attracted enough new lenders and/or rationed enough borrowers out of the market. 4. Federal Reserve and Industry Efforts to Mitigate Settlement Fails Treasury market participants have an interest in mitigating settlement fails in order to limit their net interest expenses as well as their exposure to the risk of counterparty insolvency— a risk explained in Box 2. The Federal Reserve has a separate interest in mitigating settlement fails to maintain the liquidity and efficiency of the market in which it conducts open market operations. (A high volume of fails can lead market participants to reduce, or even withdraw from, their normal activities. Such activities include dealers making markets for customers, investors lending securities to dealers to facilitate settlement of dealer sales, and arbitrageurs seeking to exploit, and thereby eliminate, price relationships that present abnormal profit opportunities.) Since 1969, the Federal Reserve has sought to mitigate settlement fails by lending Treasury securities to primary dealers to facilitate settlement of dealer sales.14 (However, the Federal Reserve lends against collateral, rather than cash, to insulate the supply of reserves available to the banking system from securities lending operations.) Pursuant to the terms and conditions of the lending program in effect in mid-2008,15 each business day at noon the Federal Reserve Bank of New York offered to lend on an overnight basis up to 90 percent of the amount of each Treasury security beneficially owned in the Federal Reserve’s System Open Market Account (SOMA), subject to an upper limit of the amount of an issue actually in Treasury market participants have an interest in mitigating settlement fails in order to limit their net interest expenses as well as their exposure to the risk of counterparty insolvency. the account.16 Primary dealers bid for a security by specifying the quantity desired (in increments of $1 million) and the overnight loan fee they were willing to pay, expressed as a rate per annum, subject to a minimum fee of 50 basis points.17 Bidding closed at 12:15 p.m. and loans were awarded to the highest bidders at their bid rates18 until all of the securities available for lending were allocated or all of the bidders had been satisfied.19 During the first six months of 2008, the Federal Reserve Bank of New York lent an average of $12.2 billion of securities per day, distributed over an average of twenty-three different issues, at an average loan fee of 61 basis points per annum. 14 In authorizing the loan of Treasury securities from the System Open Market Account in 1969, the Federal Open Market Committee stated that the action “was taken after the Manager [of the System Open Market Account] had advised that the problem of delivery failures in the Government securities market had worsened significantly over the past year, partly because private facilities for lending such securities had become inadequate; that delivery failures were markedly impairing the performance of the market; and that the functioning of the market would be improved if securities held in the System Open Market Account could be lent, for the express purpose of avoiding delivery failures, to Government securities dealers doing business with the Federal Reserve Bank of New York . . . .” (Federal Reserve Bulletin, January 1970, p. 32). 15 “SOMA Securities Lending Program Terms and Conditions (Revised),” Federal Reserve Bank of New York, August 22, 2008, available at http:// www.newyorkfed.org/newsevetns/news/ markets/2008/an082208.html. Details of the history of the Fed’s lending program appear in Fleming and Garbade (2007). 16 In order to avoid failing itself, the Fed does not agree to lend securities unless it has actual possession of the securities at the time of an auction. Thus, it will not agree to lend securities that it lent the preceding business day and that have not yet been returned. 17 A loan fee for a security is approximately equivalent to the security’s specialness spread. See Fleming and Garbade (2007). The minimum fee avoids crowding out private lenders when security loan markets are functioning normally, but it has been reduced to nearly zero when those markets are not functioning well. 18 The auction for each security is a discriminating, or multiple-price, auction. FRBNY Economic Policy Review / October 2010 53 Treasury market participants have also acted cooperatively to mitigate settlement fails and otherwise reduce the cost of settling transactions. Between 1986 and 1988, dealers in Treasury securities organized the Government Securities Clearing Corporation (GSCC) to serve as a central counterparty in interdealer transactions in Treasury and related securities. As explained earlier, multilateral net settlement through a central counterparty economizes on the quantity of securities that have to be delivered to settle a given volume of transactions. GSCC also implemented a trade confirmation protocol that essentially eliminated interdealer fails due to miscommunication between dealers, as well as a procedure for marking failed trades to current market prices that materially reduced the consequences of counterparty insolvency. The GSCC extended its net settlement system to include Treasury auction takedowns in 1994 and repurchase agreements in 1995.20 In 2002, GSCC became a wholly owned subsidiary of Depository Trust & Clearing Corporation and was renamed the Fixed Income Clearing Corporation. 5. Chronic Settlement Fails Demand to borrow a security (relative to the supply available for lending) can sometimes be large enough to drive the specials rate for the security down to near zero. Prior to May 2009, sellers would then become largely indifferent between a) failing and b) borrowing the security to avoid failing. In this extreme case, any unsatisfied demand to borrow would spill over into fails. Fails could expand further if security lenders, observing a growing incidence of settlement fails, declined to continue lending out of concern that their securities may not be returned on a timely basis.21 Fails could expand still further if, as explained in Box 3, market participants concluded that they could acquire a cheap option on a future increase in a specials rate by contracting to sell a security in a special collateral repo and then strategically failing to deliver the security. More generally, settlement fails could become chronic when the specials rate for a security was driven down to near zero. Three episodes of chronic fails have been described in the literature: in May and June of 1986,22 following the terrorist attack on the World Trade Center on September 11, 2001,23 and during the summer of 2003.24 Box 3 Strategic Fails When the specials rate for a security is close to zero in a market without a fails charge convention, a market participant with no position in the security may sometimes agree to lend the security on a term repurchase agreement and then fail, intentionally, on the starting leg of the repo. Suppose, for example, the three-week specials rate for a fiveyear note is 10 basis points and that XYZ Co. believes the specials rate will be 50 basis points in one week. If XYZ contracts (in the specials market) to borrow $50 million for three weeks against lending the note, it will owe interest of $2,917 at the end of three weeks.a It will owe this amount even if it fails to deliver the note any time during the three-week interval. XYZ Co. has effectively purchased (for $2,917, payable in three weeks) an option on an exchange of $50 million for the five-year note at any time during the next three weeks for the balance of the three-week interval. XYZ could choose to let its option expire unexercised and simply pay the $2,917 premium at the end of three weeks. However, if XYZ Co.’s expectations prove correct, it can exercise the option after one week by borrowing the five-year note for two weeks against lending $50 million at 50 basis points (earning interest of $9,722b) and delivering the note in (delayed) settlement of its earlier negotiated three-week repurchase agreement. The $50 million received from delivering the note funds the loan that allows XYZ Co. to borrow the note, and XYZ Co. has net interest earnings of $6,805 ($6,805 = $9,722 interest income, less $2,917 interest expense). More generally, a very low specials rate presents an opportunity to speculate on an increase in the rate—by lending on, and then failing on, a special collateral repurchase agreement—with limited downside exposure.c In the limit, a repo rate of zero may be viewed by some participants as a “risk-free” opportunity to intentionally fail and either profit or break even. Such practices can lead to an increase in aggregate settlement fails and the associated indirect costs discussed in Box 2. a $2,917 = (21/360) 0.10 percent per annum $50 million. b $9,722 = (14/360) 0.50 percent per annum $50 million. c Strategic fails are noted in Fleming and Garbade (2002, p. 47), Fleming and Garbade (2004, pp. 3-4), and in “Remarks by Jeff Huther, Director of the Office of Debt Management to the Bond Market Association’s Annual Meeting,” April 22, 2004, available at http://www.treas.gov/press/releases/ js1455.htm. 19 Awards were subject to three limitations: 1) no dealer could have outstanding borrowings of more than 25 percent of the amount of an issue beneficially owned in the System Open Market Account, 2) no dealer could have outstanding borrowings of more than $750 million of any single issue, and 3) no dealer could have outstanding borrowings of more than $3 billion of securities in total. 20 See Garbade and Ingber (2005) and Ingber (2006). 21 This can quickly lead to a self-reinforcing and destabilizing cycle, with lenders withdrawing collateral out of a concern that borrowers may fail to return the securities, thereby increasing the incidence of settlement fails and triggering further collateral withdrawals. 54 The Introduction of the TMPG Fails Charge 22 Cornell and Shapiro (1989). Cornell and Shapiro do not discuss fails directly, but do document a near-zero specials rate for the 9 1/4 percent Treasury bond of February 2016 and discuss the reasons for that low rate. The existence of widespread settlement fails in the issue was common knowledge among market participants at the time. 23 Fleming and Garbade (2002). 24 Fleming and Garbade (2004). See also Fleming and Garbade (2005). 5.1 The 1986 Episode An “on-the-run” Treasury security is the most recently auctioned security in a given series, such as the most recently auctioned six-month bill or ten-year note. In late April 1986, dealers began to sell the on-the-run thirty-year Treasury bond (the 9 1/4 percent bond of February 2016) short in anticipation of bidding for the new thirty-year bond that would be announced on April 30 and auctioned on May 8 for settlement on May 15.25 Such short selling in advance of an auction announcement was normal and customary dealer behavior.26 However, dealers soon found themselves unable to borrow enough 9 1/4 percent bonds to finance their short positions, in part because a significant quantity of the bonds was owned by investors who declined to lend.27 Strong dealer demand and limited supply combined to drive the special collateral repo rate for the bonds down to about 5 basis points, and dealers began to fail on their settlement obligations. Failing, however, was expensive because the Treasury general collateral repo rate was about 6.75 percent, so dealers with short positions had an economic incentive to cure their fails another way: by buying (rather than borrowing) the 9 1/4 percent Treasury bonds and delivering the purchased bonds—thereby closing out their short positions. They bid up the price of the 9 1/4 percent bonds, relative to the prices of other Treasury bonds with similar maturities, until the higher price induced holders to sell the 9 1/4 percent bonds and replace them with higher yielding substitutes, thereby allowing dealers to cover their short positions.28 25 Department of the Treasury, Securities and Exchange Commission, and Board of Governors of the Federal Reserve System (1992, p. B-1, footnote 1) (“participants sold the outstanding 9 1/4 percent bond . . . to prepare for the roll into the WI [“when-issued”] thirty-year bond.”). Following the announcement of the forthcoming auction, dealers planned to buy the on-the-run thirtyyear bond (thereby covering their previous short sales of that bond) against selling the WI thirty-year bond short. The transactions would leave them with short positions in the WI bond that they could cover in the auction. 26 Department of the Treasury, Securities and Exchange Commission, and Board of Governors of the Federal Reserve System (1992, p. 10, footnote 11) (“dealers . . . sold [the 9 1/4 percent bond] short as part of a trading strategy that had worked in the past as they prepared to bid for a new thirty-year bond.”). The sequence of shorting the on-the-run thirty-year, rolling the short into the WI thirty-year, and then bidding to buy the WI thirty-year in the auction was part of the process whereby dealers distributed new bonds to market participants. 27 Department of the Treasury, Securities and Exchange Commission, and Board of Governors of the Federal Reserve System (1992, p. 10, footnote 11; p. B-1, footnote 1) (“some institutional investors did not make the [9 1/4 percent bonds] available to the repo market” and “securities needed to [finance] short positions were not readily available to the repo market.”). 28 Cornell and Shapiro (1989, pp. 303-4) suggest that the 9 1/4 percent bond of February 2016 was overvalued by as much as 7 percent of principal value compared with one close substitute (the 9 7/8 percent bond of November 2015). 5.2 The 2001 Episode The 2001 fails episode was attributable, in the first instance, to the terrorist attack on the World Trade Center on September 11, 2001. The attack destroyed the offices of several interdealer brokers and impaired telecommunication services throughout lower Manhattan. GSCC recorded $266 billion in interdealer settlement fails on September 11 and $440 billion in interdealer fails on September 12.29 Sellers tried to borrow the securities Settlement fails began to shrink to more normal levels after the Treasury reopened the on-the-run ten-year note in an extraordinary unscheduled auction offering on Thursday, October 4, and after officials indicated that they might reopen the on-the-run five-year note as well. needed to cure their fails but holders realized that, in view of the severe operational problems, their securities might not be returned on a timely basis and they consequently declined to lend.30 The contraction in the supply of collateral pushed specials rates to near zero and settlement fails remained elevated. Daily average fails in Treasury securities reported by primary dealers to the Federal Reserve31 reached $200 billion per day during the week of September 13-19 and continued high through early October. Settlement fails were particularly high for the on-the-run five-year note (the 4 5/8 percent note 29 Fleming and Garbade (2002, p. 46). See “After Attack, Settlement Woes Still Clogging Repo Market,” Dow Jones Newswires, September 26, 2001, 9:05 (noting “a general reluctance among large portfolios to lend their securities” and “in a chain reaction, the fear of failing trades is ‘causing portfolio managers, securities lending desks and foreign central banks to hold even tighter on to their collateral,’ which is exacerbating the situation . . . .”); “Treasury Market is Faced with Incomplete Trades,” Wall Street Journal, October 3, 2001, p. C10 (“Because of the rate of fails . . . dealers are reluctant to use their securities as collateral. They are worried that they might not have securities delivered to them . . . . ”); “U.S. Sells $6 Billion in 10Year Notes to Help Overcome Shortage,” Bloomberg News, October 4, 2001, 13:16 (quoting Peter Fisher, Under Secretary of the Treasury for Domestic Finance, as saying that “the cause of the fails is [in part] the result of . . . reluctance by institutional investors to lend into a market that is suffering from extraordinarily high fails levels.”); and “U.S. Acts on Shortage of Treasuries,” New York Times, October 5, 2001, p. C1 (“With the prospect that securities might not be returned, both dealers and large investors have become unwilling to lend them in the repo market.”). 31 Fleming and Garbade (2005) describe the settlement fails data reported by primary dealers to the Federal Reserve. Unless otherwise noted, this article measures settlement fails as the daily average over weekly intervals of the average of cumulative primary dealer fails to receive Treasury securities during a week and primary dealer fails to deliver Treasury securities over the same week. The Federal Reserve does not publish data on settlement fails on a dayby-day basis. 30 FRBNY Economic Policy Review / October 2010 55 of May 2006) and the on-the-run ten-year note (the 5 percent note of August 2011). Settlement fails began to shrink to more normal levels after the Treasury reopened the on-the-run ten-year note in an extraordinary unscheduled auction offering on Thursday, October 4, and after officials indicated that they might reopen the on-the-run five-year note as well.32 Peter Fisher, the Under Secretary of the Treasury for Domestic Finance, stated that the Treasury reopened the ten-year note “to reduce the risk that . . . settlement problems turn into a much bigger problem for the Treasury market . . . .”33 Fisher went on to observe that “we have something that is self-compounding. There is some point at which your fails pile up, and that is the point at which you damage the price-discovery process and the smooth operating of the Treasury market.” The actions of Treasury officials convinced market participants that the Treasury would take unprecedented steps to facilitate settlements and maintain market liquidity. Holders of the on-the-run five- and ten-year notes began to make the notes available, and the level of fails subsided.34 $25 billion per day during the week ending June 18 to $103 billion per day during the week ending July 2, and topped out at $232 billion per day during the week ending August 20. Settlement fails persisted for months36 and were not fully resolved until the end of the year, following an offering of a new series of ten-year notes in November.37 5.4 Proposals to Mitigate Chronic Settlement Fails The 2003 episode had a strong impact on the thinking of market participants. Unlike the 1986 episode, which was shortlived and quickly forgotten, and unlike the 2001 episode, which clearly stemmed from unusual circumstances, the 2003 episode was lengthy, large-scale, and stemmed from a marketplace The 2003 [fails] episode raised the question of whether something should be done, by government officials or by private sector market participants, to mitigate chronic fails. 5.3 The 2003 Episode The 2003 fails episode was attributable, in the first instance, to a heavy volume of short sales of the on-the-run ten-year note (the 3 5/8 percent note of May 2013) in late June 2003 by market participants seeking to hedge their interest rate risk on long positions in other fixed-income securities.35 The short sales created an unusually large demand to borrow the note that drove the specials rate for the note down to zero, after which the residual, unsatisfied demand spilled over into fails. The fails became chronic when investors began to withdraw from lending the note. Daily average fails in Treasury securities reported by primary dealers to the Federal Reserve went from 32 “U.S. Sells $6 Billion in 10-Year Notes to Help Overcome Shortage,” Bloomberg News, October 4, 2001, 13:16, and “U.S. Acts on Shortage of Treasuries,” New York Times, October 5, 2001, p. C1 (both quoting the Under Secretary of the Treasury for Domestic Finance as saying that the Treasury might reopen the five-year note in the next week). 33 “U.S. Acts on Shortage of Treasuries,” New York Times, October 5, 2001, p. C1. 34 See, for example, “Remedial Reopenings and the Treasury Supply Outlook,” Money Market Observer, October 8, 2001 (“Dealers reported a dramatic reduction in the volume of fails on [October 5] after the settlement of the additional $6 billion of ten-year notes . . . .”), and “Another Emergency Treasurys Sale Looks Unlikely as Shortages that Hamstrung ‘Repo Market’ Ease,” Wall Street Journal, October 11, 2001, p. C17. 35 “Supply Dries Up Following Fall in Prices,” Financial Times, August 23, 2003, p. 27 (reporting that “Demand for Treasuries from some quarters has also risen as prices have fallen because many institutions want to borrow the securities and ‘short’ them in the expectation that prices will continue to drop. Traders say hedged positions for the [on-the-run ten-year note] now exceed the amount of Treasury securities available.”). 56 The Introduction of the TMPG Fails Charge activity—hedging—that was a very ordinary occurrence. The 2003 episode raised the question of whether something should be done, by government officials or by private sector market participants, to mitigate chronic fails. The key difference between the 1986 and 2003 episodes was the level of the Treasury general collateral repo rate. In May 1986, the overnight general collateral repo rate was about 6.75 percent. That made it costly to continue to fail even after the special collateral repo rate on the 9 1/4 percent bonds of February 2016 had been driven down to near zero and the economic incentive to avoid failing by borrowing the bonds had been eliminated. The high cost of failing incentivized short sellers to cover their short positions with outright purchases, and they bid up the price of the 9 1/4 percent bonds to a level that gave holders an economic incentive to swap out of the issue and into higher yielding substitutes. In mid-2003, however, the overnight Treasury general collateral repo rate was about 1 percent, so the cost of failing was modest. Short sellers had little incentive to cover their 36 See, for example, “Treasury Issue Remains a Headache,” Wall Street Journal, November 17, 2003, p. C13. 37 “California Standoff Dims Prospects,” Wall Street Journal, December 9, 2003, p. C17 (reporting “progress for the . . . May ten-year note, which traders said appeared to be emerging from six months of gridlock, thanks to supply that entered the market last week. The note was trading in positive territory in the repurchase-agreement market. For months, it had been stuck at 0% . . . .”). short positions with outright purchases after demand to borrow the May 2013 ten-year note had driven the specials rate on the note down to near zero and eliminated the incentive to borrow the note to avoid failing. Market participants and government officials learned from the 2003 episode that settlement fails were liable to become chronic quickly when short-term interest rates are low,38 and they began to contemplate institutional innovations to avoid, or at least mitigate, chronic settlement fails. Most discussions centered around three possibilities: unpredictable reopening program would raise borrowing costs over the long run.42 A Treasury lending facility: Like reopenings, a Treasury lending facility would involve additional issuance from the Treasury. Unlike reopenings, a Treasury lending facility would increase supply on only a temporary basis. Such a facility was put forth as a “straw man” in a Treasury white paper published Market participants and government officials learned from the 2003 episode that settlement fails were liable to become chronic quickly when short-term interest rates are low. • a regular program to reopen an issue when settlement fails in the issue become chronic, • a securities lending facility run by the Treasury Department, and • a fee to be paid by failing sellers to their counterparties to incentivize the sellers to resolve their fails. Reopenings: Reopening an issue to alleviate chronic fails was exactly what the Treasury did when it reopened the on-the-run ten-year note on October 4, 2001.39 However, Treasury officials were reluctant to institutionalize reopenings as a device to mitigate chronic fails. Three months after the 2001 reopening, Under Secretary Fisher told market participants that while “it would be imprudent of me to say that the Treasury will never again hold such an auction . . . you should not count on it, you should not expect it . . . .”40 The problem was that reopenings in response to chronic fails ran counter to “regular and predictable” issuance, a cornerstone of Treasury debt management since the 1970s.41 Treasury officials were concerned that the uncertainties engendered by an 38 See, for example, “Remarks by Jeff Huther, Director of the Office of Debt Management, to the Bond Market Association’s Annual Meeting,” Department of the Treasury, April 22, 2004, available at http://www.treas.gov/ press/releases/js1455.htm (“The heart of the pricing problem last year was, unquestionably, the low federal funds rate and the consequently low ceiling on the cost of financing a short position.”); Department of the Treasury (2006, p. 26, p. 174, footnote 2) (“The potential for chronic fails episodes thus increases in a very low interest rate environment such as that prevailing during the summer of 2003.”); and “Statement of Under Secretary for Domestic Finance Randal K. Quarles to Bond Market Association Annual Meeting,” Department of the Treasury, May 19, 2006, available at http://www.treas.gov/ press/releases/js4274.htm (“When the central bank wishes to establish very low short-term rates, the maximum degree of specialness will be quite small. During these periods, we might expect to see greater incidence of fails episodes because the cost of failing is low.”). 39 That reopening was not the first time the Treasury increased the supply of a security in response to unusual market conditions. On November 3, 1992, the Treasury announced that it would reopen the 6 3/8 percent note of August 2002, at that time the on-the-run ten-year note, “in order to alleviate an acute, protracted shortage” of the note. See “Treasury November Quarterly Financing,” Office of Financing, Department of the Treasury, November 3, 1992. 40 “Remarks by Peter R. Fisher, Under Secretary of the Treasury for Domestic Finance, Before the Bond Market Association Legal and Compliance Conference,” January 8, 2002, available at http://www.treas.gov/press/releases/ po.906.htm. in May 2006.43 The white paper was written to stimulate public discussion of mechanisms to make available “an additional, temporary supply of Treasury securities on rare occasions when market shortages threaten to impair the functioning of the market for Treasury securities and broader financial markets . . .” However, Treasury officials questioned whether the Secretary of the Treasury has statutory authority to issue securities on a temporary basis to alleviate chronic settlement fails. Federal law provides that “the Secretary of the Treasury may borrow on the credit of the United States Government amounts necessary for expenditures authorized by law and may issue bonds of the Government for the amounts borrowed.”44 Similar provisions authorize the issuance of notes and bills.45 The 2006 Treasury white paper suggested that “the Treasury would likely need to pursue new authority to issue securities for the purpose of securities lending. . . .”46 41 Garbade (2007) describes the emergence of “regular and predictable” as a Treasury debt management strategy. 42 See “Remarks of Undersecretary of the Treasury Peter R. Fisher to the Futures Industry Association, Boca Raton, Florida,” March 14, 2002, available at http://www.treas.gov/press/releases/po1098.htm. 43 Department of the Treasury (2006). The public responses to the white paper are available at http://www.ustreas.gov/offices/domestic-finance/debtmanagement/slf-comments.pdf. See also Garbade and Kambhu (2005). A Treasury lending facility was also recommended by the Treasury Borrowing Advisory Committee following the chronic fails of late September and early October 2001; see “Report to the Secretary of the Treasury from the Treasury Advisory Committee of the Bond Market Association,” October 30, 2001, available at http://www.ustreas.gov/offices/domestic-finance/debtmanagement/adv-com/reports/rpt-2001-q4.pdf (“members overwhelmingly felt that Treasury should expand their ability to enhance liquidity in the Treasury market. To accomplish this, they could set up a repo facility to help alleviate protracted shortages, in particular, large and persistent fails . . . .”). 44 31 U.S.C. § 3102 (2010). 45 See 31 U.S.C. § 3103-3104 (2010). 46 Department of the Treasury (2006, pp. 26, 178). FRBNY Economic Policy Review / October 2010 57 A fails charge: In 2002, two economists at the Federal Reserve Bank of New York suggested that “chronic fails can also be alleviated by increasing the cost of failing with a penalty fee.”47 The economists noted that such a fee would give sellers an economic incentive to borrow securities to avoid failing even when the special collateral repo rate for the securities was close to zero. They further noted that a fails charge might lead market participants to borrow securities against lending money at negative specials rates (in order to avoid the fails charge) and that such negative specials rates could attract additional securities lenders (because they would receive, rather than pay, interest on the money they borrowed against lending securities). The economists suggested that a fails charge might be set at some threshold rate minus the general collateral repo rate, with a minimum of zero.48 The fails charge would be above zero only if the general collateral repo rate was below the threshold rate and would not be higher than what was necessary to bring the total cost of failing to the threshold rate. (For example, if the threshold rate is 5 percent and the general collateral repo rate is 3 percent, the fails charge would be 2 percent and the total cost of failing would be 5 percent.) They further suggested that the fails charge could be instituted through a “goodpractice” recommendation of the Bond Market Association.49 The economists noted that a fails charge could be implemented implicitly by reducing the invoice price on a transaction each day the seller fails—a material departure from the existing convention of deferring settlement at an unchanged invoice price—but observed that “the operational burden of changing an invoice price following a delay in settlement would undoubtedly be substantial.”50 5.5 Inaction prior to the Insolvency of Lehman Following the 2003 episode of chronic settlement fails, both government officials and private sector market participants understood that chronic fails are prone to blossom in an environment of low interest rates. Several parties had identified ways to address the problem, but each of the suggestions had a material deficiency. Treasury officials asked private sector participants to address the problem, but nothing substantive came of their requests.51 No significant progress was made with 47 Fleming and Garbade (2002, p. 52). Fleming and Garbade (2002, p. 53). 49 The Bond Market Association joined with the Securities Industry Association in 2006 to form the Securities Industry and Financial Markets Association. 50 Fleming and Garbade (2002, p. 52). 48 58 The Introduction of the TMPG Fails Charge respect to addressing the problem of chronic fails before the insolvency of Lehman in the fall of 2008. 6. Chronic Settlement Fails in the Wake of the Insolvency of Lehman The announcement, early in the morning of Monday, September 15, 2008, that Lehman was insolvent triggered a “flight to safety” that, by the close of trading that day, pushed the yield on four-week Treasury bills down to 36 basis points, The first response to the rising tide of settlement fails was the decision of the Federal Reserve to relax the terms of its securities lending program. 100 basis points lower than the yield on the preceding Friday. Yields on longer term bills also moved sharply lower. By the close of trading on Wednesday, September 17, yields on fourweek bills were down to 7 basis points. Over the balance of the month, four-week-bill yields fluctuated between about 10 basis points and 100 basis points—well below the 1.50 to 1.85 percent range that had prevailed since the beginning of August (Chart 1). Greater demand for high-quality, short-term debt also drove down repo rates on Treasury collateral. The overnight Treasury general collateral repo rate averaged 90 basis points between September 15 and September 30, well below the 2 percent level that had prevailed during the preceding six weeks (Chart 2). In the wake of Lehman’s insolvency and in the midst of the ensuing flight to safety, investors became increasingly reluctant to lend Treasury securities.52 Unable to replace their maturing borrowings, dealers began to fail on their delivery obligations. 51 “Minutes of the Meeting of the Treasury Borrowing Advisory Committee,” November 4, 2008, available at http://www.treas.gov/press/releases/ hp1239.htm (stating that “Since November 2003, Treasury has repeatedly asked the private sector to address [the fails] issue proactively. On several occasions, market participants have emphatically stated that they would resolve the situation without government intervention, but such steps have not been implemented.”). See also Wrightson, Federal Reserve Data, October 17, 2008 (stating that “the repo market has managed to fend off regulatory reform in past cycles.”), and “The Treasury Market Reaches Breaking Point,” Euromoney, December 1, 2008 (quoting a former Treasury employee as saying that “It was politically difficult to convince the market to put a stop to fails to deliver in treasuries. There were some forceful voices insisting that if the Treasury got involved, they would take the incentives out of the specials market altogether. Those making their living as specialist dealers, as well as those making a living shorting securities outright, were worried about potential supply changes which would eliminate trading opportunities for them.”). Chart 1 Chart 2 Yields on Four-Week Treasury Bills Target Federal Funds Rates and Rates on Overnight General Collateral Repurchase Agreements on Treasury Securities August to October 2008 Percent 2.0 August to October 2008 2.0 2.5 Percent 1.5 2.5 1.5 Repo rates on dates prior to September 15 Dates prior to September 15 2.0 2.0 1.0 1.5 1.0 1.5 Repo rates on dates after September 14 0.5 0.5 Target federal funds rate Dates after September 14 0.5 0.0 0.5 0 August September October Source: Federal Reserve Statistical Release H.15. The fails persisted because the low general collateral repo rate left sellers with little incentive to cure the fails. Primary dealer settlement fails in Treasury securities mushroomed to an average of $253 billion per day during the week of Thursday, September 18, to Wednesday, September 24—far in excess of the level that had prevailed in August and the first half of September (Chart 3). And unlike earlier episodes, fails in the wake of Lehman’s insolvency were not concentrated in one or two issues; rather, they involved securities across the entire yield curve. The first response to the rising tide of settlement fails was the decision of the Federal Reserve to relax the terms of its securities lending program. As shown in Table 1, on Tuesday, September 23, the Fed raised the limit on total borrowings by a single dealer from $3 billion to $4 billion. Loans to primary dealers from the SOMA portfolio reached new heights but primary dealer settlement fails continued to rise, averaging $342 billion per day over the interval from September 25 to October 8 (Chart 3). On Wednesday, October 8, both the Federal Reserve and the Treasury acted in response to the continuing crisis. The Federal Reserve further eased the terms of its securities lending program by reducing the minimum loan fee from 50 basis 52 1.0 1.0 “Demand for Short-Term Treasury Debt Puts a Crimp in World-Wide Supply,” Wall Street Journal, September 25, 2008, p. C1 (reporting that “some foreign central-bank officials … are reluctant to lend out their safest collateral—U.S. Treasurys.”); “U.S. Treasury Steps Up Debt Sales to Reduce Shortages (Update 2),” Bloomberg.com, October 8, 2008, 12:43 EDT (quoting the head of interest rate strategy at Credit Suisse Securities as saying that “people are so nervous about the financial crisis that they’re holding on to their collateral and not lending it out.”); and “More Treasury Bonds on Way to Ease Crisis,” Wall Street Journal, October 9, 2008, p. A6 (reporting that “investors have been unwilling to lend Treasury securities to other market participants.”). 0.0 0 August September October November Sources: Federal Reserve Statistical Release H.15; Board of Governors of the Federal Reserve System. points to 10 basis points and by expanding the limit on total borrowings by a single dealer to $5 billion (Table 1). Treasury officials took the unprecedented step of reopening four off-therun Treasury notes, announcing at 10:40 a.m. that they would “reopen multiple securities which have created severe dislocations in the market causing acute, protracted shortages.”53 Two of the reopened notes were auctioned later the same day (at 11:30 a.m. and 1:00 p.m., respectively), and the other two notes were auctioned the following day (Table 2). The decision to reopen a substantial amount ($10 billion each) of so many different notes made clear the scale of the fails problem; the decision to auction one note with less than an hour of notice and a second note with less than three hours of notice emphasized the urgency of the situation. Although the reopenings helped to mitigate settlement fails in the issues that were reopened, aggregate primary dealer Treasury fails continued to rise, reaching a daily average level of $379 billion per day over the interval from October 9 to October 22 (Chart 3). Comments to the effect that “Treasury market functioning remains impaired” and “the repo market is not functioning” became commonplace. On October 17, a widely read market letter remarked that “the breakdown in the clearing mechanism for the Treasury market is beginning to emerge as a top-tier policy concern.”54 53 U.S. Department of the Treasury, “Statement on Treasury Market Conditions and Debt Management Actions,” October 8, 2008, available at http://www.treas.gov/press/releases//hp1186.htm. 54 Wrightson, Federal Reserve Data, October 17, 2008. FRBNY Economic Policy Review / October 2010 59 Chart 3 Daily Average (Over Weekly Intervals) Primary Dealer Settlement Fails in Treasury Securities August to October 2008 Billions of dollars 400 350 Denotes business days 300 Denotes weekends and holidays 250 200 150 100 50 0 August September October Source: Federal Reserve Bank of New York. Notes: The first square marks the Lehman insolvency on Monday, September 15. The second square marks the effective date of a revision in the terms and conditions of the Federal Reserve System Open Market Account (SOMA) securities lending program on Tuesday, September 23. The third square marks the announcement of the surprise reopening of four Treasury notes on Wednesday, October 8, and the effective date of a further revision in the terms and conditions of the SOMA securities lending program on the same day. Table 1 Terms and Conditions of Federal Reserve Security Loan Auctions Limits on Outstanding Borrowings by a Single Dealer Theoretical Amount of a Single Issue Offered (Percentage of SOMA Holdings)a Minimum Loan Fee (Basis Points) 90 September 23, 2008 Per Issue Total 50 Lesser of $750 million and 25 percent of amount beneficially owned in SOMA portfolio $3 billion 90 50 Lesser of $750 million and 25 percent of amount beneficially owned in SOMA portfolio $4 billion October 8, 2008 90 10 Lesser of $750 million and 25 percent of amount beneficially owned in SOMA portfolio $5 billion December 18, 2008b 90 1 Lesser of $750 million and 25 percent of amount beneficially owned in SOMA portfolio $5 billion Effective Date Terms and conditions prior to Lehman insolvency November 26, 2007 Post-Lehman terms and conditions Source: Federal Reserve Bank of New York. Notes: SOMA is the Federal Reserve System Open Market Account. Entries in bold indicate a change in terms. a Amount actually offered is the lesser of the theoretical amount offered and the amount of the issue actually in the SOMA account at the time of an auction. Last revision prior to the end of 2008. b 60 The Introduction of the TMPG Fails Charge Table 2 Treasury Notes Reopened in October 2008 4 1/8 Percent Note Maturing May 15, 2015 4 1/4 Percent Note Maturing August 15, 2015 4 Percent Note Maturing February 15, 2015 3 1/2 Percent Note Maturing February 15, 2018 $10 billion $10 billion $10 billion $10 billion October 8, 2008, 11:30 a.m. October 8, 2008, 1:00 p.m. October 9, 2008, 11:30 a.m. October 9, 2008, 11:30 a.m. October 15, 2008 October 15, 2008 October 15, 2008 October 15, 2008 Amount bid competitively $12.1 billion $21.1 billion $23.7 billion $23.1 billion Closing market yield on October 7, 2008 (percent) 2.87 2.98 2.79 3.57 Auction yield (percent) 3.31 3.44 3.23 3.79 Closing market yield on October 9, 2008 (percent) 3.35 3.57 3.22 3.92 Amount offered Auction date and time Issue date Sources: U.S. Treasury Department; Wall Street Journal. Note: Over the interval from October 7 to October 9, the closing market yield on the on-the-run five-year note (the 3 1/8 percent note of September 30, 2013) rose from 2.47 percent to 2.79 percent and the closing market yield on the on-the-run ten-year note (the 4 percent note of August 2018) rose from 3.50 percent to 3.81 percent. 7. The TMPG Steps Up By mid-October 2008, Treasury and Federal Reserve officials and private sector market participants understood that the volume and persistence of settlement fails in Treasury securities was a major problem, but what could or should be done was far from obvious. The four reopenings had reduced fails in the reopened notes, but speculation over whether the Treasury would reopen other chronically failing issues was contributing to unwanted volatility in the prices of other Treasury securities. Additionally, there was some indication that the reopenings had not been well received. The first auction, of $10 billion of the 4 1/8 percent notes of May 2015, attracted only $12.1 billion of tenders, and the notes were sold at a price almost 3 points below where outstanding notes of the same series traded prior to the auction. An alternative approach was to revise the market convention of postponing—without any explicit penalty and at an unchanged invoice price—a seller’s obligation to deliver Treasury securities if the seller failed to deliver the securities on a scheduled settlement date. However, precisely because the treatment of settlement fails was a matter of market convention, rather than law or regulation, it could not be changed except through widespread adoption of an alternative convention. Fortuitously, in early 2007 the Federal Reserve Bank of New York had sponsored the organization of a new forum— the Treasury Market Practices Group—for discussing Treasury market practices and for advocating the adoption of practices deemed to be in the best interests of the market.55 The TMPG is a group of private sector market professionals committed to supporting the integrity and efficiency of the market for U.S. Treasury securities. Membership includes senior business managers and legal and compliance professionals from broker-dealer firms, banks, buy-side firms, The TMPG is a group of private sector market professionals committed to supporting the integrity and efficiency of the market for U.S. Treasury securities. and other organizations involved in Treasury market infrastructure. (Box 4 identifies the membership in October 2008.) The TMPG routinely meets about eight to ten times a year to discuss trading issues and best-practice recommendations for the Treasury market and publishes “Treasury Market Best Practices,” 56 a “living document” that aims to support 55 Federal Reserve Bank of New York, “Statement on Formation of PrivateSector Treasury Market Best Practices Group,” February 9, 2007, available at http://www.newyorkfed.org/newsevents/news/markets/2007/an070209.html. See also Federal Reserve Bank of New York, “Statement Regarding New York Fed Meeting with Primary Dealers,” November 6, 2006, available at http:// www.newyorkfed.org/newsevetns/news-archive/markets/2006/an061105.html. 56 Available at http://www.newyorkfed.org/tmpg/TMPG-bestpractices_033109.pdf. FRBNY Economic Policy Review / October 2010 61 Box 4 at a crossroad. . . . At this critical juncture it is incumbent that TMPG take the leadership position on this issue and work as a group to provide practical, real time solutions. . . . Our goal as members of this committee is to support the integrity and efficiency of the U.S. Government Treasury Market. . . . Membership of the Treasury Market Practices Group in October 2008 Thomas Wipf, Chair Morgan Stanley Fran Bermanzohn Goldman Sachs (last meeting in October 2008) Arthur Certosimo The Bank of New York Mellon Daniel Dufresne Citadel Investment Group, LLC Peter Economou State Street John Fath BTG Michael Haddad Caxton Associates (last meeting in January 2009) Curt Hollingsworth Fidelity Investments James Hraska Barclays Capital Murray Pozmanter Depository Trust & Clearing Corporation Gerald Pucci BlackRock John Roberts Barclays Capital Bill Santangelo Countrywide Securities Corp. (last meeting in October 2008) Peter Stebbing Reserve Bank of Australia Nancy Sullivan The Bank of New York Mellon Matthew Zames JPMorgan Chase Subsequent Additions prior to May 1, 2009 Glenn Hadden Goldman Sachs Stuart Wexler Merrill Lynch Treasury market integrity and efficiency. Best-practice recommendations include guidelines for promoting market liquidity, for integrating compliance and trading functions in a meaningful fashion, and for managing large positions in ways that avoid adverse consequences for market liquidity. TMPG practice guidance has also addressed the efficient clearing and settlement of trades. Thus, the TMPG was well positioned in October 2008 to provide the leadership required to revise the market convention for settlement fails.57 The first meeting of the TMPG after the reopening auctions of October 8 and 9 was on Thursday, October 23. The chairman, Tom Wipf of Morgan Stanley, opened the meeting by reminding members of the urgency of the situation: To overstate the obvious, the work of today’s meeting around settlement fails in Treasuries finds our committee 57 The Association of Primary Dealers provided similar leadership in revising the market convention for the treatment of accrued interest in repurchase agreements after the 1982 failure of Drysdale Government Securities (Garbade 2006). 62 The Introduction of the TMPG Fails Charge William Dudley, Executive Vice President of the Federal Reserve Bank of New York and Manager of the Fed’s System Open Market Account, echoed Wipf’s call for leadership: “This [meeting] is happening at a critical time in the market place where leadership is important to creating confidence and stability—we believe this group can, should and will provide that leadership.” 7.1 The November 12 Recommendations During the October 23 meeting, and in a series of subsequent telephone conference calls, TMPG members discussed changes in market practices that might reduce chronic fails and limit the likelihood of a recurrence. The group unveiled its recommendations on Wednesday, November 12, 2008.58 The principal recommendation suggested that “market participants agree that the invoice price . . . on any cash or financing transaction that fails to settle on the originally scheduled date be reduced at a fails [charge] rate equal to the greater of a) 3 percent per annum minus the fed funds target rate … and b) zero.” As shown in Chart 4, this would penalize fails at a rate that starts at zero when the target federal funds rate is at or above 3 percent and rises to 3 percent as the target funds rate declines toward zero. It follows that the economic cost of failing would never fall below about 3 percent per annum.59 The TMPG concluded that the “out-of-pocket cost to the party failing to deliver securities will provide a compelling incentive to resolve fails promptly.” 58 Treasury Market Practices Group, “Treasury Market Practices Group Endorses Several Measures to Address Widespread Settlement Fails,” November 12, 2008, available at http://www.newyorkfed.org/tmpg/ PR081112.pdf. The recommendations were reported in “Treasury’s Warning Adds to Plunge in Failed Trades,” Bloomberg.com, November 12, 2008, 15:59 EST, and “Repo Experts Propose Plans to Counteract Rise in ‘Fails,’” FT.com, November 12, 2008, 20:00. 59 This follows because the sum of the target federal funds rate, which is usually at or slightly above the Treasury general collateral repo rate, and the fails charge rate would never be less than 3 percent. The TMPG could have referenced the overnight Treasury general collateral repo rate in lieu of the target federal funds rate, but the target funds rate is more familiar and more readily observable to market participants. There is no definitive and widely disseminated measure of overnight Treasury general collateral repo rates. The Federal Reserve, by comparison, publicly announces the target funds rate. The TMPG recognized that the “the introduction of [the recommended convention] raises operational, legal and other implementation issues that may vary across Treasury market participants” and promised to engage in “further analysis of these issues,” with a goal of announcing by January 5, 2009, its recommendations for implementation.61 Chart 4 TMPG Fails Charge Rate as a Function of Target Federal Funds Rate TMPG fails charge rate (percent per annum) 3.0 2.5 2.0 8. The Crisis Recedes but Support for Revising the Market Convention Persists 1.5 1.0 0.5 0 0 1 2 3 4 Target federal funds rate (percent per annum) The TMPG explicitly based its recommendation on the dysfunctionality of the existing market convention for settlement fails: Past experience—for example, during the summer of 2003—shows that settlement fails in a particular [security] may become widespread and persistent when the special collateral repo rate for that [security] nears zero. Special collateral repo rates cannot exceed the Treasury general collateral repo rate. As a result, settlement fails across a wide variety of [securities] can . . . become widespread and persistent when the Treasury general collateral repo rate is near zero— as is currently the case. The underlying problem is the Treasury market contracting convention that a seller can deliver securities after the originally scheduled settlement date at an unchanged invoice price [and] without incurring any penalty. Introduction of a dynamic fails [charge] with a finite cap rate would remedy this problem. In particular, a dynamic fails [charge] would provide an incentive for sellers to resolve fails promptly, and could lead to repo contracting conventions [that is, negative repo rates] that would give beneficial owners of Treasury securities an opportunity to earn as much as the [3 percent] cap rate in securities loan fee income regardless of the level of nominal interest rates.60 60 Treasury Market Practices Group, “Treasury Market Practices Group Endorses Several Measures to Address Widespread Settlement Fails,” November 12, 2008. 5 By the time the TMPG made its November 12 recommendation, settlement fails in the Treasury market were receding rapidly. As shown in Chart 5, primary dealer fails declined from a daily average of $379 billion during the week of October 16-22 to a daily average of $70 billion during the week of November 13-19 and averaged less than $50 billion a day in December. Support for a revised market convention for settlement fails persisted in spite of the receding volume of fails, largely because the crisis of late September and early October had given added currency to the view that the existing convention was dysfunctional. The discussion of settlement fails during the November 4, 2008, meeting of the Treasury Borrowing Advisory Committee, as well as the views expressed in a prominent market newsletter in early January 2009, illustrates the growing consensus.62 61 The TMPG made three additional recommendations on November 12: 1) that market participants undertake a study of the most efficient way to margin fails in Treasury securities (in order to reduce counterparty credit risk exposure), 2) that market participants examine whether the Fixed Income Clearing Corporation, the two clearing banks, or other interested parties might develop “new or enhanced … multilateral netting arrangements” that might reduce settlement fails, and 3) that market participants pursue consensual cash settlement of transactions in Treasury securities that have been failing for more than five days. The TMPG also expressed its support for “discussion of a standing facility by the U.S. Department of the Treasury to provide temporary new supply of specific securities at a penalty rate when settlement fails persist,” but noted that the creation of such a facility was a long-term goal and that progress on a fails charge should not be contingent on the development of a Treasury security lending facility. 62 William Dudley, Executive Vice President, Federal Reserve Bank of New York, and Manager of the System Open Market Account, stated during a public conference call on January 14, 2009, on the TMPG fails initiative that: “Although settlement fails have declined recently from record levels amid reduced trading volumes, the extremely low level of interest rates suggests that fails could again rise significantly when trading activity picks up. The fundamental incentive to deliver securities under current market conditions is simply not sufficient at very low nominal interest rates to reduce the probability of large chronic fails to acceptable levels.” FRBNY Economic Policy Review / October 2010 63 Chart 5 Daily Average (Over Weekly Intervals) Primary Dealer Settlement Fails in Treasury Securities September 2008 to April 2009 Billions of dollars 400 350 300 250 200 150 100 50 2008 Apr. 16 - Apr. 22 Apr. 2 - Apr. 8 Mar. 19 - Mar. 25 Mar. 5 - Mar. 11 Feb. 19 - Feb. 25 Feb. 5 - Feb. 11 Jan. 22 - Jan. 28 Jan. 8 - Jan. 14 Dec. 25 - Dec. 31 Dec. 11 - Dec. 17 Nov. 27 - Dec. 3 Nov. 13 - Nov. 19 Oct. 30 - Nov. 5 Oct. 16 - Oct. 22 Oct. 2 - Oct. 8 Sep. 18 - Sep. 24 Sep. 4 - Sep. 10 0 2009 Source: Federal Reserve Bank of New York. 8.1 The November Meeting of the Treasury Borrowing Advisory Committee The Treasury Borrowing Advisory Committee (TBAC) is a committee of market professionals selected to advise the Secretary of the Treasury on matters relating to Treasury debt management. At its November 4 meeting, the committee discussed the upcoming midquarter refunding and, inter alia, the fails situation. The TBAC’s discussion of settlement fails focused initially on better ways for the Treasury to reopen outstanding issues than the “snap” reopenings of October 8 and 9, but then turned to the market convention for settlement fails. Several committee members observed that investors had “little economic incentive to lend securities when general collateral [repo] rates stood at 20 basis points,” and one member suggested that “a negative [repo] rate of 200 or 300 basis points . . . would create the correct economic incentives to cause holders of securities in low interest rate environments to lend securities again.”63 64 The Introduction of the TMPG Fails Charge In its ensuing report to the Secretary of the Treasury, the TBAC expressed the view that the low level of short-term interest rates “has made the cost of failing negligible, [leaving] little desire for short-sellers to close out their positions” and noted the suggestion of one committee member “that there should be a cost in the form of a penalty rate associated with fails in a low-rate environment.” The report further noted that such a cost would encourage negative-rate repo trading, “which would allow the free market to determine the effective cost of the fail, and change the economics of securities lending.”64 63 U.S. Department of the Treasury, “Minutes of the Meeting of the Treasury Borrowing Advisory Committee of the Securities Industry and Financial Markets Association, November 4, 2008,” November 5, 2008, available at http://www.ustreas.gov/press/releases/hp1239.htm. 64 U.S. Department of the Treasury, “Report to the Secretary of the Treasury from the Treasury Borrowing Advisory Committee of the Securities Industry and Financial Markets Association, November 4, 2008,” November 5, 2008, available at http://www.ustreas.gov/press/releases/hp1238.htm. 8.2 Comments in Wrightson’s Money Market Observer The January 5, 2009, edition of Wrightson’s Money Market Observer devoted substantial space to the TMPG proposal to revise the market convention for settlement fails. The newsletter noted that creating an explicit charge for settlement fails was “the most straight-forward way to remedy the obvious structural flaws that lead to delivery logjams in today’s market,” and pointed out the anticipated benefits of restoring competitive market forces to the special collateral repo markets: “The TMPG believes (correctly, in our view) that the repo market will be more elastic—and the Treasury clearing process more efficient—if the floor on special repo rates is set low enough [that is, below zero] to preserve a spread relative to general collateral rates even in the current rate environment.” In lieu of adjusting invoice prices, several TMPG members suggested that an economically equivalent result could be obtained if a seller who makes a late delivery agrees to make a side payment to the buyer in an amount equal to what became known as the “TMPG fails charge.” The charge for a fail on a given business day would be computed as: n C = --------- .01 max 3 – R trgt 0 P , 360 (1) where: C = charge, in dollars, n = number of calendar days to the next following business day, R trgt = target federal funds rate at the close of business on the business day preceding the fail, in percent per annum, and P = total proceeds due from the buyer, in dollars. 9. Getting It Right Although a consensus had emerged in support of revising the market convention for settlement fails, the TMPG recommendation for reducing invoice prices itself needed some revision. The TMPG recommendation would have required a seller and a buyer to reduce the invoice price on a failing transaction by matching amounts on a daily basis. If one party reduced the Although a consensus had emerged in support of revising the market convention for settlement fails, the TMPG recommendation for reducing invoice prices itself needed some revision. invoice price and the other did not, or if the two parties reduced the invoice price by different amounts, any attempt by the seller to deliver securities against payment would be rejected by the buyer (because the buyer would be looking to pay a different amount than what the seller was looking to receive). TMPG members who understood the complex architecture of broker-dealer and custodian settlement systems pointed out that requiring matching daily price reductions would impose a major operational burden on market participants and could lead to an explosion in rejected deliveries (and thus in settlement fails). For example, if P = $10 million, R trgt = 1 percent, and n = three days, then C = $1,666.66.65 This procedure had the advantage of not requiring any change in existing settlement systems. The idea of replacing invoice price adjustments with side payments illustrates an important aspect of the TMPG initiative: by working collaboratively, the TMPG was able to achieve its objectives while accommodating an existing institutional structure: back-office settlement systems. The difference between a price adjustment and a side payment may seem trivial, but the success of the TMPG initiative hinged on recognizing the difference. 9.1 The January 5 Announcement On January 5, 2009, the TMPG announced that it was recommending a fails charge in the form of a side payment on transactions that failed to settle on a timely basis and that it was making several additional refinements to its November 12 recommendation.66 The three key refinements $1,666.66 = (3/360) .01 max[3 1, 0] $10,000,000. Treasury Market Practices Group, “Timeline for New Market Practices to Address Widespread Settlement Fails in U.S. Treasury Securities,” January 5, 2009, available at http://www.newyorkfed.org/tmpg/PR090105c.pdf, and Treasury Market Practices Group, “Claiming a Fails Charge for a Settlement Fail in U.S. Treasury Securities,” January 5, 2009, available at http:// www.newyorkfed.org/tmpg/PR090105a.pdf. The January 5 announcement was reported in “Repo Arena Gets a Plan on Penalties,” Wall Street Journal, January 6, 2009, p. C3, and “Penalty for Failed Trades Set to Spark New Era for US Repo,” Financial Times, January 7, 2009, p. 22. 65 66 FRBNY Economic Policy Review / October 2010 65 were 1) a statement of the process for claiming a fails charge, 2) a timeline suggesting that market participants begin claiming for settlement fails on transactions agreed to on or after May 1, 2009, and 3) replacement of the target federal funds rate (in the formula for the fails charge, equation 1 above) with a “TMPG reference rate.” The latter rate was defined as the target federal funds rate if the Federal Open Market Committee specified a target rate or the lower limit of the target band for the federal funds rate if the FOMC specified a target band.67 In the event the FOMC specified neither a target rate nor a target band, the On January 5, 2009, the TMPG announced that it was recommending a fails charge in the form of a side payment on transactions that failed to settle on a timely basis and that it was making several additional refinements to its November 12 The January 5 announcement suggested that the best way to initiate the fails charge would be for buyers to tender claims directly to sellers.70 A seller could either pay what was claimed or dispute the claim and negotiate with its counterparty over the amount due. The TMPG further suggested that if an investor employed a professional asset manager and that manager contracted to sell securities that were not delivered on a timely basis, the claim for the fails charge should be directed to the asset manager (rather than to the investor or to the investor’s custodian). This suggestion was based on the pragmatic notion that since the sale had been negotiated by the asset manager, the asset manager would be in the best position to recognize the sale and identify who was responsible for the settlement fail, be it the asset manager, the investor’s custodian, or some other party,71 or whether the claim should be left for the account of the investor.72 9.2 Trading Practice and Market Practice Recommendations recommendation. TMPG committed to recommending some other similar, readily observable, short-term interest rate as a reference rate for the fails charge formula.68 The decision to recommend a side payment (in lieu of an invoice price adjustment) required the TMPG to specify a way for buyers to collect from sellers who failed to deliver securities on a timely basis. In the case of buyers and sellers who settled through FICC, a collection process could be added to other similar processes previously implemented by FICC (such as the collections and disbursements that result from marking transactions to current market prices).69 However, the collection process was not as simple for transactions that settled bilaterally, as was the case for most transactions between dealers and their nondealer customers. 67 This charge was necessitated by the December 16, 2008, decision of the Federal Open Market Committee to establish a target range for the federal funds rate of 0 to 1/4 percent. 68 In late March 2009, the TMPG announced a slightly different form for the fails charge computation: 1 C = --------- .01 max 3 – RTMPG 0 P . 360 In this form, the charge is computed for each calendar day that a seller’s delivery obligation is failing. R TMPG is the TMPG reference rate on the business day preceding the day for which the charge is computed. See Treasury Market Practices Group, “Treasury Market Practices Group Announces Updates to Fails Charge Recommendation,” March 30, 2009, available at http://www.newyorkfed.org/tmpg/tmpg_033009.pdf. 66 The Introduction of the TMPG Fails Charge Following the January 5 announcement, TMPG members and other market participants collaborated to publish two documents providing guidance on how to implement the TMPG fails charge. The documents were important for clarifying how fails charges should be calculated and claimed and generally for enhancing the transparency of the new market convention. 69 The addition required a change in FICC rules that had to be approved by the Securities and Exchange Commission. FICC filed the proposed rule change on February 25, 2009 (Securities and Exchange Commission Release no. 3459569, March 12, 2009), and the Commission granted approval two months later (Securities and Exchange Commission Release no. 34-59802, April 20, 2009). 70 The January 5 announcement noted the possibility of setting up a central industry utility to receive and process claims, but observed that the design of such a facility raised novel questions regarding the identification of buyers and sellers and would require further consultation with market participants. 71 Some investors retain an agent to lend securities from the investor’s portfolio. Such agents are commonly called “agent sec lenders.” In most cases, an agent sec lender is obliged to reclaim securities out on loan if the investor’s asset manager decides to sell the securities. If an agent sec lender fails to reclaim securities on a timely basis and thereby causes a settlement fail, the fail may be the responsibility of the agent sec lender, rather than the asset manager or the custodian. 72 A sale of securities negotiated by an asset manager may fail to settle on a timely basis because the investor’s custodian failed to receive the same securities on an unrelated purchase. Such fails cannot be attributed to faulty behavior by the asset manager or the investor’s custodian, so the resulting fails charge would be left for the account of the investor. The investor can, of course, direct its asset manager to file a claim on the seller who failed to deliver securities to the investor. Trading practice recommendations: On January 15, 2009, the TMPG and the Securities Industry and Financial Markets Association (SIFMA) published a “U.S. Treasury Securities Fails Charge Trading Practice”73 to give market participants TMPG members and other market participants collaborated to publish two documents providing guidance on how to implement the TMPG fails charge. The documents were important for clarifying how fails charges should be calculated and claimed and generally for enhancing the transparency of the new market convention. guidance on exactly the types of transactions that were covered by, and excluded from, the TMPG fails charge. The “Trading Practice” also recommended the form of a letter that a market participant could send to counterparties, advising them of the participant’s adoption of the new policy for settlement fails, and suggested a statement that could be added to trade confirmations indicating that a transaction was subject to the fails charge. Market practice recommendations: On April 23, 2009, SIFMA published a “Treasury Market Practices Group Fails Charge Market Practice”74 that recommended procedures for buy-side firms to use in connection with the new fails charge. The recommended procedures included processes for researching and tracking fails, calculating fails charges, determining responsibility for a claim, sending and receiving claims, and accounting for claims. The procedures also included suggestions made earlier by SIFMA and adopted by the TMPG75 that claims be submitted at the beginning of a month for fails settled during the preceding month (to accommodate custodians and asset managers who structured their control 73 Treasury Market Practices Group and Securities Industry and Financial Markets Association, “U.S. Treasury Securities Fails Charge Trading Practice,” January 15, 2009, available at http://www.sifma.org/capital_markets/docs/ Fails-Charge-Trading-Practice.pdf. 74 Securities Industry and Financial Markets Association, “Treasury Market Practices Group Fails Charge Market Practice,” April 23, 2009, available at http://www.theasset manager.com/docs/AM_Custodian_IndustryProcedures_ TMPG_FailsCharge.pdf. 75 See Treasury Market Practices Group, “Treasury Market Practices Group Announces Updates to Fails Charge Recommendation,” March 30, 2009, available at http://www.newyorkfed.org/tmpg/tmpg_033009.pdf. systems around settled transactions) and be in excess of $500 per issue per settlement (to limit costly research and billing efforts to nontrivial claims). 10. Implementation The TMPG fails charge went into effect on May 1, 2009, replacing the former market convention of postponing— without any explicit penalty and at an unchanged invoice price—a seller’s obligation to deliver Treasury securities when the seller fails to deliver the securities on a scheduled settlement date. Henceforth, the cost of failing to settle a sale of Treasury securities in a timely fashion would not be less than 3 percent per annum. It would be premature to argue that the TMPG fails charge has eliminated the possibility of yet another episode of chronic settlement fails in Treasury securities; past episodes were rare to begin with and some future event may demonstrate the existence of an unsuspected flaw in the new system. It may also be the case that the 3 percent benchmark rate is too low and that chronic fails would be better mitigated with a 3 1/2 or 4 percent rate. However, there is no evidence to date that the The new convention is not yet out of its infancy, but there is reason to anticipate that the TMPG fails charge will . . . dampen future eruptions [of settlement fails]. new market convention, and the 3 percent benchmark rate, are not working. Chart 6 shows daily average settlement fails over weekly intervals from the beginning of 2009 to July 2010. Fails averaged a bit over $14.4 billion per day during the first four months of 2009, but only $4.2 billion per day since implementation of the fails charge.76 More important, the relatively modest eruptions of settlement fails that appeared during the first week of July 2009 and the first week of January 2010 quickly subsided. The new convention is not yet out of its infancy, but there is reason to anticipate that the TMPG fails charge will similarly dampen future eruptions. 76 The fails charge was never intended to eliminate all settlement fails. (Fails attributable to miscommunication or operational problems are unlikely to be eliminated by the fails charge—although they may be resolved more quickly.) Rather, the fails charge was aimed at mitigating episodes of chronic fails that can threaten market liquidity and efficiency. FRBNY Economic Policy Review / October 2010 67 Chart 6 Daily Average (Over Weekly Intervals) Primary Dealer Settlement Fails in Treasury Securities January 2009 to July 2010 Billions of dollars 35 30 After May 1, 2009 Prior to May 1, 2009 25 20 15 10 5 2009 Jul. 15 - Jul. 21 Jun. 17 - Jun. 23 May 20 - May 26 Apr. 22 - Apr. 28 Mar. 25 - Mar. 31 Feb. 25 - Mar. 3 Jan. 28 - Feb. 3 Jan. 1 - Jan. 6 Dec. 3 - Dec. 9 Nov. 5 - Nov. 11 Oct. 8 - Oct. 14 Sep. 10 - Sep. 16 Aug. 13 - Aug. 19 Jul. 16 - Jul. 22 Jun. 18 - Jun. 24 May 21 - May 27 Apr. 23 - Apr. 29 Mar. 26 - Apr. 1 Feb. 26 - Mar. 4 Jan. 29 - Feb. 4 Jan. 1 - Jan. 7 0 2010 Source: Federal Reserve Bank of New York. 11. Conclusion The TMPG fails charge initiative is important both for what it accomplished and for how it was accomplished. Substantively, the initiative revised an outmoded convention and mitigated an important dysfunctionality in a market of critical national significance. Procedurally, the initiative demonstrated how cooperation between the public and private sectors can speed innovative and efficient responses to changing circumstances. At the time of the May 1, 2009, implementation of the fails charge, the Federal Reserve Bank of New York welcomed the new convention: We applaud the dedicated efforts of the TMPG in spearheading the development and implementation of this targeted solution to the settlement fails problem. This significant milestone in the evolution of Treasury market practice demonstrates that groups, such as the TMPG, are effective in addressing deficiencies in market functioning and facilitating market best practices.77 68 The Introduction of the TMPG Fails Charge In a subsequent letter to the TMPG membership expressing his personal thanks for the Group’s dedication and commitment to making the fails charge a reality, William Dudley, now president of the Federal Reserve Bank of New York, reflected on the significance of the new market convention: The implementation of the fails charge marks a rare and significant evolution in Treasury market architecture. In my view, one would need to look back to 1982 to find a development of similar magnitude, when the collapse of Drysdale Securities led to the adoption of a new market practice to include accrued interest in repo contracts. The fails charge stands among relatively few revisions to contracting conventions in the Treasury market since the development of a liquid national market following World War I. 77 Federal Reserve Bank of New York, “New York Fed Applauds Implementation of the TMPG’s Fails Charge Recommendation,” May 1, 2009, available at http://www.newyorkfed.org/newsevents/news/markets/2009/ ma090501.html. Appendix: Additional Individuals Who Provided Feedback and Assistance in the Implementation Phase of the TMPG Fails Charge Initiative The following list attempts to include all non-TMPG (Treasury Market Practices Group) individuals who either served on formal subgroups that contributed to the timely implementation of the TMPG fails charge or participated in less formal conference calls and meetings. The authors apologize for any inadvertent omissions. David Aman Marc Baum Brandon Becker Brent Blake Gary Buki Kevin Caffrey Maria Carina Michael Cetta Brayton Cherry Ed Corral David Cosgrove Brian Crowe Craig Delany Laura Dietel Craig Dudsak Steve Dunn Joe Finan Marcellus Fisher Barbara Friedman Robert Good Simon Griffiths Olivier Grimonpont Chris Haas Janice Hamilton Robert Hennessy Eugene Ing Dyann Kiessling Bradley Koehler Marty Kruse Michael Landolfi Lourdes Leon Christine Lin Fred Lipinski Colin Lloyd Frank Lupica Diane Madera Claudia Maia Jennifer Manor Frank Martino Christopher Marzullo Jason McCann Shirley McCoy Katherine McGaugh Kevin Meagher Omar Medina Cleary Gottlieb Steen & Hamilton LLP Ramius LLC WilmerHale State Street The Bank of New York Mellon The Bank of New York Mellon Euroclear AllianceBernstein Holdings Brown Brothers Harriman & Co. JPMorgan Chase ICAP Fidelity Investments JPMorgan Chase State Street Citigroup JPMorgan Chase Morgan Stanley & Co. PIMCO New York Life Investment Management LLC Goldman, Sachs & Co. JPMorgan Chase Euroclear Merrill Lynch The Northern Trust Company The Bank of New York Mellon Depository Trust & Clearing Corporation Fidelity Investments Euroclear BNY Mellon Asset Servicing State Street Morgan Stanley & Co. Citigroup Depository Trust & Clearing Corporation Cleary Gottlieb Steen & Hamilton LLP Lord, Abbett & Co. LLC Morgan Stanley & Co. Euroclear Fidelity Investments New York Life Investment Management LLC Lord, Abbett & Co. LLC Reserve Bank of Australia JPMorgan Chase State Street Fidelity Investments Goldman Sachs Asset Management FRBNY Economic Policy Review / October 2010 69 Appendix: Additional Individuals Who Provided Feedback and Assistance in the Implementation Phase of the TMPG Fails Charge Initiative (Continued) Fatima Mehladi Eric Miller Tamara Molinary Carolyn Monroe-Koatz Penny Morgan Louis Nazarro Edward Neeck Peter Novello Elisa Nuottajarvi Paul Parseghian Judith Polzer Joseph Pomo Thomas Ponti Nancy Prior Pawan Puneet Christopher Ramsay Bill Rose Theodore Rothschild Timothy Ryan Joseph Sack Randy Snook Guido Stroemer Brian Swann Rob Toomey Diane Trupia Raymond Tyrrell Jason Ward Andrew Waskow Mark Willis Patricia Yak Lawrence Young Anthony Zook 70 Euroclear Citadel Solutions, LLC AllianceBernstein Holdings JPMorgan Chase Western Asset Management JPMorgan Chase JPMorgan Chase Lord, Abbett & Co. LLC SIFMA Prudential Investment Management Inc. JPMorgan Chase Goldman Sachs Asset Management State Street Fidelity Investments State Street IMS Citadel Investment Group, LLC BTG JPMorgan Chase SIFMA SIFMA SIFMA UBS Goldman Sachs SIFMA SIFMA Brown Brothers Harriman & Co. Fidelity Investments Goldman Sachs Merrill Lynch Credit Suisse First Boston Credit Suisse First Boston JPMorgan Chase The Introduction of the TMPG Fails Charge References Committee on Banking, Finance, and Urban Affairs. 1985. “The Federal Reserve Bank of New York Discount Window Loan of $22.6 Billion Extended to the Bank of New York.” Hearing before the Subcommittee on Domestic Monetary Policy of the Committee on Banking, Finance, and Urban Affairs. U.S. House of Representatives, 99th Cong., 1st Sess., December 12. Cornell, B., and A. Shapiro. 1989. “The Mispricing of U.S. Treasury Bonds: A Case Study.” Review of Financial Studies 2, no. 3 (fall): 297-310. Department of the Treasury. 2006. “Consideration of a Proposed Treasury Securities Lending Facility.” Federal Register 71, no. 85 (May 3): 26,174-8. Department of the Treasury, Securities and Exchange Commission, and Board of Governors of the Federal Reserve System. 1992. “Joint Report on the Government Securities Market.” January. Duffie, D. 1996. “Special Repo Rates.” Journal of Finance 51, no. 2 (June): 493-526. Fisher, M. 2002. “Special Repo Rates: An Introduction.” Federal Reserve Bank of Atlanta Economic Review 87, no. 2 (second quarter): 27-43. Fleming, M., and K. Garbade. 2002. “When the Back Office Moved to the Front Burner: Settlement Fails in the Treasury Market after 9/11.” Federal Reserve Bank of New York Economic Policy Review 8, no. 2 (November): 35-57. ———. 2003. “The Repurchase Agreement Refined: GCF Repo.” Federal Reserve Bank of New York Current Issues in Economics and Finance 9, no. 6 (June). ———. 2004. “Repurchase Agreements with Negative Interest Rates.” Federal Reserve Bank of New York Current Issues in Economics and Finance 10, no. 5 (April). ———. 2007. “Dealer Behavior in the Specials Market for U.S. Treasury Securities.” Journal of Financial Intermediation 16, no. 2 (April): 204-28. Garbade, K. 2004. “Origins of the Federal Reserve Book-Entry System.” Federal Reserve Bank of New York Economic Policy Review 10, no. 3 (December): 33-50. ———. 2006. “The Evolution of Repo Contracting Conventions in the 1980s.” Federal Reserve Bank of New York Economic Policy Review 12, no. 1 (February): 25-40. ———. 2007. “The Emergence of ‘Regular and Predictable’ as a Treasury Debt Management Strategy.” Federal Reserve Bank of New York Economic Policy Review 13, no. 1 (March): 53-71. Garbade, K., and J. Ingber. 2005. “The Treasury Auction Process: Objectives, Structure, and Recent Adaptations.” Federal Reserve Bank of New York Current Issues in Economics and Finance 11, no. 2 (February). Garbade, K., and J. Kambhu. 2005. “Why Is the U.S. Treasury Contemplating Becoming a Lender of Last Resort for Treasury Securities?” Federal Reserve Bank of New York Staff Reports, no. 223, October. Ingber, J. 2006. “A Decade of Repo Netting.” Futures & Derivatives Law Report 26, no. 2 (February): 3-12. Jordan, B., and S. Jordan. 1997. “Special Repo Rates: An Empirical Analysis.” Journal of Finance 52, no. 5 (December): 2051-72. Keane, F. 1996. “Repo Rate Patterns for New Treasury Notes.” Federal Reserve Bank of New York Current Issues in Economics and Finance 2, no. 10 (September). Sender, H. 1986. “The Day the Computers Went Down.” Institutional Investor 11, no. 3 (March): 203-7. ———. 2005. “Explaining Settlement Fails.” Federal Reserve Bank of New York Current Issues in Economics and Finance 11, no. 9 (September). The views expressed are those of the authors and do not necessarily reflect the position of the Federal Reserve Bank of New York or the Federal Reserve System. The Federal Reserve Bank of New York provides no warranty, express or implied, as to the accuracy, timeliness, completeness, merchantability, or fitness for any particular purpose of any information contained in documents produced and provided by the Federal Reserve Bank of New York in any form or manner whatsoever. FRBNY Economic Policy Review / October 2010 71