The full text on this page is automatically extracted from the file linked above and may contain errors and inconsistencies.
Second Quarter 2014 Volume 97, Issue 2 © Michael Hoffman Camden County Boathouse, New Jersey The Paper Trail of Knowledge Transfers Shadow Banking and the Crisis of 2007-08 Forecast Disagreement in the Survey of Professional Forecasters Research Rap INSIDE Issn: 0007-7011 Second QUARTER 2014 The Business Review is published four times a year by the Research Department of the Federal Reserve Bank of Philadelphia. The views expressed by the authors are not necessarily those of the Federal Reserve. We welcome your comments at PHIL. BRComments@phil.frb.org. The Paper Trail of Knowledge Transfers For a free subscription, go to www. philadelphiafed.org/research-and-data/ publications. Archived articles may be downloaded at www.philadelphiafed. org/research-and-data/publications/ business-review. To request permission to reprint articles in whole or in part, click on Permission to Reprint at www. philadelphiafed.org/research-and-data/ publications. Articles may be photocopied without permission. Microform copies may be purchased from ProQuest Information and Learning, 300 N. Zeeb Road, Ann Arbor, MI 48106. The Federal Reserve Bank of Philadelphia helps formulate and implement monetary policy, supervises banks and bank and savings and loan holding companies, and provides financial services to depository institutions and the federal government. It is one of 12 regional Reserve Banks that, together with the U.S. Federal Reserve Board of Governors, make up the Federal Reserve System. The Philadelphia Fed serves eastern Pennsylvania, southern New Jersey, and Delaware. Charles I. Plosser President and Chief Executive Officer Loretta J. Mester Executive Vice President and Director of Research Colleen Gallagher Research Publications Manager Dianne Hallowell Art Director and Manager 1 Why do firms tend to locate near other firms? Economists suspect that geographic clustering spurs innovation by letting businesses tap a climate rich in informal transfers of knowledge. By tracing links between inventors filing for patents for the same inventions, Jeffrey Lin shares new evidence supporting the idea that proximity offers businesses tangible benefits. Shadow Banking and the Crisis of 2007-08 7 In recent decades, institutions that function much like traditional banks have grown outside regulatory oversight. Yet, as Daniel Sanches explains, these so-called shadow banks are as vulnerable to runs as regular banks. Because banking crises can inflict lasting economic harm, economists are interested in tracing how the panic ensued in the shadow system. Forecast Disagreement in the Survey of Professional Forecasters 15 To enact effective policies and spend resources efficiently, firms, policymakers, and markets need accurate economic forecasts. But even though economists generally work with similar models and data, their projections often range widely. To better understand why, Keith Sill explores what the evidence and theories say about how forecasters form their views. Research Rap Abstracts of the latest working papers produced by the Research Department of the Federal Reserve Bank of Philadelphia. 25 The Paper Trail of Knowledge Transfers By Jeffrey Lin W hy do firms and inventors tend to locate in dense, costly areas? One intriguing hypothesis is that such geographic clustering lets them benefit from local knowledge spillovers. As Nobel laureate Robert Lucas has noted, the benefits of one person’s knowledge spilling over to others play a central role in economic growth and the existence of cities: “What can people be paying Manhattan or downtown Chicago rents for, if not for being near other people?” Proximity may improve the sharing of knowledge, the matching of ideas to firms, or the rate of learning.1 If dense clustering indeed confers these benefits, then that raises the possibility that individuals and firms may not be fully taking them into account when deciding where to locate, resulting in underinvestment in new ideas. A key counterargument to the importance of knowledge spillovers is that firms might prefer to keep their work secret from competitors. For example, many firms include nondisclosure and noncompete clauses in employment contracts for researchers and scientists. Yet, as Alfred Marshall suggested, knowledge is difficult to keep secret: “The mysteries of the trade become no mysteries; but are as it were in the air.” So are, in fact, knowledge spillovers an important reason why inventors tend to locate near one another? We know that other factors might also encourage firms and inventors to locate near one another. For example, Gerald A. Carlino’s 2001 Business Review article discusses such mechanisms. firms might benefit from better matches with specialized workers.2 They may benefit from the sharing of local production inputs such as cheap electrical power or hard-to-find machinery and parts. Or skilled inventors may be attracted to superior amenities such as restaurants, shopping, or safety. A key challenge, then, is to account for these alternative explanations so that we do not erroneously infer that knowledge spillovers are empirically important. To explore this challenge, I review the empirical literature regarding evidence of knowledge spillovers contained in patent citations and nonpatent data. I then describe the evidence that Ina 1 2 See my 2009 Business Review article. Jeffrey Lin is an economic advisor and economist at the Federal Reserve Bank of Philadelphia. The views expressed in this article are not necessarily those of the Federal Reserve. This article and other Philadelphia Fed reports and research are available at www. philadelphiafed.org/research-and-data/publications. www.philadelphiafed.org Ganguli, Nick Reynolds, and I found using a novel measure — cases of simultaneous invention that result in rival claims known as patent interferences. First, let us look at what researchers have found by studying routine patent applications. EVIDENCE FROM PATENT CITATIONS At its most basic, the challenge of verifying the existence of knowledge spillovers was observed by Paul Krugman — namely, that knowledge flows are invisible: “They leave no paper trail by which they may be measured or tracked.” Adam Jaffe, Manuel Trajtenberg, and Rebecca Henderson tackled this problem by observing that the flow of knowledge from one inventor to another could, in fact, be tracked using patent citations.3 Their paper and ones that followed have provided the best evidence to date that local knowledge spillovers might be one important mechanism contributing to the geographic proximity of inventors.4 They exploit the fact that patents include citations to older patents. If a new patent cites a previous patent, this citation is evidence that the older patent contains knowledge upon which the new patent relies. Though a citation to a nearby inventor is at first glance evidence that knowledge has passed from the earlier inventor to the citing inventor, it does Robert M. Hunt’s 2001 Business Review article discusses patents as a measure of knowledge production. 3 4 See also the work by Jaffe (1986) and Keller (2002), and the surveys by Rosenthal and Strange (2004) and Audretsch and Feldman (2004). Business Review Q2 2014 1 not necessarily indicate that geographic proximity has facilitated this transfer of knowledge. It might simply be the case that inventors are located nearby for some other reasons besides the opportunity to take advantage of knowledge spillovers, such as to be near some common physical input to the invention process. For example, if a patent awarded to researchers at a University of Pennsylvania hospital is cited in a subsequent application for a patent awarded to researchers at Temple University, that might be because of local knowledge spillovers. But it could also be the case that these patentees are near each other simply because many hospitals are needed to serve the large population of the Philadelphia area, and proximity offers no advantage in transmitting knowledge. Expected proximity. To address this inference problem, Jaffe and his coauthors develop a clever matching strategy. They measure the distance between an earlier “originating” patent and a subsequent “citing” patent that references the originating patent as an important knowledge input. Then they compare this distance to the distance between the originating patent and a matched “control” patent. The control patent is similar to the matched citing patent in terms of the date of invention and technological classification, but it does not cite the matched originating patent. Thus, the control patent represents the expected proximity of inventors working in the same research field and time period, not conditioned on a citation link. If the inventors of the citation-linked patent pair are observed in closer proximity versus this benchmark, then this is strong evidence that a local knowledge spillover has occurred, especially since we have accounted for the underlying geographic distribution of research activity and hence other reasons why inventors might be located together. In fact, Jaffe and his coauthors find that 2 Q2 2014 Business Review originating patentees are much more likely to be from the same metropolitan area as citing patentees, compared with a matched control patentee. Despite this clever study design, subsequent researchers have identified several limitations of this analysis. First, there are the standard drawbacks to using patent data: Not all inventions are patented, and some patents do not represent valuable or worthwhile inventions. More recent papers have tried to correct these problems by, for example, measuring the quality of patents based on patent renewals or subsequent citations. Second, many patent citations are actually added by patent examiners, not inventors. Thus, citations may not actually represent true knowledge flows for inventors, but rather noise introduced by the patent office.5 Third, Peter Thompson and Melanie Fox-Kean note that Jaffe’s results are sensitive to the selection of an appropriate control patent. By varying how broadly the technology classifications and dates of application are specified for the sample of matched control patents, Thompson and FoxKean found that imperfect matching explained a significant part of Jaffe’s original result. A final issue, which Jaffe and his coauthors acknowledged in their original paper, is that many knowledge inputs are not actually reflected in citations. This is significant because we might expect that geographic proximity is especially important for the transfer of tacit, operational knowledge — that is, knowledge that is not necessarily codified in a written patent application. This is the kind of knowledge transmitted in hallways and over coffee, rather than through literature searches of previous work. Because many more knowledge spillovers may occur than are reflected in citations, Jaffe and his coauthors suggest that their findings may actually represent a lower bound for the occurrence of knowledge spillovers among inventors. 5 See the papers by Juan Alcácer and Michelle Gittelman and by Peter Thompson (2006). 6 See the survey article by David Audretsch and Maryann Feldman. EVIDENCE FROM RESEARCH AND NONPATENT DATA Other papers have sidestepped patents altogether.6 Bruce Weinberg found that physicists who moved to cities where Nobel laureates were already working were more likely to begin their own Nobel Prize-winning work there. Gerald A. Carlino, Jake K. Carr, Robert M. Hunt, and Tony E. Smith have shown that research and development labs are highly geographically concentrated, substantially more so than the corresponding industry concentration patterns. In my previous work, I showed that new activities related to the implementation of new knowledge are concentrated in metropolitan areas with highly educated populations. Finally, Petra Moser has shown localization among prize-winning inventors at World’s Fairs in the 19th century, although these patterns weakened over time. Of course, as with inventors, scientists may locate near each other for reasons besides knowledge spillovers. Thus, we cannot be certain whether an increase in their productivity might stem from knowledge spillovers from nearby scientists or from some other reason. Fabian Waldinger investigated local knowledge spillovers among scientists in Germany. Waldinger relies on evidence from the expulsion of Jewish and certain other scientists from Germany under the Nazis. Some university departments experienced many expulsions, while other departments had not employed Jewish scholars and were therefore unaffected. If knowledge www.philadelphiafed.org spillovers are important, one might expect the productivity of the remaining scientists in the affected departments to decline following the expulsion of their Jewish colleagues. Waldinger finds that the publishing activity of the scientists whose departments suffered losses did not decline compared with that of other scientists. Thus, he concludes that there is no evidence for local knowledge spillovers among German scientists in this period. EVIDENCE FROM PATENT INTERFERENCES In my work with Ina Ganguli and Nick Reynolds, I have been using patent interferences to try to provide new evidence on the relevance of local knowledge spillovers for invention. Patent interferences are especially valuable for measuring local spillovers of tacit or uncodified knowledge that is missing in traditional patent studies. Patent interferences are a unique historical feature of the U.S. patent system. Until 2011, the United States had a “first to invent” rule for assigning priority of invention, versus the “first to file” rule more common in the rest of the world.7 When the U.S. patent office received applications from multiple parties with identical claims at roughly the same time, it was obligated to investigate the competing claims to determine which party was entitled to patent protection. This investigation, known as a patent interference proceeding, determined who had conceived of the invention and reduced it to practice first. Typically, the parties submitted dated laboratory notebooks, testimony by associates, and media reports as evidence of first invention. There are many famous examples 7 More details about the patent interference proceedings can be found in Calvert and Sofocleous (1982), Cohen and Ishii (2006), de Simone et al. (1963), and Kingston (2004). www.philadelphiafed.org of patent interferences in U.S. history, including Alexander Graham Bell’s and Elisha Gray’s simultaneous invention of the telephone. Because Bell’s and Gray’s applications arrived at the patent office on the same day and contained nearly identical claims, an interference proceeding was declared. Eventually, Bell was determined to have conceived of the idea and reduced it to practice first, and he was awarded the patent. Knowledge in common. Importantly for economists, patent interferences create a record of instances when the same invention is created by inventors working independently, a phenomenon that is highly suggestive interference is evidence of a knowledge spillover among the inventors. Several details about the interference process support the argument that these proceedings are a good measure of common, independent knowledge inputs. First, interferences were declared by a patent examiner specializing in a particular technological area. Thus, interfering claims were likely to be detected. (In some cases, the examiner was alerted to a possible interference by one of the applicants. Note that an interference is different from patent infringement, in which the holder of an existing patent sues an infringing party. Private parties cannot sue for an interference.) Patent interferences are especially valuable for measuring local spillovers of tacit or uncodified knowledge that is missing in traditional patent studies. of common knowledge inputs. In other words, inventors involved in an interference are likely to have command of similar knowledge. For example, interfering inventors may have similar backgrounds in chemistry, or they may have similar knowledge of market conditions. This is especially true if certain inventions require highly specific knowledge. For example, for Bell and Gray to have both invented the telephone contemporaneously, they must have had similar knowledge about electrical conductivity and the properties of various conductive metals, as well as similar expectations about market demand for a device that transmitted voices in real time. For Jon Merz and Michelle Henry, a patent interference is an indication that “discovery has become ordinary.” That is, its occurrence suggests that certain knowledge is shared among several inventors. In other words, a patent Second, interferences involved parties with roughly the same date of patent application. An inventor who delayed filing an application in order to conceal an invention would lose the priority contest. Thus, interferences are less likely to reflect secrecy or other legal strategies of the participants and are more likely to reflect genuinely independent, simultaneous inventions versus infringements or patent “racing” by inventors who believed that rival applications were imminent. Third, during the interference proceedings, circumstances that suggested stealing, espionage, or collaborative invention typically led to dismissal with prejudice. In other words, worker poaching and espionage that is independent of shared knowledge are unlikely explanations for the bulk of cases of patent applications interfering with each other. Note that recruiting other firms’ researchers or spying Business Review Q2 2014 3 on them does seem to involve shared knowledge inputs — a shared desire to solve a common problem, for example. In addition, patent judges had been compelled by statute to rule against applicants found to have stolen a competitor’s idea, deterring would-be spies from pursuing an interference. In fact, in our examination of case decisions, no more than a small handful of judgments mention espionage as a relevant factor in decisions. Fourth, competing claims that are similar but not identical did not result in a completed interference case. Fifth, the patent office verified that interfering parties had independent financial interests (for example, that they were not different branches of a multinational conglomerate); otherwise, the case was dismissed. Thus, interferences are not the result of knowledge sharing within organizations. Finally, the separate applications were required to have been made roughly at the same time, often within a year. Thus, copying subsequent to publication and disclosure of an older patent are unlikely to have occurred. Interferences improve on traditional patent studies in a number of ways. One, interferences involve patents that are more valuable than the average patent. Since an interference requires parties to actively contest for priority, it is unlikely that inventors would spend time or money in pursuit of a worthless patent. Two, we have information on patent interferences over a long period, from the 19th century to 2011. Three, patent interferences do not rely on citations to prove common knowledge inputs and are thus not subject to some of the weaknesses noted earlier. Specifically, while patent citations necessarily capture the spillover only of written, publicly available knowledge, simultaneous invention is evidence that some kind of spillover, whether written or not, is likely to 4 Q2 2014 Business Review have occurred. Thus, interferences capture spillovers of tacit knowledge. As I noted earlier, we might expect that tacit knowledge is especially sensitive to geographic proximity. If so, then we expect results on the localization of interferences to be stronger than on the localization of citations. We have constructed a database of over 1,000 interference cases from the early 1980s to 2011 from the U.S. patent office. This database includes the names of the inventors involved in the interference, their patent and application numbers, and the date of the interference. We match this information to a database of inventor locations (based on their Zip codes) produced by the Harvard Business School. Testing geographic concentration. If local knowledge spillovers are important, one possible test is to see whether patent interferences, as measures of shared, possibly tacit knowl- edge, are more likely to occur between inventors who are located close to each other versus those located farther apart. The black line in Figure 1 shows this pattern for only the interference cases involving pairs of U.S.-based inventors. The horizontal axis measures the distance in miles as the crow flies between the observed locations of the two parties involved in a patent interference. (Since a single patent application can be made on behalf of multiple inventors, we measure the minimum distance between inventors of the different parties to the interference.) The vertical axis shows the percent of interfering inventor pairs in our database that are separated by at most the distance indicated by the horizontal axis. Thus, as we move to the right, we accumulate our inventor pairs until 100 percent of our pairs are within 4,258 miles — the maximum distance observed between two interfering U.S. inventors. FIGURE 1 Interfering Inventors More Geographically Concentrated Cumulative percent of patent pairs 100 90 80 70 60 50 40 30 Interfering patents 20 Control patents 10 0 0 500 1,000 1,500 2,000 2,500 3,000 3,500 4,000 Distance between inventors (miles) Sources: U.S. Patent and Trademark Office, Harvard University Patent Network Dataverse. www.philadelphiafed.org Despite the large possible range of distances between inventors, the black line shows that 20 percent of interfering inventors are within only 100 miles of each other, and half of interfering inventor pairs are within 540 miles of each other. While this is a good starting point for showing that proximity matters for shared knowledge inputs, it still might be true that inventors are located close to each other to take advantage of some other factor. Similar to the logic of the patent citation studies, we can compare the localization of interferences with the localization of noninterfering patents in the same technology classification and year. In that way, we can control for the underlying distribution of research activity that doesn’t rely on common knowledge inputs, as interferences do. For each pair of interfering patents, we selected up to 10 control patents. Our goal was to control for the underlying geographic distribution of inventive activity by selecting patents that were similar to the interfering patents but not involved in the interference case. We selected control patents based on two criteria. First, a control patent had to share at least one of the many possible three-digit technological classification codes assigned by the patent office that the two interfering patents had shared. Second, the control patent’s application date had to fall www.philadelphiafed.org between the application dates of the two interfering patents. If no eligible control patent was found, we then expanded the selection window incrementally by 10 days before the earlier interfering application and 10 days after the later interfering application until an eligible control patent was found. Finally, we randomly chose one of the two interfering patents to match with the control patent. We then compared the distance between the interfering inventors with the distance between the randomly selected interfering inventor and the control inventor. The results. The gold line in Figure 1 shows our results for the proximity of interfering inventors to control patent inventors. It represents the expected distribution of distances between inventors working in technology fields and time periods similar to those of our sample of patent interferences, but it is not conditioned on an interfering link between inventors. Comparing the distribution of distances between interfering inventors with the control distribution of noninterfering inventors, it is clear that interfering inventors are more geographically concentrated. While 20 percent of interfering inventors are within 100 miles of each other, less than 1 percent of noninterfering inventors are within 100 miles of each other. And while half of interfering inventors are within 540 miles of each other, the same is true of less than 21 percent of noninterfering inventors. Interfering inventors are especially more likely to be geographically concentrated at small distances. For example, more than 3 percent of interfering U.S. inventors are in the same Zip code, versus none of the noninterfering inventor-control pairs. Eleven percent of interfering inventors are within 15 miles of each other, compared with less than 1 percent of noninterfering inventors. These results are consistent with a growing literature documenting that knowledge spillovers attenuate rapidly with distance. CONCLUSION Although local knowledge spillovers are of central interest to economists, the evidence to date on their existence is mixed. Patterns in our data on patent interferences suggest that inventors working independently but using common knowledge inputs are substantially more geographically concentrated than other inventors working in the same field and time period who are not linked by common knowledge inputs. These results suggest that localized knowledge spillovers may be especially salient for forms of tacit or uncodified knowledge, which is difficult to observe using citations but more likely detectable from interferences. BR Business Review Q2 2014 5 REFERENCES Alcácer, Juan, and Michelle Gittelman. “Patent Citations as a Measure of Knowledge Flows: The Influence of Examiner Citations,” Review of Economics and Statistics, 88:4 (2006), pp. 774-779. Cohen, Linda R., and Jun Ishii. “Competition, Innovation and Racing for Priority at the U.S. Patent and Trademark Office,” University of California, Irvine, Working Paper (2006). Audretsch, David B., and Maryann P. Feldman. “Knowledge Spillovers and the Geography of Innovation,” in J. Vernon Henderson and Jacques-François Thisse, eds., Handbook of Regional and Urban Economics, 4. Amsterdam: Elsevier, 2004, pp. 2,713-2,739. De Simone, Daniel V., James B. Gambrell, and Charles F. Gareau. “Characteristics of Interference Practice,” Journal of the Patent Office Society, 45 (1963), pp. 503-591. Audretsch, David B., and Maryann P. Feldman. “R&D Spillovers and the Geography of Innovation and Production,” American Economic Review 86:3 (1996), pp. 630-640. Azoulay, Pierre, Joshua Graff Zivin, and Jialan Wang. “Superstar Extinction,” Quarterly Journal of Economics, 125:2 (2010), pp. 549-589. Borjas, George J., and Kirk B. Doran. “The Collapse of the Soviet Union and the Productivity of American Mathematicians,” Quarterly Journal of Economics, Oxford University Press, 127:3 (2012), pp. 1,143-1,203. Calvert, Ian A., and Michael Sofocleous. “Three Years of Interference Statistics,” Journal of the Patent Office Society, 64:12 (1982), pp. 699-707. Carlino, Gerald A., Satyajit Chatterjee, and Robert M. Hunt. “Urban Density and the Rate of Invention,” Journal of Urban Economics, 61 (2007), pp. 389-419. Carlino, Gerald A. “Knowledge Spillovers: Cities’ Role in the New Economy,” Federal Reserve Bank of Philadelphia Business Review (Fourth Quarter 2001). Carlino, Gerald A., Jake K. Carr, Robert M. Hunt, and Tony E. Smith. “The Agglomeration of R&D Labs,” Federal Reserve Bank of Philadelphia Working Paper 12-22 (2012). Chapin, Seymour L. “Patent Interferences and the History of Technology: A HighFlying Example,” Technology and Culture, 12:3 (1971), pp. 414-416. Duranton, Gilles, and Diego Puga. “MicroFoundations of Urban Agglomeration Economies,” in Vernon Henderson and Jacques-Francois Thisse, eds., Handbook of Regional and Urban Economics, 1. Amsterdam: Elsevier, 2004, pp. 2,063-2,117. Duranton, Gilles, and Diego Puga. “Nursery Cities: Urban Diversity, Process Innovation, and the Life-Cycle of Products,” American Economic Review, 91:5 (2001), pp. 1,454-1,477. Glaeser, Edward L. “Learning in Cities,” Journal of Urban Economics, 46 (1999), pp. 254-277. Hunt, Robert M. “You Can Patent That? Are Patents on Computer Programs and Business Methods Good for the New Economy?” Federal Reserve Bank of Philadelphia Business Review (First Quarter 2001). Jaffe, Adam B. “Technological Opportunity and Spillovers of R&D: Evidence from Firms’ Patents, Profits, and Market Value,” American Economic Review, 76:5 (1986), pp. 984-1,001. Jaffe, Adam B., Manuel Trajtenberg, and Rebecca Henderson. “Geographic Localization of Knowledge Spillovers as Evidenced by Patent Citations,” Quarterly Journal of Economics, 108:3 (1993), pp. 577-598. Keller, Wolfgang. “Geographic Localization of International Technology Diffusion,” American Economic Review, 92:1 (2002) pp. 120-142. Kingston, William R. “Light on Simultaneous Invention from U.S. Patent Office ‘Interference’ Records,” World Patent Information, 26:3 (2004), pp. 209-220. Krugman, Paul. Geography and Trade. Cambridge: MIT Press, 1991. 6 Q2 2014 Business Review Lin, Jeffrey. “Technological Adaptation, Cities, and New Work,” Review of Economics and Statistics, 93:2 (2011), pp. 554-574. Lin, Jeffrey. “Urban Productivity Advantages from Job Search and Matching,” Federal Reserve Bank of Philadlpelphia Business Review (First Quarter 2011). Lucas, Robert E. “On the Mechanics of Economic Development,” Journal of Monetary Economics, 22 (1988), pp. 3-42. Marshall, Alfred. Principles of Economics. London: Macmillan, 1920. Merz, Jon F., and Michelle R. Henry. “The Prevalence of Patent Interferences in Gene Technology,” Nature Biotechnology, 22:2 (2004), pp. 153-154. Moser, Petra. “Did Patents Weaken the Localization of Innovations? Evidence from World’s Fairs,” Journal of Economic History, 71:2 (2011), pp. 363-381. Romer, Paul M. “Endogenous Technological Change,” Journal of Political Economy, 98:5 (1990), pp. 71-102. Rosenthal, Stuart S., and William C. Strange. “Evidence on the Nature and Sources of Agglomeration Economies,” in J. Vernon Henderson and Jacques-François Thisse, eds., Handbook of Regional and Urban Economics 4. Amsterdam: Elsevier, 2004, pp. 2,119-2,171. Thompson, Peter. “Patent Citations and the Geography of Knowledge Spillovers: Evidence from Inventor- and ExaminerAdded Citations,” Review of Economics and Statistics, 88:2 (May 2006), pp. 383-388. Thompson, Peter, and Melanie Fox-Kean. “Patent Citations and the Geography of Knowledge Spillovers: A Reassessment,” American Economic Review, 95:1 (March 2005), pp. 450-460. Waldinger, Fabian. “Peer Effects in Science: Evidence from the Dismissal of Scientists in Nazi Germany,” Review of Economic Studies, 79 (2012), pp. 838-861. Weinberg, Bruce A. “Geography and Innovation: Evidence from Nobel Laureate Physicists,” mimeo (2007). www.philadelphiafed.org Shadow Banking and the Crisis of 2007-08 By Daniel Sanches S ome economists have noted that recessions accompanied by banking crises tend to be deeper and more difficult to recover from than other recessions — even those associated with other types of financial crises. For instance, the bursting of the dot.com bubble in 2001 was a very important financial event that was not accompanied by a protracted recession. The potential of banking crises to do lasting economic harm led policymakers to adopt safeguards in the 1930s that have essentially eliminated traditional banking panics in the U.S. Although the Great Recession of 2007-09 was associated with a protracted financial market disruption — and the failures of some large banks like Washington Mutual and IndyMac — we did not observe widespread withdrawals from commercial banks, as in a traditional banking crisis. However, economists Gary Gorton and Andrew Metrick show that it can be viewed as a banking crisis that originated in the shadow banking system. In the last 30 years, institutions very similar in function to traditional banks have grown outside regulatory oversight. One lesson of the financial crisis is that these institutions are as vulnerable to panics as traditional banks because they are subject to similar risks. ECONOMIC FALLOUT OF BANKING CRISES Banking crises can harm the economy. Financial crises are usually associated with bad economic outcomes — recessions. One particular kind of financial crisis to which economists have devoted a lot of attention is the type that originates in the banking sector. A banking crisis is a widespread withdrawal of funds from depository institutions — that is, a run on the liabilities of a large number of banks.1 Like other financial crises, banking crises are usually associated with economic downturns, and there is evidence that banking crises often worsen economic downturns as weaknesses at the banks spill over into financial problems for households and firms. When financial events affect consumption and investment decisions by households and nonfinancial firms, economists say that they have real effects. Many researchers have provided evidence that banking crises can make 1 A common form of bank liability is the demand deposit contract — a typical checking account that most people have at a commercial bank. Daniel Sanches is a senior economist at the Federal Reserve Bank of Philadelphia. The views expressed in this article are not necessarily those of the Federal Reserve. This article and other Philadelphia Fed reports and research are available at www.philadelphiafed.org/ research-and-data/publications. www.philadelphiafed.org economic contractions more severe and more protracted, with various studies emphasizing different channels. In a highly influential book providing a systematic account of banking crises in the United States from 1867 to 1960, Milton Friedman and Anna Schwartz identify the reduction in the wealth of bank shareholders and the decline in the money supply that usually follow a business downturn accompanied by a banking crisis as the main causes of a further drop in real economic activity. A decline in the money supply has real effects because households and firms need money to pay for their purchases. Thus, a decline in the money supply leads to a decline in transactions and real economic activity.2 Looking at the banking crises of the 1930s, Ben Bernanke identified the increased cost of intermediation services — the costs that banks incur when assessing the creditworthiness of borrowers — following the recurrent banking crises of the early 1930s as causing a significant reduction in the flow of funds from lenders to borrowers through the banking system. This constricted flow of credit impeded the real economy’s recovery from the Great Depression. Michael Bordo, Barry Eichengreen, Daniela Klingebiel, and Maria Soledad Martinez-Peria provide crosscountry evidence of the real effects of banking crises over a period of 120 years. They also find that recessions that are accompanied by banking crises are more severe than those that are For more on this subject, see my 2012 Business Review article, “The Optimum Quantity of Money.” 2 Business Review Q2 2014 7 not. Moreover, they provide evidence that banking crises plague advanced and developing economies equally, confirming the view that a banking crisis is not just a concern for countries at low levels of economic development. Anatomy of a banking crisis. Although every banking crisis is different, those that occurred up to and including the Great Depression follow a similar pattern. Let me briefly describe the typical sequence of events that leads to a banking crisis. Bad news arrives. Usually at the peak of an economic expansion, bad news about the quality of the assets held by a group of banks (or a major bank) leads to larger withdrawals than usual. For instance, a failed attempt by the Knickerbocker Trust Company to corner the copper market and the subsequent decision of a major bank to no longer clear checks issued by the Knickerbocker triggered a run on the Knickerbocker on October 18, 1907, sparking the Panic of 1907. The fact that it was one of the largest depository institutions in New York contributed to the public’s perception that other banks could also be in distress. Banks sell assets to meet the increase in withdrawals. To meet the higher demand for cash, a bank initially draws down its cash reserves. But its reserves may not be enough if the withdrawal process quickly intensifies. The bank can also sell some of its assets to cover withdrawals. Selling assets causes asset prices to fall. If many banks are trying to sell assets at the same time, the assets can be sold only at a large discount. Think of what would happen if four neighbors on your block put their houses up for sale on the same day as you did. All else equal, you would have to lower your price to get anyone to buy. Financial asset markets work the same way. For easily marketable fixed-income assets such as Treasury securities or certain corporate bonds traded in large 8 Q2 2014 Business Review markets, buyers can still be found by selling at a discount. However, a large fraction of a bank’s assets consists of mortgages and commercial and industrial loans made to households and firms whose creditworthiness is unknown to the wider market, which means the bank would probably find few if any buyers. And anyone willing to purchase the loan would demand a substantial discount to compensate for the lack of information about the borrower’s creditworthiness. Thus, selling assets on short notice may be extremely costly for a bank. Depositors begin a run on healthy banks. Banks facing large withdrawals may borrow in the interbank market, where banks routinely borrow reserves Banks suspend convertibility. The final step comes as banks react to widespread withdrawals. One way to stop the drain on funds is to temporarily suspend the convertibility of deposits into cash — banks may simply lock their doors — in an attempt to preserve capital until depositors calm down and things get back to normal. Strictly speaking, this is a breach of the demand deposit contract.4 This description of a typical banking crisis clearly reveals why banks are fragile: They fund illiquid assets with deposits that can be withdrawn at will. Economists usually refer to this practice as maturity transformation. It is important to mention that this role played by banks has a value for society. Although every banking crisis is different, those that occurred up to and including the Great Depression follow a similar pattern. from each other. But if banks want to borrow more reserves than usual, they must pay a higher interest rate to the lending bank. Larger discounts in asset markets and higher interest rates in interbank markets are usually signs of financial strain. If widespread distrust of banks causes depositors to withdraw their funds even from healthy banks, a line is crossed. The number of banks that want to sell assets increases, resulting in even steeper discounts, and the number of banks that want to borrow in the interbank market also increases, making it harder for each borrower bank to obtain a loan. As this process intensifies, we have a full-scale panic.3 3 We can think of these withdrawals as a way for depositors to monitor their banks. That is, by withdrawing their money, depositors are checking whether the bank is healthy enough to pay. This might explain why people decide to withdraw their funds even from banks initially viewed as safe and sound. People have a preference for holding highly liquid assets — assets that are easy to sell without taking a loss — but the most profitable investments take a long time to pay off. Banks offer demand deposit contracts that give people ready access to their funds and a higher rate of return than they would get by holding liquid assets directly. Banks are able to offer a higher rate of return to depositors because they pool resources in such a way that permits them to invest a significant fraction of their assets in higher-yielding, long-term projects such as mortgages and other types of long-term loans. Normally, funding illiquid assets with short-term liabilities works fine. But when depositors begin to worry about losses, a bank run may ensue. In the second half of the 19th century, the decision to suspend convertibility was usually coordinated by private bank associations. 4 www.philadelphiafed.org U.S. bank runs essentially disappeared in the 1930s. The introduction of federal deposit insurance in 1933 with the creation of the Federal Deposit Insurance Corporation (FDIC) ended the banking crises that had been recurrent events in the U.S. even before the Great Depression. The government’s deposit guarantees largely relieved depositors of the need to constantly monitor the health of banks. In turn, the government has undertaken the monitoring of banks through regulation and supervision. But regulations are not costless. FDIC premiums, capital requirements, and regulatory restrictions on bank portfolios increase banks’ costs. These costs are informally referred to as regulatory taxes. And banks, like any other firm, have a strong incentive to avoid taxes. THE RISE OF SHADOW BANKING The Great Recession in the U.S. was associated with a severe financial crisis, but we did not observe people rushing to their banks to withdraw their deposits. However, a closer look suggests that the crisis was not very different from a typical banking crisis, except that it was triggered outside the traditional banking sector. According to Gary Gorton and Andrew Metrick, the financial crisis can be viewed as a banking crisis that originated in the shadow banking system.5 The shadow banking system is a set of institutions that carry out functions very similar to those of traditional banks but that are largely unregulated. They perform the same kind of maturity transformation traditionally performed by commercial banks. Thus, the shadow banking system, despite its somewhat unwholesomesounding name, provides a useful service to society. This is to say that shadow banking is not necessarily a bad thing. The problem is that, under certain circumstances, these financial institutions can become fragile — that is, subject to panics. An important fact about the shadow banking system is that it has grown significantly in the last 30 years. For instance, Gorton and Metrick estimate that just before the financial crisis of 2007-08, the assets of the shadow banking system were at least as large as the assets of commercial banks.6 Another important fact about the shadow banking system is that it has grown outside the oversight of regulators. Why did this happen? As banking and finance in general have expanded in recent decades, part of that growth has occurred in the shadow system, largely to avoid the costs associated with regulation.7 As I will now explain, the shadow banking system works pretty much like a typical commercial bank even though the parties involved in the transactions are not the bankers and depositors that we typically have in mind. For the most part, I will follow Gorton and Metrick and focus on the market for repurchase agreements (or repos) as the main cause of the panic in the shadow banking system and one of the centers of the financial crisis. But the shadow banking system also includes other markets and institutions such as asset-backed commercial paper in which the same basic structure (risky, illiquid assets funded by shortterm liabilities) recurs. The repo market. The repo market is a market for short-term, mainly overnight, collateralized loans. To understand why repos work pretty much According to Gorton and Metrick, this is probably an underestimate because this comparison involves the assets of only a fraction of the shadow banking system. 6 For more on the rise of shadow banking, see the review by Tobias Adrian and Adam Ashcraft. 7 5 See also chapter 2 in Gary Gorton’s 2010 book. www.philadelphiafed.org like banking and to see why the repo crisis was actually a banking crisis, it is necessary to look at how repo transactions work. Let me start by identifying the “depositors,” the repo lenders. These are largely institutional investors such as pension funds and large corporations that need some place to invest large amounts of money for short periods. They also want to obtain higher yields than those offered by regulated commercial banks. Most important, these institutional investors want their funds to be safe.8 One alternative is the repo market. A firm can make an overnight loan to a borrower. To make the loan safe, the firm receives collateral usually in the form of government bonds, which are liquid and fluctuate little in value over short periods. If the borrower is unable to return the funds, the lending firm will simply seize the collateral. Provided that the value of the underlying collateral does not change significantly over short periods, a repo transaction is safe for the repo lender. Like a bank depositor, the repo lender has ready access to its money and has the opportunity to reallocate its funds toward some other use on a daily basis. Thus, a repo transaction offers the firm both the convenience of having ready access to its funds and a level of safety not much different from that of a federally insured demand deposit. Until 2011, large commercial depositors could not receive interest on their short-term deposits, another motivation for them to seek an alternative place to park their funds.9 When the 8 It is also important to mention that the amounts these institutional investors wish to deposit are typically larger than the maximum amount insured by the FDIC. 9 As part of the Dodd-Frank Act, the Federal Reserve Board in July 2011 repealed Regulation Q, which had prohibited banks from paying interest on corporate checking accounts. Business Review Q2 2014 9 repo borrower repurchases the security from the repo lender, he or she also pays interest to the lender. As should be clear by now, the “banker” in the repo transaction is the repo borrower, which typically is an investment bank or the broker-dealer arm of a large bank holding company. These institutions use the funds they borrow in the repo market to finance a wide range of activities, some of them quite risky. As long as the repo is collateralized by a Treasury security, it is not fragile in the same sense as traditional banking because the asset that collateralizes the repo is highly liquid and can be easily sold. If the repo borrower can’t repay on time, the repo lender can simply take the collateral and sell it for cash. This is basically how the shadow banking system works. Depositors (institutional investors and large corporations) need a place to park liquid funds that provides them with ready access to their money, pays an interest rate higher than that offered by traditional banks, and spares them the expense and hassle of managing their own cash.10 Bankers (investment banks and broker-dealer firms) are willing to provide such a product in the form of repo transactions. Finally, safe collateral such as U.S. Treasury bonds are essential to make this financial transaction work. The growth of the repo market increased the demand for collateral. The growth of the repo market prior to the financial crisis of 2007-08 was extraordinary. The volume of repo As Robert Lucas puts it: “In a monetary economy, it is in everyone’s private interest to try to get someone else to hold non-interest-bearing cash and reserves. But someone has to hold it all, so all of these efforts must simply cancel out. All of us spend several hours per year in this effort, and we employ thousands of talented and highly trained people to help us. These personhours are simply thrown away, wasted on a task that should not have to be performed at all.” 10 10 Q2 2014 Business Review transactions reported by primary dealers (those who trade directly with the Federal Reserve System) had grown from roughly $2 trillion in 1997 to $7 trillion in 2008. This estimate, of course, leaves out unreported transactions. Gorton and Metrick estimate that the overall size of the repo market just before the financial crisis was roughly the same as the size of the traditional banking sector as measured by total assets.11 As we have seen, Treasury securities play an important role in the functioning of the shadow banking system. However, repo markets are not the only source of demand for Treasury securities. They are also used as collateral in derivative markets and settlement systems. Furthermore, many foreign governments, especially the central banks of developing countries such as China, demand Treasury securities because they are safe and highly liquid.12 For instance, in 2005 only 48.6 percent of total U.S. debt was privately held, according to the Federal Reserve Bank of San Francisco. About a third of that privately held debt was held in reserve by foreign central banks, which means that only about a third of total U.S. debt (or $2.6 trillion) was available for private transactions. Unlike for other goods and services, higher demand for Treasury securities doesn’t automatically provide an incentive to increase supply. The supply of government bonds is determined by government borrowing, a direct consequence of fiscal policy. For instance, the decision to reduce the fiscal deficit in the U.S. in the 1990s and early 2000s may have contributed to a shortage of government bonds available for repo transactions. One piece of indirect evidence that government bonds were in short supply is the practice in financial markets known as rehypothecation, which simply means that traders can use the same collateral to secure more than one transaction. To the extent that this practice had become widespread before the crisis of 2007-08, traders may have had an incentive to develop other methods to conduct a growing number of transactions with a limited amount of good collateral.13 Mortgage-backed securities helped satisfy the demand for collateral. The solution to the shortage of good collateral was found in another form of financial innovation that had evolved significantly since the 1980s: securitization. Commercial banks make many loans to consumers and firms. Instead of holding these loans on its own balance sheet, a bank can sell them to a shell company the bank creates and manages for this purpose, called a special purpose vehicle (SPV). The SPV funds the acquisition of these assets (mortgages, car loans, credit card receivables, etc.) by issuing asset-backed securities (ABS) that, as the name implies, are backed by the loans the SPV holds and that become the SPV’s liabilities when it sells them to investors in the capital markets.14 Figure 1 shows how commercial banks fund loans through securitization. Most important, this organizational form allows financial institutions to increase the scale of their overall operations without increasing their balance sheets, which would require them to increase their regulatory capital. Thus, setting up an SPV is a way of avoiding 11 They also cite a range of estimates by other economists of the same order of magnitude. 13 Foreign demand for Treasury securities increased significantly in the 1990s and 2000s, the flip side of the large trade surpluses run by China and some other developing countries. 14 See the chapter by Gary Gorton and Nicholas S. Souleles. 12 For more on the role of rehypothecation, see the 2011 Business Review article by Cyril Monnet. www.philadelphiafed.org FIGURE 1 Securitization and Shadow Banking STEP 1: STEP 2: STEP 3: Bank originates loans. Bank creates SPV to purchase loan portfolio. SPV issues ABS to fund portfolio purchase. CASH BORROWERS capital requirements, which increases a financial institution’s overall degree of leverage by raising its total assets relative to its capital.15 If carefully chosen, a portfolio of loans backing the ABS can be safe and predictable. Thus, by making it possible to bundle individual loans and sell claims on the loan portfolio on the market, this form of financial innovation offers an alternative to using deposits to fund banks’ illiquid assets. When carefully executed, securitization is extremely valuable for both banks and investors. The housing boom in the U.S. in the 2000s was financed in this way. The large increase in the number and size of mortgage loans created a large supply of a particular type of ABS called mortgage-backed securities (MBS). As the name suggests, MBS are ABS that bundle mortgages. Given the growth of the repo market and the relative scarcity of government bonds, the use of ABS as collateral in the repo market seemed to be a reasonable solution to the shortage of collateral. Gorton and Metrick have argued that the use of ABS as collateral in the In this article, I emphasize avoiding regulatory taxes as a motivation for securitization. See Ronel Elul’s article for an account of the efficiency benefits of securitization. 15 www.philadelphiafed.org SPV BANK LOANS CASH CASH LOANS repo market had increased significantly prior to the financial crisis. As I discuss below, this is still a controversial claim. Despite a wealth of anecdotal evidence, we have no precise estimates of the share of repo transactions that used ABS as collateral. CRISIS IN THE SHADOW BANKING SYSTEM ABS as repo collateral created the conditions for a banking panic. As long as the repo was collateralized by Treasury securities, lenders (depositors) didn’t have to worry about the borrower’s risk of default or about the value of the underlying collateral. But this changed when the repo was collateralized by ABS. In 2007, house prices in the U.S. started to decline, raising concerns that homeowners could start defaulting on their mortgages in large numbers. In turn, lenders with repo collateralized by MBS started worrying about potential losses. What was the reason for their concern? After all, as I have argued, when carefully executed, securitization can generate a safe asset for investors, and indeed many MBS were built to be nearly riskless under normal conditions. Usually, ABS are designed to be safe. ABS reduce credit risk in two ways: diversification and overcollateralization. For instance, pooling mort- INVESTORS ABS gages that had been originated in cities all over the U.S. is one way to create diversification. Under normal circumstances, large numbers of homeowners in all regions of the country are very unlikely to default on their mortgages at the same time. Overcollateralization simply involves pooling enough mortgages to guarantee that it can generate enough cash flow to make the promised payments to investors even if some of the borrowers default. The amount of overcollateralization required to make an MBS safe usually depends on certain fundamental market indicators, including the trend in house prices. Significantly, the statistical models used to design, price, and provide credit ratings for MBS estimated default rates based on data collected during periods of generally rising house prices and during periods when housing price declines were localized.16 Bad news arrived. Now consider a scenario in which investors expect house prices to rise and, contrary to their expectations, house prices begin to fall, and keep falling. That is what happened in the U.S. in 2007. When house prices fell for several consecu- 16 Indeed, Christopher Foote, Kristopher Gerardi, and Paul Willen have documented that people had overly optimistic beliefs about house prices. Business Review Q2 2014 11 tive months, an increasing number of investors believed that the average rate of default on any given pool of mortgages was going to rise. Their fears became more concrete when in the summer of 2007, two hedge funds sponsored by Bear Stearns that had invested heavily in subprime mortgages filed for bankruptcy and BNP Paribas suspended withdrawals from three money market mutual funds that were exposed to subprime mortgages. An important indicator of their fears was that the ABX index, a measure of the risk of default on subprime MBS, began to rise. This raised concerns that many SPVs were not holding enough collateral to generate sufficient cash flow to make good on the promised payments to investors. Another reason to have doubts about the true value of MBS was that many investors did not know where the risks were concentrated. Although many MBS were wisely built to be nearly riskless, several classes of MBS contained a disproportionate fraction of mortgages that had been extended to people of dubious creditworthiness. And the risk of these subprime mortgages was particularly sensitive to the decline in housing prices. Repo lenders ran on repo borrowers, including healthy borrowers. A depositor with serious doubts about the underlying value of the collateral can do two things: either ask for more collateral or simply not renew the repo. Both actions can be interpreted as a decision to withdraw funds from the shadow banking system, much like the decision bank depositors make to withdraw funds from their bank when they believe they might not be able to get all their money out. Repo lenders initially asked for more collateral, but ultimately they simply refused to renew their loans. In other words, the repo market froze.17 Because investors could not tell safe MBS from risky MBS in most cases, 12 Q2 2014 Business Review they withdrew their funds even from shadow banks that probably had safe MBS to secure repos. This problem was severe enough to turn the initial panic into a systemic event — a banking crisis. Thus, the financial crisis was not very different from the banking crises of old. Investors in the repo market behaved pretty much like bank depositors did during U.S. banking crises before 1933. And the outcome was certainly very similar. The initial banking crisis spread to other financial markets, and several financial firms either failed or had to be rescued by the federal government to prevent further failures. A caveat. Gorton and Metrick’s explanation for the events that sparked the 2007-08 financial crisis depends on the claim that the fraction of the repo market that used ABS as collateral was large enough to generate a systemic event. But this claim has been a source of controversy among financial economists. For instance, Arvind Krishnamurthy, Dmitry Orlov, and Stefan Nagel have argued that a relatively small share of repo transactions in which money market mutual funds and securities lenders were the repo lenders was collateralized by ABS prior to the 2007-08 crisis. However, these authors focus on a relatively small segment of the repo market, the triparty repo market, while Gorton and Metrick study the larger bilateral repo market, for which there is as yet no direct evidence about the collateral used in transactions.18 Furthermore, Krishnamurthy and coauthors note that while the share of the transactions collateralized 17 See also Yaron Leitner’s article explaining why markets freeze. See also Benjamin Lester’s article for a discussion of regulatory interventions in response to market freezes. Triparty repo transactions take place between two counterparties intermediated by a dealer bank. Bilateral repo transactions occur between two counterparties without an intermediary. 18 by ABS was modest, such transactions were more concentrated among a small number of large banks that experienced significant problems. So focusing on average shares may be misleading. Nonetheless, the details of Gorton and Metrick’s account of developments in the repo market will remain a source of controversy until researchers can collect more complete data. Moreover, some evidence suggests that the financial crisis was actually triggered in another part of the shadow banking system. See the accompanying discussion, Crisis in the ABCP Market. SHADOW BANKING PANIC MAY HAVE DEEPENED RECESSION It is still too early to fully disentangle the relative importance of the various factors that led to a particularly deep recession and a particularly slow recovery. But like many earlier recessions associated with banking crises, the crisis in the shadow banking system may have played a significant role in the depth of the downturn and the slow recovery. The crisis in the shadow banking system has significantly reduced the ability of commercial banks to originate and renew loans, creating ongoing problems for households and firms that rely on bank loans. Some economists have even argued that the effects of the collapse can persist for an extended period. For instance, Viral Acharya has argued that traditional lenders cannot easily fill the role that shadow banks had played in providing credit to the economy. This void has certainly contributed to the delay in restoring the flow of credit to a volume consistent with that of a recovery from a typical recession that had not been accompanied by a banking crisis. The shadow banking system has not fully recovered from the financial crisis. Even though it has continued to operate with government support, it is www.philadelphiafed.org unclear whether the volume of operations will return to that observed prior to the crisis anytime soon, or whether it should. Since the crisis, the government-sponsored enterprises Fannie Mae and Freddie Mac have carried out nearly all securitizations in housing markets. At this point it is unclear whether the private sector will ever play the same role in the creation of securitized assets that it had before the crisis. CONCLUSION One lesson of the financial crisis is that institutions quite similar to banks tend to rise up outside of regulatory purview. This is an important matter because this shadow banking system is fragile and subject to panics. And banking panics — regardless of where they occur — have pernicious economic repercussions. This potential for economic harm had led some economists before the crisis to propose tighter regulation of the shadow banking system. In the aftermath, policymakers were working to write new rules. BR Crisis in the ABCP Market S ome economists have argued that problems in another segment of the shadow banking system can be identified as the prime cause of the 2007-08 financial crisis. For example, in his discussion of Gorton and Metrick’s account of the crisis, Andrei Shleifer has provided evidence that the contraction in the asset-backed commercial paper (ABCP) market happened before the contraction in the repo market. Thus, he suggests that problems in the ABCP market may have triggered the financial crisis. Commercial paper is a short-term debt instrument that both financial and nonfinancial firms use to finance ongoing operations. Financial firms issue commercial paper to fund a wide array of activities, including the purchase of long-term securities such as MBS. One form of funding through the issuance of commercial paper that has increased significantly in the last 20 years is ABCP. A financial firm can set up an SPV to purchase a portfolio of securities by issuing commercial paper on the capital markets. ABCP maturities can vary from one day (as in a typical repo transaction) to 90 days. The typical maturity of ABCP is 30 days. Again, we have something that looks like a bank, but it operates outside the regulatory system. The main investors in ABCP are money market mutual funds. Similar to the investors in the repo market, money market mutual funds also need a convenient place to invest some of their resources for short periods. These investors also want their investments to be safe and to yield an attractive return. Provided that the assets securing ABCP are of sufficiently high quality, such an investment vehicle is fairly safe, at least under normal market conditions. The short-term duration of ABCP gives investors an opportunity to “withdraw” their funds in case they decide to invest elsewhere or in case they have doubts about the quality of the assets securing ABCP. The issuers of ABCP are SPVs that are sponsored by large financial institutions, including traditional commercial banks.a The SPVs allow these institutions to fund a wide array of securities at any moment. The short duration of ABCP means that an SPV has to roll over its debt every time an ABCP matures. Many SPVs used the proceeds from the sale of ABCP to invest in MBS (i.e., the collateral backing ABCP were MBS). As we have seen, the perception of MBS as a safe debt instrument can suddenly change once the trend in house prices becomes clearly downward. Starting in the summer of 2007, many investors stopped refinancing maturing ABCP because of potential exposure to subprime mortgages via MBS. A full-scale panic ensued as the spread on overnight ABCP over the federal funds interest rate (the rate of interest on unsecured loans in the interbank market in the U.S.) increased from 10 basis points to 150 basis points. The outstanding amount of ABCP shrank steadily after the summer of 2007, despite several interventions by the Federal Reserve System in the form of liquidity facilities, offering short-term credit to banks to refinance maturing ABCP. The ABCP market also provides another example of financial transactions carried out outside the oversight of regulators that are very similar to traditional banking. Thus, a closer look at the crisis in the ABCP market has also demonstrated that it was not very different from previous banking crises. Perhaps the most balanced view is that while the financial crisis began in the shadow banking system, it had many epicenters. Furthermore, the structural similarities among many of the institutions in the shadow banking system — illiquid assets funded by short-term liabilities — and the trigger for the crisis — the decline in housing prices — tell much the same story. A sponsor financial institution usually provides credit guarantees to the SPV. For a detailed description of the ABCP market, see the paper by Marcin Kacperczyk and Philipp Schnabl. a www.philadelphiafed.org Business Review Q2 2014 13 REFERENCES Acharya, Viral. “Understanding Financial Crises: Theory and Evidence from the Crisis of 2007-2008,” NBER Reporter (April 2013). Adrian, Tobias, and Adam B. Ashcraft. “Shadow Banking: A Review of the Literature,” Federal Reserve Bank of New York Staff Report 580 (October 2012). Bernanke, Ben. “Nonmonetary Effects of the Financial Crisis in the Propagation of the Great Depression,” American Economic Review, 73 (1983), pp. 257-276. Bordo, Michael, Barry Eichengreen, Daniela Klingebiel, and Maria Soledad Martinez-Peria. “Is the Crisis Problem Growing More Severe?” Economic Policy, 16 (2001), pp. 51-82. Foote, Christopher, Kristopher Gerardi, and Paul Willen. “Why Did So Many People Make So Many Ex Post Bad Decisions? The Causes of the Foreclosure Crisis,” Public Policy Discussion Papers 12:2, Federal Reserve Bank of Boston (2012). Friedman, Milton, and Anna Schwartz. A Monetary History of the United States, 18671960. Princeton University Press, 1963. Gorton, Gary B. Slapped by the Invisible Hand: The Panic of 2007. New York: Oxford University Press, 2010. Gorton, Gary B., and Andrew Metrick. “Securitized Banking and the Run on Repo,” Journal of Financial Economics, 104 (2012), pp. 425-451. Elul, Ronel. “The Economics of Asset Securitization,” Federal Reserve Bank of Philadelphia Business Review (Third Quarter 2005). Gorton, Gary B., and Nicholas S. Souleles. “Special Purpose Vehicles and Securitization,” in Rene Stulz and Mark Carey, eds., The Risks of Financial Institutions. University of Chicago Press, 2006. Federal Reserve Bank of San Francisco, “Who Are the Largest Holders of U.S. Public Debt?” http://www.frbsf.org/education/publications/doctor-econ/2005/july/ public-national-debt. Kacperczyk, Marcin, and Philipp Schnabl. “When Safe Proved Risky: Commercial Paper During the Financial Crisis of 20072009,” Journal of Economic Perspectives, 24 (2010), pp. 29-50. 14 Q2 2014 Business Review Krishnamurthy, Arvind, Stefan Nagel, and Dmitry Orlov. “Sizing Up Repo,” NBER Working Paper 17768 (2012). Leitner, Yaron. “Why Do Markets Freeze?” Federal Reserve Bank of Philadelphia Business Review (Second Quarter 2011). Lester, Benjamin. “Breaking the Ice: Government Interventions in Frozen Markets,” Federal Reserve Bank of Philadelphia Business Review (Fourth Quarter 2013). Lucas, Robert E., Jr. “Inflation and Welfare,” Econometrica, 68 (2000), pp. 247-274. Monnet, Cyril. “Rehypothecation,” Federal Reserve Bank of Philadelphia Business Review (Fourth Quarter 2011). Sanches, Daniel. “The Optimum Quantity of Money,” Federal Reserve Bank of Philadelphia Business Review (Fourth Quarter 2012). Schleifer, Andrei. “Comments on Gorton and Metrick: Regulating the Shadow Banking System,” Brookings Papers on Economic Activity, 2 (2010), pp. 298-303. www.philadelphiafed.org Forecast Disagreement in the Survey of Professional Forecasters By Keith Sill M any people engaged in activities related to business, financial markets, and policymaking closely follow economic forecasts. Our interest in forecasts stems from the fact that, to an important degree, the decisions we make today are influenced by our expectations about the economy. Accurate forecasts lead to better decision-making and more efficient use of economic resources, and so there is a clear benefit to identifying good forecasts. An important resource for evaluating the predictions and performance of professional forecasters is the Survey of Professional Forecasters, conducted by the Philadelphia Fed Research Department’s Real-Time Data Research Center. The SPF is a quarterly survey that asks a panel of professional forecasters about their projections for a range of economic variables, including output growth, unemployment, inflation, and interest rates. When examining the SPF data, it becomes clear that professional forecasters have wide-ranging views about the future evolution of the economy. This is perhaps a bit surprising, since the statistical methods that underlie good forecasting models are well known, and professional forecasters by and large have access to the same data on the economy’s past performance. With forecasters having similar tools and data to work with, why do we observe this wide dispersion in their projections?1 Are expectations wide-ranging because of differences in models and methods used to make the forecasts? Or does the wide disagreement stem from how different forecasters process and analyze information and then use it as an input into their forecast-generation process? To design and implement effective economic policies, it is important to understand how expectations are formed. One way to do so is to study forecast disagreement. In this article we will examine some features of the forecasts that underlie the SPF and discuss what theories and evidence tell us about forecaster behavior and how expectations about the economy are formed and evolve over time. For a discussion on measuring the accuracy of the survey’s forecasts, which is beyond the scope of this article, see Stark (2010). 1 Keith Sill is a vice president and director of the Real-Time Data Research Center at the Federal Reserve Bank of Philadelphia. The views expressed in this article are not necessarily those of the Federal Reserve. This article and other Philadelphia Fed reports and research are available at www.philadelphiafed.org/research-and-data/ publications. www.philadelphiafed.org THE SURVEY’S DESIGN The SPF is the oldest quarterly survey of macroeconomic forecasts in the United States, having been initiated in 1968 under the leadership of Victor Zarnowitz at the American Statistical Association and the National Bureau of Economic Research. After conducting what was then known as the ASA-NBER Quarterly Economic Outlook Survey for 22 years, the ASANBER turned the survey over to the Federal Reserve Bank of Philadelphia in 1990, at which time it was renamed the Survey of Professional Forecasters. The Philadelphia Fed’s Real-Time Data Research Center now conducts the SPF. A panel of professional forecasters (there are usually around 45 respondents per survey) is asked to give projections for a range of major macroeconomic variables over various time horizons.2 To maintain high quality, the SPF screens its participants. Most have had advanced training in economic theory and statistics and use statistical models to generate their projections. To keep the integrity of the survey high, participation is limited to those employed by firms or paid by clients to generate forecasts now or in the past. Because of these criteria and the types of individuals who participate in the SPF, we 2 The survey results are released to the public free of charge at 10 a.m. on the second or third Friday of the second month of each quarter. The release schedule and the results of current and past surveys, as well as the underlying data, including anonymized individual forecaster projections, are available at http://philadelphiafed. org/research-and-data/real-time-center/surveyof-professional-forecasters. Business Review Q2 2014 15 can surmise that fairly sophisticated models and statistical methods underlie their projections. SPF participants use their models to forecast quarterly values of major macroeconomic variables for up to five quarters, including the current quarter, and annual projections up to three years ahead. In addition, the SPF asks for long-term annual averages for headline and core inflation, real GDP growth, productivity growth, and stock and bond returns. A somewhat unusual feature of the survey is that, instead of asking participants just for single forecasts for output growth, inflation, and the unemployment rate, it asks them to assign probabilities to different outcomes and so gives a more comprehensive picture of the forecasters’ views of the future.3 SPF FORECAST DISAGREEMENT For the most part, the main, or “headline,” forecast numbers reported in the SPF are the median values across forecasters. Each median, though, belies the variation that exists among individual projections for key variables that describe the macroeconomy. In fact, the range of forecast values underlying the median can be substantial, and it changes over time. At times, the forecasters show more agreement and at other times more disagreement in their projections. The Real-Time Data Research Center website provides data on SPF forecast disagreement for the variables that are regularly reported in the SPF. The measure of disagreement that is reported is the difference between the 75th percentile and the 25th percentile of the forecasts, which is called the interquartile range. In other words, 3 See my 2012 article for more on forecast uncertainty and the forecast probabilities that are reported in the SPF. That article also presents some evidence on how forecast disagreement affects the macroeconomy. 16 Q2 2014 Business Review suppose there are 100 separate forecasts for annual real GDP growth in 2014. Order the forecasts from highest value to lowest value, and take the difference between the 75th slot and the 25th slot as the measure of disagreement. We measure disagreement in this way in order to ensure that any outliers among the forecasts, perhaps due to mistaken entries in the respondent questionnaires, do not unduly influence the measure of disagreement. Figures 1 through 3 show plots of disagreement measured by the interquartile range from the center’s website for real GDP growth, GDP price index inflation, and the unemployment rate. Each measure of disagreement is for the four-quarters-ahead forecast as of the date on the horizontal axis. Roughly speaking, this suggests that about 50 percent of the forecasts fall within a range of about 0.4 percentage point below to about 0.4 percentage point above the median forecast. The other 50 percent of the forecasts are even further away from the median. Consequently, the disagreement among the forecasters seems not too large but nonetheless represents a significant difference between the top and bottom of the distribution. By way of comparison, the standard deviation of quarterly real GDP growth from the first quarter of 1991 to the third quarter of 2013 was about 2.5 percentage points at an annual rate. Recall, though, that the measure of disagreement shown in Figures 1 through 3 is somewhat conservative. The range of forecast values underlying the median can be substantial, and it changes over time. The charts show that disagreement generally tended to be higher in the survey’s early years — the late 1970s and early 1980s — compared with the latter half of the sample. Broadly speaking, this pattern of declining disagreement tracks the period known as the Great Moderation from 1984 to 2008, when the overall volatility of the economic data was lower than in the pre-1984 period.4 How do we interpret the data in Figures 1 through 3? Take the case of disagreement for real GDP growth. Since the early 1990s, the disagreement for forecasts of real GDP growth four quarters ahead has bounced around in a range of 0.5 percentage point to 1.5 percentage points, with an average of 0.86 percentage point. For more on the Great Moderation, see my 2004 article, “What Accounts for the Postwar Decline in Economic Volatility?” 4 To calculate it, we make no use of the forecasts in the top and bottom 25 percent of the distribution — which, if included, would widen the disagreement. Indeed, this potentially wide disagreement is part of the reason that the SPF generally reports median rather than average forecasts. The median is the midmost forecast when forecasts are ranked from high to low. So, unlike with the average forecast, the effect of outliers is discounted. If we use all the forecasts to calculate the standard deviation across projections of four-quarters-ahead real GDP growth, we obtain Figure 4. For the most part, the standard deviation measure tracks the interquartile range measure fairly closely, though it is clearly more volatile, especially early in the sample. This volatility may partly reflect reporting errors by members of the forecast panel. Sometimes an SPF respondent will submit www.philadelphiafed.org FIGURES 1–3 Significant Dispersion for Key Indicators Figure 1: Real GDP Growth Rate Dispersion by Quarter Interquartile range of 4-quarters-ahead forecasts, Percent 5.0 4.5 4.0 3.5 3.0 2.5 2.0 1.5 1.0 0.5 0 1974 1978 1982 1986 1990 1994 1998 2002 2006 2010 Figure 2: Unemployment Rate Dispersion by Quarter Interquartile range of 4-quarters-ahead forecasts, Percent 1.6 1.4 1.2 1.0 0.8 0.6 0.4 0.2 0 1974 1978 1982 1986 1990 1994 1998 2002 2006 2010 Figure 3: GDP Deflator Inflation Rate Dispersion by Quarter Interquartile range of 4-quarters-ahead forecasts, Percent 3.0 2.5 2.0 1.5 1.0 0.5 a forecast that seems extreme relative to those of the other respondents. If the outlier appears to be an error, that forecast is removed. In general, though, it is difficult to identify reporting errors versus actual views. For example, in the survey for the first quarter of 2013, the interquartile range for the four-quarters-ahead real GDP growth projections is 0.6, while the standard deviation is 1.7. However, one forecaster had entered a four-quarters-ahead real GDP growth forecast of 12.6 percent — which might have been a reporting error. If we exclude that forecast, the standard deviation falls to 0.832, which is much closer to the interquartile range. The median and the interquartile range are less affected by such outliers. We can see this in Figure 5, which plots the forecasts for four-quartersahead real GDP growth from the SPF for the first quarter of 2013. The gold dots are observations in the upper 25 percent and lower 25 percent tails of the distribution. We see two outliers, one calling for 12.6 percent growth and one calling for a 1.5 percent contraction. We cannot say for sure that these were reporting errors, but it seems possible. Excluding those two, the remainder fall in a range of about 2 percent to 4.2 percent, while the central tendency ranges from 2.5 percent to 3.1 percent. Disagreement still significant. If we use all the forecasts to calculate the standard deviation, we can construct confidence intervals for the forecasters. A confidence interval indicates the probability of a forecast falling within a certain range. Typically, a 95 percent confidence interval is plus or minus two standard deviations around the mean estimate.5 In the case of forecasts for real GDP growth four quarters 0 1974 1978 1982 1986 1990 1994 1998 2002 2006 Source: Survey of Professional Forecasters, Federal Reserve Bank of Philadelphia. www.philadelphiafed.org 2010 5 This is the case if the observations are drawn from a normal distribution. Business Review Q2 2014 17 FIGURE 4 Standard Deviation Generally Tracks GDP Forecast Range Interquartile range of 4-quarters-ahead forecasts, Percent 8 7 Interquartile range Standard deviation 6 5 4 3 2 Figure 7 shows the forecasts for GDP deflator inflation from the 2013 first quarter SPF. Again, two look suspicious — both calling for significant deflation — and would have a large impact on the standard deviation of the forecasts. Excluding the deflation outliers, the forecasts range from about 1.3 percent for inflation four quarters ahead to almost 4 percent. The central tendency is narrow at 1.7 percent to 2.3 percent. But we again see that professional forecasters can have strikingly different views about how the economy will evolve over the next 12 months. 1 0 1974 1978 1982 1986 1990 1994 1998 2002 2006 2010 Sources: Survey of Professional Forecasters, Federal Reserve Bank of Philadelphia; author’s calculations. ahead, this calculation suggests that we can estimate that 95 percent of the forecasts will, on average, be in a range of about 1.7 percentage points above to 1.7 percentage points below the mean of the forecasts (the average standard deviation of the four-quarters-ahead projections from the first quarter of 1992 to the first quarter of 2013 is 0.85). So, if the average forecast was 3 percent, we would estimate that 95 percent of the forecasts for four-quarters-ahead real GDP growth would fall in a range of about 1.3 percent to 4.7 percent. This range highlights that the SPF forecasters typically have fairly divergent views on how real output growth is likely to evolve in the nottoo-distant future. When looking at the interquartile ranges of the forecasts for the unemployment rate and GDP deflator inflation, we also see a tendency toward a decline in average disagreement in the post-1990 sample. As it had for real GDP growth, disagreement among these forecasts increased following the 18 Q2 2014 Business Review recession that began in December 2007. Disagreement for the unemployment rate forecast is a bit lower than that for inflation or output growth, possibly because the high persistence in the unemployment rate makes it a somewhat easier variable to forecast. Nevertheless, the range of views on future unemployment rates is significant. Take the results of the survey taken in the first quarter of 2013. Figure 6 plots the forecasts for the four-quarters-ahead unemployment rate from high to low. The gold dots again denote the upper and lower quartiles of the distribution. We see that while the interquartile range was fairly small at around 0.5, the range of overall forecasts was larger. Some forecasters thought that the economy would make little if any progress on the unemployment rate (the median current-quarter forecast for the 2013 first quarter unemployment rate was 7.8 percent). But some thought unemployment would fall below 7 percent. Most thought it would be in a range of 7.3 percent to 7.6 percent. DIFFERENT MODELS AND METHODS The forecasters who make up the SPF panel use a variety of statistical models to help them make their projections, and this variety of models surely plays a role in forecast disagreement. But how large a role might that be? Their models generally fall into one of two major categories: reducedform models and structural models. Reduced-form models impose little, if any, economic theory to refine their structure. For example, one of the simplest forecasting models for real GDP growth is to suppose that current GDP growth is related to past GDP growth in a linear fashion. To forecast a greater range of variables, the model can be expanded by adding more lagged variables to form a system of equations that relates current values of variables such as output growth, inflation, and interest rates to their lagged values. A forecaster using such a system of equations chooses which variables to have in the system as well as the number of lags of variables to use. Once those choices are made, the model can be estimated using historical data and then used to generate forecasts. One does not need to bring much economic theory to bear when specifying such a reduced-form model, since there are typically no restrictions on the estiwww.philadelphiafed.org FIGURES 5–7 Strikingly Different Year-Ahead Projections Figure 5: Real GDP Growth Rate Forecasts Q1 2013 4-quarters-ahead forecasts, Percent 14 12 10 8 6 4 2 0 -2 -4 0 10 20 Top and bottom quartiles Interquartile range 30 40 50 Individual survey responses Figure 6: Unemployment Rate Forecasts Q1 2013 4-quarters-ahead forecasts, Percent 8.2 8.0 7.8 7.6 7.4 7.2 7.0 6.8 6.6 0 10 20 Top and bottom quartiles Interquartile range 30 40 50 Individual survey responses Figure 7: GDP Deflator Inflation Rate Forecasts Q1 2013 4-quarters-ahead forecasts, Percent 5 4 3 2 1 0 -1 -2 -3 -4 0 10 20 Top and bottom quartiles Interquartile range 30 40 Individual survey responses Source: Survey of Professional Forecasters, Federal Reserve Bank of Philadelphia. www.philadelphiafed.org 50 mated coefficients and no restrictions on how the model’s variables interact with each other. Another approach to building a forecast model is to use economic theory to restrict the way in which the variables interact. For example, one might stipulate that the relationship between household consumption and hours worked in the labor market is related to the real wage in some particular way, or that the relationship between current and future consumption is tied in a specific way to the interest rate. Models that impose economic theory on the data’s interrelationships are called structural models. Like reduced-form models, structural models relate values of variables to their lagged values. But structural models use economic theory to impose complex relationships among those variables. Structural models are especially useful for honing the story behind the forecast. For example, such models may indicate that output growth will rise because of higher demand today, or that inflation is expected to fall because firms expect the marginal costs of production to fall in the future. These kinds of stories that are consistent with economic theory are typically more difficult to tease out when using reduced-form models. There is another element that usually is an important part of the forecast process: judgment. Often forecasters will examine the historical errors in the equations of their models — that is, how much the predicted value from the equation differed from that actual value in the data. If the forecast is persistently missing on the high or low side, the forecaster may alter the equation away from the estimated values a bit so that its predictions are more in line with the most recent data observations. This is a judgment-based adjustment of the forecast whereby forecasters subjectively alter the predictions generated by the statistical model to Business Review Q2 2014 19 bring them more into conformity with the recent behavior of the economy and their own views on how the future is likely to unfold. Typically, data, models, and judgment are combined to produce the final forecast. If forecasters are using different models on which to base their forecasts, then we might expect to see disagreement in those forecasts. Some models might be more accurate descriptions of the economy than others. If one forecasting methodology is consistently better than another and the data are able to discriminate among the models, we should see bad models being driven out over time by good models. Likewise, we should see that forecasters using the best models should consistently produce better forecasts than other forecasters do. That is, if model heterogeneity is the most important reason that forecasts differ and the data are informative about the models, then we should be able to identify forecasters and models that reliably outperform their peers. Evidence casts doubt. However, some evidence from the forecast evaluation literature casts some doubt that model heterogeneity is the key element behind forecast disagreement. First, one of the most robust findings from the forecasting literature is that mean forecasts systematically outperform individual forecasts. That is to say that over time, a more accurate forecast can be had by taking the average of many different forecasts rather than sticking with one individual forecaster’s projections. If one forecaster and his or her model were consistently producing better forecasts, then we wouldn’t expect to see such a gain from averaging.6 Why does this forecast averaging tend to work so well compared with any one forecast over time? If there was one, known, true model of 6 See the 2006 article and references therein by Allan Timmermann. 20 Q2 2014 Business Review the economy, then the forecasts from that model would dominate alternative models. But we don’t have the true model of the economy. Yet, because the economy is so complex, different models may capture different features of the economy in a successful way. At certain times, some of those features may be more important for successful forecasts than at other times. So by averaging a wide range of forecasts, we can incorporate many features of the economy that are impossible to capture shocks should generate persistently better forecasts as long as the shocks a particular model is good at analyzing are still important drivers of the dynamics of the economy. We might then expect that some forecasters will give more accurate forecasts for several reporting periods in a row. However, another finding from the empirical forecasting literature is that the best forecaster in any one period is no more likely to be the best in the next period.7 So, although economic shocks are per- By averaging a wide range of forecasts, we can incorporate many features of the economy that are impossible to capture in any single model and on average make more accurate forecasts. in any single model and on average make more accurate forecasts. Aside from forecast averaging, is there other evidence that suggests model differences may not be the key factor underlying forecast dispersion? Models can differ in how they incorporate economic shocks, which are defined as unpredictable disturbances to the economy from events such as the outbreak of war or an unexpected surge in global commodity prices. For example, some models might be better at predicting how the economy will respond to oil price shocks, while other models might be good at predicting how the economy will respond to fiscal policy shocks. Therefore, depending on the specific mix of shocks hitting the economy at any one time, some models may produce the most accurate forecasts for one period, only to have their relative predictive power reversed when a different set of shocks hits. However, shocks and their effects tend to persist. That is, the impact tends to decline slowly as the shocks fully work their way through the economy. This persistence implies that models that are especially accurate for particular sistent, forecast performance is not. The findings that average forecasts tend to outperform individual forecasts and that top forecasters don’t stay on top for long suggest that differences in models may not be the most important element behind forecast disagreement. But as is often true in economics, the case is not so clear-cut. Recent research by Andrew Patton and Allan Timmermann examines how forecast disagreement changes with the forecast horizon. Using survey data from Consensus Economics, they find greater disagreement among longer-term forecasts than near-term forecasts. Because variables such as real GDP growth and inflation tend to return to their mean values over time, the observation that long-horizon forecasts show more disagreement is consistent with the idea that differences in economic models are an important factor. This is because different models might be calibrated to yield different long-run averages for key variables, and model-based long-run forecasts will 7 See the 2007 article by Michael Bryan and Linsey Molloy. www.philadelphiafed.org FIGURE 8 No Wider Variation in Longer-Term Forecasts Range of 1-Quarter-Ahead GDP Forecasts Forecasts for real GDP growth, Percent 8 6 4 2 0 -2 2010 Q1 2011 Q1 2012 Q1 2013 Q1 Range of 4-Quarters-Ahead Forecasts Forecasts for real GDP growth, Percent 8 6 4 2 0 -2 2010 Q1 2011 Q1 2012 Q1 2013 Q1 Source: Survey of Professional Forecasters, Federal Reserve Bank of Philadelphia. Note: In the upper chart, one outlier forecast from the 2012Q3 survey was eliminated. In the lower chart, one outlier from each of the following surveys was eliminated: 2010Q1, 2012Q3, and 2013Q1. reflect those differences. This pattern is more difficult to discern in recent SPF projections for real GDP growth. However, the longest forecasts we can use from the SPF are looking only four quarters ahead from the quarter in which the survey is conducted, which is more in the realm of near-term forecasts. With that caveat, Figure 8 plots individual forecasts of one-quarter-ahead and four-quartersahead real GDP growth from the first quarter of 2010 to the second quarter of 2013. Looking at the panels in the www.philadelphiafed.org figure, there is no clear tendency for the longer-horizon forecasts to show more disagreement than the shorterhorizon forecasts over this period.8 If we were to extend the data sample back to 1970, we would see the same basic pattern: There is no obvious increase in disagreement when we move from one8 Patton and Timmermann were able to compare forecast horizons as short as one month with those as long as 24 months. In addition, they used the Consensus Forecast survey for 1991 to 2008, which typically had about 25 forecasters in the panel. quarter-ahead to four-quarters-ahead real GDP forecasts in the SPF. In considering how near-term forecasts might differ from long-term forecasts, it is important to note that near-term forecast disagreement is more likely to be influenced by current information that is used to kick off the forecast. As noted earlier, a forecast combines models with data and judgment to generate a projection. The timeliness of the data that a forecaster has in hand when making a forecast and the extent to which the forecaster discerns the true state of the economy from that data are critical elements when projecting the economy’s future state. Perhaps the data and information analysis that go into forecasts are also key drivers of forecast disagreement.9 AN IMPERFECT SIGNAL As new data become available and are put into the forecasting models, the projections are modified, sometimes dramatically. So how forecasters respond to the arrival of new information is an important factor in the forecast-generation process. When constructing models of the economy, economists often assume for simplicity’s sake that people costlessly receive all the information they need to make their decisions. In reality, though, information is costly to process and often subject to revision over time. Take the release of quarterly real GDP data. An initial estimate is released in the month after the end of the quarter, a second release two months after the end of the quarter, and a final release three months after the end of the quarter. But even that’s not really the “final” release, since the estimate will again be revised in July for the next several years, and then again every five years or so with benchmark revisions 9 Note, though, that without some discipline, models with heterogeneous forecasting rules can rationalize any disagreement outcome. Business Review Q2 2014 21 to the national income and product accounts.10 Forecasters looking at today’s data to put into their models realize that they have only an imperfect signal of the true state of the economy at any point in time. In contrast to the case of full information, we can say that there are information “frictions” that make the true state of the economy difficult and costly to assess. There are two prominent theories of imperfect information in macroeconomics that have different implications for what we should observe in forecast surveys such as the SPF. The first, often referred to as the sticky-information theory, is described in a 2002 paper by Greg Mankiw and Ricardo Reis. In this theory, economic agents such as forecasters are assumed to update the information they use to make decisions and forecasts randomly — with a certain probability each period that is independent of economic conditions or past decisions. It is as if each day people play an information lottery. If they win the lottery, they go ahead and update their view of the world based on current data. And the assumption is that they receive full information about the state of the economy when they update. If they lose, they don’t update their information and continue to make decisions and forecasts based on stale data. Clearly, this is an extreme view of the world and is unlikely to be strictly true. But if there is a fixed cost to acquiring and processing new information, then households and firms 10 The most recent benchmark revisions were conducted in the summer of 2013. According to the Bureau of Economic Analysis, comprehensive revisions encompass (1) updated definitions and classifications to more accurately portray the evolving U.S. economy, (2) changes in presentations to make the NIPA tables more informative, and (3) statistical changes that introduce improved methodologies and newly available and revised source data. See http:// www.bea.gov/newsreleases/national/gdp/ gdpnewsrelease.htm. 22 Q2 2014 Business Review will update their information only infrequently.11 Another prominent theory of information frictions assumes instead that agents monitor the flow of data and update their information continually, but that the information they receive about economic fundamentals is contaminated by random “noise” that obscures the signals they are interested in. For example, take the case of a monetary policymaker who is concerned about the behavior of inflation when setting interest rates. At a point in time, the policymaker has observations on current and past inflation and tries to discern the trend in inflation from transitory movements that are likely to dissipate over time. Inflation may be higher today because of, say, a temporary weather shock that affects food prices. The policymaker is more likely concerned with the underlying trend rate of inflation but must make some inference about that unobserved trend from the underlying data. So, he or she doesn’t necessarily possess the full information. More generally, time and resources must be spent to best estimate the desired information, and because time and resources are costly, there are tradeoffs in deciding how to process information. Information frictions have implications for forecast behavior, including forecast disagreement. How do forecasters respond to economic shocks when there are information rigidities? In the case of sticky information, not all forecasters are updating their projections in response to the shock at the same time.12 Since the respondents are surveyed at the same time, we would hypothesize that the average forecast 11 That the presence of fixed costs of updating information can rationalize the sticky-information model is derived in a 2006 article by Ricardo Reis. It is plausible, for instance, that an economist employed by a bank or other firm to generate 12 will be somewhat inertial and slow to adjust to the shock. Consequently, forecast errors will persistently be above or below zero (depending on whether the shock is positive or negative) for some time, though eventually everyone updates his or her information and the average forecast error returns to zero.13 Similarly, in the case of imperfect information, the change in the average forecast in response to the shock is inertial — in this case only a fraction of the signal about economic fundamentals is incorporated into the current estimate of the underlying state of the economy so that adjustment to the data is only partial. This slow response is reflected in the persistence of forecasters’ beliefs about the underlying state of the economy, which in turn leads to a somewhat inertial adjustment of the forecasts to shocks to economic fundamentals. Effect of shocks on disagreement. So, both imperfect information theories suggest that forecast errors might be persistently below or above zero in response to shocks (but that the errors converge to zero over time). But what about forecast disagreement? Here, the two theories offer different predictions. The sticky-information theory predicts that disagreement will rise in response to shocks. This happens because not all forecasters are updating their information sets and forecasts after a shock, so the forecasts of those who do update may move further away from the forecasts of those who don’t: Disagreement increases. The noisy-information theory predicts that disagreement should not respond to shocks. In this theory, forecasters conforecasts largely for internal use might update those forecasts less frequently than would a forecaster who primarily sells forecasts to a wide clientele. Key references for the macroeconomic implications of imperfect information literature include Lucas (1972), Sims (2003), and Woodford (2003). 13 www.philadelphiafed.org tinually monitor and react to the flow REFERENCES of data.14 The idea is easier to grasp if we assume that forecasters are using the same model. Then, if they are also monitoring the data continually, the disagreement among forecasts arises from idiosyncratic differences in how the forecasters process the information that ends up being fed into their models. For example, suppose the data for the model include the real interest rate, which is not observed directly but must instead be inferred by subtracting the expected rate of inflation from the nominal interest rate. Depending on how they measure expected inflation, different forecasters can arrive at different measures of the real interest rate. As long as the ways that individual forecasters process information do not themselves respond to shocks, then the dispersion of the forecasts will not vary in response to shocks. To sum up then, both imperfectinformation theories predict that forecast errors should show some persistence, but they have different implications for forecast disagreement. The sticky-information theory suggests that disagreement rises in response to shocks, while the noisy-information theory suggests that it doesn’t.15 IS THERE EMPIRICAL EVIDENCE? Is there empirical evidence on the role of imperfect information in forecaster behavior? Recent work by Olivier Coibion and Yuriy Gorodnichenko examines whether the implications This assumes that the dispersion of the idiosyncratic noise shocks that forecasters receive does not respond to economic fundamentals such as inflation or output. However, if forecasters who receive the same signal interpret it differently, then forecast dispersion can be correlated with economic shocks. See Coibion and Gorodnichenko (2012) for details. of imperfect-information theories are found in the forecast data. They use a variety of surveys, including the SPF, to investigate how forecast errors and disagreement respond to shocks and whether that response can be rationalized by imperfect-information theories. Assessing the response of economic variables to shocks can be tricky because shocks themselves are often unobserved and so have to be information problems are resolved. Indeed, Coibion and Gorodnichenko generally find serially correlated forecast errors in response to a variety of shocks, especially TFP and oil price shocks. For example, after an inflationary shock, they find a predictable sequence of serially correlated positive inflation forecast errors. Over time, though, these errors converge back to zero — just as the theory predicts One key prediction of the imperfect information theories is that forecast errors should show some persistence in response to shocks. identified from the data. Coibion and Gorodnichenko identify several shocks, including monetary policy shocks, total factor productivity (TFP) shocks, oil shocks, and fiscal policy shocks, and then use regression methods to see how inflation forecast errors and forecast disagreement respond to these shocks.16 For the most part, the forecast horizon under investigation is one year ahead. Recall that one key prediction of the imperfect information theories is that forecast errors should show some persistence in response to shocks. If a shock hits the economy and forecasters either don’t incorporate it quickly into their forecasts or else have a hard time extracting the relevant signals from the data, then their forecasts are likely to over- or under-shoot until these they should. If there were no information problems confronting forecasters, we would expect that forecast errors would in turn not show predictable patterns in the data. Coibion and Gorodnichenko also examine how forecast disagreement responds to shocks. Recall that this response has the potential to distinguish between the sticky-information and noisy-information theories. After examining how disagreement changes in response to many different shocks that hit the economy over the past 30 years or so, they conclude that, on balance, structural shocks do not seem to notably increase disagreement across forecasters. This finding gives an edge to the noisy-information theory, though for other dimensions of the data, the sticky-information theory does better.17 16 Total factor productivity is the residual in accounting for economic output after the contributions of labor and capital inputs have been measured. It can be viewed as the contribution of technological change to output growth. A monetary policy shock can be thought of as the surprise component of the monetary policy instrument. It is the difference between a realized policy outcome — usually a short-term interest rate — and the rate that had been predicted by a specific model. 17 14 The two theories have additional predictions for the data besides those we have discussed here. See the 2012 paper by Coibion and Gorodnichenko for a fuller exposition. 15 www.philadelphiafed.org Some of Coibion and Gorodnichenko’s findings, though, are more consistent with the sticky-information theory. For example, the convergence rate of forecast errors is just as rapid for monetary policy shocks as it is for TFP shocks. Under the noisy-information theory, if TFP shocks were more important for determining productivity and economic growth, one might expect forecasters to pay more attention to these shocks, which implies that forecast error convergence would differ among shocks. Business Review Q2 2014 23 It seems that elements of both theories are present in the data. This evidence indicates that imperfect-information theories are consistent with the forecast data, though discriminating between the two theories is difficult. Is imperfect information important or insignificant in the data? Coibion and Gorodnichenko also try to answer this question using data from the SPF on inflation forecasts. In the context of the sticky-information model, their estimates imply an average duration of six to seven months between information updates on the part of forecasters.18 In the context of imperfect-information models, the estimates imply that new information receives less than half the weight it would if there was no imperfect information and forecasters did not have to extract the relevant signals. Thus, the evidence is consistent with imperfect information having a significant role in how forecasters process new information and how often they acquire it. 18 It seems puzzling that professional forecasters would update their forecasts infrequently, but other researchers have also found patterns in the data consistent with this implication. Two such studies are Clements (2012) and Andrade and Le Bihan (2010). CONCLUSION Forecast disagreement is large and varies over time. Differences in how forecasters see the economy evolving in the future can be attributed to the models they use and the way they monitor and incorporate into their forecasts the heavy, continual flow of information on the state of the economy. Empirical evidence from forecast surveys suggests that both model heterogeneity and imperfect information about the economy play roles in the wide dispersion in professional forecasts. BR REFERENCES Andrade, Philippe, and Hervé Le Bihan. “Inattentive Professional Forecasters,” Banque de France Working Paper 307 (2010). Bryan, Michael F., and Linsey Molloy. “Mirror, Mirror, Who’s the Best Forecaster of Them All?” Federal Reserve Bank of Cleveland Economic Commentary (2007). Clements, Michael P. “Do Professional Forecasters Pay Attention to Data Releases?” International Journal of Forecasting, 28 (April-June 2012), pp. 297-308. Coibion, Olivier, and Yuriy Gorodnichenko. “What Can Survey Forecasts Tell Us About Information Rigidities?” Journal of Political Economy, 120:1 (2012) pp. 116-159. Coibion, Olivier, and Yuriy Gorodnichenko. “Information Rigidity and the Expectations Formation Process: A Simple Framework and New Facts,” National Bureau of Economic Research Working Paper 16537 (2010). 24 Q2 2014 Business Review Lucas, Robert E. “Expectations and the Neutrality of Money,” Journal of Economic Theory, 4 (April 1972), pp. 103-124. Mankiw, N. Gregory, and Ricardo Reis. “Sticky Information Versus Sticky Prices: A Proposal to Replace the New Keynesian Phillips Curve.” Quarterly Journal of Economics, 117 (November 2002), pp. 1,295-1,328. Patton, Andrew J., and Allan Timmermann. “Why Do Forecasters Disagree? Lessons from the Term Structure of Cross-Sectional Disagreement,” Journal of Monetary Economics, 57:7 (2010), pp. 803-820. Reis, Ricardo. “Inattentive Producers,” Review of Economic Studies, 73 (July 2006), pp. 793-821. Timmermann, Allan. “Forecast Combinations,” in G. Elliott, C. Granger, and A. Timmermann, eds., Handbook of Economic Forecasting, 1:1 (2006), pp. 135-196. Sill, Keith. “What Accounts for the Postwar Decline in Economic Volatility?” Federal Reserve Bank of Philadelphia Business Review (Fourth Quarter 2004). Sill, Keith. “Measuring Economic Uncertainty Using the Survey of Professional Forecasters,” Federal Reserve Bank of Philadelphia Business Review (Fourth Quarter 2012). Sims, Christopher A. “Implications of Rational Inattention,” Journal of Monetary Economics, 50 (April 2003), pp. 665-690. Stark, Tom. “Realistic Evaluation of RealTime Forecasts in the Survey of Professional Forecasters,” Federal Reserve Bank of Philadelphia Special Report (May 2010). Woodford, Michael. “Imperfect Common Knowledge and the Effects of Monetary Policy,” in P. Aghion, R. Frydman, J. Stiglitz, and M. Woodford, eds., Knowledge, Information, and Expectations in Modern Macroeconomics in Honor of Edmund Phelps, Princeton, N.J.: Princeton University Press (2003). www.philadelphiafed.org Research Rap Abstracts of research papers produced by the economists at the Philadelphia Fed Economists and visiting scholars at the Philadelphia Fed produce papers of interest to the professional researcher on banking, financial markets, economic forecasting, the housing market, consumer finance, the regional economy, and more. More abstracts may be found at www.philadelphiafed.org/ research-and-data/publications/research-rap/. You can find their full working papers at http://www.philadelphiafed.org/research-and-data/publications/working-papers/. The Continuing Power of the Yield Spread in Forecasting Recessions The authors replicate the main results of Rudebusch and Williams (2009), who show that the use of the yield spread in a probit model can predict recessions better than the Survey of Professional Forecasters. Croushore and Marsten investigate the robustness of their results in several ways: extending the sample to include the 2007-09 recession, changing the starting date of the sample, changing the ending date of the sample, using rolling windows of data instead of just an expanding sample, and using alternative measures of the “actual” value of real output. The results show that the Rudebusch-Williams findings are robust in all dimensions. Working Paper 14-5. Dean Croushore, University of Richmond and Federal Reserve Bank of Philadelphia Visiting Scholar; Katherine Marsten, University of Richmond. Continuous Markov Equilibria with Quasi-Geometric Discounting The authors prove that the standard quasigeometric discounting model used in dynamic consumer theory and political economics does not possess continuous Markov perfect equilibria (MPE) if there is a strictly positive lower bound on wealth. The authors also show that, at points of discontinuity, the decision maker strictly prefers lotteries over the next period’s assets. The authors then extend the standard model to have lotteries and establish the existence of an MPE with continuous decision www.philadelphiafed.org rules. The models with and without lotteries are numerically compared, and it is shown that the model with lotteries behaves more in accord with economic intuition. Working Paper 14-6. Satyajit Chatterjee, Federal Reserve Bank of Philadelphia; Burcu Eyigungor, Federal Reserve Bank of Philadelphia. The Economics of Debt Collection: Enforcement of Consumer Credit Contracts In the U.S., third-party debt collection agencies employ more than 140,000 people and recover more than $50 billion each year, mostly from consumers. Informational, legal, and other factors suggest that original creditors should have an advantage in collecting debts owed to them. Then, why does the debt collection industry exist and why is it so large? Explanations based on economies of scale or specialization cannot address many of the observed stylized facts. The authors develop an application of common agency theory that better explains those facts. The model explains how reliance on an unconcentrated industry of third-party debt collection agencies can implement an equilibrium with more intense collections activity than creditors would implement by themselves. The authors derive empirical implications for the nature of the debt collection market and the structure of the debt collection industry. A welfare analysis shows that, under certain conditions, an equilibrium in which creditors rely on third-party debt collectors can generate more credit supply and aggregate borrower surplus than an equilibrium where lenders collect debts owed to them on their own. There are, however, Business Review Q2 2014 25 situations where the opposite is true. The model also suggests a number of policy instruments that may improve the functioning of the collections market. Working Paper 14-7. Viktar Fedaseyeu, Bocconi University and Federal Reserve Bank of Philadelphia Visiting Scholar; Robert M. Hunt, Federal Reserve Bank of Philadelphia. Foreclosure Delay and Consumer Credit Performance The deep housing market recession from 2008 through 2010 was characterized by a steep increase in the number of foreclosures. Foreclosure timelines — the length of time between initial mortgage delinquency and completion of foreclosure — also expanded significantly, averaging up to three years in some states. Most individuals undergoing foreclosure are experiencing serious financial stress. However, extended foreclosure timelines enable mortgage defaulters to live in their homes without making housing payments until the completion of the foreclosure process, thus providing a liquidity benefit. This paper tests whether the resulting liquidity was used to help cure nonmortgage credit delinquency. The authors find a significant relationship between longer foreclosure timelines and household performance on nonmortgage consumer credit during and after the foreclosure process. Their results indicate that a longer period of nonpayment of housing-related expenses results in higher cure rates on delinquent nonmortgage debts and improved household balance sheets. Foreclosure delay may have mitigated the impact of the economic downturn on credit card default. However, credit card performance may deteriorate in the future as the current foreclosure backlog is cleared and the affected households once again incur housing expenses. Working Paper 14-8. Paul Calem, Federal Reserve Bank of Philadelphia; Julapa Jagtiani, Federal Reserve Bank of Philadelphia; William W. Lang, Federal Reserve Bank of Philadelphia. Competing for Order Flow in OTC Markets The authors develop a model of a two-sided asset market in which trades are intermediated by dealers and are bilateral. Dealers compete to attract order flow by posting the terms at which they execute trades, which can include prices, quantities, and execution times, and investors direct their orders toward dealers that offer the most attractive terms of trade. Equilibrium outcomes have the following properties. First, investors face a trade-off between trading costs and speeds of execution. Second, the asset market is endogenously segmented in the sense that investors with different asset valuations and different asset holdings will trade at different speeds and different costs. For example, under a Leontief technology to match investors and dealers, per unit trading costs decrease with the size of the trade, in accordance with the evidence from the market for corporate bonds. Third, 26 Q2 2014 Business Review dealers’ implicit bargaining powers are endogenous and typically vary across sub-markets. Finally, the authors obtain a rich set of comparative statics both analytically, by studying a limiting economy where trading frictions are small, and numerically. For instance, the authors find that the relationship between trading costs and dealers’ bargaining power can be hump-shaped. Working Paper 14-9. Benjamin Lester, Federal Reserve Bank of Philadelphia; Guillaume Rocheteau, University of California–Irvine; Pierre-Olivier Weill, University of California–Los Angeles. Forecasting Credit Card Portfolio Losses in the Great Recession: A Study in Model Risk Credit card portfolios represent a significant component of the balance sheets of the largest US banks. The charge-off rate in this asset class increased drastically during the Great Recession. The recent economic downturn offers a unique opportunity to analyze the performance of credit risk models applied to credit card portfolios under conditions of economic stress. Specifically, the authors evaluate three potential sources of model risk: model specification, sample selection, and stress scenario selection. Their analysis indicates that model specifications that incorporate interactions between policy variables and core account characteristics generate the most accurate loss projections across risk segments. Models estimated over a time frame that includes a significant economic downturn are able to project levels of credit loss consistent with those experienced during the Great Recession. Models estimated over a time frame that does not include a significant economic downturn can severely underpredict credit loss in some cases, and the level of forecast error can be significantly impacted by model specification assumptions. Higher credit-score segments of the portfolio are proportionally more severely impacted by downturn economic conditions and model specification assumptions. The selection of the stress scenario can have a dramatic impact on projected loss. Working Paper 14-10. José J. Canals-Cerdá, Federal Reserve Bank of Philadelphia; Sougata Kerr, Federal Reserve Bank of Philadelphia. Misallocation, Informality, and Human Capital: Understanding the Role of Institutions The aim of this paper is to quantify the role of formalsector institutions in shaping the demand for human capital and the level of informality. The authors propose a firm dynamics model where firms face capital market imperfections and costs of operating in the formal sector. Formal firms have a larger set of production opportunities and the ability to employ skilled workers, but informal firms can www.philadelphiafed.org avoid the costs of formalization. These firm-level distortions give rise to endogenous formal and informal sectors and, more importantly, affect the demand for skilled workers. The model predicts that countries with a low degree of debt enforcement and high costs of formalization are characterized by relatively lower stocks of skilled workers, larger informal sectors, low allocative efficiency, and measured TFP. Moreover, the authors find that the interaction between entry costs and financial frictions (as opposed to the sum of their individual effects) is the main driver of these differences. This complementarity effect derives from the introduction of skilled workers, which prevents firms from substituting labor for capital and in turn moves them closer to the financial constraint. Working Paper 14-11. Pablo N. D’Erasmo, University of Maryland and Federal Reserve Bank of Philadelphia; Hernan J. Moscoso Boedo, University of Virginia; Asli Senkal, University of Virginia. banks. Banks accumulate securities like Treasury bills and undertake short-term borrowing when there are cash flow shortfalls. A nontrivial size distribution of banks arises out of endogenous entry and exit, as well as banks’ buffer stocks of securities. The authors test the model using business cycle properties and the bank lending channel across banks of different sizes studied by Kashyap and Stein (2000). They find that a rise in capital requirements from 4% to 6% leads to a substantial reduction in exit rates of small banks and a more concentrated industry. Aggregate loan supply falls and interest rates rise by 50 basis points. The lower exit rate causes the tax/output rate necessary to fund deposit insurance to drop in half. Higher interest rates, however, induce higher loan delinquencies as well as a lower level of intermediated output. Working Paper 14-13. Dean Corbae, University of Wisconsin–Madison and National Bureau of Economic Research; Pablo N. D’Erasmo, Federal Reserve Bank of Philadelphia. Market Exposure and Endogenous Firm Volatility over the Business Cycle The authors propose a theory of endogenous firm-level volatility over the business cycle based on endogenous market exposure. Firms that reach a larger number of markets diversify market-specific demand risk at a cost. The model is driven only by total factor productivity shocks and captures the business cycle properties of firm-level volatility. Using a panel of U.S. firms (Compustat), the authors empirically document the countercyclical nature of firm-level volatility. They then match this panel to Compustat’s Segment data and the U.S. Census’s Longitudinal Business Database (LBD) to show that, consistent with their model, measures of market reach are procyclical, and the countercyclicality of firm-level volatility is driven mostly by those firms that adjust the number of markets to which they are exposed. This finding is explained by the negative elasticity between various measures of market exposure and firm-level idiosyncratic volatility the authors uncover using Compustat, the LBD, and the Kauffman Firm Survey. Working Paper 14-12. Ryan Decker, University of Maryland; Pablo N. D’Erasmo, University of Maryland and Federal Reserve Bank of Philadelphia; Hernan J. Moscoso Boedo, University of Virginia. Trade Adjustment Dynamics and the Welfare Gains from Trade The authors build a micro-founded two-country dynamic general equilibrium model in which trade responds more to a cut in tariffs in the long run than in the short run. The model introduces a time element to the fixed-variable cost trade-off in a heterogeneous producer trade model. Thus, the dynamics of aggregate trade adjustment arise from producer-level decisions to invest in lowering their future variable export costs. The model is calibrated to match salient features of new exporter growth and provides a new estimate of the exporting technology. At the micro level, the authors find that new exporters commonly incur substantial losses in the first three years in the export market and that export profits are back-loaded. At the macro level, the slow export expansion at the producer level leads to sluggishness in the aggregate response of exports to a change in tariffs, with a long-run trade elasticity that is 2.9 times the short-run trade elasticity. The authors estimate the welfare gains from trade from a cut in tariffs, taking into account the transition period. While the intensity of trade expands slowly, consumption overshoots its new steady-state level, so the welfare gains are almost 15 times larger than the long-run change in consumption. Models without this dynamic export decision underestimate the gains to lowering tariffs, particularly when constrained to also match the gradual expansion of aggregate trade flows. Working Paper 14-14. George Alessandria, Federal Reserve Bank of Philadelphia; Horag Choi, Monash University; Kim Ruhl, New York University Stern School of Business. Capital Requirements in a Quantitative Model of Banking Industry Dynamics The authors develop a model of banking industry dynamics to study the quantitative impact of capital requirements on bank risk taking, commercial bank failure, and market structure. They propose a market structure where big, dominant banks interact with small, competitive fringe www.philadelphiafed.org Business Review Q2 2014 27 TM Now Available: The Federal Reserve Historical Inventory and History Web Gateway The Federal Reserve’s inventory of historical collections offers students, researchers, and others online access to thousands of documents and artifacts related to the Fed’s 100-year history from sources across the Federal Reserve System, universities, and private collections. To view the inventory, go to http://www. federalreserve.gov/newsevents/press/other/other20120530a1. pdf. More content and features will be added over time. Do you know of materials that should be included? Information may be submitted at http://www.federalreserve.gov/apps/contactus/ feedback.aspx. Also available is the new Federal Reserve History Web Gateway. Designed to encourage deeper reflection on the Fed’s role in the nation’s economy, the gateway presents biographies, timelines, photographs, and more that tell the story of the Fed’s purpose, key economic events affecting U.S. history, and the people who shaped the Fed. Go to http://www.federalreservehistory.org/. On December 23, 1913, President Woodrow Wilson signed the Federal Reserve Act, establishing the Federal Reserve System as the U.S. central bank. Its mission is to conduct the nation’s monetary policy; supervise and regulate banks; maintain the stability of the financial system; and provide financial services to depository institutions, the U.S. government, and foreign official institutions. The Federal Reserve Board opened for business on August 10, 1914, and on November 16, 1914, the 12 regional Reserve Banks were open for business. Congress designed the Fed with a decentralized structure. The Federal Reserve Bank of Philadelphia — serving eastern Pennsylvania, southern New Jersey, and Delaware — is one of 12 regional Reserve Banks that, together with the seven-member Board of Governors in Washington, D.C., make up the Federal Reserve System. The Board, appointed by the President of the United States and confirmed by the Senate, represents the public sector, while the Reserve Banks and the local citizens on their boards of directors represent the private sector. The Research Department of the Philadelphia Fed supports the Fed’s mission through its research; surveys of firms and forecasters; reports on banking, markets, and the regional and U.S. economies; and publications such as the Business Review. PRESORTED STANDARD U.S. POSTAGE PAID Federal R eserve Bank of Philadelphia PHILADELPHIA, PA PERMIT #583 Ten Independence Mall Philadelphia, Pennsylvania 19106-1574 ADDRESS SERVICE REQUESTED www.philadelphiafed.org Past and current Business Review articles can be downloaded for free from our website. There you will also find data and other information on the regional and national economy, consumer finance issues, resources for teachers, information on our community development initiatives, the latest research publications from the Federal Reserve Bank of Philadelphia, and more. explore and learn