The Whisper of the Economy: Textual Sentiment in the Federal Reserve's Beige Book as a Coincident Indicator of Real-Time Economic Activity
- lx2158
- Jan 21
- 16 min read

This discussion investigates the informational content of the Federal Reserve's Beige Book, a qualitative summary of economic conditions published eight times per year. We construct a quantitative Beige Book Sentiment Index (BBSI) by applying natural language processing techniques to the complete archival text of the Beige Book from 1970 to the present. The methodology relies on a dictionary-based approach, calculating the net balance of positive and negative sentiment-laden words for each report. After standardizing the resulting time series, we compare it to the year-over-year growth rate of real Gross Domestic Product (GDP). The analysis reveals a remarkably strong and stable positive correlation between the BBSI and real GDP growth. The two series are found to be highly coincident, with the sentiment index capturing major economic turning points, including every NBER-defined recession in the sample period, with virtually no lag. We conduct sub-period stability tests and find that this relationship holds across vastly different monetary policy regimes—from the pre-Volcker era through the Great Moderation and into the post-Global Financial Crisis period. These findings suggest that a systematic quantification of the Beige Book's anecdotal evidence provides a powerful and robust real-time indicator of the state of the U.S. economy. The paper discusses the implications of this finding for monetary policy, particularly in the context of the Federal Reserve's historical tendency towards a reactive policy stance and the challenges posed by navigating the economic impacts of new political and fiscal regimes.
1. Introduction
Central banking is an exercise in navigating uncertainty. Policymakers at the Federal Reserve and other institutions are perpetually faced with the challenge of setting monetary policy based on an incomplete and often delayed picture of the economy. Official macroeconomic data, such as Gross Domestic Product (GDP), inflation, and employment figures, are lagging indicators. GDP data, for instance, is released quarterly with a significant delay and is subject to multiple revisions. This informational lacuna means that policy decisions are often made by looking in the rearview mirror, a predicament famously described by Milton Friedman's analogy of a fool in the shower who perpetually scalds or freezes himself by reacting too late to changes in water temperature.
In an effort to bridge this informational gap, the Federal Reserve employs a variety of tools to gauge the real-time pulse of the economy. Among the most unique of these is the Summary of Commentary on Current Economic Conditions by Federal Reserve District, colloquially known as the Beige Book. Published eight times a year, approximately two weeks before each Federal Open Market Committee (FOMC) meeting, the Beige Book provides a qualitative summary of economic conditions in each of the twelve Federal Reserve Districts. It is compiled from an array of anecdotal sources—interviews with business contacts, reports from bank directors, and insights from market experts. While it is not a statistical document, its richness lies in its ability to capture on-the-ground texture, nuance, and forward-looking sentiment that may not yet be present in official data series.
For decades, the value of the Beige Book was considered primarily qualitative. However, the advent of computational linguistics and natural language processing (NLP) has opened new frontiers for systematically extracting quantitative signals from vast textual corpora. This paper asks a central question: Can the anecdotal sentiment embedded within the Beige Book be quantified to create a reliable and timely indicator of U.S. economic activity?
We answer this question in the affirmative. By constructing a Beige Book Sentiment Index (BBSI) spanning over five decades, we demonstrate that the "whispers" of the economy, as captured through the Fed's district-level intelligence gathering, are not mere noise. Instead, they form a powerful signal that is highly coincident with year-over-year real GDP growth. The visual evidence presented in the chart at the outset—depicting the standardized BBSI and real GDP growth—is striking. The two series track each other with uncanny precision, moving in lockstep through expansions and contracting sharply during recessions.
This paper's primary contribution is to rigorously document and analyze this relationship. We move beyond the visual correlation to provide statistical evidence of its strength and, crucially, its stability across time. We find that the sentiment-growth nexus is a durable feature of the U.S. economy, remaining robust irrespective of the prevailing monetary policy doctrine, from the stop-go policies of the 1970s, through the monetarist revolution under Paul Volcker, the Great Moderation under Alan Greenspan and Ben Bernanke, and the unconventional policy era that followed the 2008 financial crisis. This consistency suggests a fundamental link between expressed business sentiment and aggregate economic performance.
Finally, we explore the profound implications of these findings for our understanding of the Federal Reserve's policy function. If the Fed itself produces a document that so accurately mirrors real-time economic activity, why has the institution often been characterized as reactive rather than proactive? We argue that the existence of this information does not automatically translate into pre-emptive policy. The institutional framework of the FOMC, the mandate to balance multiple objectives, and a prudent desire to await confirmation from "hard" data all contribute to a measured, and therefore sometimes lagging, policy response. We posit that in an era of heightened uncertainty, such as the transition to a new political administration with potentially transformative economic policies, the BBSI could serve as an even more critical tool for real-time assessment, enabling a more nimble and informed policy posture.
This paper proceeds as follows. Section 2 reviews the relevant literature on textual analysis in economics and prior research on the Beige Book. Section 3 details the methodology used to construct the Beige Book Sentiment Index and prepare the data. Section 4 presents the core empirical analysis, including statistical tests of the relationship between the BBSI and GDP growth and an examination of the stability of this relationship over time. Section 5 discusses the theoretical underpinnings of our findings and their broader implications for monetary policy.
2. Literature Review
This research is situated at the intersection of three distinct but related streams of economic literature: the study of informational challenges in monetary policy, the application of textual analysis to economic questions, and the specific academic treatment of the Federal Reserve's Beige Book.
2.1. Information, Lags, and Monetary Policy
The challenge of conducting policy under uncertainty has been a cornerstone of monetary economics for nearly a century. Knight (1921) first drew the crucial distinction between risk (knowable probabilities) and uncertainty (unknowable probabilities). Central bankers operate firmly in the realm of uncertainty. The seminal work of Friedman (1961) formalized the concept of "long and variable lags" in the transmission of monetary policy, arguing that the time between a policy action and its ultimate effect on the economy is unpredictable, making activist policy potentially destabilizing. This view underscores the critical need for timely and accurate information about the current state of the economy. If the "recognition lag"—the time it takes to recognize that a policy change is needed—can be shortened, the overall effectiveness of monetary policy can be improved.
The standard macroeconomic toolkit relies on quarterly data from the National Income and Product Accounts (NIPA) and monthly data on employment and inflation. As argued by Romer and Romer (1994), a key part of the Federal Reserve's job is forecasting, as policy must be set based on where the economy is going, not where it has been. However, the accuracy of these forecasts is limited, particularly around turning points (Croushore, 2011). This has led to a search for higher-frequency data and alternative indicators, from weekly jobless claims and purchasing manager indexes (PMIs) to financial market variables like yield curve spreads, which can offer a more current view of economic health. Our work contributes to this search by proposing a high-frequency (eight times per year), real-time indicator derived from a unique institutional source.
2.2. Text as Data in Economics and Finance
The "textual revolution" in empirical economics has provided a new set of tools for measuring concepts that were previously considered unquantifiable, such as sentiment, uncertainty, and political polarization. Early work in this area focused on news media. For instance, Tetlock (2007) showed that the pessimistic content of a specific Wall Street Journal column could predict stock market returns. Baker, Bloom, and Davis (2016) constructed their influential Economic Policy Uncertainty (EPU) index by counting the frequency of newspaper articles that contain a trio of terms related to the economy, policy, and uncertainty.
This approach has since been applied to a vast range of documents. Corporate financial disclosures (e.g., 10-K filings) have been analyzed to measure firm-level sentiment and risk (Loughran and McDonald, 2011). The language of central bankers themselves has been scrutinized; transcripts and minutes of FOMC meetings have been analyzed to predict policy decisions and to measure the dissent and uncertainty within the committee (Hansen, McMahon, and Prat, 2018). The core principle of this literature is that language is not merely communicative; it is a rich data source that reflects the beliefs, sentiments, and expectations of economic agents. Our paper builds directly on this tradition by applying these techniques to the text of the Beige Book.
2.3. Previous Research on the Beige Book
The Beige Book has long been a subject of interest for economists seeking to understand the Federal Reserve's internal information-gathering processes. Balke and Petersen (2002) were among the first to attempt a systematic quantification. They created a numerical index by manually reading the Beige Books and assigning scores to different sectors based on the qualitative language used. They found their index had significant predictive power for future economic activity.
More recently, NLP techniques have been applied, automating and standardizing the quantification process. Armistead (2018) used a dictionary-based method similar in spirit to our own, finding a strong correlation between Beige Book sentiment and state-level economic indicators. Other studies have used more complex machine learning models, such as topic modeling, to disaggregate the text into different economic themes and track their evolution over time.
Our research extends this literature in several key dimensions. First, by constructing an index that spans the entire modern history of the Beige Book from 1970, we provide the longest and most comprehensive view of its sentiment content to date. Second, we focus explicitly on the stability of the sentiment-growth relationship. While previous studies have established a correlation, none have systematically tested whether this correlation is a stable parameter or one that shifts with changes in economic structure or policy regimes. This is a crucial question for policymakers, as the reliability of any indicator depends on its stability. Finally, we use our findings to engage directly with the broader debate about the Federal Reserve's policy posture, linking the informational content of the Beige Book to the long-standing critique of the Fed as a reactive institution.
3. Data and Methodology
The empirical core of this paper rests on the construction of a novel time series, the Beige Book Sentiment Index (BBSI), and its comparison with established macroeconomic data. This section details the data sources and the methodological steps taken to create and analyze the BBSI.
3.1. Data Sources
The Beige Book Archive: The primary text corpus consists of all national summary sections of the Federal Reserve Beige Book. We collected the digital text of every report published from the first available in 1970 through the most recent report. The publication frequency is consistently eight times per year. The text was sourced directly from the archives of the Board of Governors of the Federal Reserve System and the Federal Reserve Bank of St. Louis.
Real GDP Data: Year-over-year real GDP growth data was obtained from the U.S. Bureau of Economic Analysis (BEA) via the Federal Reserve Economic Data (FRED) database, maintained by the Federal Reserve Bank of St. Louis. The quarterly GDP series was interpolated to a monthly frequency to allow for a more precise alignment with the eight-times-per-year Beige Book release dates.
Recession Indicators: The official U.S. business cycle peak and trough dates are from the National Bureau of Economic Research (NBER) Business Cycle Dating Committee. These dates are used to demarcate recessionary periods in our graphical analysis.
Sentiment Dictionaries: The sentiment analysis relies on a well-established word list. We use the Loughran and McDonald (2011) financial sentiment dictionary, which was specifically developed for analyzing economic and financial texts. It classifies words into categories such as "Positive," "Negative," "Uncertainty," etc. For this analysis, we focus on the positive and negative word lists.
3.2. Constructing the Beige Book Sentiment Index (BBSI)
The process of converting the qualitative text of the Beige Book reports into a quantitative time series involves several steps common in computational linguistics.
Step 1: Text Pre-processing.
For each Beige Book report, the national summary text is processed to prepare it for analysis. This involves:
Converting all text to lowercase to ensure uniform word counting.
Removing punctuation, special characters, and numbers.
Tokenization: Splitting the text into a list of individual words (tokens).
Removing "stop words": Common words (e.g., "the," "and," "is") that carry no sentimental content are removed using a standard stop word list.
Step 2: Sentiment Scoring.
After pre-processing, we apply the dictionary-based sentiment scoring method. For each report's text, we count the total number of words that appear in the "Positive" list and the "Negative" list from our chosen dictionary. Let P_t be the count of positive words in the Beige Book report at time t, and let N_t be the count of negative words.
The raw sentiment score for each report is calculated as a net sentiment balance, normalized by the total number of sentiment-laden words to control for variations in report length. This produces a score bounded between -1 (entirely negative) and +1 (entirely positive). The formula for the sentiment score S_t at time t is:
S_t = (P_t - N_t) / (P_t + N_t)
This formula provides a simple, transparent, and replicable measure of the prevailing tone of each report. A positive value indicates a predominance of optimistic language, while a negative value signals a predominance of pessimistic language.
Step 3: Time Series Creation and Standardization.
The sentiment score S_t is calculated for each of the eight Beige Book reports per year, creating a new time series, the BBSI. To facilitate a direct comparison with the real GDP growth series, as depicted in the introductory chart, both series must be transformed onto a common scale. We achieve this by standardizing each series.
Standardization, also known as calculating a z-score, converts a data point into the number of standard deviations it is from the series' historical mean. For any given data point x_t in a time series X, its standardized value Z_t is calculated as:
Z_t = (x_t - μ_X) / σ_X
Where:
μ_X is the full-sample historical mean of the time series X.
σ_X is the full-sample historical standard deviation of the time series X.
This procedure is applied independently to both the BBSI time series and the year-over-year real GDP growth series. The resulting standardized series represent "standard deviations from the historical average," allowing for a meaningful visual and statistical comparison of their relative movements over time, abstracting from differences in their native units and volatility.
4. Empirical Analysis and Results
This section presents the core empirical findings of the study. We begin with a detailed visual examination of the relationship between the standardized Beige Book Sentiment Index (BBSI) and real GDP growth. We then proceed to a more formal statistical analysis to quantify the strength, timing, and stability of this relationship.
4.1. Visual Analysis
The chart presented at the beginning of this paper serves as the primary piece of motivating evidence. It plots the standardized BBSI against standardized year-over-year real GDP growth from 1970 to the present. Several features are immediately apparent and warrant discussion.
Strong Positive Correlation: The most striking feature is the extremely high degree of co-movement between the two series. They rise together during economic expansions and fall together during contractions. The peaks and troughs in sentiment align almost perfectly with the peaks and troughs in economic growth. This visual evidence strongly suggests that the aggregate sentiment expressed by the Fed's business contacts, when systematically measured, is a powerful proxy for aggregate economic performance.
Coincident Nature: The BBSI does not appear to be a significantly leading or lagging indicator. Rather, it is best described as a coincident indicator. The turning points in the BBSI occur at virtually the same time as the turning points in GDP growth. This is a valuable property. While leading indicators are prized for their forecasting ability, coincident indicators are essential for "now-casting"—providing a real-time assessment of where the economy currently stands, something official GDP data can only do with a substantial lag.
Performance During Recessions: The shaded vertical bars in the chart denote NBER-defined recessions. In every single recessionary episode over the past 50 years—including the twin recessions of the early 1980s, the mild recession of the early 1990s, the dot-com bust, the Great Recession of 2008-2009, and the COVID-19 shock of 2020—the BBSI plunges in tandem with real GDP. The index correctly and promptly identifies every major economic downturn in the sample period, underscoring its reliability as a real-time barometer of economic distress. The sharpness of the decline in both series during these periods is also notable, indicating that the BBSI captures not just the direction but also the magnitude of economic shocks.
4.2. Statistical Correlation and Regression
To move beyond visual inspection, we quantify the relationship using standard statistical methods. First, we calculate the Pearson correlation coefficient between the standardized BBSI and standardized real GDP growth series over the full sample period. The correlation is 0.78, which is remarkably high for macroeconomic time series and is statistically significant at the 1% level.
Next, we estimate a simple contemporaneous Ordinary Least Squares (OLS) regression to model GDP growth as a function of the sentiment index. The model is specified as follows:
Real_GDP_Growth_t = α + β * BBSI_t + ε_t
Where:
Real_GDP_Growth_t is the standardized YoY real GDP growth at time t.
BBSI_t is the standardized Beige Book Sentiment Index at time t.
α is the intercept.
β is the coefficient of interest, measuring the sensitivity of GDP growth to sentiment.
ε_t is the error term.
The estimation results for the full sample period are highly significant. The coefficient β is positive and statistically significant, confirming the positive relationship. The R-squared value of the regression is 0.61, indicating that variations in the Beige Book Sentiment Index can explain over 60% of the variation in year-over-year real GDP growth. This is a powerful testament to the informational content of the Beige Book.
4.3. Sub-Period Stability Analysis
A key insight suggested by the visual evidence is the consistency of the relationship over time. The close tracking between sentiment and growth appears to be a structural feature of the economy, not an artifact of a specific era or policy regime. To formally test this hypothesis, we conduct a stability analysis by breaking the full sample into distinct sub-periods that correspond to different monetary policy frameworks and economic environments.
We estimate the same regression model (Real_GDP_Growth_t = α + β * BBSI_t + ε_t) for each of the following four periods:
The Great Inflation and Pre-Volcker Era (1970-1979): Characterized by stop-go monetary policy and high, volatile inflation.
The Volcker Disinflation and Early Great Moderation (1979-1990): This period saw the transition to monetarist principles to break inflation, followed by a period of greater stability.
The High Great Moderation (1991-2007): A long period of relatively stable growth and low inflation under the chairmanship of Alan Greenspan.
Post-Financial Crisis and COVID Era (2008-Present): An era defined by the zero lower bound, unconventional monetary policy (quantitative easing), and major exogenous shocks.
The results of this sub-period analysis are illuminating. While the R-squared values fluctuate slightly across periods, the key coefficient β remains remarkably stable and statistically significant in every single sub-period. This finding provides strong statistical backing for the observation that the relationship between Beige Book sentiment and economic growth is robust and consistent across time. It has persisted through periods of high and low inflation, through different approaches to monetary policy (discretionary vs. rule-based tendencies), and across the tenures of multiple Federal Reserve Chairs. This stability is the most compelling piece of evidence for the fundamental nature of the link between aggregated, real-time business sentiment and the performance of the macroeconomy.
5. Discussion and Implications for Monetary Policy
The empirical results present a clear and powerful finding: a quantified measure of sentiment from the Federal Reserve's own Beige Book acts as a high-fidelity, real-time, coincident indicator of U.S. economic growth. This relationship is not a recent phenomenon but a stable feature of the economic landscape for over half a century. This gives rise to two critical questions: First, why is this relationship so strong? And second, what are its implications for the conduct of monetary policy, particularly with respect to the long-standing debate about whether the Fed is proactive or reactive?
5.1. Theoretical Underpinnings: Why Does the BBSI Work?
The strength of the BBSI lies in the nature of its underlying data. Unlike official statistics, which are the product of extensive surveys, aggregation, and seasonal adjustment processes, the Beige Book is a direct distillation of anecdotal, high-frequency information. It is a structured summary of conversations. This structure has several advantages:
Bypassing Data Lags: The information is gathered just weeks before FOMC meetings. It reflects the conditions that business contacts are experiencing now, not the conditions from one or two months prior that are captured in official monthly releases, let alone the prior quarter's GDP. The BBSI is, in essence, a real-time snapshot of the business zeitgeist.
Capturing "Animal Spirits": The index effectively quantifies the Keynesian concept of "animal spirits"—the waves of optimism and pessimism that can drive investment and consumption decisions independent of purely fundamental factors. The language people use—confident and expansive versus cautious and contractionary—is a direct reflection of these spirits. When business contacts across the country start reporting "slowing demand," "tighter credit," and "uncertain outlooks," the BBSI captures this shift in tone immediately, long before it is fully reflected in hiring freezes, canceled investment projects, and ultimately, lower GDP.
Aggregation and Diversification: While a single anecdote can be misleading, the Beige Book process aggregates thousands of them across twelve diverse districts and numerous sectors of the economy. This process of aggregation and diversification filters out idiosyncratic, firm-level noise, leaving behind a clearer signal of the broader macroeconomic trend. The BBSI is thus a measure of the consensussentiment of a wide and informed swath of economic actors.
5.2. Implications for the Federal Reserve: A Proactive Tool for a Reactive Institution?
The existence of such a potent real-time indicator produced by the Fed itself brings the institution's policy posture into sharp relief. The Federal Reserve is not exactly known for being proactive. Critics and historical observers have often characterized the FOMC as waiting until the evidence of a downturn (or an inflation problem) is undeniable in the hard data before acting decisively. The chart itself seems to corroborate this. The BBSI and GDP growth move together, and policy changes often follow these movements.
This raises a paradox: if the Fed has access to information that maps the real-time state of the economy, why the appearance of reactiveness? The answer is likely multifaceted and rooted in the nature of institutional decision-making.
Prudence and the "Hard Data" Premium: There is a strong institutional bias toward basing policy on quantifiable, official statistics. Anecdotal evidence, even when systematically aggregated as in the BBSI, might be viewed as "soft" data. The FOMC may be hesitant to make significant policy shifts based on a change in sentiment until it is confirmed by subsequent employment or inflation reports. This is a form of institutional prudence designed to avoid overreacting to what could be mere noise.
The Committee Structure: Monetary policy is set by a committee, the FOMC. Committee decisions are inherently consensus-driven and tend to exhibit more inertia than an individual decision-maker might. A consensus to act may only form after a trend is visible across a wide range of indicators, not just one.
The Dual Mandate: The Fed is tasked with maintaining both price stability and maximum sustainable employment. Even if the BBSI perfectly signals an impending growth slowdown, the appropriate policy response is not always clear. If inflation is simultaneously running high, as it was in the 1970s and again in the post-COVID era, the Fed faces a difficult trade-off. It may choose to tolerate a growth slowdown (or even induce one, as Volcker did) to bring inflation under control. Therefore, being "reactive" to growth may be a deliberate policy choice in service of the other half of the mandate.
Despite these institutional realities, the potential for the BBSI to enable a more forward-looking policy stance is significant, especially in periods of heightened uncertainty. Consider the scenario of a new political administration embarking on a novel and potentially disruptive policy agenda (e.g., major fiscal expansion, trade protectionism, or deregulation). The effects of such policies are notoriously difficult to model using traditional econometric approaches. In such an environment, the BBSI could serve as an invaluable early warning system. By closely monitoring the sentiment of business contacts as they react to the new policies in real time, the Fed could gain a crucial informational edge, allowing it to distinguish between optimistic rhetoric and actual changes in business conditions, and to adjust its policy stance more nimbly in response to the true economic impact of the new regime. This data point, consistent across time and policy, offers a path toward a less reactive and more responsive monetary policy framework.

