Two Approaches to Macroeconomic Forecasting.
Webb, Roy H.
Following World War II, the quantity and quality of macroeconomic data expanded dramatically. The most important factor was the regular
publication of the National Income and Product Accounts, which contained
hundreds of consistently defined and measured statistics that summarized
overall economic activity. As the data supply expanded, entrepreneurs
realized that a market existed for applying that increasingly
inexpensive data to the needs of individual firms and government
agencies. And as the price of computing power plummeted, it became
feasible to use large statistical macroeconomic models to process the
data and produce valuable services. Businesses were eager to have
forecasts of aggregates like gross domestic product, and even more eager
for forecasts of narrowly defined components that were especially
relevant for their particular firms. Many government policymakers were
also enthusiastic at the prospect of obtaining forecasts that quantified
the most likely effects of policy actions.
In the 1960s large Keynesian macroeconomic models seemed to be
natural tools for meeting the demand for macroeconomic forecasts.
Tinbergen (1939) had laid much of the statistical groundwork, and Klein
(1950) built an early prototype Keynesian econometric model with 16
equations. By the end of the 1960s there were several competing models,
each with hundreds of equations. A few prominent economists questioned
the logical foundations of these models, however, and macroeconomic
events of the 1970s intensified their concerns. At the time, some
economists tried to improve the existing large macroeconomic models, but
others argued for altogether different approaches. For example, Sims
(1980) first criticized several important aspects of the large models
and then suggested using vector autoregressive (VAR) models for
macroeconomic forecasting. While many economists today use VAR models,
many others continue to forecast with traditional macroeconomic models.
This article first describes in more detail the traditional and VAR
approaches to forecasting. It then examines why both forecasting methods
continue to be used. Briefly, each approach has its own strengths and
weaknesses, and even the best practice forecast is inevitably less
precise than consumers would like. This acknowledged imprecision of
forecasts can be frustrating, since forecasts are necessary for making
decisions, and the alternative to a formal forecast is an informal one
that is subject to unexamined pitfalls and is thus more likely to prove
inaccurate.
1. TRADITIONAL LARGE MACROECONOMIC MODELS
These models are often referred to as Keynesian since their basic
design takes as given the idea that prices fail to clear markets, at
least in the short run. In accord with that general principle, their
exact specification can be thought of as an elaboration of the textbook
IS-LM model augmented with a Phillips curve. A simple version of an
empirical Keynesian model is given below:
[C.sub.t] = [[alpha].sub.1] + [[beta].sub.11]([Y.sub.t] - [T.sub.t]
+ [[epsilon].sub.1,t] (1)
[I.sub.t] = [[alpha].sub.2] + [[beta].sub.21]([R.sub.t] -
[[[pi].sup.e].sub.t+1]) + [[epsilon].sub.2,t] (2)
[M.sub.t] = [[alpha].sub.3] + [[beta].sub.31][Y.sub.t] +
[[beta].sub.32][R.sub.t] + [[epsilon].sub.3,t] (3)
[[pi].sub.t] = [[alpha].sub.4] + [[beta].sub.41]
[Y.sub.t]/[[Y.sup.p].sub.t] + [[epsilon].sub.4,t] (4)
[[[pi].sup.e].sub.t+1] = [[theta].sub.51][[pi].sub.t] +
[[theta].sub.52][[pi].sub.t-1] (5)
Y [equivalent to] [C.sub.t] + [I.sub.t] + [G.sub.t]. (6)
Equation (1) is the consumption function, in which real consumer
spending C depends on real disposable income Y - T. In equation (2),
business investment spending I is determined by the real interest rate R
- [[pi].sup.e]. Equation (3) represents real money demand M, which is
determined by real GDP Y and the nominal interest rate [R.sub.t] [1] In
equation (4), inflation is determined by GDP relative to potential GDP )
[Y.sup.p]; in this simple model, this equation plays the role of the
Phillips curve. [2] And in equation (5), expected inflation [[pi].sup.e]
during the next period is assumed to be a simple weighted average of
current inflation and the previous period's inflation. Equation (6)
is the identity that defines real GDP as the sum of consumer spending,
investment spending, and government spending G. In the stochastic equations, [epsilon] is an error term and [alpha] and [beta] are
coefficients that can be estimated from macro data, usually by ordinary
least squares regressions. The [theta] coefficients in equation (5) are
assumed rather than estimated. [3]
One can easily imagine more elaborate versions of this model. Each
major aggregate can be divided several times. Thus consumption could be
divided into spending on durables, nondurables, and services, and
spending on durables could be further divided into purchases of autos,
home appliances, and other items. Also, in large models there would be
equations that describe areas omitted from the simple model above, such
as imports, exports, labor demand, and wages. None of these additions
changes the basic character of the Keynesian model.
To use the model for forecasting, one must first estimate the
model's coefficients, usually by ordinary least squares. In
practice, estimating the model as written would not produce satisfactory
results. This could be seen in several ways, such as low [R.sup.2]
statistics for several equations, indicating that the model fits the
data poorly. There is an easy way to raise the statistics describing the
model's fit, however. Most macroeconomic data series in the United
States are strongly serially correlated, so simply including one or more
lags of the dependent variable in each equation will substantially boost
the reported [R.sup.2] values. For example, estimating equation (2)
above from 1983Q1 through 1998Q4 yields an [R.sup.2] of 0.02, but adding
the lagged dependent variable raised it to 0.97. What has happened is
that investment has grown with the size of the economy. The inclusion of
any variable with an upward trend will raise the reported [R.sup.2]
statistic. The lagged dependent variable is a convenie nt example of a
variable with an upward trend, but many other variables could serve
equally well. This example illustrates that simply looking at the
statistical fit of an equation may not be informative, and economists
now understand that other means are necessary to evaluate an empirical
equation or model. At the time the Keynesian models were being
developed, however, this point was often not appreciated.
Once the model's coefficients have been estimated, a
forecaster would need future time paths for the model's exogenous variables. In this case the exogenous variables are those determined by
government policy-G, T, and M-and potential GDP, which is determined
outside the model by technology. And although the money supply is
ultimately determined by monetary policy, the Federal Reserve's
policy actions immediately affect the federal funds rate. Thus rather
than specifying a time path for the money supply, analysts would
estimate the money demand equation and then rearrange the terms in order
to put the interest rate on the left side. The future time path for
short-term interest rates then became a key input into the forecasting
process, although its source was rarely well documented.
Next, one could combine the estimated model with the recent data
for endogenous variables and future time paths for exogenous variables
and produce a forecast. With most large Keynesian models that initial
forecast would require modification. [4] The reason for modifying the
forecast is to factor in information that was not included in the model.
For example, suppose that the model predicted weak consumer spending for
the current quarter, but an analyst knew that retail sales grew rapidly
in the first two months of the quarter. Or suppose that the analyst
observes that consumer spending had been more robust than the model had
predicted for the last several quarters. Also, the model's forecast
might display some other property that the analyst did not believe, such
as a continuously falling ratio of consumer spending to GDP. These are
all examples of information that could lead an analyst to raise the
forecast for consumer spending above the model's prediction. To
change the forecast an analyst would use "add factors," which
are additions to the constant terms in the equations above. Thus if one
wanted to boost predicted consumer spending by $100 billion in a
particular quarter, the analyst would add that amount to the constant
term for that quarter. In the model given above, there are four constant
terms represented by the [alpha] coefficients. To forecast ahead eight
quarters, one could consider 32 possible add factors that could modify
the forecast. Add factors have long been a key part of the process that
uses Keynesian models to produce forecasts and are still important. For
example, an appendix to a recent forecast by Data Resources, a leading
econometric forecasting service that uses a Keynesian model, lists over
10,000 potential add factors.
2. CRITICISMS OF KEYNESIAN MODELS FOR FORECASTING
One of the most critical components of an economywide model is the
linkage between nominal and real variables. The Phillips curve relation
between wage or price growth and unemployment rates provided that key
linkage for Keynesian macroeconomic models. The Phillips curve was
discovered, however, as an empirical relationship. Thus when it was
first incorporated in Keynesian models, it did not have a firm
theoretical foundation in the sense that it was not derived from a model
of optimizing agents. Milton Friedman (1968) criticized the simple
Phillips curve, similar to equation (5), at the time that it appeared to
be consistent with the unemployment and inflation rates that had been
observed in the 1950s and the 1960s. His concern was that the Phillips
curve may at times appear to give a stable relation between the amount
of slack in the economy and the inflation rate. But suppose that the
Federal Reserve were to ease monetary policy in an attempt to
permanently raise output above potential. The model above ign ores the
fact that people would eventually figure out the new policy strategy,
and thus, according to Friedman's logic, an expectations formation
equation such as (5) would no longer hold. In the long run, he argued,
an attempt to hold output above potential would fail; expectations would
fully adjust to the new policy and output would return to potential, but
inflation would be permanently higher.
Friedman's verbal exposition was very influential, but it did
not contain a fully specified analytical model. Using a formal model
that captured Friedman's insight, Lucas (1972) introduced rational
expectations to macroeconomic analysis as a key element for constructing
a dynamic macro model. Among the important conclusions of that paper, he
demonstrated that a Phillips curve could fit previously observed data
well but would not be valid if the monetary policy process were to
change. The book that contained the Lucas paper also contained several
papers that presented long-run Phillips curves from leading Keynesian
models; a representative result of those models was that a 4 percent
rate of unemployment corresponded to 3.5 percent inflation and that
higher inflation would give lower unemployment (Christ 1972).
Propositions in economics are rarely tested decisively. In this
case, though, it was soon clear that the simple Phillips curve was not a
stable, dependable relation. In the fourth quarter of 1972 the
unemployment rate averaged 5.4 percent and consumer inflation over the
previous four quarters was 3.3 percent. By the third quarter of 1975,
unemployment had risen to 8.9 percent; the inflation rate, however, did
not fall but instead rose to 11.0 percent.
In retrospect, one can identify many problems with the Keynesian
models of that period. Some could be resolved without making wholesale
change to the models. For example, most models were changed to
incorporate a natural rate of unemployment in the long run, thereby
removing the permanent trade-off between unemployment and inflation.
Also, most large Keynesian models were expanded to add an energy sector,
so that exogenous oil price changes could be factored in. But some of
the criticisms called for a fundamental change in the strategy of
building and using macroeconomic models.
One of the most influential was the Lucas (1976) critique. Lucas
focused on the use of econometric models to predict the effects of
government economic policy. Rather than thinking of individual policy
actions in isolation, he defined policy to mean a strategy in which
specific actions are chosen in order to achieve well-defined goals. As
an example of this meaning of policy, consider the possibility that the
Federal Reserve changed interest rates during the early 1960s in order
to keep GDP close to potential and inflation low. That behavior could be
represented as a reaction function such as equation (7):
[R.sub.t] = [R.sub.t-1] + [[beta].sub.61]
[Y.sub.t]/[[Y.sup.p].sub.t] + [[beta].sub.62][[pi].sub.t] +
[[epsilon].sub.6,t]. (7)
Now suppose that the reaction function changed in the late 1960s
and that less importance was placed on achieving a low rate of
inflation. One can imagine replacing equation (7) with the new reaction
function; however, Lucas argued that even with the new reaction
function, a model would not give reliable policy advice. The reason is
that the parameters of all the other equations reflect choices that were
made when the previous policy rule was in effect. Under the new policy
rule the parameters could well be significantly different in each
equation above. This result is easiest to see in equation (6), which
describes the formation of expectations of inflation in a manner that
might be reasonable for a period when the monetary authority was
stabilizing inflation. Individuals could do better, though, if the
monetary policy strategy was in the process of changing substantially.
During that period an analyst who wanted to produce reliable conditional
forecasts would need to replace equation (6), even if the model as a
whole continued to provide useful short-term forecasts of overall
economic activity. As Lucas (1976, p. 20) put it, "the features
which lead to success in short-term forecasting are unrelated to
quantitative policy evaluation,...[T]he major econometric models are
(well) designed to perform the former task only, and ... simulations
using these models can, in principle, provide no useful information as
to the actual consequences of alternative economic policies."
This critique presented a difficult challenge for macroeconomic
model builders. Every macroeconomic model is a simplification of a very
complex economy, and the Keynesian models are no exception. One of the
key elements of Keynesian models is that prices do no adjust
instantaneously to equate supply and demand in every market. The reasons
underlying sluggish price adjustment are not usually modeled, however.
Thus the models cannot answer the question of to what extent, in
response to a policy change, the sluggishness of price adjustment would
change. The Lucas critique challenged the reliability of policy advice
from models that could not answer such a basic question.
Analysts continue to offer policy advice based on Keynesian models
and also other macroeconomic models that are subject to the Lucas
critique. These analysts are in effect discounting the relevance of the
possibility that their estimated coefficients could vary under the type
of policy change analyzed by Lucas. For a succinct example of the
reasoning that would allow the use of Keynesian models for policy
analysis, consider the counterargument given by Tobin (1981, p. 392),
"Lucas's famous 'critique' is a valid point ...
[but] the critique is not so devastating that macroeconomic
model-builders should immediately close up shop. The public's
perception of policy regimes is not so precise as to exclude
considerable room for discretionary policy moves that the public would
see neither as surprises nor as signals of a systematic change in
regime. Moreover, behavioral 'rules of thumb,' though not
permanent, may persist long enough for the horizons of macroeconomic
policy-makers." Sims (1982) gave a lengthier defe nse of
traditional policy analysis.
Authors such as Lucas and Sargent (1979) and Sims (1980) also
criticized Keynesian models for not being based on intertemporal
optimizing behavior of individuals. At the time they recommended
different strategies for model building. Since that time, however, there
have been notable improvements in the economic theory embodied in
Keynesian models. For example, in the Federal Reserve Board's
FRB/US model, it is possible to simulate the model under the assumption
that the expectations of individuals are the same as the entire
model's forecasts (Brayton et al. 1997). And many modelers have
successfully derived individual equations from optimizing dynamic
models. Still, Keynesian models continue to be based on unmodeled
frictions such as sluggish price adjustment. It is therefore not
surprising that economists have explored alternative methods of
forecasting and policy analysis. One important method was proposed by
Sims (1980) and is discussed in the next section.
3. VAR MODELS
VAR models offer a very simple method of generating forecasts.
Consider the simplest reasonable forecast imaginable, extrapolating the
recent past. In practice, a reasonably accurate forecast for many data
series from the United States over the past half century can be made by
simply predicting that the growth rate observed in the previous period
will continue unchanged. One could do better, though, by substituting a
weighted average of recent growth rates for the single period last
observed. That weighted average would be an autoregressive (AR)
forecast, and these are often used by economists, at least as
benchmarks. Only slightly more complicated is the idea that, instead of
thinking of an autoregressive forecast of a single variable, one could
imagine an autoregressive forecast of a vector of variables. The
advantage of such a VAR relative to simpler alternatives would be that
it allowed for the possibility of multivariate interaction. The simplest
possible VAR is given below in equations (8) and (9), with only two
variables and only one lagged value used for each variable; one can
easily imagine using longer lag lengths and more variables:
[R.sub.t] = [a.sub.11][R.sub.t-1] + [a.sub.12][p.sub.t-1] +
[u.sub.1,t] (8)
[p.sub.t] = [a.sub.21][R.sub.t-1] + [a.sub.22][p.sub.t-1] +
[u.sub.2,t] (9)
Because of the extreme simplicity of the VAR model, it may seem
unlikely to produce accurate forecasts. Robert Litterman (1986),
however, issued a series of forecasts from small VAR models that
incorporated from six to eight variables. The results, summarized in
Table 1, are root mean squared errors (RMSEs), that is, e = [square root
of][[sigma].sub.t][([A.sub.t] - [P.sub.t]).sup.2], where e is the RMSE,
A is the actual value of a macroeconomic variable, and P is the
predicted value. One caveat is that the data summarized in this table
cover a relatively short time period, and thus it is a statistically
small sample. Over that period, in comparison with forecasts from
services using large Keynesian models, the VAR forecasts were more
accurate for real GNP more than one quarter ahead, less accurate for
inflation, and of comparable accuracy for nominal GNP and the interest
rate.
In another study, Lupoletti and Webb (1986) also compared VAR
forecasts to those of commercial forecasting services over a longer time
period than in the previous comparison. A different caveat applies to
their results, shown in Table 2. They studied simulated forecasts versus
actual forecasts from the forecasting services. While the details [5] of
the simple model were not varied to obtain more accurate forecasts, it
is inevitable in such studies that if the VAR forecasts had been
significantly less accurate, then the results probably would not have
seemed novel enough to warrant publication. That said, their
five-variable VAR model produced forecasts that, for four and six
quarters ahead, were of comparable accuracy to those of the commercial
forecasting services. The commercial services predicted real and nominal
GNP significantly more accurately for one and two quarters ahead, which
probably indicates the advantage of incorporating current data into a
forecast by using add factors. [6]
The VAR model studied by Lupoletti and Webb has five variables,
each with six lags. With a constant term, each equation contains 31
coefficients to be estimated--a large number relative to the length of
postwar U.S. time series. Although there are methods to reduce the
effective number of coefficients that need to be estimated, the number
of coefficients still rises rapidly as the number of variables is
increased. Thus as a practical matter, any VAR model will contain only a
fairly small number of variables. As a result, a VAR model will always
ignore potentially valuable data. How, then, is it possible for them to
produce relatively accurate forecasts? One possibility is that there is
only a limited amount of information in all macroeconomic time series
that is relevant for forecasting broad aggregates like GDP or its price
index and that a shrewdly chosen VAR model can capture much of that
information.
At best, then, a VAR model is a satisfactory approximation to an
underlying structure that would be better approximated by a larger, more
complex model. That more complex model would include how government
policymakers respond to economic events. The VAR approximation will be
based on the average response over a particular sample period. A
forecast from a VAR model will thus be an unconditional forecast in that
it is not conditioned on any particular sequence of policy actions but
rather on the average behavior of policymakers observed in the past. A
forecast from a Keynesian model, however, usually is based on a
particular sequence of policy actions and is referred to as a
conditional forecast--that is, conditional on that particular sequence.
Despite the Lucas critique, many users of Keynesian models seek to
determine the consequences of possible policy actions by simulating
their model with different time paths of policy actions. But, although
the Lucas critique was discussed above in reference to Keynesi an
models, it is equally valid for VAR models. To help emphasize this
point, the next section reviews some details of using a VAR model for
forecasting.
Forecasting with VAR Models
To forecast with the VAR model summarized in equations (8) and (9),
one would estimate the [a.sub.ij] coefficients, usually by ordinary
least squares, and calculate period t values based on data for period t
- 1. One can then use the period t forecasts to calculate forecasts for
period t + 1; for example, inflation forecasts in the above model would
be
[p.sub.t+1] = [a.sub.21][R.sub.t] + [a.sub.22][P.sub.t] +
[u.sub.2,t+1]
= ([a.sub.21][a.sub.11] + [a.sub.22][a.sub.21])[R.sub.t-1] +
([a.sub.21][a.sub.12] + [[a.sup.2].sub.22])[p.sub.t-1] +
[a.sub.21][u.sub.1,t] + [a.sub.22][u.sub.2,t] + [u.sub.2,t+1], (10)
where the second line in (10) was obtained by taking the first line
and substituting the right-hand sides of (8) and (9) for the estimated
values of [R.sub.t] and [p.sub.t], respectively. The above procedure can
be repeated as many times as needed to produce as long a forecast as
desired.
It is often assumed that the realizations of unknown error
terms--[u.sub.1,t], [u.sub.2,t], and [u.sub.2,t+1]--will all equal zero.
One can discard that assumption to incorporate information that was not
used to estimate the model. Suppose the above model uses monthly data,
and at the beginning of a month one knows last month's average
interest rate but not the inflation rate, which the Labor Department will release two weeks later. One could simply substitute the realized
interest rate for the estimated rate in the calculations above; in
equation (10) that would mean plugging in the realized value of
[u.sub.1,t]. Since the errors in a VAR are usually contemporaneously correlated, a realization of [u.sub.1,t] will also provide information
about [u.sub.2,t]. Specifically, the variances and covariances of the
error terms are taken from the variance-covariance matrix that was
estimated through period t - 1 when the [a.sub.ij] coefficients were
estimated; the expected value of [u.sub.2,t] is then the ratio of th e
estimated covariance of [u.sub.1] and [u.sub.2] to the estimated
variance of [u.sub.1] times the realization of [u.sub.1,t]. This
expected value of [u.sub.2,t] can then also be included in equations (8)
and (9) in order to forecast inflation in periods t and t + 1. One can
easily apply this basic method for forecasting with a VAR, and the
refinement for incorporating partial data for a period, to more
complicated models with longer lags, more variables, and deterministic terms such as constants, time trends, and dummy variables.
With this background in mind, imagine that the true structure of
the economy is given by the Keynesian model of equations (1) through (6)
along with the monetary reaction function (7). Now suppose that the VAR
model represented by equations (8) and (9) is estimated. Algebraic manipulation [7] yields the estimated coefficients of the VAR model as
functions of the underlying structural coefficients and error terms in
equations (8[minutes]) and (9[minutes]):
[[pi].sub.t] = [B.sub.1,t] + [A.sub.11][[pi].sub.t-1] +
[A.sub.12][R.sub.t-1] + [U.sub.1,t] (8[minutes])
[R.sub.t] = [B.sub.2,t] + [A.sub.21][[pi].sub.t-1] +
[A.sub.22][R.sub.t-1] + [U.sub.2,t], (9[minutes])
where
[B.sub.1,t] = [([[alpha].sub.1] + [[alpha].sub.2] -
[[beta].sub.11][T.sub.t] + [G.sub.t] + (1 -
[[beta].sub.11])([[alpha].sub.3] -
[M.sub.t])/[[beta].sub.21][[theta].sub.51] + ([[beta].sub.21] -
[[alpha].sub.4] [[beta].sub.61]/[[beta].sub.41])]/[delta]
[A.sub.11] = [[theta].sub.52][[pi].sub.t-1]/[[theta].sub.51][delta]
[A.sub.12] = [[beta].sub.21] + (1 -
[[beta].sub.11])[[beta].sub.32]/[[beta].sub.21][delta]
[U.sub.1,t] = [[e.sub.1,t] + [e.sub.2,t] = (1 -
[[beta].sub.11])[e.sub.3,t]]/[[beta].sub.21][[theta].sub.51][delta]
[B.sub.2,t] = [[[alpha].sub.1] + [[alpha].sub.2] + [[alpha].sub.3]
(1 - [[beta].sub.11]) + [[beta].sub.61]/[[beta].sub.41] [[alpha].sub.4]
+ [G.sub.t] - [T.sub.t] - (1 -
[[beta].sub.11])[M.sub.t]][[beta].sub.41][[beta].sub.62] +
[[beta].sub.61]/[[beta].sub.21][[beta].sub.41][[theta].sub.51][delta]
[A.sub.21] = -[[beta].sub.21][[theta].sub.51]([[beta].sub.41][[beta].sub.62] + [[beta].sub.61])/[[beta].sub.21][[beta].sub.41][[theta].sub.51][delta ]
[A.sub.22] = 1/[delta]
[U.sub.2,t] = [[[epsilon].sub.1,t] + [[epsilon].sub.2,t] +
[[epsilon].sub.3,t](1 - [[beta].sub.11])][[beta].sub.41][[beta].sub.62]
+ [[beta].sub.61]/[[beta].sub.21][[beta].sub.41][[theta].sub.51][delta]
+ [[epsilon].sub.4,t][[beta].sub.61]/[[beta].sub.41][delta] +
[[epsilon].sub.5,t][[delta].sup.-1]
[delta] = (1 - [[beta].sub.21] + (1 -
[[beta].sub.11])[[beta].sub.32]/[[beta].sub.21])([[beta].sub.62] +
[[beta].sub.61]/[[beta].sub.41]).
Viewing the model as equations (8[minutes]) and (9[minutes])
reveals the problematic nature of conditional forecasting with the
model. Suppose an analyst wishes to study the effect of a tighter
monetary policy on the inflation rate by first obtaining a baseline
forecast from the VAR model and then raising the interest rate
prediction by a full percentage point for the next quarter. This step
would be accomplished by feeding in a particular nonzero value for
[u.sub.2,t+1] in equation (10). However, note that in terms of the
underlying structure, the error term [U.sub.2,t] is a complicated
composite of the five error terms from the equations of the underlying
model. Yet for policy analysis it would be necessary to identify that
composite error term as a monetary policy disturbance. [8]
An identification that ignores the distinction between VAR errors,
the [u.sub.i,t]s, and the underlying structural errors, such as the
[[epsilon].sub.j,t]'s in the example above, can lead to absurd
results. Suppose one simulates a tighter monetary policy in the model
presented above by forcing the VAR model to predict higher interest
rates; the outcome is a higher inflation prediction. The reason is that,
in the quarterly macroeconomic time series of the last 50 years, the
dominant shocks to interest rates and inflation have been aggregate
demand shocks, and a positive aggregate demand shock raises interest
rates, inflation, output, and employment. The VAR model captures these
correlations. Asking the model to simulate a higher interest rate path
will lead it to predict a higher inflation path as well. Now a clever
user can tinker with the model--adding variables, changing the dates
over which the model was estimated, and so forth--and eventually develop
a VAR model that yields a lower inflation path in resp onse to higher
interest rates. At this point, though, the model would add little value
beyond reflecting the user's prior beliefs.
To recap, VAR models are unsuited to conditional forecasting
because a VAR residual tends to be such a hodgepodge. In addition, the
models are vulnerable to the Lucas critique. Suppose that the monetary
authority decided to put a higher weight on its inflation target and a
lower weight on its output target and that its new reaction function
could be represented by (7[minutes]):
[R.sup.t] = [R.sub.t-1] + ([[beta].sub.61] -
[phi])[Y.sub.t]/[[Y.sup.p].sub.t] + ([[beta].sub.62] +
[phi])[[pi].sub.t] + [[epsilon].sub.6,t] (7[minutes])
The interpretation of the VAR's coefficients in terms of the
underlying structural coefficients would also change, with each instance
of [[beta].sub.61] changing to [[beta].sub.61] - [phi] and each instance
of [[beta].sub.62] changing to [[beta].sub.62] + [phi]. Thus following a
discrete change in the monetary strategy, the VAR's coefficients
would be systematically biased and even the accuracy of its
unconditional forecasts would be compromised.
Some authors, including Sims (1982), have questioned whether large
policy changes in the United States have resulted in meaningful
parameter instability in reduced forms such as VARs. One of the most
dramatic changes in estimated coefficients in VAR equations for U.S.
data occurred in an inflation equation. Table 3 is reproduced from Webb
(1995) and shows significant changes in an inflation equation's
coefficients estimated in different subperiods. [9] The subperiods,
moreover, were determined by the author's review of minutes of the
Federal Open Market Committee in order to find monetary policy actions
that could indicate a discrete change in the monetary strategy. The
results are thus consistent with the view that the monetary reaction
function changed substantially in the mid-1960s and again in the early
1980s and that the changes in the economic structure played havoc with a
VAR price equation's coefficients.
This section has thus presented two separate reasons for
distrusting conditional forecasts from VAR models. First, their small
size guarantees that residuals will be complicated amalgamations, and no
single residual can be meaningfully interpreted as solely resulting from
a policy action. Second, applying the Lucas critique to VAR models
implies that a VAR model's coefficients would be expected to change
in response to a discrete policy change.
Several researchers who have recognized these deficiencies but were
unwilling to give up the simplicity of the VAR approach have turned to
structural VARs, or SVARs. [10] These models attempt to apply both
economic theory that is often loosely specified and statistical
assumptions to a VAR in order to interpret the residuals and conduct
meaningful policy analysis. In many studies key statistical assumptions
are that the economy is accurately described by a small number of
equations containing stochastic error terms, and that these structural
errors are uncorrelated across equations. The economic restrictions vary
considerably from model to model; the common feature is that just enough
restrictions are introduced so that the reduced-form errors, such as in
equations (7[minutes]) and (8[minutes]) above, can be used to estimate
the structural errors. For example, two of the restrictions used in a
widely cited paper by Blanchard (1989) were (1) that reduced-form GDP
errors were equal to structural aggregate demand errors, and (2) that
reduced-form unemployment errors, given output, were equal to structural
supply errors. After presenting those and other restrictions, the author
noted "There is an obvious arbitrariness to any set of
identification restrictions, and the discussion above is no
exception" (p. 1150).
It is often the case that a reader will find an identifying
assumption of an SVAR somewhat questionable. A major difficulty of the
SVAR approach is that there is no empirical method for testing a
restriction. Moreover, if different models give different results, there
are no accepted performance measures that can be used to identify
superior performance. Since there are millions of possible SVARS that
could be based on the last half century of U.S. macroeconomic data,
their results will not be persuasive to a wide audience until a method
is found to separate the best models from the rest. [11]
4. FINAL THOUGHTS ON CONDITIONAL FORECASTING
This article has discussed two approaches to macroeconomic
forecasting. Both approaches have produced econometric models that fit
observed data reasonably well, and both have produced fairly accurate
unconditional forecasts. The VAR approach was found unsuitable for
conditional forecasting and policy analysis. There is a wide division
within the economics profession on the usefulness of large Keynesian
models for policy analysis. At one extreme are those who accept the
Lucas critique as a fatal blow and accordingly see little value in using
Keynesian models for policy analysis. At the other extreme are analysts
who are comfortable with traditional Keynesian models. In the middle are
many economists with some degree of discomfort at using the existing
Keynesian models, in part due to the features that allow the models to
fit the historical data well but may not remain valid in the event of a
significant policy change. But policy analysis will continue, formally
or informally, regardless of economists' comfort with the models
and with the strategies for using them. Decisions on the setting of
policy instruments will continue to be made and will be based on some
type of analysis.
One possibility is that policy analysis and economic forecasting will be seen as two different problems requiring two different types of
models. Economists have constructed a large number of small structural
models that can be quantified and used for policy analysis. A large
number of statistical approaches to forecasting are available as well.
It is not necessary that the same model be used for both.
Keynesian models, though, are still widely used for policy
analysis, and there are actions that model builders could take to
enhance the persuasiveness of their results. One would be to publish two
forecasts on a routine basis--the usual forecast with add factors
incorporating the modelers' judgment and a mechanical forecast with
no add factors. In that way a user could easily distinguish the
judgmental content from the pure model forecast. For example, if one
wanted to determine the possible impact of a tax cut on consumption, one
would want to consider whether large add factors in a consumption
equation such as equation (1) above were needed to achieve satisfactory
results.
It would also be helpful for forecast consumers to know how much a
model's specification has changed over time. Of course one hopes
that new developments in economics are incorporated into models and that
poorly performing specifications are discarded. As a result, some
specification changes are to be expected. But if one saw that the
consumption equation of a large model had been in a state of flux for
several years, the numerous changes could signify that the model's
analysis of a tax cut's effect on consumption was based on an
unstable foundation.
In addition, it would be helpful to see more analysis of forecast
errors. At a minimum, each forecast should be accompanied by confidence
intervals for the most important variables stating the likely range of
results. As the ex post errors indicate in Tables 1 and 2, these
confidence intervals could be quite wide. For example, real GDP growth
has averaged 2.8 percent over the last 30 years. In Table 2, the RMSE
for four-quarter predictions of real growth from the best commercial
forecasting service was 2.2 percent. Thus if a model predicted real
growth to be the 2.8 percent average, and one used that RMSE as an
approximate standard deviation of future forecast errors, then one would
expect actual outcomes to be outside of a wide 0.6 to 5.0 percent range
about 30 percent of the time. Now suppose that an exercise in policy
analysis with that model revealed a difference of 1.0 percent for real
GDP growth over the next year; a user might not consider that difference
very meaningful, given the relatively large im precision of the
model's GDP forecast.
Finally, it would be especially helpful to have a detailed analysis
of errors in a manner relevant for policy analysis. For example,
continuing with the predicted effect of a tax cut, the model's
predictions could be stated in the form of a multiplier that related the
tax cut to a predicted change in real growth. That multiplier would be a
random variable that could be statistically analyzed in the context of
the whole model, and the user could be told the sampling distribution of
that statistic. Also, one would want data on how well the model
predicted the effects of tax cuts that had actually occurred in the
past.
The unifying theme of these recommendations is for model builders
to open the black box that generates forecasts. Until this supplementary
information routinely accompanies the output of large forecasting
models, many will see an exercise in policy evaluation as having
unknowable properties and value it accordingly.
[email protected]. The views expressed are the author's
and not necessarily those of the Federal Reserve Bank of Richmond or the
Federal Reserve System.
(1.) The same letter is used for GDP and personal income since in
this simple model there are no elements such as depreciation or indirect
business taxes that prevent gross national product from equaling
national income or personal income.
(2.) In this article the role of the Phillips curve is to
empirically relate the inflation rate and a measure of slack in the
economy. In a typical large Keynesian model, the Phillips curve would be
an equation that relates wage growth to the unemployment rate, with an
additional equation that relates wage growth to price changes and
another relating the unemployment rate to GDP relative to potential.
(3.) The coefficients are assumed, rather than estimated, due to
the problematic nature of existing data on actual expectations of
inflation.
(4.) Not every Keynesian model required modification, however. Fair
(1971), for example, presented a model that has evolved over time but
has not made use of the add factors defined below.
(5.) In this case the authors could have changed the start date of
the regressions used to estimate the VAR model's coefficients, the
choice of variables (monetary base versus Ml or M2, for example), or the
number of lag lengths. In addition, this model was unrestricted, whereas
most VAR forecasters use restrictions to reduce the effective number of
estimated coefficients; experimenting with methods of restricting
parameters would have lowered the average errors of the VAR forecasts.
(6.) For example, an analyst might note that labor input, measured
as employee hours, was increasing rapidly in a quarter in which GDP was
forecast to rise slowly. The unexpected increase in employee hours could
indicate that labor demand had risen due to unexpectedly rapid GDP
growth. If other data were consistent with that line of reasoning, the
analyst would then increase the constant terms in the equations
determining GDP for the current quarter and quite possibly the next
quarter as well. Since statistical agencies release important new data
every week, there are many such opportunities for skilled analysts to
improve forecast accuracy by informally incorporating the latest data.
(7.) In brief, substitute equations (1) and (2) into (6), solve for
Y, then substitute the resulting expression for Y into equation (3), and
rearrange terms so that [[pi].sub.t] is on the left. Next, solve
equations (4) and (7) for Y/[Y.sup.p], equate the resulting expressions,
and rearrange terms so that [R.sub.t] is on the left. The resulting two
equations for [[pi].sub.t] and [R.sub.t] can be solved for each variable
as an expression containing lagged values of [pi] and R, exogenous
variables, structural error terms, and underlying structural
coefficients.
(8.) This point is not newusee Cooley and LeRoy (1985),
(9.) Consider, for example, the sum of coefficients on the nominal
variables--inflation, the monetary base, and the nominal interest rate.
In the early period the sum is 0.17, rising to 1.23 in the middle
period, and then falling to 0.80 in the final period.
(10.) A clear exposition of the SVAR approach is given by Sarte
(1999).
(11.) Other authors have argued that SVAR results are not robust,
including Cooley and Dwyer (1998) and Cecchetti and Rich (1999).
REFERENCES
Blanchard, Olivier Jean. "A Traditional Interpretation of
Macroeconomic Fluctuations," American Economic Review, vol. 79
(December 1989), pp. 1146-64.
Brayton, Flint, Eileen Mauskopf, David Reifschneider, Peter
Tinsley, and John Williams. "The Role of Expectations in the FRB/US
Macroeconomic Model," Federal Reserve Bulletin, vol. 83 (April
1997), pp. 227-45.
Cecchetti, Stephen G., and Robert W. Rich. "Structural
Estimates of the U.S. Sacrifice Ratio," Federal Reserve Bank of New
York, 1999.
Christ, Carl. "Discussion," in The Econometrics of Price
Determination. Washington: Board of Governors of the Federal Reserve
System, 1972, pp. 330-31.
Cooley, Thomas F., and Mark Dwyer. "Business Cycle Analysis
without Much Theory: A Look at Structural VARs," Journal of
Econometrics, vol. 83 (March-April 1998), pp. 57-88.
----, and Stephan F. LeRoy. "Atheoretical Macroeconomics: A
Critique," Journal of Monetary Economics, vol. 16 (November 1985),
pp. 283-308.
Fair, Ray. A Short-Run Forecasting Model of the United States
Economy. Lexington, Mass.: D.C. Heath, 1971.
Friedman, Milton. "The Role of Monetary Policy," American
Economic Review, vol. 58 (March 1968), pp. 1-17.
Klein, Lawrence. Economic Fluctuations in the United States
1921-1941. New York: Wiley, 1950.
Litterman, Robert. "Forecasting with Bayesian Vector
Autoregressions--Five Years of Experience," Journal of Business and
Economic Statistics, vol. 4 (January 1986), pp. 25-38.
Lucas, Robert. "Econometric Policy Evaluation: A
Critique," in Karl Brunner and Allan H. Meltzer, eds., The Phillips
Curve and Labor Markets, Carnegie-Rochester Conference Series on Public
Policy, vol. 1 (1976), pp. 19-46.
----. "Econometric Testing of the Natural Rate
Hypothesis," in The Econometrics of Price Determination.
Washington: Board of Governors of the Federal Reserve System, 1972, pp.
50-59.
----, and Thomas Sargent. "After Keynesian
Macroeconomics," Federal Reserve Bank of Minneapolis Quarterly
Review, vol. 3 (Spring 1979), pp. 1-16.
Lupoletti, William, and Roy Webb. "Defining and Improving the
Accuracy of Macroeconomic Forecasts: Contributions from a VAR
Model," Journal of Business, vol. 59 (April 1986), pp. 263-85.
McNees, Stephen. "Forecasting Accuracy of Alternative
Techniques: A Comparison of U.S. Macroeconomic Forecasts," Journal
of Business and Economic Statistics, vol. 4 (January 1986), pp. 5-16.
Sarte, Pierre-Daniel G. "An Empirical Investigation of
Fluctuations in Manufacturing Sales and Inventory within a Sticky-Price
Framework," Federal Reserve Bank of Richmond Economic Quarterly,
vol. 85 (Summer 1999), pp. 61-84.
Sims, Christopher A. "Policy Analysis with Econometric
Models," Brookings Papers on Economic Activity, 1:1982, pp.
107-165.
_____. "Macroeconomics and Reality," Econometrica, vol.
48 (January 1980), pp. 1-48.
Tinbergen, J. Business Cycles in the United States: 1912-1932.
League of Nations, 1939.
Tobin, James. "Comments," in J. Kmenta and J. B. Ramsey,
eds., Large-Scale Macro-Econometric Models. New York: North-Holland,
1981, pp. 391-92.
Webb, Roy. "Forecasts of Inflation from VAR Models,"
Journal of Forecasting, vol. 14 (May 1995), pp. 267-85.
Average Forecast Errors from Forecasts Made
in the Early 1980s
Variable:
Forecast Horizon
(Quarters) Chase DRI WEFA BVAR
Real GNP:
1 2.4 2.0 3.1 2.8
2 2.6 2.3 2.6 2.1
4 2.7 2.5 2.4 1.9
8 2.0 2.0 1.7 1.3
GNP deflator:
1 1.4 1.4 1.9 2.5
2 1.0 1.1 1.5 2.5
4 1.4 1.4 1.7 3.3
8 2.2 2.2 2.4 4.1
Nominal GNP:
1 3.2 2.7 3.7 3.6
2 3.2 2.7 3.6 3.3
4 3.6 3.2 3.8 4.0
8 3.6 3.6 2.4 3.5
Treasury bill rate:
1 0.2 0.1 0.4 0.1
2 1.8 1.9 1.8 1.7
4 3.3 3.2 3.2 2.9
8 2.9 3.7 1.1 3.7
Notes: Data are root mean squared errors (RMSEs) from postsample
forecasts. Forecasts are from 1980Q2 to l985Ql. Forecasts of real GNP,
the GNP deflator, and nominal GNP are percentage changes from the
previous quarter, and forecasts of the Treasury bill rate are cumulative
changes in the quarterly average level. Data are from McNees (1986).
Forecasts from WEFA were made in mid-quarter, and the others were made
one month later.
Average Forecast Errors from Simulated Forecasts
Variable:
Forecast Horizon
(Quarters) Chase DRI WEFA VAR
Real GNP:
1 4.1 4.0 4.2 5.3
2 3.1 3.1 2.9 4.1
4 2.5 2.5 2.2 2.8
6 2.3 2.3 1.9 2.4
GNP deflator:
1 1.8 2.0 1.9 2.3
2 1.9 2.0 1.9 2.0
4 2.2 2.1 2.0 2.1
6 2.5 2.4 2.2 2.4
Nominal GNP:
1 5.1 4.6 4.9 6.0
2 4.1 3.6 3.8 4.7
4 3.5 3.0 3.0 3.3
6 3.3 2.7 2.6 3.1
Treasury bill rate:
1 1.5 1.4 ... 1.3
2 2.2 2.1 ... 2.1
4 2.9 2.6 ... 2.8
6 3.5 3.2 ... 3.5
Notes: Data are root mean squared errors (RMSEs) from postsample
forecasts. Ranges for RMSEs are: one-quarter forecasts, 1970:4-1983:4;
two-quarter forecasts, 1971:1-1983:4; four-quarter forecasts,
1971:3-1983:4; and six-quarter forecasts, 1972:1-1983:4. The VAR
forecasts are simulated forecasts, as described in the text. Forecasts
of real GNP, the GNP deflator, and nominal GNP are cumulative percentage
changes, and forecasts of the Treasury bill rate are for its quarterly
average level.
Regression Results for Several Time Periods
1952Q2 to 1966Q4 [R.sub.2] = -0.08
[p.sub.t] = 0.28 - 0.08[p.sub.t-1] + 0.08[p.sub.t-2] +
0.11[p.sub.t-3] + 0.07[r.sub.t-1] + 0.02[c.sub.t-1] - 0.0l[m.sub.t-1] +
0.05[y.sub.t-1]
(0.06) (-0.54) (0.51) (0.72) (0.24) (0.28) (-0.01) (0.76)
1967Q1 to 1981Q2 [R.sub.2] = 0.57
[p.sub.t] = -2.78 + 0.30[p.sub.t-1] - 0.04[p.sub.t-2] +
0.04[p.sub.t-3] + 0.33[r.sub.t-1] + 0.02[c.sub.t-1] + 0.60[m.sub.t-1] -
0.08[y.sub.t-1]
(-0.54) (2.30) (-0.28) (0.27) (2.56) (0.26) (4.87) (-1.52)
1981Q3 to 1990Q4 [R.sub.2] = 0.51
[p.sub.t] = -8.87 + 0.21[p.sub.t-1] + 0.09[p.sub.t-2] +
0.20[p.sub.t-3] + 0.20[r.sub.t-1] + 0.10[c.sub.t-1] + 0.10[m.sub.t-1] -
0.15[y.sub.t-1]
(-1.54) (1.16) (0.53) (1.15) (1.07) (1.68) (1.02) (-0.21)
1952Q2 to 1990Q4 [R.sub.2] = 0.59
[p.sub.t] = -3.84 + 0.30[p.sub.t-1] + 0.23[p.sub.t-2] +
0.22[p.sub.t-3] + 0.005[r.sub.t-1] + 0.05[c.sub.t-1] + 0.17[m.sub.t-1] -
0.22[y.sub.t-1]
(-1.42) (6.38) (2.89) (2.71) (0.07) (1.54) (2.95) (-0.59)
Note: Coefficients were estimated by ordinary least squares;
t-statistics are in parentheses.