Econometric Theory/Serial Correlation

From testwiki
Revision as of 05:38, 14 February 2008 by 205.175.121.71 (talk) (Causes of Autocorrelation)
(diff) ← Older revision | Latest revision (diff) | Newer revision → (diff)
Jump to navigation Jump to search

There are times, especially in time-series data, that the CLR assumption of corr(ϵt,ϵt1)=0 is broken. This is known in econometrics as Serial Correlation or Autocorrelation. This means that corr(ϵt,ϵt1)0 and there is a pattern across the error terms. The error terms are then not independently distributed across the observations and are not strictly random.


Examples of Autocorrelation

corr(ϵt,ϵt1)>0

Positive Autocorrelation





















corr(ϵt,ϵt1)<0

Negative Autocorrelation






















Functional Form

When the error term is related to the previous error term, it can be written in an algebraic equation. ϵt=ρϵt1+ut where ρ is the autocorrelation coefficient between the two disturbance terms, and u is the disturbance term for the autocorrelation. This is known as an Autoregressive Process. 1</rho=corr(ϵt,ϵt1)<1The u is needed within the equation because although the error term is less random, it still has a slight random effect.

Serial Correlation of the Nth Order

The Autoregressive model:

  • First order Autoregressive Process:ϵt=ρϵt1+ut
    • This is known as the first order autoregression, due to the error term only depending on the previous error term.
    • It is commonly displayed in textbooks as AR(1)
  • nth order Autoregressive Process:ϵt=ρ1ϵt1+ρ2ϵt2++ρnϵtn+ut
    • Known as AR(n)


Causes of Autocorrelation

  1. Spacial Autocorrelation

corr(ϵt,ϵt1)0 Spacial Autocorrelation occurs when the two errors are specially and/or geographically related. In simpler terms, they are "next to each." Examples: The city of St. Paul has a spike of crime and so they hire additional police. The following year, they found that the crime rate decreased significantly. Amazingly, the city of Minneapolis, had not adjusted the police force, finds that they have a increase in the crime rate over the same period.

  • Note: this type of Autocorrelation occurs over cross-sectional samples.
  1. Inertia/Time to Adjust
    1. This often occurs in Macro, time series data. The US interest rate unexpectedly increases and so there is an associated change in exchange rates with other countries. Reaching a new equilibrium could take some time.
  2. Prolonged Influences
    1. This is again a Macro, time series issue dealing with economic shocks. It is now expected that the US interest rate will increase. ##The associated exchange rates will slowly adjust up-until the announcement by the Federal Reserve and may overshoot the equilibrium.
  3. Data Smoothing/Manipulation
    1. Using functions to smooth data will bring autocorrelation into the disturbance terms
  4. Misspecification
    1. A regression will often show signs of autocorrelation when there are omitted variables. Because the missing independent variable now exists in the disturbance term, we get a disturbance term that looks like: ϵt=β2X2+ut when the correct specification is Yt=β0+β1X1+β2X2+ut

Consequences of Autocorrelation

The main problem with Autocorrelation is that it may make a model look better than it actually is.

list of consequences

  1. Coefficients are still unbiased E(ϵt)=0,cov(Xt,ut)=0
  2. True variance of β^ is increased by the presence of Autocorrelation.
  3. Estimated Variance of β^ is smaller due to Autocorrelation (biased downward).
  4. A decrease with SE(β^) and an increase of the t-stats. This results in the estimator looking more accurate than it actually is.
  5. R² becomes inflated.

All of these problems result in hypothesis tests becoming invalid.

Autocorrelation in data. 2 runs, but the real OLS, which we would have never found, is somewhere in the middle.


Testing for AC

  1. . View a graph of the Dependant variable against the error term (AKA, a residual scatter-plot).
  2. . Durbin-Watson test.
    1. . Assume ϵt=ϵt1ρ+ut
    2. . test H(0): ρ = 0 (no AC) against H(1): ρ > 0 (one-tailed test)
    3. . Test Stat DW=(ϵtϵt1)2ϵ2=22ρ
  • Any value under D(L) (in the D-W table) rejects the null hypothesis and AC exists.
  • Any value between D(L) and D(W) leaves us with no conclusion of AC.
  • Any value larger than D(W) accepts the null hypothesis and AC does not exist.

The Durbin-Watson distribution

  • Note, this is one tail test, To get the other tail. Use 4 - DW as the test stat instead.