Econometric Theory/Serial Correlation
There are times, especially in time-series data, that the CLR assumption of is broken. This is known in econometrics as Serial Correlation or Autocorrelation. This means that and there is a pattern across the error terms. The error terms are then not independently distributed across the observations and are not strictly random.
Examples of Autocorrelation
Functional Form
When the error term is related to the previous error term, it can be written in an algebraic equation. where ρ is the autocorrelation coefficient between the two disturbance terms, and u is the disturbance term for the autocorrelation. This is known as an Autoregressive Process. The u is needed within the equation because although the error term is less random, it still has a slight random effect.
Serial Correlation of the Nth Order
The Autoregressive model:
- First order Autoregressive Process:
- This is known as the first order autoregression, due to the error term only depending on the previous error term.
- It is commonly displayed in textbooks as AR(1)
- nth order Autoregressive Process:
- Known as AR(n)
Causes of Autocorrelation
- Spacial Autocorrelation
Spacial Autocorrelation occurs when the two errors are specially and/or geographically related. In simpler terms, they are "next to each." Examples: The city of St. Paul has a spike of crime and so they hire additional police. The following year, they found that the crime rate decreased significantly. Amazingly, the city of Minneapolis, had not adjusted the police force, finds that they have a increase in the crime rate over the same period.
- Note: this type of Autocorrelation occurs over cross-sectional samples.
- Inertia/Time to Adjust
- This often occurs in Macro, time series data. The US interest rate unexpectedly increases and so there is an associated change in exchange rates with other countries. Reaching a new equilibrium could take some time.
- Prolonged Influences
- This is again a Macro, time series issue dealing with economic shocks. It is now expected that the US interest rate will increase. ##The associated exchange rates will slowly adjust up-until the announcement by the Federal Reserve and may overshoot the equilibrium.
- Data Smoothing/Manipulation
- Using functions to smooth data will bring autocorrelation into the disturbance terms
- Misspecification
- A regression will often show signs of autocorrelation when there are omitted variables. Because the missing independent variable now exists in the disturbance term, we get a disturbance term that looks like: when the correct specification is
Consequences of Autocorrelation
The main problem with Autocorrelation is that it may make a model look better than it actually is.
list of consequences
- Coefficients are still unbiased
- True variance of is increased by the presence of Autocorrelation.
- Estimated Variance of is smaller due to Autocorrelation (biased downward).
- A decrease with and an increase of the t-stats. This results in the estimator looking more accurate than it actually is.
- R² becomes inflated.
All of these problems result in hypothesis tests becoming invalid.
Testing for AC
- . View a graph of the Dependant variable against the error term (AKA, a residual scatter-plot).
- . Durbin-Watson test.
- . Assume
- . test H(0): ρ = 0 (no AC) against H(1): ρ > 0 (one-tailed test)
- . Test Stat
- Any value under D(L) (in the D-W table) rejects the null hypothesis and AC exists.
- Any value between D(L) and D(W) leaves us with no conclusion of AC.
- Any value larger than D(W) accepts the null hypothesis and AC does not exist.
- Note, this is one tail test, To get the other tail. Use 4 - DW as the test stat instead.