Breusch–Godfrey test

Statistical hypothesis test for the presence of serial correlation

In statistics, the Breusch–Godfrey test is used to assess the validity of some of the modelling assumptions inherent in applying regression-like models to observed data series.[1][2] In particular, it tests for the presence of serial correlation that has not been included in a proposed model structure and which, if present, would mean that incorrect conclusions would be drawn from other tests or that sub-optimal estimates of model parameters would be obtained.

The regression models to which the test can be applied include cases where lagged values of the dependent variables are used as independent variables in the model's representation for later observations. This type of structure is common in econometric models.

The test is named after Trevor S. Breusch and Leslie G. Godfrey.

Background

The Breusch–Godfrey test is a test for autocorrelation in the errors in a regression model. It makes use of the residuals from the model being considered in a regression analysis, and a test statistic is derived from these. The null hypothesis is that there is no serial correlation of any order up to p.[3]

Because the test is based on the idea of Lagrange multiplier testing, it is sometimes referred to as an LM test for serial correlation.[4]

A similar assessment can be also carried out with the Durbin–Watson test and the Ljung–Box test. However, the test is more general than that using the Durbin–Watson statistic (or Durbin's h statistic), which is only valid for nonstochastic regressors and for testing the possibility of a first-order autoregressive model (e.g. AR(1)) for the regression errors.[citation needed] The BG test has none of these restrictions, and is statistically more powerful than Durbin's h statistic.[citation needed] The BG test is considered to be more general than the Ljung-Box test because the latter requires the assumption of strict exogeneity, but the BG test does not. However, the BG test requires the assumptions of stronger forms of predeterminedness and conditional homoscedasticity.

Procedure

Consider a linear regression of any form, for example

Y t = β 1 + β 2 X t , 1 + β 3 X t , 2 + u t {\displaystyle Y_{t}=\beta _{1}+\beta _{2}X_{t,1}+\beta _{3}X_{t,2}+u_{t}\,}

where the errors might follow an AR(p) autoregressive scheme, as follows:

u t = ρ 1 u t 1 + ρ 2 u t 2 + + ρ p u t p + ε t . {\displaystyle u_{t}=\rho _{1}u_{t-1}+\rho _{2}u_{t-2}+\cdots +\rho _{p}u_{t-p}+\varepsilon _{t}.\,}

The simple regression model is first fitted by ordinary least squares to obtain a set of sample residuals u ^ t {\displaystyle {\hat {u}}_{t}} .

Breusch and Godfrey[citation needed] proved that, if the following auxiliary regression model is fitted

u ^ t = α 0 + α 1 X t , 1 + α 2 X t , 2 + ρ 1 u ^ t 1 + ρ 2 u ^ t 2 + + ρ p u ^ t p + ε t {\displaystyle {\hat {u}}_{t}=\alpha _{0}+\alpha _{1}X_{t,1}+\alpha _{2}X_{t,2}+\rho _{1}{\hat {u}}_{t-1}+\rho _{2}{\hat {u}}_{t-2}+\cdots +\rho _{p}{\hat {u}}_{t-p}+\varepsilon _{t}\,}

and if the usual Coefficient of determination ( R 2 {\displaystyle R^{2}} statistic) is calculated for this model:

R 2 := j = 1 T p ( u T j u ^ T j ) 2 j = 1 T p ( u T j u ¯ ) 2 {\displaystyle R^{2}:={\frac {\sum _{j=1}^{T-p}(u_{T-j}-{\hat {u}}_{T-j})^{2}}{\sum _{j=1}^{T-p}(u_{T-j}-{\bar {u}})^{2}}}} ,

where u ¯ {\displaystyle {\bar {u}}} stands for the arithmetic mean over the last n = T p {\displaystyle n=T-p} samples, where T {\displaystyle T} is the total number of observations and p {\displaystyle p} is the number of error lags used in the auxiliary regression.

The following asymptotic approximation can be used for the distribution of the test statistic:

n R 2 χ p 2 , {\displaystyle nR^{2}\,\sim \,\chi _{p}^{2},\,}

when the null hypothesis H 0 : { ρ i = 0  for all  i } {\displaystyle {H_{0}:\lbrace \rho _{i}=0{\text{ for all }}i\rbrace }} holds (that is, there is no serial correlation of any order up to p). Here n is

Software

  • In R, this test is performed by function bgtest, available in package lmtest.[5][6]
  • In Stata, this test is performed by the command estat bgodfrey.[7][8]
  • In SAS, the GODFREY option of the MODEL statement in PROC AUTOREG provides a version of this test.
  • In Python Statsmodels, the acorr_breusch_godfrey function in the module statsmodels.stats.diagnostic [9]
  • In EViews, this test is already done after a regression, at "View" → "Residual Diagnostics" → "Serial Correlation LM Test".
  • In Julia, the BreuschGodfreyTest function is available in the HypothesisTests package.[10]
  • In gretl, this test can be obtained via the modtest command, or under the "Test" → "Autocorrelation" menu entry in the GUI client.


See also

References

  1. ^ Breusch, T. S. (1978). "Testing for Autocorrelation in Dynamic Linear Models". Australian Economic Papers. 17: 334–355. doi:10.1111/j.1467-8454.1978.tb00635.x.
  2. ^ Godfrey, L. G. (1978). "Testing Against General Autoregressive and Moving Average Error Models when the Regressors Include Lagged Dependent Variables". Econometrica. 46: 1293–1301. JSTOR 1913829.
  3. ^ Macrodados 6.3 Help – Econometric Tools[permanent dead link]
  4. ^ Asteriou, Dimitrios; Hall, Stephen G. (2011). "The Breusch–Godfrey LM test for serial correlation". Applied Econometrics (Second ed.). New York: Palgrave Macmillan. pp. 159–61. ISBN 978-0-230-27182-1.
  5. ^ "lmtest: Testing Linear Regression Models". CRAN.
  6. ^ Kleiber, Christian; Zeileis, Achim (2008). "Testing for autocorrelation". Applied Econometrics with R. New York: Springer. pp. 104–106. ISBN 978-0-387-77318-6.
  7. ^ "Postestimation tools for regress with time series" (PDF). Stata Manual.
  8. ^ Baum, Christopher F. (2006). "Testing for serial correlation". An Introduction to Modern Econometrics Using Stata. Stata Press. pp. 155–158. ISBN 1-59718-013-0.
  9. ^ Breusch-Godfrey test in Python http://statsmodels.sourceforge.net/devel/generated/statsmodels.stats.diagnostic.acorr_breush_godfrey.html?highlight=autocorrelation Archived 2014-02-28 at the Wayback Machine
  10. ^ "Time series tests". juliastats.org. Retrieved 2020-02-04.

Further reading

  • Godfrey, L. G. (1988). Misspecification Tests in Econometrics. Cambridge, UK: Cambridge. ISBN 0-521-26616-5.
  • Godfrey, L. G. (1996). "Misspecification Tests and Their Uses in Econometrics". Journal of Statistical Planning and Inference. 49 (2): 241–260. doi:10.1016/0378-3758(95)00039-9.
  • Maddala, G. S.; Lahiri, Kajal (2009). Introduction to Econometrics (Fourth ed.). Chichester: Wiley. pp. 259–260.
  • v
  • t
  • e
Continuous data
Center
Dispersion
Shape
Count data
Summary tables
Dependence
Graphics
Study design
Survey methodology
Controlled experiments
Adaptive designs
Observational studies
Statistical theory
Frequentist inference
Point estimation
Interval estimation
Testing hypotheses
Parametric tests
Specific tests
  • Z-test (normal)
  • Student's t-test
  • F-test
Goodness of fit
Rank statistics
Bayesian inference
Correlation
Regression analysis
Linear regression
Non-standard predictors
Generalized linear model
Partition of variance
Categorical
Multivariate
Time-series
General
Specific tests
Time domain
Frequency domain
Survival
Survival function
Hazard function
Test
Biostatistics
Engineering statistics
Social statistics
Spatial statistics
  • Category
  • icon Mathematics portal
  • Commons
  • WikiProject