The Chow test (), proposed by econometrician Gregory Chow in 1960, is a test of whether the true coefficients in two linear regressions on different data sets are equal. In econometrics, it is most commonly used in time series analysis to test for the presence of a structural break at a period which can be assumed to be known a priori (for instance, a major historical event such as a war). In program evaluation, the Chow test is often used to determine whether the independent variables have different impacts on different subgroups of the population. Suppose that we model our data as If we split our data into two groups, then we have and The null hypothesis of the Chow test asserts that , , and , and there is the assumption that the model errors are independent and identically distributed from a normal distribution with unknown variance. Let be the sum of squared residuals from the combined data, be the sum of squared residuals from the first group, and be the sum of squared residuals from the second group. and are the number of observations in each group and is the total number of parameters (in this case 3, i.e. 2 independent variables coefficients + intercept). Then the Chow test statistic is The test statistic follows the F-distribution with and degrees of freedom. The same result can be achieved via dummy variables. Consider the two data sets which are being compared. Firstly there is the 'primary' data set i={1,...,} and the 'secondary' data set i={+1,...,n}. Then there is the union of these two sets: i={1,...,n}. If there is no structural change between the primary and secondary data sets a regression can be run over the union without the issue of biased estimators arising. Consider the regression: Which is run over i={1,...,n}. D is a dummy variable taking a value of 1 for i={+1,...,n} and 0 otherwise. If both data sets can be explained fully by then there is no use in the dummy variable as the data set is explained fully by the restricted equation.

About this result
This page is automatically generated and may contain information that is not correct, complete, up-to-date, or relevant to your search query. The same applies to every other page on this website. Please make sure to verify the information with EPFL's official sources.
Related concepts (1)
Linear regression
In statistics, linear regression is a linear approach for modelling the relationship between a scalar response and one or more explanatory variables (also known as dependent and independent variables). The case of one explanatory variable is called simple linear regression; for more than one, the process is called multiple linear regression. This term is distinct from multivariate linear regression, where multiple correlated dependent variables are predicted, rather than a single scalar variable.

Graph Chatbot

Chat with Graph Search

Ask any question about EPFL courses, lectures, exercises, research, news, etc. or try the example questions below.

DISCLAIMER: The Graph Chatbot is not programmed to provide explicit or categorical answers to your questions. Rather, it transforms your questions into API requests that are distributed across the various IT services officially administered by EPFL. Its purpose is solely to collect and recommend relevant references to content that you can explore to help you answer your questions.