In statistics, the reduced chi-square statistic is used extensively in goodness of fit testing. It is also known as mean squared weighted deviation (MSWD) in isotopic dating and variance of unit weight in the context of weighted least squares.
Its square root is called regression standard error, standard error of the regression, or standard error of the equation
(see )
It is defined as chi-square per degree of freedom:
where the chi-squared is a weighted sum of squared deviations:
with inputs: variance , observations O, and calculated data C.
The degree of freedom, , equals the number of observations n minus the number of fitted parameters m.
In weighted least squares, the definition is often written in matrix notation as
where r is the vector of residuals, and W is the weight matrix, the inverse of the input (diagonal) covariance matrix of observations. If W is non-diagonal, then generalized least squares applies.
In ordinary least squares, the definition simplifies to:
where the numerator is the residual sum of squares (RSS).
When the fit is just an ordinary mean, then equals the sample standard deviation.
As a general rule, when the variance of the measurement error is known a priori, a indicates a poor model fit. A indicates that the fit has not fully captured the data (or that the error variance has been underestimated). In principle, a value of around indicates that the extent of the match between observations and estimates is in accord with the error variance. A indicates that the model is "over-fitting" the data: either the model is improperly fitting noise, or the error variance has been overestimated.
When the variance of the measurement error is only partially known, the reduced chi-squared may serve as a correction estimated a posteriori.
In geochronology, the MSWD is a measure of goodness of fit that takes into account the relative importance of both the internal and external reproducibility, with most common usage in isotopic dating.
This page is automatically generated and may contain information that is not correct, complete, up-to-date, or relevant to your search query. The same applies to every other page on this website. Please make sure to verify the information with EPFL's official sources.
In the academic or industrial world, to optimize a system, it is necessary to establish strategies for the experimental approach. The DOE allows you to choose the best set of measurement points to min
This course addresses the relationship between specific technological features and the learners' cognitive processes. It also covers the methods and results of empirical studies on this topic: do stud
The p-Laplacian problem -del & sdot; ((mu + |del u|(p-2))del u) = f is considered, where mu is a given positive number. An anisotropic a posteriori residual-based error estimator is presented. The error estimator is shown to be equivalent, up to higher ord ...
Walter De Gruyter Gmbh2024
, ,
Dynamic downscaling of atmospheric forcing data to the hectometer resolution has shown increases in accuracy for landsurface models, but at great computational cost. Here we present a validation of a novel intermediate complexity atmospheric model, HICAR, ...
Frontiers Media Sa2024
,
This paper presents a comparison of machine learning (ML) methods used for three-dimensional localization of partial discharges (PD) in a power transformer tank. The study examines ML and deep learning (DL) methods, ranging from support vector machines (SV ...
In statistics, the number of degrees of freedom is the number of values in the final calculation of a statistic that are free to vary. Estimates of statistical parameters can be based upon different amounts of information or data. The number of independent pieces of information that go into the estimate of a parameter is called the degrees of freedom. In general, the degrees of freedom of an estimate of a parameter are equal to the number of independent scores that go into the estimate minus the number of parameters used as intermediate steps in the estimation of the parameter itself.
In statistics and optimization, errors and residuals are two closely related and easily confused measures of the deviation of an observed value of an element of a statistical sample from its "true value" (not necessarily observable). The error of an observation is the deviation of the observed value from the true value of a quantity of interest (for example, a population mean). The residual is the difference between the observed value and the estimated value of the quantity of interest (for example, a sample mean).
Weighted least squares (WLS), also known as weighted linear regression, is a generalization of ordinary least squares and linear regression in which knowledge of the unequal variance of observations (heteroscedasticity) is incorporated into the regression. WLS is also a specialization of generalized least squares, when all the off-diagonal entries of the covariance matrix of the errors, are null.