In the statistics, the Deming regression , named W.C. Deming , is a kind of that tries to find the best smoothing line for a two-dimensional data set. Regression differs from in that it takes into account the in the observation in both the x axis and the y axis. Regression is a special case of the method of least complete squares , which considers any number of indicators and has a more complex structure of errors.
The Deming regression is equivalent to the maximum likelihood estimate for , in which the errors of the two variables are considered independent and have a normal distribution , and the ratio of their variances, δ , is known [1] . In practice, this ratio can be estimated from the source data. However, the regression procedure does not take into account possible errors in estimating the variance relations.
Deming's regression is only slightly more complicated than . Most of the statistical packages used in clinical chemistry provide a Deming regression.
The model was originally proposed by Adcock [2] , who considered the case δ = 1, and then was considered in a more general form by Kummellem [3] with arbitrary δ . However, their ideas remained largely unnoticed for more than 50 years, until Kupmans [4] revived them and later distributed Deming [5] . The book of the latter has become so popular in clinical chemistry and related fields that the method in these areas was called the Deming regression [6] .
Content
Specification
Suppose that the data ( y i , x i ) are values obtained during measurements of the "true" values ( y i * , x i * ) that lie on the regression line:
where the errors ε and η are independent and the ratio of their variances is known:
In practice, the variance of parameters and often unknown, which complicates the assessment . Note that when the measurement method and the same, these dispersions are most likely equal, so in this case .
We are trying to find the direct "best smoothing"
such that the weighted sum of squares of residuals is minimal [7]
Solution
The solution can be expressed in terms of second order moments. That is, we first calculate the following values (all sums are taken by i = 1: n ):
Finally, the least squares estimation parameters will be [8] :
Orthogonal Regression
In case of equality of error variances, i.e. when , the Deming regression becomes an orthogonal regression — it minimizes the sum of squares of . In this case, we denote each sample point z j on the complex plane (ie, the point ( x j , y j ) of the sample is written as z j = x j + iy j , where i is the imaginary unit ). Let Z be the sum of squared differences from the sample points to the center of gravity (also represented in the complex coordinates). The center of gravity is the average of the sampling points. Then [9] :
- If Z = 0, then any line passing through the center of gravity is a line of the best orthogonal smoothing.
- If Z ≠ 0, the line of the best orthogonal smoothing passes through the center of gravity and is parallel to the vector from the origin to {\ displaystyle {\ sqrt {Z}}} .
Coolidge gave a trigonometric interpretation of the best orthogonal smoothing line in 1913 [10] .
Applications
In the case of three non - collinear points on the plane, the triangle formed by these points has a single Steiner ellipse inscribed that touches the sides of the triangle at mid-points. The main axis of this ellipse will be the orthogonal regression of these three vertices [11] .
Notes
- ↑ Linnet, 1993 .
- ↑ Adcock, 1878 .
- ↑ Kummell, 1879 .
- ↑ Koopmans, 1937 .
- ↑ Deming, 1943 .
- ↑ Cornbleet, Gochman, 1979 , p. 432–438.
- ↑ Fuller, 1987 , p. ch.1.3.3.
- ↑ Glaister, 2001 , p. 104-107.
- ↑ Minda, Phelps, 2008 , p. 679–689, Theorem 2.3.
- ↑ Coolidge, 1913 , p. 187–190.
- ↑ Minda, Phelps, 2008 , p. 679–689, Corollary 2.4.
Literature
- RJ Adcock. A problem in least squares // The Analyst. - Annals of Mathematics, 1878. - Vol. 5 , no. 2 - pp . 53–54 . - DOI : 10.2307 / 2635758 .
- JL Coolidge. The American Mathematical Monthly . - 1913. - V. 20 , no. 6 - p . 187–190 . - DOI : 10.2307 / 2973072 .
- PJ Cornbleet, N. Gochman. Incorrect Least – Squares Regression Coefficients // Clin. Chem .. - 1979. - V. 25 , no. 3 - p . 432–438 . - PMID 262186 .
- WE Deming. Statistical adjustment of data. - Wiley, NY (Dover Publications edition, 1985), 1943. - ISBN 0-486-64685-8 .
- Wayne A. Fuller. Measurement error models. - John Wiley & Sons, Inc., 1987. - ISBN 0-471-86187-1 .
- P. Glaister. Least squares revisited // The Mathematical Gazette . - 2001. - Vol. 85 March . - p . 104-107 .
- TC Koopmans. Linear regression analysis of economic time series. - DeErven F. Bohn, Haarlem, Netherlands, 1937.
- CH Kummell. Reduction of Observation Determinations. - Annals of Mathematics, 1879. - Vol. 6 , no. 4 - pp . 97–105 . - DOI : 10.2307 / 2635646 .
- K. Linnet. Evaluation of regression procedures for method comparison studies // Clinical Chemistry. - 1993. - V. 39 , no. 3 - p . 424–432 . - PMID 8448852 .
- D. Minda, S. Phelps. Triangles, ellipses, and cubic polynomials // American Mathematical Monthly . - 2008. - Vol. 115 , no. 8 - p . 679–689 .