Degrees of freedom

来源:互联网 发布:linux退出文本编辑 编辑:程序博客网 时间:2024/04/30 11:08

Degrees of freedom in linear models[edit]
The demonstration of the t and chi-squared distributions for one-sample problems above is the simplest example where degrees-of-freedom arise. However, similar geometry and vector decompositions underlie much of the theory of linear models, including linear regression and analysis of variance. An explicit example based on comparison of three means is presented here; the geometry of linear models is discussed in more complete detail by Christensen (2002).[4]

Suppose independent observations are made for three populations,X1,,Xn, Y1,,Ynand Z1,,Zn. The restriction to three groups and equal sample sizes simplifies notation, but the ideas are easily generalized.

The observations can be decomposed as

XiYiZi=M¯+(X¯M¯)+(XiX¯)=M¯+(Y¯M¯)+(YiY¯)=M¯+(Z¯M¯)+(ZiZ¯)

where X¯,Y¯,Z¯are the means of the individual samples, and M¯=(X¯+Y¯+Z¯)/3 is the mean of all 3n observations. In vector notation this decomposition can be written as

X1XnY1YnZ1Zn

= M¯
111111

+
X¯M¯X¯M¯Y¯M¯Y¯M¯Z¯M¯Z¯M¯

+
X1X¯XnX¯Y1Y¯YnY¯Z1Z¯ZnZ¯
.
The observation vector, on the left-hand side, has 3n degrees of freedom. On the right-hand side, the first vector has one degree of freedom (or dimension) for the overall mean. The second vector depends on three random variables, X¯M¯,Y¯M¯and ZM. However, these must sum to 0 and so are constrained; the vector therefore must lie in a 2-dimensional subspace, and has 2 degrees of freedom. The remaining 3n − 3 degrees of freedom are in the residual vector (made up of n − 1 degrees of freedom within each of the populations).

Sum of squares and degrees of freedom

In statistical testing problems, one usually isn’t interested in the component vectors themselves, but rather in their squared lengths, or Sum of Squares. The degrees of freedom associated with a sum-of-squares is the degrees-of-freedom of the corresponding component vectors.

The three-population example above is an example of one-way Analysis of Variance. The model, or treatment, sum-of-squares is the squared length of the second vector,

SSTr=n(X¯M¯)2+n(Y¯M¯)2+n(Z¯M¯)2

with 2 degrees of freedom. The residual, or error, sum-of-squares is
SSE=ni=1(XiX¯)2+ni=1(YiY¯)2+ni=1(ZiZ¯)2
with 3(n−1) degrees of freedom. Of course, introductory books on ANOVA usually state formulae without showing the vectors, but it is this underlying geometry that gives rise to SS formulae, and shows how to unambiguously determine the degrees of freedom in any given situation.

Under the null hypothesis of no difference between population means (and assuming that standard ANOVA regularity assumptions are satisfied) the sums of squares have scaled chi-squared distributions, with the corresponding degrees of freedom. The F-test statistic is the ratio, after scaling by the degrees of freedom. If there is no difference between population means this ratio follows an F distribution with 2 and 3n − 3 degrees of freedom.

In some complicated settings, such as unbalanced split-plot designs, the sums-of-squares no longer have scaled chi-squared distributions. Comparison of sum-of-squares with degrees-of-freedom is no longer meaningful, and software may report certain fractional ‘degrees of freedom’ in these cases. Such numbers have no genuine degrees-of-freedom interpretation, but are simply providing an approximate chi-squared distribution for the corresponding sum-of-squares. The details of such approximations are beyond the scope of this page.

0 0
原创粉丝点击