Partition of sums of squares
This article is about the partition of sums of squares in statistics. For other uses, see Sum of squares
This article needs attention from an expert in Statistics
. Please add a reason
or a talk
parameter to this template to explain the issue with the article. WikiProject Statistics
may be able to help recruit an expert. (November 2008)
The partition of sums of squares
is a concept that permeates much of inferential statistics
and descriptive statistics
. More properly, it is the partitioning of sums of squared deviations or errors
. Mathematically, the sum of squared deviations is an unscaled, or unadjusted measure of dispersion
(also called variability
). When scaled for the number of degrees of freedom
, it estimates the variance
, or spread of the observations about their mean value. Partitioning of the sum of squared deviations into various components allows the overall variability in a dataset to be ascribed to different types or sources of variability, with the relative importance of each being quantified by the size of each component of the overall sum of squares.
The distance from any point in a collection of data, to the mean of the data, is the deviation. This can be written as
is the ith data point, and is the estimate of the mean. If all such deviations are squared, then summed, as in
, this gives the "sum of squares" for these data.
When more data are added to the collection the sum of squares will increase, except in unlikely cases such as the new data being equal to the mean. So usually, the sum of squares will grow with the size of the data collection. That is a manifestation of the fact that it is unscaled.
In many cases, the number of degrees of freedom
is simply the number of data in the collection, minus one. We write this as n
− 1, where n
is the number of data.
Scaling (also known as normalizing) means adjusting the sum of squares so that it does not grow as the size of the data collection grows. This is important when we want to compare samples of different sizes, such as a sample of 100 people compared to a sample of 20 people. If the sum of squares was not normalized, its value would always be larger for the sample of 100 people than for the sample of 20 people. To scale the sum of squares, we divide it by the degrees of freedom, i.e., calculate the sum of squares per degree of freedom, or variance. Standard deviation
, in turn, is the square root of the variance.
Partitioning the sum of squares in linear regression
where this equation is equivalent to each of the following forms:
is the value estimated by the regression line having
as the estimated coefficients
The requirement that the model includes a constant or equivalently that the design matrix contains a column of ones ensures that
The proof can also be expressed in vector form, as follows:
The elimination of terms in the last line, used the fact that
Note that the residual sum of squares can be further partitioned as the lack-of-fit sum of squares
plus the sum of squares due to pure error.
Last edited on 3 May 2021, at 06:03
Content is available under CC BY-SA 3.0
unless otherwise noted.