Another way of describing the degrees of freedom, as per William Gosset, has been stated as “The general idea: given that n pieces of data x1, x2, … xn, you use one ‘degree of freedom’ when you compute μ, leaving n-1 independent pieces of information.”2
As summarized by one of my former professors, he stated that the degrees of freedom reflect the total number of observations minus the number of population parameters that are being estimated by a sample statistic. Since we assume populations are infinite and cannot be feasibly used to generate parameters, we rely on samples to generate statistical inferences back to the original population; that is given that the sampling techniques are both random and representative.
This may seem very elementary, but from my own experiences, degrees of freedom have not been the easiest of concept to comprehend – especially for the novice. A definition that can also be representative of the concept of the degrees of freedom can be summarized as “equal to the number of independent pieces of information concerning the variance.3 For a random sample from a population, the number of degrees of freedom is equal to the sample size minus one.
A numerical example of this approach might help with the above definition. The values would reflect the actual observations of the data set. This example is simply for illustration purposes. Given that we have eight (8) data points that sum up to 60, we can randomly and independently assign values to seven (7) of them. For instance, we may record them as: 8, 9, 7, 6, 7, 10, and 7. The seven values would have the freedom to be any numbers, yet the eighth number would have to be some fixed value to get a total of 60 (in this case the value would have to be 6. Hence, the degrees of freedom are (n-1) or (8–1) = 7. There are 7 numbers that can take on any values, but only one number will make the equation (the sum of the values, in this case) hold true.
Dr. W. Edwards Deming had a slightly different take on the degrees of freedom. He ascertained that: “Degrees of freedom is a term denoting the proper devisor (example, n-1) required under a 2d* moment of a sample drawn without replacement to give an unbiased estimate of a varience.4 He further went on to define “The number of degrees of freedom, as has been explained, is the devisor required under the 2d moment (ns2) to get an unbiased estimate of σ2. Thus ns2/(n-1) is an unbiased estimate of σ2, and n-1 is the number of degrees of freedom in the estimate.”5
What really inspired me to write this article about the impact of the degrees of freedom was a conversation I had with my wife, Nancy. She has a Ph. D. in Physiology and is a former professor at Ohio Northern University (retired). One of the faculties she provided Biology training for was the School of Pharmacy. She was heading to her class, when she called me on the telephone and asked if I had an ‘easy’ way of explaining the degrees of freedom. I gave her the description I use in my classes:
This explanation seems to have satisfied many a student of mine and places the emphasis on a common statistical concept that, without getting into confidence intervals; the larger the sample size, the more confident we can be of our estimates. To summarize this idea in a slightly different way - if our sampling technique is random and representative, the likelihood that we have a good estimator of a parameter is greater with larger sample sizes.
I have attempted to address the various approaches to the degrees of freedom and hopefully my simplistic approach to the rationale behind what we are trying to accomplish can shed some light on future explanations of such a vital part of statistical analysis.
*Note: 2d refers to the 2nd. moment about the mean, another way of describing the variance.
1). Gonick, L. and Smith, W. (1993), The Cartoon Guide to Statistics, Harper Collins Publishers, pg. 22
2). Breyfogle, Forrest W. III (1946), Implementing Six Sigma, John Wiley & Sons, pg. 1105
3). Upton, Graham and Cook, Ian (2002), Dictionary of Statistics, Oxford University Press, pg. 100
4). Deming, William Edwards (1950), Some Theory of Sampling, Dover Publications, Inc., pg. 352
5). Ibid., pg. 541