For 
 samples of a variate having a distribution with known Mean 
, the ``population
variance'' (usually called ``variance'' for short, although the word ``population'' should be added when needed to
distinguish it from the Sample Variance) is defined by
where 
  | 
(2) | 
 
But since 
 is an Unbiased Estimator for the Mean
  | 
(3) | 
 
it follows that the variance
  | 
(4) | 
 
The population Standard Deviation is then defined as
  | 
(5) | 
 
A useful identity involving the variance is
  | 
(6) | 
 
Therefore,
If the population Mean is not known, using the sample mean 
 instead of the population mean 
 to
compute
  | 
(9) | 
 
gives a Biased Estimator of the population variance.  In such cases, it is appropriate to use a 
Student's t-Distribution instead of a Gaussian Distribution.  However, it turns out (as discussed
below) that an Unbiased Estimator for the population variance is given by
  | 
(10) | 
 
The Mean and Variance of the sample standard deviation for a distribution with population mean 
 and
Variance are
The quantity 
 has a Chi-Squared Distribution.
For multiple variables, the variance is given using the definition of Covariance,
A linear sum has a similar form:
These equations can be expressed using the Covariance Matrix.
To estimate the population Variance from a sample of 
 elements with a priori unknown Mean (i.e.,
the Mean is estimated from the sample itself), we need an Unbiased Estimator for 
.  This is 
given by the k-Statistic 
, where
  | 
(15) | 
 
and 
 is the Sample Variance
  | 
(16) | 
 
Note that some authors prefer the definition
  | 
(17) | 
 
since this makes the sample variance an Unbiased Estimator for the population variance.
When computing numerically, the Mean must be computed before 
 can be determined. This requires storing the set of
sample values.  It is possible to calculate 
 using a recursion relationship involving only the last sample as
follows.  Here, use 
 to denote 
 calculated from the first 
 samples (not the 
th Moment)
  | 
(18) | 
 
and 
 denotes the value for the sample variance 
 calculated from the first 
 samples.  The first few
values calculated for the Mean are
Therefore, for 
, 3 it is true that
  | 
(22) | 
 
Therefore, by induction,
and
  | 
(26) | 
 
for 
, so
Working on the first term,
Use (24) to write
  | 
(29) | 
 
so
  | 
(30) | 
 
Now work on the second term in (27),
  | 
(31) | 
 
Considering the third term in (27),
But 
  | 
(33) | 
 
so
Plugging (30), (31), and (34) into (27),
so
  | 
(36) | 
 
To find the variance of 
 itself, remember that
  | 
(37) | 
 
and
  | 
(38) | 
 
Now find 
.
Working on the first term of (39),
The second term of (39) is known from k-Statistic,
  | 
(41) | 
 
as is the third term,
Combining (39)-(42) gives
so plugging in (38) and (43) gives
Student calculated the Skewness and Kurtosis of the distribution of 
 as
and conjectured that the true distribution is Pearson Type III Distribution
  | 
(47) | 
 
where
This was proven by R. A. Fisher.  
The distribution of 
 itself is given by
  | 
(50) | 
 
  | 
(51) | 
 
where 
  | 
(52) | 
 
The Moments are given by
  | 
(53) | 
 
and the variance is
An Unbiased Estimator of 
 is 
.  Romanovsky showed that
  | 
(55) | 
 
See also Correlation (Statistical), Covariance, Covariance Matrix, k-Statistic, Mean, 
Sample Variance
References
Press, W. H.; Flannery, B. P.; Teukolsky, S. A.; and Vetterling, W. T.  ``Moments of a Distribution: Mean,
  Variance, Skewness, and So Forth.''  §14.1 in
  Numerical Recipes in FORTRAN: The Art of Scientific Computing, 2nd ed.  Cambridge, England:
  Cambridge University Press, pp. 604-609, 1992.
© 1996-9 Eric W. Weisstein 
1999-05-26