| 释义 | VarianceFor  samples of a variate having a distribution with known Mean  , the ``populationvariance'' (usually called ``variance'' for short, although the word ``population'' should be added when needed todistinguish it from the Sample Variance) is defined by 
 where
 
 But since|  | (2) | 
 is an Unbiased Estimator for the Mean 
 it follows that the variance|  | (3) | 
 The population Standard Deviation is then defined as|  | (4) | 
 A useful identity involving the variance is|  | (5) | 
 Therefore,|  | (6) | 
 
 
 If the population Mean is not known, using the sample mean  instead of the population mean  tocompute 
 gives a Biased Estimator of the population variance.  In such cases, it is appropriate to use a Student's t-Distribution instead of a Gaussian Distribution.  However, it turns out (as discussedbelow) that an Unbiased Estimator for the population variance is given by|  | (9) | 
 |  | (10) | 
 The Mean and Variance of the sample standard deviation for a distribution with population mean  andVariance are 
 The quantity
  has a Chi-Squared Distribution. 
 For multiple variables, the variance is given using the definition of Covariance,
 A linear sum has a similar form:These equations can be expressed using the Covariance Matrix.
 
 To estimate the population Variance from a sample of  elements with a priori unknown Mean (i.e.,the Mean is estimated from the sample itself), we need an Unbiased Estimator for  .  This is given by the k-Statistic  , where 
 and|  | (15) | 
 is the Sample Variance 
 Note that some authors prefer the definition|  | (16) | 
 since this makes the sample variance an Unbiased Estimator for the population variance.|  | (17) | 
 When computing numerically, the Mean must be computed before  can be determined. This requires storing the set ofsample values.  It is possible to calculate  using a recursion relationship involving only the last sample asfollows.  Here, use  to denote  calculated from the first  samples (not the  th Moment) 
 and|  | (18) | 
 denotes the value for the sample variance  calculated from the first  samples.  The first fewvalues calculated for the Mean are 
 Therefore, for
  , 3 it is true that 
 Therefore, by induction,|  | (22) | 
 and
 
 for|  | (26) | 
 , so 
 Working on the first term,
 
 Use (24) to write
 
 so|  | (29) | 
 Now work on the second term in (27),|  | (30) | 
 Considering the third term in (27),But|  | (31) | 
 soPlugging (30), (31), and (34) into (27),|  | (33) | 
 so
 
 |  | (36) | 
 To find the variance of  itself, remember that 
 and|  | (37) | 
 Now find|  | (38) | 
 . 
 Working on the first term of (39),
 
 The second term of (39) is known from k-Statistic,
 
 as is the third term,|  | (41) | 
 Combining (39)-(42) gives
 
 so plugging in (38) and (43) gives
 
 Student calculated the Skewness and Kurtosis of the distribution of
  as 
 and conjectured that the true distribution is Pearson Type III Distribution
 
 where|  | (47) | 
 This was proven by R. A. Fisher.
 
 The distribution of  itself is given by 
 |  | (50) | 
 where|  | (51) | 
 The Moments are given by|  | (52) | 
 and the variance is|  | (53) | 
 An Unbiased Estimator of
  is  .  Romanovsky showed that 
 See also Correlation (Statistical), Covariance, Covariance Matrix, k-Statistic, Mean, Sample Variance|  | (55) | 
References
 Press, W. H.; Flannery, B. P.; Teukolsky, S. A.; and Vetterling, W. T.  ``Moments of a Distribution: Mean,  Variance, Skewness, and So Forth.''  §14.1 in  Numerical Recipes in FORTRAN: The Art of Scientific Computing, 2nd ed.  Cambridge, England:  Cambridge University Press, pp. 604-609, 1992.
 |