When we measure the variability of a set of data, there are two closely linked statistics related to this: the variance and standard deviation, which both indicate how spread-out the data values are and involve similar steps in their calculation. However, the major difference between these two statistical analyses is that the standard deviation is the square root of the variance.

In order to understand the differences between these two observations of statistical spread, one must first understand what each represents: Variance represents all data points in a set and is calculated by averaging the squared deviation of each mean while the standard deviation is a measure of spread around the mean when the central tendency is calculated via the mean.

As a result, the variance can be expressed as the average squared deviation of the values from the means or [squaring deviation of the means] divided by the number of observations and standard deviation can be expressed as the square root of the variance.

## Construction of Variance

To fully understand the difference between these statistics we need to understand the calculation of the variance. The steps to calculating the sample variance are as follows:

- Calculate the sample mean of the data.
- Find the difference between the mean and each of the data values.
- Square these differences.
- Add the squared differences together.
- Divide this sum by one less than the total number of data values.

The reasons for each of these steps are as follows:

- The mean provides the center point or average of the data.
- The differences from the mean help to determine the deviations from that mean. Data values that are far from the mean will produce a greater deviation than those that are close to the mean.
- The differences are squared because if the differences are added without being squared, this sum will be zero.
- The addition of these squared deviations provides a measurement of total deviation.
- The division by one less than the sample size provides a sort of mean deviation. This negates the effect of having many data points each contribute to the measurement of spread.

As stated before, the standard deviation is simply calculated by finding the square root of this result, which provides the absolute standard of deviation regardless of a total number of data values.

## Variance and Standard Deviation

When we consider the variance, we realize that there is one major drawback to using it. When we follow the steps of the calculation of the variance, this shows that the variance is measured in terms of square units because we added together squared differences in our calculation. For example, if our sample data is measured in terms of meters, then the units for a variance would be given in square meters.

In order to standardize our measure of spread, we need to take the square root of the variance. This will eliminate the problem of squared units, and gives us a measure of the spread that will have the same units as our original sample.

There are many formulas in mathematical statistics that have nicer looking forms when we state them in terms of variance instead of standard deviation.