Beyond measures of central tendency (mean, median), statistics uses measures of dispersion and distribution shape to describe a data set.
The standard deviation (σ) is a number that quantifies the amount of variation or dispersion of a set of data values.
It is calculated as the square root of the variance, which is the average of the squared differences from the Mean.
The normal distribution, often called the bell curve, is a very common and important probability distribution in statistics. It is symmetric about the mean. Many natural phenomena, such as height, blood pressure, and measurement errors, are approximately normally distributed.
For a normal distribution, a predictable percentage of the data falls within a certain number of standard deviations from the mean:
This rule is a powerful tool for quickly assessing the spread of data and identifying outliers in a normally distributed data set.
If a data set has a very small standard deviation, what does this tell you about the data points?
In a perfect normal distribution, what is the relationship between the mean, median, and mode?
The scores on a test are normally distributed with a mean of 80 and a standard deviation of 5. According to the empirical rule, what percentage of students scored between 75 and 85?