In this topic, we'll focus on the empirical rule — a section of the probability theory. This statistical rule helps to create risk models that are associated with certain decisions and alternative outcomes. For example, investors use empirical rule formula to predict future stock prices, marketing managers can reduce the likelihood of failure in product sales, and casinos can develop games of chance, such that their profit is maximized.
Probability
Basically, the probability is the degree to which something can happen. Probabilities can be expressed as proportions that range from to . Moreover, they can be expressed as percentages ranging from to . Roughly speaking, indicates the impossibility of the future event, and indicates certainty.
For example, when a coin is tossed, there are two possible outcomes: heads or tails . So, the probability of the coin landing heads as well as tails is or
To determine the probability of a single event, first, we need to know the total number of favorable and possible outcomes. Let's say we need to determine the probability of getting when throwing a die. In total, we have six possible outcomes: . Obviously, there's only one favorable outcome for us: . Now that we know the number of outcomes, we use the probability formula:
The probability of rolling a when throwing a die can be calculated as . However, in these examples, we know exactly the number of possible outcomes, and it's rather small. But what if we have millions of outcomes? Or we even don't know the exact number of outcomes? That's when the law of large numbers (LLN) comes into play.
Chebyshev's inequality
LLN states that if the same experiment or study is repeated independently many times, the average of the results of the trials must be close to the expected value. Note that the independence of the experiments is critical. LLN needs to have a set of random variables not influencing one another. Getting coins from a bag one after another while knowing, for example, the number of gold and silver ones there, won't be considered independent experiments because each time the probability of taking out another silver coin depends on the number of silver coins taken out previously.
Chebyshev's inequality refers to the law of large numbers. The theorem states that a certain proportion of any data set must fall within a particular range around the mean value. This is determined by the standard deviation of the data. The mean is referred to as the average of all the numbers in the data set. Chebyshev's theorem is significant because it applies to a wide distribution range. It can be stated mathematically in the form of an inequality, which is why the theorem is often referred to as Chebyshev's inequality.
Let be a random variable with a finite mean and a finite standard deviation and let be any positive number. The probability that the difference between and is more than standard deviations can't be more than .
Now, let's look at an example. Suppose a class takes a test. The average score is and the standard deviation is . What is the proportion of scores that fall between and ?
The mean is . and are points below and above the mean, respectively. The standard deviation is . Consequently, you want to determine the proportion of scores that fall within standard deviations of the mean.
or
Here we calculated the probability that any student's score falls more than standard deviations away from the mean (less than or more than ). But we need the opposite: the proportion of scores that fall less than standard deviations away, so . Therefore, , which means that at least of the scores will fall within the range of .
Three-sigma rule
However, a lot of data follows the pattern of a normal distribution. In this case, we usually use the much more convenient or the . It is sometimes called the empirical rule because it originally came from observations and its name translates as "based on observation". You can use this so-called rule of thumb when you are told your data is normal, nearly normal, or if you have a normal, bell-shaped distribution with a single peak. Basically, it's a version of Chebyshev's inequality that gives more precise predictions. Let's take a look at it:
-
About of the values fall within one standard deviation of the mean:
-
About of the values fall within two standard deviations from the mean:
-
About of the values fall within three standard deviations from the mean:
For example, we have observed daily sales in a store for some time. The range of values follows a normal distribution with an average value of dollars and a standard deviation of dollars. Then, according to the rule, sales lower than and higher than are practically impossible events. This means that it makes no sense to consider these sales as potentially possible.
However, it is often important to get precise results. Let's look at another example. The intelligence quotient (IQ) scores of human population are normally distributed with a mean of and a size of deviation equal to . What percentage of people have an IQ between and ? According to the empirical rule:
of people have an IQ between and (IQ more than and less than ).
of people have an IQ between and (IQ more than and less than ).
of people have an IQ between and (IQ more than and less than ).
But we don't need data about the interval of IQ between and , only about . This area is to the right of the mean if we look at the empirical rule graph.
In this case, to get the answer, we halve the percentage: .
In this case, to get the answer, we halve the empirical rule percentages: .
Conclusion
In this topic, we got acquainted with the notion of probability. We familiarized ourselves with different methods that are used to calculate the probability of any given event, some of them are:
- The probability formula: is applied in straightforward cases with a few outcomes and looks as follows:
- Chebyshev's theorem: it is particularly useful in situations when the mean and variance are known, as it can be applied to any probability distribution in that case.
- The empirical rule: is used to predict probable outcomes for a random variable that follows a normal distribution.