Probability is the mathematical study of measuring uncertainty. Probabilities are classically determined when their numerical values are based upon an enumeration of every possible outcome.
In classical probability, we call the process which generates outcomes a statistical experiment. A list of all possible outcomes of a statistical experiment is called a sample space. We especially desire that the outcomes in our sample space be equally likely.
Caution: A sample space is really a population of outcomes, not a sample of outcomes.
When examined classically, the probability that an event will occur will be equal to the ratio of the number of outcomes producing that event, to the total number of possible outcomes for that experiment (that is, the size of the sample space). More specifically, if $A$ is the name of an event, $f$ is the frequency with which that event occurs in the sample space, and $N$ is the size of the sample space, then:
$P(A) = \dfrac{f}{N}$ |
Suppose our statistical experiment involves rolling one die. Since the die has 6 sides, there are six possible outcomes in the sample space. We can write the sample space as the set ${1,2,3,4,5,6}$. We can also create a probability distribution, which is basically a frequency distribution with the frequency column replaced by a column of probabilities. For rolling one die, the frequency distribution is:
Outcome on the Die | Probability |
1 | $\dfrac16$ |
2 | $\dfrac16$ |
3 | $\dfrac16$ |
4 | $\dfrac16$ |
5 | $\dfrac16$ |
6 | $\dfrac16$ |
We will let $x$ represent the outcome on the die. Then:
The basic rule was used in almost all parts of the preceding example. Notice especially how important the "little" words were. Be sure you understand what "at most" and "at least" mean. Also notice the distinction between "and" and "or".
The last part of the example, where we looked for an outcome of both 4 and 5, turned out to have a probability of zero. We call the events "outcome of 4" and "outcome of 5" mutually exclusive events, which means they cannot happen simultaneously. Usually, the outcomes of a sample space, and therefore in a probability distribution, will be constructed so that they are mutually exclusive.
Suppose we now roll two dice. The outcomes of this experiment depend on the two separate outcomes of each die, so there are two independent variables. We can display the 36 outcomes in a table.
(1,1) | (1,2) | (1,3) | (1,4) | (1,5) | (1,6) |
(2,1) | (2,2) | (2,3) | (2,4) | (2,5) | (2,6) |
(3,1) | (3,2) | (3,3) | (3,4) | (3,5) | (3,6) |
(4,1) | (4,2) | (4,3) | (4,4) | (4,5) | (4,6) |
(5,1) | (5,2) | (5,3) | (5,4) | (5,5) | (5,6) |
(6,1) | (6,2) | (6,3) | (6,4) | (6,5) | (6,6) |
The sample space can help us determine the following probabilities.
There are many other types of classical probability problems besides rolling dice. Examples include flipping coins, drawing cards from a deck, guessing on a multiple choice test, selecting jellybeans from a bag, choosing people for a committee, and so on. Any situation that can be analyzed to determine its component parts, with the probabilities computed on the basis of those parts, is fodder for classical probability.