Appendix A - Basic Probability Concepts
In this appendix definitions and some example calculations are presented which will aid in our discussions. This is not meant to be a comprehensive introduction to the topic. It is primarily meant to serve as a means for introducing notation and terminology for the course. This example is a variation of one given by David Griffiths in Intoduction to Quantum Mechanics (David J. Giffiths’s book [4]).
Example: Suppose that in some room, there are four people with the following heights:
- 1 person is 1.5 meters tall
- 1 person is 1.6 meters tall
- 2 people are 1.8 meters tall
Let stand for the number of people. We might write their heights as , , . The total number of people is
where runs over all values and (in this case) we are rounding to the nearest tenth of a meter. It is easily seen that .
Now if I draw a name out of a hat that contains each person's name once, I will get the name of a person who is 1.6 meters tall with probability . (We assume that each person has a unique name and that it appears once and only once in the hat.) We write this as
and we would generally write for any value
Now since we are going to get someone's name when we draw, we must have
which is easy enough to check.
There are several aspects of this probability distribution that we might like to know. Here are some that are particularly useful:
- The most probable value for the height is 1.8 meters.
- The median is 1.7 meters (two people above and two below).
- The average (or mean) is given by
(A.1) |
Note that the mean and the median do not have to be the same. If there is an odd number of values, the median is the middle number in the list; if even, it is the mean of the two middle values. It is mere coincidence that they are the same here. The bracket, , is the standard notation for finding the average value of a function. This is done by calculating
For the average this is just
Note: The average value is called the expectation value in quantum mechanics. This can be misleading because it is not the most probable, nor is it ''what to expect.''
When one would like to discuss the properties of a particular probability distribution, describing it takes some effort. It is not enough to know the average, median, and most probable values; a lot of details of the probability distribution remain unknown to us if these are all we are given. What else would one like to know? Without describing it entirely, one may like to know more about the ''shape'' of the distribution. For example, how spread out is it?
The most important measure of this is the variance, which is the standard deviation squared ( ). The variance is defined as (in terms of our variable )
(A.2) |
where . This can also be written as
(A.3) |
Stirling's Formula
For large , the following approximation is quite useful: