Today's Topics : Probability

Math (P)refresher Day 5:
Great Expectations, Part I
September 2000
Today's Topics : Probability Conditional Probability and Independence Bayes' Rule
1
Probability
Probability: Many events or outcomes are random. In every day speech, we say that we are
about the outcome of random events. Probability is a formal model of uncertainty
which provides a measure of uncertainty governed by a particular set of rules. A dierent
model of uncertainty would, of course, have a dierent set of rules and measures. Our focus on
probability is justied because it has proven to be a particularly useful model of uncertainty.
Sample Space: A set or collection of all possible outcomes. Outcomes in the set can be
discrete elements or points along a continuous interval.
Examples:
1. Discrete: the numbers on a die, the number of possible wars that could occur each year,
whether a vote cast is republican or democrat.
2. Continuous: GNP, arms spending, age.
Probability Distribution: A probability distribution on a sample space S is a specication
of numbers Pr(A) that satisfy Axioms 1{3.
Axioms of Probability: It's necessary to assign a number Pr(A) to each event A in the
sample space S .
1. Axiom: For any event A, Pr(A) 0.
2. Axiom: Pr(S ) = 1
3. Axiom:
1 For
any1innite sequence of disjoint events A1 ; A2 ; : : : ,
S
P
Pr
Ai = Pr(Ai )
uncertain
i=1
i=1
Basic Theorems of Probability:
1. Pr(;) = 0
2. Forany sequence
of n disjoint events A ; A ; : : : ; An ,
n P
n
S
Pr
Ai = Pr(Ai )
i
i
3. Pr(A) = 1 ; Pr(A)
1
=1
2
=1
Much of the material and examples for this lecture are taken from Wackerly, Mendenhall, & Scheaer (1996)
Mathematical Statistics with Applications, Degroot (1985) Probability and Statistics, Morrow (1994) Game Theory
for Political Scientists, King (1989) Unifying Political Methodology, and Ross (1987) Introduction to Probability and
Statistics for Scientists and Engineers.
2
Math (P)refresher: Great Expectations
4. For any event A, 0 Pr(A) 1.
5. If A B , then Pr(A) Pr(B ).
6. For any two events A and B , Pr(A [ B ) = Pr(A) + Pr(B ) ; Pr(A \ B )
Examples: Let's assume we have an evenly-balanced, six-sided die. Then,
1. Sample space S = f1; 2; 3; 4; 5; 6g
2. Pr(1) = = Pr(6) = 1=6
3. Pr(;) = Pr(7) = 0
4. Pr (f1; 3; 5g) = 1=6 + 1=6 + 1=6 = 1=2
5. Pr f1; 2g = Pr (f3; 4; 5; 6g) = 2=3
6. Let B = S and A = f1; 2; 3; 4; 5g B . Then Pr(A) = 5=6 < Pr(B ) = 1.
7. Let A = f1; 2; 3g and B = f2; 4; 6g. Then A [ B = f1; 2; 3; 4; 6g, A \ B = f2g, and
Pr(A [ B ) = Pr(A) + Pr(B ) ; Pr(A \ B )
= 3=6 + 3=6 ; 1=6
= 5=6
2
Conditional Probability and Independence
Conditional Probability: The conditional probability Pr(AjB ) of an event A is the probability of A, given that another event B has occurred. It is calculated as
A \ B)
Pr(AjB ) = Pr(Pr(
B)
Example: Assume A and B occur with the following frequencies:
and let nab + nab + nab + nab = N . Then
1.
2.
3.
4.
5.
A A
B nab nab
B nab nab
Pr(A) nabN+nab
Pr(B ) nabN+nab
Pr(A \ B ) nNab
Pr(AjB ) Pr(Pr(AB\B) ) = nabn+abnab
Pr(B jA) Pr(Pr(AA\B) ) = nabn+abnab
Example: A six-sided die is rolled. What is the probability of a 1, given the outcome
is an
odd number? Let A = f1g, B = f1; 3; 5g, and A \ B = f1g. Then, Pr(AjB ) = AB\B =
= = 1=3.
=
Multiplicative Law of Probability: The probability of the intersection of two events A
and B is
Pr(A \ B ) = Pr(A) Pr(B jA) = Pr(B ) Pr(AjB )
Pr(
Pr(
1 6
1 2
which follows directly from the denition of conditional probability.
)
)
3
Math (P)refresher: Great Expectations
Independence: If the occurrence or nonoccurrence of either events A and B have no eect
on the occurrence or nonoccurrence of the other, then A and B are independent. If A and B
are independent, then
1. Pr(AjB ) = Pr(A)
2. Pr(B jA) = Pr(B )
3. Pr(A \ B ) = Pr(A) Pr(B )
Calculating the Probability of an Event Using the Event-Composition Method:
The event-composition method for calculating the probability of an event A involves expressing A as a composition involving the unions and/or intersections of other events. Then use
the laws of probability to to nd Pr(A). The steps used in the event-composition method are:
1. Dene the experiment.
2. Identify the general nature of the sample points.
3. Write an equation expressing the event of interest A as a composition of two or more
events, using unions, intersections, and/or complements.
4. Apply the additive and multiplicative laws of probability to the compositions obtained
in step 3 to nd Pr(A).
Examples:
1.
2.
3.
4.
3
Bayes' Rule
Law of Total Probability: Let S be the sample space of some experiment and let the
disjoint k events B ; : : : ; Bk partition S . If A is some other event in S , then the events
AB ; AB ; : : : ; ABk will form a partition of A and we can write A as
A = (AB ) [ [ (ABk )
Since the k events are disjoint,
1
1
2
1
Pr(A) =
=
k
X
i=1
k
X
i=1
Pr(ABi )
Pr(Bi ) Pr(AjBi )
Sometimes it is easier to calculate the conditional probabilities and sum them than it is to
calculate Pr(A) directly.
4
Math (P)refresher: Great Expectations
Bayes Rule: Assume that events B ; : : : ; Bk form a partition of the space S . Then
ABj ) = Pr(Bj ) Pr(AjBj )
Pr(Bj jA) = Pr(
k
Pr(A)
P
Pr(Bi ) Pr(AjBi )
1
i=1
If there are only two states of B , then this is just
B1 ) Pr(AjB1 )
Pr(B1 jA) = Pr(B ) Pr(Pr(
AjB1 ) + Pr(B2 ) Pr(AjB2 )
1
Bayes rule determines the posterior probability of a state or type Pr(Bj jA) by calculating the
probability Pr(ABj ) that both the event A and the state Bj will occur and dividing it by the
probability that the event will occur regardless of the state (by summing across all Bi ).
Often Bayes' rule is used when one wants to calculate a posterior probability about the \state"
or type of an object, given that some event has occurred. The states could be something like
Normal/Defective, Normal/Diseased, Democrat/Republican, etc. The event on which one
conditions could be something like a sampling from a batch of components, a test for a
disease, or a question about a policy position.
Prior and Posterior Probabilities: In the above, Pr(B1 ) is often called the prior probability, since it's the probability of B1 before anything else is known. Pr(B1 jA) is called the
posterior probability, since it's the probability after other information is taken into account.
Examples:
1. A test for cancer correctly detects it 90% of the time, but incorrectly identies a person
as having cancer 10% of the time. If 10% of all people have cancer at any given time,
what is the probability that a person who tests positive actually has cancer?
Let C = fCancerg, C = fNo cancerg, + = fTests positiveg, and ; = fTests negativeg.
Then
Pr(C ) Pr(+jC )
:9)
Pr(C j+) =
= (:1)(:(9):1)(
+ (:9)(:1) = 0:5
Pr(C ) Pr(+jC ) + Pr(C ) Pr(+jC )
2. In Boston, 30% of the people are conservatives, 50% are liberals, and 20% are independents. In the last election, 65% of conservatives, 82% of liberals, and 50% of independents
voted. If a person in Boston is selected at random and we learn that s/he did not vote
last election, what is the probability s/he is a liberal?
Let C , L, and I represent conservatives, liberals, and independents, respectively. And,
let V = fVotedg and V = fDidn't voteg. Then,
Pr(L) Pr(V jL)
Pr(C ) Pr(V jC ) + Pr(L) Pr(V jL) + Pr(I ) Pr(V jI )
:18)
= (:3)(:35) +((::5)(
5)(:18) + (:2)(:5)
= 18
59
Pr(LjV ) =