1. Sample Space and Probability Part IV: Pascal Triangle and Bernoulli Trials ECE 302 Spring 2012 Purdue University, School of ECE Prof. Ilya Pollak ConnecGon between Pascal triangle and probability theory: Number of successes in a sequence of independent Bernoulli trials • A Bernoulli trial is any probabilisGc experiment with two possible outcomes Ilya Pollak ConnecGon between Pascal triangle and probability theory: Number of successes in a sequence of independent Bernoulli trials • A Bernoulli trial is any probabilisGc experiment with two possible outcomes, e.g., – Will CiGgroup become insolvent during next 12 months? – Democrats or Republicans in the next elecGon? – Will Dow Jones go up tomorrow? – Will a new drug cure at least 80% of the paGents? Ilya Pollak ConnecGon between Pascal triangle and probability theory: Number of successes in a sequence of independent Bernoulli trials • A Bernoulli trial is any probabilisGc experiment with two possible outcomes, e.g., – Will CiGgroup become insolvent during next 12 months? – Democrats or Republicans in the next elecGon? – Will Dow Jones go up tomorrow? – Will a new drug cure at least 80% of the paGents? • Terminology: someGmes the two outcomes are called “success” and “failure.” • Suppose the probability of success is p. What is the probability of k successes in n independent trials? Ilya Pollak Probability of k successes in n independent Bernoulli trials • n independent coin tosses, P(H) = p Ilya Pollak Probability of k successes in n independent Bernoulli trials • n independent coin tosses, P(H) = p • E.g., P(HTTHHH) = p(1-‐p)(1-‐p)p3 = p4(1-‐p)2 Ilya Pollak Probability of k successes in n independent Bernoulli trials • n independent coin tosses, P(H) = p • E.g., P(HTTHHH) = p(1-‐p)(1-‐p)p3 = p4(1-‐p)2 • P(specific sequence with k H’s and (n-‐k) T’s) = pk (1-‐p)n-‐k Ilya Pollak Probability of k successes in n independent Bernoulli trials • • • • n independent coin tosses, P(H) = p E.g., P(HTTHHH) = p(1-‐p)(1-‐p)p3 = p4(1-‐p)2 P(specific sequence with k H’s and (n-‐k) T’s) = pk (1-‐p)n-‐k P(k heads) = (number of k-‐head sequences) ·∙ pk (1-‐p)n-‐k Ilya Pollak Probability of k successes in n independent Bernoulli trials • • • • n independent coin tosses, P(H) = p E.g., P(HTTHHH) = p(1-‐p)(1-‐p)p3 = p4(1-‐p)2 P(specific sequence with k H’s and (n-‐k) T’s) = pk (1-‐p)n-‐k P(k heads) = (number of k-‐head sequences) ·∙ pk (1-‐p)n-‐k Ilya Pollak An interesGng property of binomial coefficients Since P(zero H's) + P(one H) + P(two H's) + … + P(n H's) = 1, n ⎛ ⎞ n k it follows that ∑⎜ ⎟ p (1− p) n−k = 1. k k= 0 ⎝ ⎠ Another way to show the same thing is to realize that n ⎛ ⎞ n k n−k n n p (1− p) = ( p + (1− p)) = 1 = 1. ∑⎜⎝ k ⎟⎠ k= 0 Ilya Pollak Binomial probabiliGes: illustraGon Ilya Pollak Binomial probabiliGes: illustraGon Ilya Pollak Ilya Pollak Ilya Pollak Ilya Pollak Ilya Pollak Ilya Pollak Ilya Pollak Ilya Pollak Ilya Pollak Ilya Pollak Ilya Pollak Ilya Pollak Ilya Pollak Ilya Pollak Ilya Pollak Comments on binomial probabiliGes and the bell curve • Summing many independent random contribuGons usually leads to the bell-‐shaped distribuGon. Ilya Pollak Comments on binomial probabiliGes and the bell curve • Summing many independent random contribuGons usually leads to the bell-‐shaped distribuGon. • This is called the central limit theorem (CLT). • We have not yet covered the tools to precisely state the CLT, but we will later in the course. Ilya Pollak Comments on binomial probabiliGes and the bell curve • Summing many independent random contribuGons usually leads to the bell-‐shaped distribuGon. • This is called the central limit theorem (CLT). • We have not yet covered the tools to precisely state the CLT, but we will later in the course. • The behavior of the binomial distribuGon for large n shown above is a manifestaGon of the CLT. Ilya Pollak InteresGngly, we get the bell curve even for asymmetric binomial probabiliGes Ilya Pollak Ilya Pollak Ilya Pollak Ilya Pollak Ilya Pollak Ilya Pollak Ilya Pollak Ilya Pollak Ilya Pollak Ilya Pollak Ilya Pollak Ilya Pollak Ilya Pollak Ilya Pollak Ilya Pollak Ilya Pollak Ilya Pollak Ilya Pollak Ilya Pollak Ilya Pollak Ilya Pollak Ilya Pollak Ilya Pollak Ilya Pollak Ilya Pollak Ilya Pollak Ilya Pollak Ilya Pollak Ilya Pollak Ilya Pollak Ilya Pollak Ilya Pollak Ilya Pollak Ilya Pollak Ilya Pollak This tells us how to empirically esGmate the probability of an event! • To esGmate the probability p based on n flips, divide the observed number of H’s by the total number of experiments: k/n. • To see the distribuGon of k/n for any n, simply rescale the x-‐axis in the distribuGon of k. • This distribuGon will tell us – What we should expect our esGmate to be, on average, and – What error we should expect to make, on average Ilya Pollak Ilya Pollak Ilya Pollak Ilya Pollak Ilya Pollak Note: o for 50 flips, the most likely outcome is the correct one, 0.8 o it’s also close to the “average” outcome o it’s very unlikely to make a mistake of more than 0.2 Ilya Pollak Ilya Pollak Ilya Pollak If p=0.8, when estimating based on 1000 flips, it’s extremely unlikely to make a mistake of more than 0.05. Ilya Pollak If p=0.8, when estimating based on 1000 flips, it’s extremely unlikely to make a mistake of more than 0.05. • Hence, when the goal is to forecast a two-way election, and the actual p is reasonably far from 1/2, polling a few hundred people is very likely to give accurate results. Ilya Pollak If p=0.8, when estimating based on 1000 flips, it’s extremely unlikely to make a mistake of more than 0.05. • Hence, when the goal is to forecast a two-way election, and the actual p is reasonably far from 1/2, polling a few hundred people is very likely to give accurate results. • However, o independence is important; o getting a representative sample is important (for a country with 300M population, this is tricky!) o when the actual p is extremely close to 1/2 (e.g., the 2000 presidential election in Florida or the 2008 senatorial election in Minnesota), pollsters’ forecasts are about as accurate as a random guess. Ilya Pollak The 2008 Franken-‐Coleman elecGon • Franken 1,212,629 votes • Coleman 1,212,317 votes • In our analysis, we will disregard third party candidate who got 437,505 votes (he actually makes pre-‐elecGon polling even more complicated) • EffecGvely, p ≈ 0.500064 Ilya Pollak ProbabiliGes for fracGons of Franken vote in pre -‐elecGon polling based on n=2.5M (more than all Franken and Coleman votes combined) • Even though we are unlikely to make an error of more than 0.001, this is not enough because p-0.5=0.000064! • Note: 42% of the area under the bell curve is to the left of 1/2. • When the election is this close, no poll can accurately predict the outcome. • In fact, the noise in the voting process itself (voting machine malfunctions, human errors, etc) becomes very important in determining the outcome. Ilya Pollak EsGmaGng the probability of success in a Bernoulli trial: summary • As the number n of independent experiments increases, the empirical fracGon of occurrences of success becomes close to the actual probability of success, p. • The error goes down proporGonately to n1/2. I.e., error aler 400 trials is twice as small as aler 100 trials. • This is called the law of large numbers. • This result will be precisely described later in the course. Ilya Pollak
© Copyright 2024