Probabilistic Reasoning I Chapter 14.1–3, Bayesian Networks Chapter 14.1–3, Bayesian Networks 1 Warm-up: People are bad at probabilities Suppose that you are worried that you might have a rare disease. You decide to get tested, and suppose that the testing methods for this disease are correct 99 percent of the time (in other words, if you have the disease, it shows that you do with 99 percent probability, and if you don’t have the disease, it shows that you do not with 99 percent probability). Suppose this disease is actually quite rare, occurring randomly in the general population in only one of every 10,000 people. If your test results come back positive, what are your chances that you actually have the disease? Do you think it is approximately: (a) .99, (b) .90, (c) .10, or (d) .01? Chapter 14.1–3, Bayesian Networks 2 Outline ♦ Syntax of Bayesian Networks ♦ Semantics of Bayesian Networks ♦ Efficient representation of conditional distributions Chapter 14.1–3, Bayesian Networks 3 Example Weather Cavity Toothache Catch W eather is independent of the other variables T oothache and Catch are conditionally independent given Cavity Topology of network encodes conditional independence assertions Each node is annotated with probability information The combination of the topology and the conditional distributions suffices to specify the full joint distribution for all the variables Chapter 14.1–3, Bayesian Networks 4 Bayesian networks A Bayesian network is a directed graph in which each node is annotated with quantitative probability information. Syntax: a set of nodes, one per random variable a directed, acyclic graph (link ≈ “directly influences”) a conditional distribution for each node given its parents: P(Xi|P arents(Xi)) that quantifies the effects of the parents on Xi Example: I’m at work, neighbor John calls to say my alarm is ringing, but neighbor Mary doesn’t call. Sometimes it’s set off by minor earthquakes. Is there a burglar? Given the evidence of who has or has not called, we would like to estimate the probability of a burglary Random variables: Causal knowledge: Chapter 14.1–3, Bayesian Networks 5 Example I’m at work, neighbor John calls to say my alarm is ringing, but neighbor Mary doesn’t call. Sometimes it’s set off by minor earthquakes. Is there a burglar? Variables: Burglar, Earthquake, Alarm, JohnCalls, M aryCalls Network topology reflects “causal” knowledge: – A burglar can set the alarm off – An earthquake can set the alarm off – The alarm can cause Mary to call – The alarm can cause John to call Laziness and Ignorance: MaryListeningMusic, JohnTelefonRings, ConfusingJohn Chapter 14.1–3, Bayesian Networks 6 Conditional probability table P(E) P(B) Burglary B E P(A|B,E) T T F F T F T F .95 .94 .29 .001 JohnCalls Earthquake .001 .002 Alarm A P(J|A) T F .90 .05 A P(M|A) MaryCalls T F .70 .01 Sumarising uncertainty: i.e., the value 0.9 includes also the uncertainty for JohnTelefonRings and ConfusingJohn How many probabilities do we need? Chapter 14.1–3, Bayesian Networks 7 14.2 Semantics of Bayesian Networks What does a Bayesian network mean? representation of the joint probability distribution (global semantics) collection of conditional independence statements (local semantics) Global semantics defines the full joint distribution as the product of the local conditional distributions: B n P (x1, . . . , xn) = Πi = 1P (xi|parents(Xi)) E A J M Recall: generic entry in the joint distribution is the probability of a conjunction of particular assignments to each variable, P (X1 = x1 ∧ ... ∧ Xn = xn). e.g., P (j ∧ m ∧ a ∧ ¬b ∧ ¬e) = Chapter 14.1–3, Bayesian Networks 8 Global semantics “Global” semantics defines the full joint distribution as the product of the local conditional distributions: B n P (x1, . . . , xn) = Πi = 1P (xi|parents(Xi)) e.g., P (j ∧ m ∧ a ∧ ¬b ∧ ¬e) E A J M = P (j|a)P (m|a)P (a|¬b, ¬e)P (¬b)P (¬e) = 0.9 × 0.7 × 0.001 × 0.999 × 0.998 ≈ 0.00063 Chapter 14.1–3, Bayesian Networks 9 Constructing Bayesian networks Need a method such that a series of locally testable assertions of conditional independence guarantees the required global semantics 1. Choose an ordering of variables X1, . . . , Xn 2. For i = 1 to n add Xi to the network select parents from X1, . . . , Xi−1 such that P(Xi|P arents(Xi)) = P(Xi|X1, . . . , Xi−1) This choice of parents guarantees the global semantics (equivalence between full joint distribution and bayesian networks): n P(X1, . . . , Xn) = Πi = 1P(Xi|X1, . . . , Xi−1) (chain rule) n = Πi = 1P(Xi|P arents(Xi)) (by construction) Chapter 14.1–3, Bayesian Networks 10 Example Suppose we choose the ordering M , J, A, B, E MaryCalls JohnCalls P (J|M ) = P (J)? Chapter 14.1–3, Bayesian Networks 11 Example Suppose we choose the ordering M , J, A, B, E MaryCalls JohnCalls Alarm P (J|M ) = P (J)? No P (A|J, M ) = P (A|J)? P (A|J, M ) = P (A)? Chapter 14.1–3, Bayesian Networks 12 Example Suppose we choose the ordering M , J, A, B, E MaryCalls JohnCalls Alarm Burglary P (J|M ) = P (J)? No P (A|J, M ) = P (A|J)? P (A|J, M ) = P (A)? No P (B|A, J, M ) = P (B|A)? P (B|A, J, M ) = P (B)? Chapter 14.1–3, Bayesian Networks 13 Example Suppose we choose the ordering M , J, A, B, E MaryCalls JohnCalls Alarm Burglary Earthquake P (J|M ) = P (J)? No P (A|J, M ) = P (A|J)? P (A|J, M ) = P (A)? No P (B|A, J, M ) = P (B|A)? Yes P (B|A, J, M ) = P (B)? No P (E|B, A, J, M ) = P (E|A)? P (E|B, A, J, M ) = P (E|A, B)? Chapter 14.1–3, Bayesian Networks 14 Example Suppose we choose the ordering M , J, A, B, E MaryCalls JohnCalls Alarm Burglary Earthquake P (J|M ) = P (J)? No P (A|J, M ) = P (A|J)? P (A|J, M ) = P (A)? No P (B|A, J, M ) = P (B|A)? Yes P (B|A, J, M ) = P (B)? No P (E|B, A, J, M ) = P (E|A)? No P (E|B, A, J, M ) = P (E|A, B)? Yes Chapter 14.1–3, Bayesian Networks 15 Compactness and node ordering MaryCalls JohnCalls Alarm Burglary Earthquake Assessing conditional probabilities is hard in noncausal directions (i.e, P (Earthquake|Burglary, Alarm) Any order will work, but the resulting network will be more compact if the variables are ordered such that causes precede effects Network is less compact: For burglary net: For full joint distribution: Chapter 14.1–3, Bayesian Networks 16 Compactness Network is less compact: 1 + 2 + 4 + 2 + 4 = 13 numbers needed For burglary net, 1 + 1 + 4 + 2 + 2 = 10 numbers For full joint distribution 25 − 1 = 31 What about the node ordering: M,J,E,B,A? How many probabilities are required? A Conditional Probability Table (CPT) for Boolean Xi with k Boolean parents has 2k rows for the combinations of parent values B E A Each row requires one number p for Xi = true (the number for Xi = f alse is just 1 − p) J If each variable has no more than k parents, the complete network requires O(n · 2k ) numbers I.e., grows linearly with n, vs. O(2n) for the full joint distribution (k << n) Chapter 14.1–3, Bayesian Networks 17 M Local semantics Local semantics: each node is conditionally independent of its non-descendants given its parents U1 Um ... X Z 1j Z nj Y1 ... Yn JohnCalls is indepedent from: Chapter 14.1–3, Bayesian Networks 18 Markov blanket Each node is conditionally independent of all others given its Markov blanket: parents + children + children’s parents U1 Um ... X Z 1j Z nj Y1 ... Yn Bulgary is indepedent from: Chapter 14.1–3, Bayesian Networks 19 Example: Car diagnosis Initial evidence: car won’t start Testable variables (green), “broken, so fix it” variables (orange) Hidden variables (gray) ensure sparse structure, reduce parameters battery age battery dead battery meter lights fanbelt broken alternator broken no charging battery flat oil light no oil gas gauge no gas car won’t start fuel line blocked starter broken dipstick Chapter 14.1–3, Bayesian Networks 20 Example: Car insurance SocioEcon Age GoodStudent ExtraCar Mileage RiskAversion VehicleYear SeniorTrain MakeModel DrivingSkill DrivingHist Antilock DrivQuality Airbag Ruggedness CarValue HomeBase AntiTheft Accident Theft OwnDamage Cushioning MedicalCost OtherCost LiabilityCost OwnCost PropertyCost Chapter 14.1–3, Bayesian Networks 21 14.3 Efficient representation of cond. distribution CPT grows exponentially with number of parents CPT becomes infinite with continuous-valued parent or child Deterministic nodes are the simplest case: its value is specified by the value of parent X = f (P arents(X)) for some function f E.g., logical relationships N orthAmerican ⇔ Canadian ∨ U S ∨ M exican E.g., numerical relationships among continuous variables ∂Level = inflow + precipitation - outflow - evaporation ∂t Chapter 14.1–3, Bayesian Networks 22 Efficient representation of cond. distribution Noisy-OR distributions model multiple noninteracting causes Example: F ever ⇔ Cold ∨ F lu ∨ M alaria Allows for uncertainty about the ability of each parent to cause the child to be true (the causal relationship between parent and child may be inhibited) Example: a patient could have a Cold, but not exhibit a Fever Assumptions: 1) Parents U1 . . . Uk include all causes (can add leak node) 2) Independent failure probability qi for each cause alone j ⇒ P (X|U1 . . . Uj , ¬Uj+1 . . . ¬Uk ) = 1 − Πi = 1qi Chapter 14.1–3, Bayesian Networks 23 Efficient representation of cond. distribution Fever is false if and only if all its parents are inhibited. The probability of this is the product of the inhibition probabilities q for each parent qcold = P (¬f ever|cold, ¬f lu, ¬malaria) = 0.6 qf lu = P (¬f ever|¬cold, f lu, ¬malaria) = 0.2 qmalaria = P (¬f ever|¬cold, ¬f lu, malaria) = 0.1 Cold F F F F T T T T F lu F F T T F F T T M alaria F T F T F T F T P (F ever) 0.0 0.9 0.8 0.98 0.4 0.94 0.88 P (¬F ever) 1.0 0.1 0.2 0.02 = 0.2 × 0.1 0.6 0.06 = 0.6 × 0.1 0.12 = 0.6 × 0.2 Number of parameters linear in number of parents Chapter 14.1–3, Bayesian Networks 24 Hybrid (discrete+continuous) networks Discrete (Subsidy? and Buys?); continuous (Harvest and Cost) Subsidy? Harvest Cost Buys? Option 1: discretization—possibly large errors, large CPTs Option 2: finitely parameterized canonical families Option 3: define conditional distribution with a collection of instances 1) Continuous variable, discrete+continuous parents (e.g., Cost) 2) Discrete variable, continuous parents (e.g., Buys?) Chapter 14.1–3, Bayesian Networks 25 Continuous child variables Need one conditional density function for child variable given continuous parents, for each possible assignment to discrete parents Most common is the linear Gaussian model, e.g.,: P (Cost = c|Harvest = h, Subsidy? = true) = N (ath + bt, σt)(c) 2 1 1 c − (ath + bt) √ = exp − 2 σt σt 2π Cost decreases as supply increases Chapter 14.1–3, Bayesian Networks 26 Discrete variable with continuous parents Probability of Buys? given Cost should be a “soft” threshold: 1 P(Buys?=false|Cost=c) 0.8 0.6 0.4 0.2 0 0 2 4 6 Cost c 8 10 12 Probit distribution uses integral of Gaussian: Rx N (0, 1)(x)dx Φ(x) = −∞ P (Buys? = true | Cost = c) = Φ((−c + µ)/σ) Chapter 14.1–3, Bayesian Networks 27 Why the probit? 1. It’s sort of the right shape 2. Can view as hard threshold whose location is subject to noise Cost Cost Noise Buys? Chapter 14.1–3, Bayesian Networks 28 Discrete variable contd. Sigmoid (or logit) distribution also used in neural networks: P (Buys? = true | Cost = c) = 1 1 + exp(−2 −c+µ σ ) Sigmoid has similar shape to probit but much longer tails: 1 0.9 P(Buys?=false|Cost=c) 0.8 0.7 0.6 0.5 0.4 0.3 0.2 0.1 0 0 2 4 6 Cost c 8 10 12 Chapter 14.1–3, Bayesian Networks 29 Summary A Bayesian network is a directed acyclic graph whose nodes correspond to random variables; each node has a conditional distribution for the node, given its parents Bayesian networks provide a concise way to represent conditional independence relationships in the domain A Bayesian network specifies a full joint distribution; each joint entry is defined as the product of the corresponding entries in the local conditional distributions. A Bayesian network is often exponentially smaller than an explicitly enumerated joint distribution. Hybrid Bayesian networks, which include both discrete and continuous variables, use a variety of canonical distributions. Chapter 14.1–3, Bayesian Networks 30 Detecting pregnancy based on three tests The first is a scanning test which has a false positive of 1% and a false negative of 10%. The second is a blood test, which detects progesterone with a false positive of 10a false negative of 30%. The third test is a urine test, which also detects progesterone with a false positive of 10% and a false negative of 20%. The probability of a detectable progesterone level is 90% given pregnancy, and 1% given no pregnancy. The probability of pregnancy after two positive tests is 0.364%. Construct the Bayesian network and specify the conditional probability tables. Chapter 14.1–3, Bayesian Networks 31
© Copyright 2024