Lecture8 - bays1
Lecture8 - bays1
Bayes’ Nets
Fall 2022
[These slides were created by Dan Klein and Pieter Abbeel for CS188 Intro to AI at UC Berkeley. All CS188 materials are available at http://ai.berkeley.edu.]
Review: Probabilistic Inference
§ Probabilistic inference: compute a desired
probability from other known probabilities (e.g.
conditional from joint)
§ Step 1: Select the § Step 2: Sum out H to get joint § Step 3: Normalize
entries consistent of Query and evidence
with the evidence
1
⇥
Z
Review: Inference by Enumeration
S T W P
§ P(W)?
summer hot sun 0.30
summer hot rain 0.05
summer cold sun 0.10
§ P(W | winter)? summer cold rain 0.05
winter hot sun 0.10
winter hot rain 0.05
winter cold sun 0.15
winter cold rain 0.20
§ P(W | winter, hot)?
Review: Inference by Enumeration
§ Obvious problems:
§ Worst-case time complexity O(dn)
§ Space complexity O(dn) to store the joint distribution
Review: The Product Rule
§ Sometimes have conditional distributions but want the joint
Review: The Product Rule
§ Example:
D W P D W P
wet sun 0.1 wet sun 0.08
R P
dry sun 0.9 dry sun 0.72
sun 0.8
wet rain 0.7 wet rain 0.14
rain 0.2
dry rain 0.3 dry rain 0.06
Review: The Chain Rule
§ This says that their joint distribution factors into a product two
simpler distributions
§ Another form:
§ We write:
T P
hot 0.5
cold 0.5
T W P T W P
hot sun 0.4 hot sun 0.3
hot rain 0.1 hot rain 0.2
cold sun 0.2 cold sun 0.3
cold rain 0.3 cold rain 0.2
W P
sun 0.6
rain 0.4
Example: Independence
§ N fair, independent coin flips:
§ Equivalent statements:
§ P(Toothache | Catch , Cavity) = P(Toothache | Cavity)
§ P(Toothache, Catch | Cavity) = P(Toothache | Cavity) P(Catch | Cavity)
§ One can be derived from the other easily
Conditional Independence
§ Unconditional (absolute) independence very rare (why?)
§ Trivial decomposition:
§ Arcs: interactions
§ Similar to CSP constraints
§ Indicate “direct influence” between variables
§ Formally: encode conditional independence
(more later)
X1 X2 Xn
R R
T T
§ Why is an agent using model 2 better?
Example: Traffic II
§ Let’s build a causal graphical model!
§ Variables
§ T: Traffic
§ R: It rains
§ L: Low pressure
§ D: Roof drips
§ B: Ballgame
§ C: Cavity
Example: Alarm Network
§ Variables
§ B: Burglary
§ A: Alarm goes off
§ M: Mary calls
§ J: John calls
§ E: Earthquake!
Bayes’ Net Semantics
Bayes’ Net Semantics
§ A set of nodes, one per variable X
§ Example:
Probabilities in BNs
§ Why are we guaranteed that setting
à Consequence:
X1 X2 Xn
+r 1/4
R
-r 3/4
+r +t 3/4
T -t 1/4
-r +t 1/2
-t 1/2
Example: Alarm Network
B P(B) E P(E)
Burglary Earthqk +e 0.002
+b 0.001
-b 0.999 -e 0.998
Alarm
B E A P(A|B,E)
+b +e +a 0.95
John Mary
calls calls +b +e -a 0.05
+b -e +a 0.94
A J P(J|A) A M P(M|A) +b -e -a 0.06
+a +j 0.9 +a +m 0.7 -b +e +a 0.29
+a -j 0.1 +a -m 0.3 -b +e -a 0.71
-a +j 0.05 -a +m 0.01 -b -e +a 0.001
-a -j 0.95 -a -m 0.99 -b -e -a 0.999
Example: Traffic
§ Causal direction
+r 1/4
R
-r 3/4
+r +t 3/16
+r -t 1/16
+r +t 3/4
-r +t 6/16
T -t 1/4
-r -t 6/16
-r +t 1/2
-t 1/2
Example: Reverse Traffic
§ Reverse causality?
+t 9/16
T
-t 7/16
+r +t 3/16
+r -t 1/16
+t +r 1/3
-r +t 6/16
R -r 2/3
-r -t 6/16
-t +r 1/7
-r 6/7
Causality?
§ When Bayes’ nets reflect the true causal patterns:
§ Often simpler (nodes have fewer parents)
§ Often easier to think about
§ Often easier to elicit from experts