Bayesian belief networks - University of Pittsburghmilos/courses/cs1571/Lectures/Class20.pdf• We...

26
1 M. Hauskrecht CS 1571 Intro to AI CS 1571 Introduction to AI Lecture 20 Milos Hauskrecht [email protected] 5329 Sennott Square Bayesian belief networks M. Hauskrecht CS 1571 Intro to AI Modeling uncertainty with probabilities Defining the full joint distribution makes it possible to represent and reason with uncertainty in a uniform way We are able to handle an arbitrary inference problem Problems: Space complexity. To store a full joint distribution we need to remember numbers. n – number of random variables, d – number of values Inference (time) complexity. To compute some queries requires . steps. Acquisition problem. Who is going to define all of the probability entries? ) (d n O ) (d n O

Transcript of Bayesian belief networks - University of Pittsburghmilos/courses/cs1571/Lectures/Class20.pdf• We...

Page 1: Bayesian belief networks - University of Pittsburghmilos/courses/cs1571/Lectures/Class20.pdf• We want to represent the probability distribution of events: – Burglary, Earthquake,

1

M. HauskrechtCS 1571 Intro to AI

CS 1571 Introduction to AILecture 20

Milos Hauskrecht

[email protected]

5329 Sennott Square

Bayesian belief networks

M. HauskrechtCS 1571 Intro to AI

Modeling uncertainty with probabilities

• Defining the full joint distribution makes it possible to represent and reason with uncertainty in a uniform way

• We are able to handle an arbitrary inference problem

Problems:

– Space complexity. To store a full joint distribution we need to remember numbers.

n – number of random variables, d – number of values

– Inference (time) complexity. To compute some queries requires . steps.

– Acquisition problem. Who is going to define all of the probability entries?

)(dnO

)(dnO

Page 2: Bayesian belief networks - University of Pittsburghmilos/courses/cs1571/Lectures/Class20.pdf• We want to represent the probability distribution of events: – Burglary, Earthquake,

2

M. HauskrechtCS 1571 Intro to AI

Bayesian belief networks (BBNs)

Bayesian belief networks.

• Represent the full joint distribution over the variables more compactly with a smaller number of parameters.

• Take advantage of conditional and marginal independencesamong random variables

• A and B are independent

• A and B are conditionally independent given C)()(),( BPAPBAP

)|()|()|,( CBPCAPCBAP )|(),|( CAPBCAP

M. HauskrechtCS 1571 Intro to AI

Alarm system example• Assume your house has an alarm system against burglary.

You live in the seismically active area and the alarm system can get occasionally set off by an earthquake. You have two neighbors, Mary and John, who do not know each other. If they hear the alarm they call you, but this is not guaranteed.

• We want to represent the probability distribution of events:

– Burglary, Earthquake, Alarm, Mary calls and John calls

Burglary

JohnCalls

Alarm

Earthquake

MaryCalls

Causal relations

Page 3: Bayesian belief networks - University of Pittsburghmilos/courses/cs1571/Lectures/Class20.pdf• We want to represent the probability distribution of events: – Burglary, Earthquake,

3

M. HauskrechtCS 1571 Intro to AI

Bayesian belief network

Burglary Earthquake

JohnCalls MaryCalls

Alarm

1. Directed acyclic graph• Nodes = random variables

Burglary, Earthquake, Alarm, Mary calls and John calls• Links = direct (causal) dependencies between variables.

The chance of Alarm is influenced by Earthquake, The chance of John calling is affected by the Alarm

M. HauskrechtCS 1571 Intro to AI

Bayesian belief network

2. Local conditional distributions • relate variables and their parents

Burglary Earthquake

JohnCalls MaryCalls

Alarm

P(B) P(E)

P(A|B,E)

P(J|A) P(M|A)

Page 4: Bayesian belief networks - University of Pittsburghmilos/courses/cs1571/Lectures/Class20.pdf• We want to represent the probability distribution of events: – Burglary, Earthquake,

4

M. HauskrechtCS 1571 Intro to AI

Bayesian belief network

Burglary Earthquake

JohnCalls MaryCalls

Alarm

B E T F

T T 0.95 0.05T F 0.94 0.06F T 0.29 0.71F F 0.001 0.999

P(B)

0.001 0.999

P(E)

0.002 0.998

A T F

T 0.90 0.1F 0.05 0.95

A T F

T 0.7 0.3F 0.01 0.99

P(A|B,E)

P(J|A) P(M|A)

T F T F

M. HauskrechtCS 1571 Intro to AI

Bayesian belief networks (general)

Two components:

• Directed acyclic graph

– Nodes correspond to random variables

– (Missing) links encode independences

• Parameters

– Local conditional probability distributions

for every variable-parent configuration

))(|( ii XpaXP

A

B

MJ

E),( SSB

)( iXpa - stand for parents of Xi

Where:

B E T F

T T 0.95 0.05T F 0.94 0.06F T 0.29 0.71F F 0.001 0.999

P(A|B,E)

Page 5: Bayesian belief networks - University of Pittsburghmilos/courses/cs1571/Lectures/Class20.pdf• We want to represent the probability distribution of events: – Burglary, Earthquake,

5

M. HauskrechtCS 1571 Intro to AI

Full joint distribution in BBNs

Full joint distribution is defined in terms of local conditional distributions (obtained via the chain rule):

))(|(),..,,(,..1

21

ni

iin XpaXXXX PP

M

A

B

J

E

),,,,( FMTJTATETBP

Example:

)|()|(),|()()( TAFMPTATJPTETBTAPTEPTBP

Then its probability is:

Assume the following assignmentof values to random variables

FMTJTATETB ,,,,

M. HauskrechtCS 1571 Intro to AI

Bayesian belief networks (BBNs)

Bayesian belief networks

• Represent the full joint distribution over the variables more compactly using the product of local conditionals.

• But how did we get to local parameterizations?

Answer:

• Chain rule +

• Graphical structure encodes conditional and marginal independences among random variables

• A and B are independent

• A and B are conditionally independent given C

• The graph structure implies the decomposition !!!

)()(),( BPAPBAP

)|()|()|,( CBPCAPCBAP )|(),|( CAPBCAP

Page 6: Bayesian belief networks - University of Pittsburghmilos/courses/cs1571/Lectures/Class20.pdf• We want to represent the probability distribution of events: – Burglary, Earthquake,

6

M. HauskrechtCS 1571 Intro to AI

Independences in BBNs

3 basic independence structures:

Burglary

JohnCalls

Alarm

Burglary

Alarm

Earthquake

JohnCalls

Alarm

MaryCalls

1. 2. 3.

M. HauskrechtCS 1571 Intro to AI

Independences in BBNs

1. JohnCalls is independent of Burglary given Alarm

Burglary

JohnCalls

Alarm

Burglary

Alarm

Earthquake

JohnCalls

Alarm

MaryCalls

1. 2. 3.

)|(),|( AJPBAJP )|()|()|,( ABPAJPABJP

Page 7: Bayesian belief networks - University of Pittsburghmilos/courses/cs1571/Lectures/Class20.pdf• We want to represent the probability distribution of events: – Burglary, Earthquake,

7

M. HauskrechtCS 1571 Intro to AI

Independences in BBNs

2. Burglary is independent of Earthquake (not knowing Alarm) Burglary and Earthquake become dependent given Alarm !!

Burglary

JohnCalls

Alarm

JohnCalls

Alarm

MaryCalls

1. 3.

)()(),( EPBPEBP

Burglary

Alarm

Earthquake

2.

M. HauskrechtCS 1571 Intro to AI

Independences in BBNs

3. MaryCalls is independent of JohnCalls given Alarm

Burglary

JohnCalls

Alarm

Burglary

Alarm

Earthquake

1. 2.

JohnCalls

Alarm

3.

MaryCalls

)|(),|( AJPMAJP

)|()|()|,( AMPAJPAMJP

Page 8: Bayesian belief networks - University of Pittsburghmilos/courses/cs1571/Lectures/Class20.pdf• We want to represent the probability distribution of events: – Burglary, Earthquake,

8

M. HauskrechtCS 1571 Intro to AI

Independences in BBN

• BBN distribution models many conditional independence relations among distant variables and sets of variables

• These are defined in terms of the graphical criterion called d-separation

• D-separation and independence– Let X,Y and Z be three sets of nodes– If X and Y are d-separated by Z, then X and Y are

conditionally independent given Z• D-separation :

– A is d-separated from B given C if every undirected path between them is blocked with C

• Path blocking– 3 cases that expand on three basic independence structures

M. HauskrechtCS 1571 Intro to AI

Undirected path blockingA is d-separated from B given C if every undirected path

between them is blocked

A BC

Page 9: Bayesian belief networks - University of Pittsburghmilos/courses/cs1571/Lectures/Class20.pdf• We want to represent the probability distribution of events: – Burglary, Earthquake,

9

M. HauskrechtCS 1571 Intro to AI

Undirected path blockingA is d-separated from B given C if every undirected path

between them is blocked

A BC

M. HauskrechtCS 1571 Intro to AI

Undirected path blockingA is d-separated from B given C if every undirected path

between them is blocked

• 1. Path blocking with a linear substructure

Z in C

X Y

X in A Y in B

Z

A BC

Page 10: Bayesian belief networks - University of Pittsburghmilos/courses/cs1571/Lectures/Class20.pdf• We want to represent the probability distribution of events: – Burglary, Earthquake,

10

M. HauskrechtCS 1571 Intro to AI

Undirected path blocking

A is d-separated from B given C if every undirected path between them is blocked

• 2. Path blocking with the wedge substructure

Z in CX Y

X in A Y in B

Z

M. HauskrechtCS 1571 Intro to AI

Undirected path blocking

A is d-separated from B given C if every undirected path between them is blocked

• 3. Path blocking with the vee substructure

Z or any of its descendants not in C

X Y

X in A Y in B

Z

Page 11: Bayesian belief networks - University of Pittsburghmilos/courses/cs1571/Lectures/Class20.pdf• We want to represent the probability distribution of events: – Burglary, Earthquake,

11

M. HauskrechtCS 1571 Intro to AI

Independences in BBNs

• Earthquake and Burglary are independent given MaryCalls F

• Burglary and MaryCalls are independent (not knowing Alarm) F

• Burglary and RadioReport are independent given Earthquake T

• Burglary and RadioReport are independent given MaryCalls F

Burglary

JohnCalls

Alarm

Earthquake

MaryCalls

RadioReport

M. HauskrechtCS 1571 Intro to AI

Bayesian belief networks (BBNs)

Bayesian belief networks

• Represents the full joint distribution over the variables more compactly using the product of local conditionals.

• So how did we get to local parameterizations?

• The decomposition is implied by the set of independences encoded in the belief network.

))(|(),..,,(,..1

21

ni

iin XpaXXXX PP

Page 12: Bayesian belief networks - University of Pittsburghmilos/courses/cs1571/Lectures/Class20.pdf• We want to represent the probability distribution of events: – Burglary, Earthquake,

12

M. HauskrechtCS 1571 Intro to AI

Full joint distribution in BBNs

M

A

B

J

E

),,,,( FMTJTATETBP

)()(),|()|()|( TEPTBPTETBTAPTAFMPTATJP

),,,(),,,|( FMTATETBPFMTATETBTJP ),,,()|( FMTATETBPTATJP

),,(),,|( TATETBPTATETBFMP ),,()|( TATETBPTAFMP

),(),|( TETBPTETBTAP )()( TEPTBP

Rewrite the full joint probability using the product rule:

M. HauskrechtCS 1571 Intro to AI

# of parameters of the full joint:

Parameter complexity problem

• In the BBN the full joint distribution is defined as:

• What did we save?

Alarm example: 5 binary (True, False) variables

Burglary

JohnCalls

Alarm

Earthquake

MaryCalls

))(|(),..,,(,..1

21

ni

iin XpaXXXX PP

3225

3112 5 One parameter is for free:

Page 13: Bayesian belief networks - University of Pittsburghmilos/courses/cs1571/Lectures/Class20.pdf• We want to represent the probability distribution of events: – Burglary, Earthquake,

13

M. HauskrechtCS 1571 Intro to AI

# of parameters of the full joint:

Parameter complexity problem

• In the BBN the full joint distribution is defined as:

• What did we save?

Alarm example: 5 binary (True, False) variables

Burglary

JohnCalls

Alarm

Earthquake

MaryCalls

))(|(),..,,(,..1

21

ni

iin XpaXXXX PP

3225

3112 5 One parameter is for free:

# of parameters of the BBN: ?

M. HauskrechtCS 1571 Intro to AI

Bayesian belief network.

Burglary Earthquake

JohnCalls MaryCalls

Alarm

B E T F

T T 0.95 0.05T F 0.94 0.06F T 0.29 0.71F F 0.001 0.999

P(B)

0.001 0.999

P(E)

0.002 0.998

A T F

T 0.90 0.1F 0.05 0.95

A T F

T 0.7 0.3F 0.01 0.99

P(A|B,E)

P(J|A) P(M|A)

T F T F

• In the BBN the full joint distribution is expressed using a set of local conditional distributions

Page 14: Bayesian belief networks - University of Pittsburghmilos/courses/cs1571/Lectures/Class20.pdf• We want to represent the probability distribution of events: – Burglary, Earthquake,

14

M. HauskrechtCS 1571 Intro to AI

Bayesian belief network.

Burglary Earthquake

JohnCalls MaryCalls

Alarm

B E T F

T T 0.95 0.05T F 0.94 0.06F T 0.29 0.71F F 0.001 0.999

P(B)

0.001 0.999

P(E)

0.002 0.998

A T F

T 0.90 0.1F 0.05 0.95

A T F

T 0.7 0.3F 0.01 0.99

P(A|B,E)

P(J|A) P(M|A)

T F T F

• In the BBN the full joint distribution is expressed using a set of local conditional distributions

2 2

8

4 4

M. HauskrechtCS 1571 Intro to AI

# of parameters of the full joint:

Parameter complexity problem

• In the BBN the full joint distribution is defined as:

• What did we save?

Alarm example: 5 binary (True, False) variables

Burglary

JohnCalls

Alarm

Earthquake

MaryCalls

))(|(),..,,(,..1

21

ni

iin XpaXXXX PP

3225

3112 5 One parameter is for free:

# of parameters of the BBN:

20)2(2)2(22 23

10)1(2)2(22 2 One parameter in every conditional is for free:

Page 15: Bayesian belief networks - University of Pittsburghmilos/courses/cs1571/Lectures/Class20.pdf• We want to represent the probability distribution of events: – Burglary, Earthquake,

15

M. HauskrechtCS 1571 Intro to AI

Model acquisition problem

The structure of the BBN• typically reflects causal relations

(BBNs are also sometime referred to as causal networks)• Causal structure is intuitive in many applications domain and it

is relatively easy to define to the domain expert

Probability parameters of BBN• are conditional distributions relating random variables and

their parents • Complexity is much smaller than the full joint• It is much easier to obtain such probabilities from the expert or

learn them automatically from data

M. HauskrechtCS 1571 Intro to AI

BBNs built in practice

• In various areas:– Intelligent user interfaces (Microsoft)– Troubleshooting, diagnosis of a technical device– Medical diagnosis:

• Pathfinder (Intellipath)• CPSC• Munin• QMR-DT

– Collaborative filtering– Military applications– Business and finance

• Insurance, credit applications

Page 16: Bayesian belief networks - University of Pittsburghmilos/courses/cs1571/Lectures/Class20.pdf• We want to represent the probability distribution of events: – Burglary, Earthquake,

16

M. HauskrechtCS 1571 Intro to AI

Diagnosis of car engine

• Diagnose the engine start problem

M. HauskrechtCS 1571 Intro to AI

Car insurance example

• Predict claim costs (medical, liability) based on application data

Page 17: Bayesian belief networks - University of Pittsburghmilos/courses/cs1571/Lectures/Class20.pdf• We want to represent the probability distribution of events: – Burglary, Earthquake,

17

M. HauskrechtCS 1571 Intro to AI

(ICU) Alarm network

M. HauskrechtCS 1571 Intro to AI

CPCS• Computer-based Patient Case Simulation system (CPCS-PM)

developed by Parker and Miller (University of Pittsburgh)

• 422 nodes and 867 arcs

Page 18: Bayesian belief networks - University of Pittsburghmilos/courses/cs1571/Lectures/Class20.pdf• We want to represent the probability distribution of events: – Burglary, Earthquake,

18

M. HauskrechtCS 1571 Intro to AI

QMR-DT

• Medical diagnosis in internal medicine

• (U Pittsburgh)

Bipartite network of disease/findings relations

Based on QMR system built at U Pittsburgh

M. HauskrechtCS 1571 Intro to AI

Inference in Bayesian network

• Bad news: – Exact inference problem in BBNs is NP-hard (Cooper)– Approximate inference is NP-hard (Dagum, Luby)

• But very often we can achieve significant improvements• Assume our Alarm network

• Assume we want to compute:

Burglary

JohnCalls

Alarm

Earthquake

MaryCalls

)( TJP

Page 19: Bayesian belief networks - University of Pittsburghmilos/courses/cs1571/Lectures/Class20.pdf• We want to represent the probability distribution of events: – Burglary, Earthquake,

19

M. HauskrechtCS 1571 Intro to AI

Inference in Bayesian networksComputing:Approach 1. Blind approach.• Sum out all un-instantiated variables from the full joint, • express the joint distribution as a product of conditionals

Computational cost:Number of additions: 15Number of products: 16*4=64

)( TJP

)()(),|()|()|(, , , ,

eEPbBPeEbBaAPaAmMPaATJPFTb FTe FTa FTm

),,,,(, , , ,

mMTJaAeEbBPFTb FTe FTa FTm

)( TJP

M. HauskrechtCS 1571 Intro to AI

Inference in Bayesian networks

Approach 2. Interleave sums and products

• Combines sums and product in a smart way (multiplications by constants can be taken out of the sum)

Computational cost: ?

)( TJP

)(),|()()|()|(,, . ,

eEPeEbBaAPbBPaAmMPaATJPFTeFTb FTa FTm

FTb FTeFTmFTa

eEPeEbBaAPbBPaAmMPaATJP, ,,,

)(),|()()|()|(

)()(),|()|()|(, , , ,

eEPbBPeEbBaAPaAmMPaATJPFTb FTe FTa FTm

}{}{

)()(xx

xfaxaf

Page 20: Bayesian belief networks - University of Pittsburghmilos/courses/cs1571/Lectures/Class20.pdf• We want to represent the probability distribution of events: – Burglary, Earthquake,

20

M. HauskrechtCS 1571 Intro to AI

Inference in Bayesian networks

Approach 2. Interleave sums and products

• Combines sums and product in a smart way (multiplications by constants can be taken out of the sum)

Computational cost:

Number of additions: ?

FTb FTeFTmFTa

eEPeEbBaAPbBPaAmMPaATJP, ,,,

)(),|()()|()|(

1

M. HauskrechtCS 1571 Intro to AI

Inference in Bayesian networks

Approach 2. Interleave sums and products

• Combines sums and product in a smart way (multiplications by constants can be taken out of the sum)

Computational cost:

Number of additions: ?

FTb FTeFTmFTa

eEPeEbBaAPbBPaAmMPaATJP, ,,,

)(),|()()|()|(

2*1

Page 21: Bayesian belief networks - University of Pittsburghmilos/courses/cs1571/Lectures/Class20.pdf• We want to represent the probability distribution of events: – Burglary, Earthquake,

21

M. HauskrechtCS 1571 Intro to AI

Inference in Bayesian networks

Approach 2. Interleave sums and products

• Combines sums and product in a smart way (multiplications by constants can be taken out of the sum)

Computational cost:

Number of additions: ?

FTb FTeFTmFTa

eEPeEbBaAPbBPaAmMPaATJP, ,,,

)(),|()()|()|(

2*2*1

M. HauskrechtCS 1571 Intro to AI

Inference in Bayesian networks

Approach 2. Interleave sums and products

• Combines sums and product in a smart way (multiplications by constants can be taken out of the sum)

Computational cost:

Number of additions: ?

FTb FTeFTmFTa

eEPeEbBaAPbBPaAmMPaATJP, ,,,

)(),|()()|()|(

2*2*12*1

Page 22: Bayesian belief networks - University of Pittsburghmilos/courses/cs1571/Lectures/Class20.pdf• We want to represent the probability distribution of events: – Burglary, Earthquake,

22

M. HauskrechtCS 1571 Intro to AI

Inference in Bayesian networks

Approach 2. Interleave sums and products

• Combines sums and product in a smart way (multiplications by constants can be taken out of the sum)

Computational cost:

Number of additions: ?

FTb FTeFTmFTa

eEPeEbBaAPbBPaAmMPaATJP, ,,,

)(),|()()|()|(

2*2*12*12*11

M. HauskrechtCS 1571 Intro to AI

Inference in Bayesian networks

Approach 2. Interleave sums and products

• Combines sums and product in a smart way (multiplications by constants can be taken out of the sum)

Computational cost:

Number of additions: 1+2*[1+1+2*1]=9

FTb FTeFTmFTa

eEPeEbBaAPbBPaAmMPaATJP, ,,,

)(),|()()|()|(

2*2*12*12*11

Page 23: Bayesian belief networks - University of Pittsburghmilos/courses/cs1571/Lectures/Class20.pdf• We want to represent the probability distribution of events: – Burglary, Earthquake,

23

M. HauskrechtCS 1571 Intro to AI

Inference in Bayesian networks

Approach 2. Interleave sums and products

• Combines sums and product in a smart way (multiplications by constants can be taken out of the sum)

Computational cost:

Number of products: ?

FTb FTeFTmFTa

eEPeEbBaAPbBPaAmMPaATJP, ,,,

)(),|()()|()|(

1

M. HauskrechtCS 1571 Intro to AI

Inference in Bayesian networks

Approach 2. Interleave sums and products

• Combines sums and product in a smart way (multiplications by constants can be taken out of the sum)

Computational cost:

Number of products: ?

FTb FTeFTmFTa

eEPeEbBaAPbBPaAmMPaATJP, ,,,

)(),|()()|()|(

2*2 *2*1

Page 24: Bayesian belief networks - University of Pittsburghmilos/courses/cs1571/Lectures/Class20.pdf• We want to represent the probability distribution of events: – Burglary, Earthquake,

24

M. HauskrechtCS 1571 Intro to AI

Inference in Bayesian networks

Approach 2. Interleave sums and products

• Combines sums and product in a smart way (multiplications by constants can be taken out of the sum)

Computational cost:

Number of products: 2*[2+2*(1+2*1)]=16

FTb FTeFTmFTa

eEPeEbBaAPbBPaAmMPaATJP, ,,,

)(),|()()|()|(

2*2 *2*12*2*12*2

M. HauskrechtCS 1571 Intro to AI

Inference in Bayesian networks

Approach 2. Interleave sums and products

• Combines sums and product in a smart way (multiplications by constants can be taken out of the sum)

Computational cost:

Number of additions: 1+2*[1+1+2*1]=9

Number of products: 2*[2+2*(1+2*1)]=16

)( TJP

)(),|()()|()|(,, . ,

eEPeEbBaAPbBPaAmMPaATJPFTeFTb FTa FTm

FTb FTeFTmFTa

eEPeEbBaAPbBPaAmMPaATJP, ,,,

)(),|()()|()|(

)()(),|()|()|(, , , ,

eEPbBPeEbBaAPaAmMPaATJPFTb FTe FTa FTm

Page 25: Bayesian belief networks - University of Pittsburghmilos/courses/cs1571/Lectures/Class20.pdf• We want to represent the probability distribution of events: – Burglary, Earthquake,

25

M. HauskrechtCS 1571 Intro to AI

Variable elimination

• Variable elimination:

– Similar idea but interleave sum and products one variable at the time during inference

– E.g. Query requires to eliminate A,B,E,M and this can be done in different order

)( TJP

)()(),|()|()|(, , , ,

eEPbBPeEbBaAPaAmMPaATJPFTb FTe FTa FTm

)( TJP

M. HauskrechtCS 1571 Intro to AI

Variable elimination

Assume order: M, E, B,A to calculate

)()(),|()|()|(, , , ,

eEPbBPeEbBaAPaAmMPaATJPFTb FTe FTa FTm

)( TJP

FTmFTb FTe FTa

aAmMPeEPbBPeEbBaAPaATJP,, , ,

)|()()(),|()|(

1)()(),|()|(, , ,

FTb FTe FTa

eEPbBPeEbBaAPaATJP

FTb FTeFTa

eEPeEbBaAPbBPaATJP, ,,

)(),|()()|(

),()()|(,

1,

bBaAbBPaATJPFTbFTa

FTa FTb

bBaAbBPaATJP,

1,

),()()|(

FTa

aAaATJP,

2 )()|( )( TJP

Page 26: Bayesian belief networks - University of Pittsburghmilos/courses/cs1571/Lectures/Class20.pdf• We want to represent the probability distribution of events: – Burglary, Earthquake,

26

M. HauskrechtCS 1571 Intro to AI

Inference in Bayesian network

• Exact inference algorithms:– Variable elimination– Recursive decomposition (Cooper, Darwiche)– Symbolic inference (D’Ambrosio)– Belief propagation algorithm (Pearl)– Clustering and joint tree approach (Lauritzen,

Spiegelhalter)– Arc reversal (Olmsted, Schachter)

• Approximate inference algorithms:– Monte Carlo methods:

• Forward sampling, Likelihood sampling– Variational methods

Book

Book

Book