2018 10-708 (CMU) Probabilistic Graphical Models {Lecture 5} [Algorithms for Exact Inference]

 Not in the frontier of research, but the results are used commonly now.

X_{k+1} - X_n are known,

to calculate the joint probability, we have to do inference

 

 Recent research: on the approximate inference teches

approx:

1) optimization-based

2) sampling-based

 

Compare the computational complexity:

    NAIVE way: K^n

    Chain rule:  n*K^2

    n=4

 

Chain rule derivation:

 

 

 

 

 

 

 

marginalizing out the rest 

 P(a) P(b) P(c|b) ...... P(h|e,f) => a,b,c,d,e,f,g,h (elimation sequence)

 

introduce a term m_h(e,f) to make e and f dependent

not introducing any dependency here

Different elimination sequence will lead to different computational complexity.

It's dependent on how large the new clique is

In one step, if you connect every vertex, then you are in trouble.

 

  

if there's a loop in graph, you can't view message passing as a variable to another variable, but clique to clique.

猜你喜欢

转载自www.cnblogs.com/ecoflex/p/10231273.html