2018 10-708 (CMU) Probabilistic Graphical Models {Lecture 5} [Algorithms for Exact Inference]
Posted ecoflex
tags:
篇首语:本文由小常识网(cha138.com)小编为大家整理,主要介绍了2018 10-708 (CMU) Probabilistic Graphical Models {Lecture 5} [Algorithms for Exact Inference]相关的知识,希望对你有一定的参考价值。
Not in the frontier of research, but the results are used commonly now.
X_{k+1} - X_n are known,
to calculate the joint probability, we have to do inference
Recent research: on the approximate inference teches
approx:
1) optimization-based
2) sampling-based
Compare the computational complexity:
NAIVE way: K^n
Chain rule: n*K^2
n=4
Chain rule derivation:
marginalizing out the rest
P(a) P(b) P(c|b) ...... P(h|e,f) => a,b,c,d,e,f,g,h (elimation sequence)
introduce a term m_h(e,f) to make e and f dependent
not introducing any dependency here
Different elimination sequence will lead to different computational complexity.
It‘s dependent on how large the new clique is
In one step, if you connect every vertex, then you are in trouble.
if there‘s a loop in graph, you can‘t view message passing as a variable to another variable, but clique to clique.
以上是关于2018 10-708 (CMU) Probabilistic Graphical Models {Lecture 5} [Algorithms for Exact Inference]的主要内容,如果未能解决你的问题,请参考以下文章
2018 10-708 (CMU) Probabilistic Graphical Models {Lecture 5} [Algorithms for Exact Inference]
2018 10-708 (CMU) Probabilistic Graphical Models {Lecture 15} [Mean field Approximation]
2018 10-708 (CMU) Probabilistic Graphical Models {Lecture 21} [A Hybrid: Deep Learning and Graphical
Probabilistic Graphical Models 10-708, Spring 2017