Excerpt from all over the place, only if their own study notes, Wuguai, invasion deleted!
In more detail, probabilistic graphical model includes a naive Bayes model, maximum entropy, hidden Markov model, CRFs, themes models, machine learning many scenes have a wide range of applications.
FIG probability model, the data (sample) by the equation represented Modeling:
- Indicate nodes, i.e., the random variable (in here, may be a token or a label), in particular, with a random variable modeling, note now represent the number of random variables (corresponding to an imaginary Sequence, it contains many the token), the distribution of the random variable;
- Represent the edges, that is, the probability of dependence. Specifically ye understood or interpreted to the specific binding of CRF graph HMM or later.
Probabilistic graphical model can be divided into two types: directed and non-directed graph.
There vs undirected graph directed graph
The figure can be seen, Bayesian networks (belief networks) are directed and undirected Markov network. Therefore, there is a Bayesian network for modeling the one-way data dependent, Markov network between the entities for interdependent model. In particular, the performance of their core differences in how to find , that represents how the joint probability.
Directed graph
For a directed graph model, so seek joint probability:
For example, for the following random variables that have to graph (note that this figure I drew is quite broad):
This should represent their joint probability:
It should be well understood.
Undirected graph
For undirected graphs, I think in general it refers to the information Markov network (note that this figure I drew is relatively broad).
If a graph is too large, it can be broken down by a factor of the product written into a number of joint probabilities. Ye decomposing it, is divided into a plurality of FIG. "Small group", note that each group must be the "maximum group" (which is attached to any two points of a specific ...... Well no explanation is the largest connected sub- Figure), there are:
Which formula should not be difficult to understand it, is to allow the normalization of results counted probability.
Therefore, like the above undirected graph:
Which is one of the largest group of joint probability random variables are on, and generally exponential function:
Well, this thing called the tube 势函数
. Note that if there are to see the shadow of CRF.
The probability that the joint probability distribution may be represented in FIG factorization is:
Note that the understanding here is quite important to note recursive process, knock on the blackboard, which is CRF beginning!