| Part of a series on |
| Machine learning anddata mining |
|---|
Learning with humans |
Model diagnostics |
This article includes a list ofgeneral references, butit lacks sufficient correspondinginline citations. Please help toimprove this article byintroducing more precise citations.(May 2017) (Learn how and when to remove this message) |
Agraphical model orprobabilistic graphical model (PGM) orstructured probabilistic model is aprobabilistic model for which agraph expresses theconditional dependence structure betweenrandom variables. Graphical models are commonly used inprobability theory,statistics—particularlyBayesian statistics—andmachine learning.
Generally, probabilistic graphical models use a graph-based representation as the foundation for encoding a distribution over a multi-dimensional space and a graph that is a compact orfactorized representation of a set of independences that hold in the specific distribution. Two branches of graphical representations of distributions are commonly used, namely,Bayesian networks andMarkov random fields. Both families encompass the properties of factorization and independences, but they differ in the set of independences they can encode and the factorization of the distribution that they induce.[1]

The undirected graph shown may have one of several interpretations; the common feature is that the presence of an edge implies some sort of dependence between the corresponding random variables. From this graph, we might deduce that B, C, and D are allconditionally independent given A. This means that if the value of A is known, then the values of B, C, and D provide no further information about each other. Equivalently (in this case), the joint probability distribution can be factorized as:
for some non-negative functions.

If the network structure of the model is adirected acyclic graph, the model represents a factorization of the jointprobability of all random variables. More precisely, if the events are then the joint probability satisfies
where is the set of parents of node (nodes with edges directed towards). In other words, thejoint distribution factors into a product of conditional distributions. For example, in the directed acyclic graph shown in the Figure this factorization would be
Any two nodes areconditionally independent given the values of their parents. In general, any two sets of nodes are conditionally independent given a third set if a criterion calledd-separation holds in the graph. Local independences and global independences are equivalent in Bayesian networks.
This type of graphical model is known as a directed graphical model,Bayesian network, or belief network. Classic machine learning models likehidden Markov models,neural networks and newer models such asvariable-order Markov models can be considered special cases of Bayesian networks.
One of the simplest Bayesian Networks is theNaive Bayes classifier.

The next figure depicts a graphical model with a cycle. This may be interpreted in terms of each variable 'depending' on the values of its parents in some manner. The particular graph shown suggests a joint probability density that factors as
but other interpretations are possible.[2]


The framework of the models, which provides algorithms for discovering and analyzing structure in complex distributions to describe them succinctly and extract the unstructured information, allows them to be constructed and utilized effectively.[1] Applications of graphical models includecausal inference,information extraction,speech recognition,computer vision, decoding oflow-density parity-check codes, modeling ofgene regulatory networks, gene finding and diagnosis of diseases, andgraphical models for protein structure.