Graphical model

{{short description|Probabilistic model}}

{{about|the representation of probability distributions using graphs|the computer graphics journal|Graphical Models{{!}}Graphical Models}}

{{Machine learning|Structured prediction}}

{{More footnotes|date=May 2017}}

A graphical model or probabilistic graphical model (PGM) or structured probabilistic model is a probabilistic model for which a graph expresses the conditional dependence structure between random variables. Graphical models are commonly used in probability theory, statistics—particularly Bayesian statistics—and machine learning.

Types of graphical models

Generally, probabilistic graphical models use a graph-based representation as the foundation for encoding a distribution over a multi-dimensional space and a graph that is a compact or factorized representation of a set of independences that hold in the specific distribution. Two branches of graphical representations of distributions are commonly used, namely, Bayesian networks and Markov random fields. Both families encompass the properties of factorization and independences, but they differ in the set of independences they can encode and the factorization of the distribution that they induce.{{cite book

|author=Koller, D.

|author2=Friedman, N.

|title=Probabilistic Graphical Models

|url=http://pgm.stanford.edu/

|publisher=MIT Press

|location=Massachusetts

|year=2009

|pages=1208

|isbn=978-0-262-01319-2

|author2-link=Nir Friedman

|author-link=Daphne Koller

|archive-url=https://web.archive.org/web/20140427083249/http://pgm.stanford.edu/

|archive-date=2014-04-27

|url-status=dead

}}

=Undirected Graphical Model=

File:Examples of an Undirected Graph.svg

The undirected graph shown may have one of several interpretations; the common feature is that the presence of an edge implies some sort of dependence between the corresponding random variables. From this graph, we might deduce that B, C, and D are all conditionally independent given A. This means that if the value of A is known, then the values of B, C, and D provide no further information about each other. Equivalently (in this case), the joint probability distribution can be factorized as:

:P[A,B,C,D] = f_{AB}[A,B] \cdot f_{AC}[A,C] \cdot f_{AD}[A,D]

for some non-negative functions f_{AB}, f_{AC}, f_{AD}.

=Bayesian network=

{{main|Bayesian network}}

File:Example of a Directed Graph.svg

If the network structure of the model is a directed acyclic graph, the model represents a factorization of the joint probability of all random variables. More precisely, if the events are X_1,\ldots,X_n then the joint probability satisfies

:P[X_1,\ldots,X_n]=\prod_{i=1}^nP[X_i|\text{pa}(X_i)]

where \text{pa}(X_i) is the set of parents of node X_i (nodes with edges directed towards X_i). In other words, the joint distribution factors into a product of conditional distributions. For example, in the directed acyclic graph shown in the Figure this factorization would be

:P[A,B,C,D] = P[A]\cdot P[B | A]\cdot P[C | A] \cdot P[D|A,C].

Any two nodes are conditionally independent given the values of their parents. In general, any two sets of nodes are conditionally independent given a third set if a criterion called d-separation holds in the graph. Local independences and global independences are equivalent in Bayesian networks.

This type of graphical model is known as a directed graphical model, Bayesian network, or belief network. Classic machine learning models like hidden Markov models, neural networks and newer models such as variable-order Markov models can be considered special cases of Bayesian networks.

One of the simplest Bayesian Networks is the Naive Bayes classifier.

=Cyclic Directed Graphical Models=

File:Graph model.svg

The next figure depicts a graphical model with a cycle. This may be interpreted in terms of each variable 'depending' on the values of its parents in some manner.

The particular graph shown suggests a joint probability density that factors as

:P[A,B,C,D] = P[A]\cdot P[B]\cdot P[C,D|A,B],

but other interpretations are possible.

{{cite book

|first=Thomas |last=Richardson

|chapter=A discovery algorithm for directed cyclic graphs

|title=Proceedings of the Twelfth Conference on Uncertainty in Artificial Intelligence

|year=1996

|publisher=Morgan Kaufmann Pub.

|isbn=978-1-55860-412-4

}}

=Other types=

File:Tan corral.png

  • Targeted Bayesian network learning (TBNL) File:Tbnl corral.jpg
  • A factor graph is an undirected bipartite graph connecting variables and factors. Each factor represents a function over the variables it is connected to. This is a helpful representation for understanding and implementing belief propagation.
  • A clique tree or junction tree is a tree of cliques, used in the junction tree algorithm.
  • A chain graph is a graph which may have both directed and undirected edges, but without any directed cycles (i.e. if we start at any vertex and move along the graph respecting the directions of any arrows, we cannot return to the vertex we started from if we have passed an arrow). Both directed acyclic graphs and undirected graphs are special cases of chain graphs, which can therefore provide a way of unifying and generalizing Bayesian and Markov networks.{{cite journal|last=Frydenberg|first=Morten|year=1990|title=The Chain Graph Markov Property|journal=Scandinavian Journal of Statistics|volume=17|issue=4|pages=333–353|mr=1096723|jstor=4616181 }}

  • An ancestral graph is a further extension, having directed, bidirected and undirected edges.{{cite journal

|first1=Thomas |last1=Richardson |first2=Peter |last2=Spirtes

|title=Ancestral graph Markov models

|journal=Annals of Statistics

|volume=30 |issue=4 |year=2002 |pages=962–1030

|doi=10.1214/aos/1031689015

|mr=1926166 | zbl = 1033.60008

|citeseerx=10.1.1.33.4906}}

Applications

The framework of the models, which provides algorithms for discovering and analyzing structure in complex distributions to describe them succinctly and extract the unstructured information, allows them to be constructed and utilized effectively. Applications of graphical models include causal inference, information extraction, speech recognition, computer vision, decoding of low-density parity-check codes, modeling of gene regulatory networks, gene finding and diagnosis of diseases, and graphical models for protein structure.

See also

Notes

{{reflist}}

Further reading

=Books and book chapters=

  • {{cite book

| last = Barber

| first = David

| title = Bayesian Reasoning and Machine Learning

| publisher = Cambridge University Press

| year = 2012

| isbn = 978-0-521-51814-7

}}

  • {{cite book

| last = Bishop

| first = Christopher M.

| author-link = Christopher Bishop

| title = Pattern Recognition and Machine Learning

| publisher = Springer

| year = 2006

| url = https://www.springer.com/us/book/9780387310732

| isbn=978-0-387-31073-2

| chapter= Chapter 8. Graphical Models

| chapter-url=https://www.microsoft.com/en-us/research/wp-content/uploads/2016/05/Bishop-PRML-sample.pdf

| pages=359–422

| mr=2247587

}}

  • {{cite book

|author=Cowell, Robert G.

|author2=Dawid, A. Philip |author3=Lauritzen, Steffen L. |author4-link=David Spiegelhalter |author4=Spiegelhalter, David J.

|title=Probabilistic networks and expert systems |publisher=Springer |location=Berlin |year=1999 |isbn=978-0-387-98767-5 |mr=1697175 |ref=cowell |author2-link=Philip Dawid }} A more advanced and statistically oriented book

  • {{cite book |author=Jensen, Finn |title=An introduction to Bayesian networks |publisher=Springer |location=Berlin |year=1996 |isbn=978-0-387-91502-9 }}
  • {{Cite book

|first=Judea |last=Pearl |author-link = Judea Pearl

| year = 1988

| title = Probabilistic Reasoning in Intelligent Systems

|url=https://archive.org/details/probabilisticrea00pear |url-access=registration | edition = 2nd revised

| location = San Mateo, CA

| publisher = Morgan Kaufmann

| mr = 0965765

|isbn = 978-1-55860-479-7

}} A computational reasoning approach, where the relationships between graphs and probabilities were formally introduced.

=Journal articles=

  • {{Cite journal

| author = Edoardo M. Airoldi |author-link1=Edoardo Airoldi

| title = Getting Started in Probabilistic Graphical Models

| journal = PLOS Computational Biology

| volume = 3

| issue = 12

| pages = e252

| year = 2007

| doi = 10.1371/journal.pcbi.0030252

| pmid = 18069887

| pmc = 2134967

|arxiv=0706.2040

|bibcode=2007PLSCB...3..252A

|doi-access=free

}}

  • {{Cite journal | last1 = Jordan | first1 = M. I. | author-link=Michael I. Jordan| doi = 10.1214/088342304000000026 | title = Graphical Models | journal = Statistical Science | volume = 19 | pages = 140–155| year = 2004 | doi-access = free }}
  • {{Cite journal|last=Ghahramani|first=Zoubin|date=May 2015|title=Probabilistic machine learning and artificial intelligence|journal=Nature|language=en|volume=521|issue=7553|pages= 452–459|doi=10.1038/nature14541|pmid=26017444|bibcode=2015Natur.521..452G|s2cid=216356|url=https://www.repository.cam.ac.uk/handle/1810/248538|doi-access=free}}

=Other=

  • [http://research.microsoft.com/en-us/um/people/heckerman/tutorial.pdf Heckerman's Bayes Net Learning Tutorial]
  • [http://www.cs.ubc.ca/~murphyk/Bayes/bnintro.html A Brief Introduction to Graphical Models and Bayesian Networks]
  • [http://www.cedar.buffalo.edu/~srihari/CSE574 Sargur Srihari's lecture slides on probabilistic graphical models]