Generated by GPT-5-mini| Graphical models | |
|---|---|
| Name | Graphical models |
| Field | Statistics, Computer Science |
| Introduced | 1980s |
| Notable | Judea Pearl, Michael I. Jordan, David Spiegelhalter |
Graphical models are probabilistic models that use graph structures to represent dependencies among random variables, combining graph theory with probability theory. They provide a compact, visual way to express conditional independence and factorization properties useful in machine learning, statistics, and artificial intelligence. Influential contributors and institutions accelerated development, and graphical models underpin many modern algorithms in perception, bioinformatics, and natural language processing.
Graphical models originated from work by researchers like Judea Pearl, David Spiegelhalter, Michael I. Jordan, Geoffrey Hinton, and Radford M. Neal and were developed within groups at institutions such as University of California, Los Angeles, University of California, Berkeley, Cambridge University, Massachusetts Institute of Technology, and Stanford University. Early milestones include the formalization of Bayesian networks and Markov random fields, with foundational texts and conferences at venues like NeurIPS, ICML, UAI (conference), and AAAI Conference on Artificial Intelligence. The formalism influenced applied work in projects at Bell Labs, IBM Research, Microsoft Research, and Google Research. Key awards and recognitions include the Turing Award winners who advanced related fields and honors such as the Royal Society fellowships granted to contributors.
Graphical models split into directed and undirected forms exemplified by frameworks attributed to researchers at Causality Institute-style centers and labs. Directed models include Bayesian networks popularized by Judea Pearl and applied in systems at NHS (England) clinical decision support pilots and industrial settings like Siemens diagnostics. Undirected models include Markov random fields used in computer vision projects at MIT CSAIL and in statistical physics studies connected to work at Princeton University and Los Alamos National Laboratory. Other variants include conditional random fields used in natural language processing work at Google DeepMind and Facebook AI Research, factor graphs used in coding theory research at Bell Labs and Caltech, and chain graphs developed in theoretical studies at ETH Zurich and University of Toronto.
A graphical model represents variables as vertices with edges encoding dependency relations, formalized in texts produced by authors at Oxford University Press and Cambridge University Press and taught in courses at Harvard University and Yale University. Semantics relate to d-separation in directed acyclic graphs discussed by Judea Pearl and Markov properties introduced in probabilistic graphical literature from Columbia University and University of Chicago. Factorization theorems trace to probabilists connected with Princeton University and Columbia University. Graphical model representations are used in software developed by teams at Stanford NLP Group, Scikit-learn contributors, and libraries from TensorFlow and PyTorch communities.
Inference tasks—marginalization, conditioning, MAP estimation—use algorithms pioneered by researchers affiliated with Carnegie Mellon University and University of California, Berkeley. Exact inference includes variable elimination and junction tree algorithms studied in seminars at Institute for Advanced Study and labs at Bell Labs. Approximate inference includes belief propagation from work at HP Labs and loopy belief propagation used in projects at Google Research, variational methods developed in groups at Columbia University and University of Cambridge, and Monte Carlo methods such as Markov chain Monte Carlo advanced at Los Alamos National Laboratory and University of Washington. Optimization-based inference includes expectation-maximization methods applied in bioinformatics projects at Broad Institute and convex relaxations researched at Microsoft Research.
Learning involves structure learning and parameter estimation; seminal algorithms are associated with research groups at Carnegie Mellon University, University College London, and ETH Zurich. Maximum likelihood estimation and Bayesian parameter learning are applied in clinical studies at Johns Hopkins University and genomics projects at Wellcome Sanger Institute. Structure discovery techniques include score-based search used in workshops at UAI (conference) and constraint-based methods traced to collaborations between Stanford University and UC Berkeley. Regularization and sparsity priors originate from statisticians connected with Princeton University and machine learning teams at Google Brain.
Graphical models power applications in computer vision research at MIT Media Lab and ETH Zurich, speech recognition systems developed by IBM Research and Microsoft Research, and natural language processing projects from Stanford NLP Group and Google DeepMind. In bioinformatics, graphical approaches are used by teams at European Bioinformatics Institute and Broad Institute for gene regulatory network inference. They support robotics work at Carnegie Mellon University and Toyota Research Institute and financial risk modeling in groups at Goldman Sachs and Morgan Stanley. Examples include diagnostic expert systems deployed in trials by World Health Organization partners and recommender systems prototyped at Netflix.
Limitations include computational intractability highlighted in theoretical analyses from Institute for Advanced Study and identifiability issues examined by researchers at Columbia University and University of California, Berkeley. Extensions address causality through frameworks advanced by Judea Pearl and collaborators at University of California, Los Angeles and incorporate deep learning through hybrid models researched at DeepMind, Facebook AI Research, and Google Brain. Recent trends link graphical models with probabilistic programming initiatives from Stan Development Team and Pyro (software), and interdisciplinary projects at Wellcome Trust and European Research Council centers continue to expand applicability.
Category:Probabilistic models