Latent Variable Models

Christopher M. Bishop

Research output: Chapter in Book/Report/Conference proceedingChapter

Abstract

A powerful approach to probabilistic modelling involves supplementing a set of observed variables with additional latent, or hidden, variables. By defining a joint distribution over visible and latent variables, the corresponding distribution of the observed variables is then obtained by marginalization. This allows relatively complex distributions to be expressed in terms of more tractable joint distributions over the expanded variable space. One well-known example of a hidden variable model is the mixture distribution in which the hidden variable is the discrete component label. In the case of continuous latent variables we obtain models such as factor analysis. The structure of such probabilistic models can be made particularly transparent by giving them a graphical representation, usually in terms of a directed acyclic graph, or Bayesian network. In this chapter we provide an overview of latent variable models for representing continuous variables. We show how a particular form of linear latent variable model can be used to provide a probabilistic formulation of the well-known technique of principal components analysis (PCA). By extending this technique to mixtures, and hierarchical mixtures, of probabilistic PCA models we are led to a powerful interactive algorithm for data visualization. We also show how the probabilistic PCA approach can be generalized to non-linear latent variable models leading to the Generative Topographic Mapping algorithm (GTM). Finally, we show how GTM can itself be extended to model temporal data.

Original languageEnglish
Title of host publicationLearning in Graphical Models
EditorsMichael I. Jordan
PublisherMIT Press
Pages371–403
Number of pages33
ISBN (Print)9780262600323
Publication statusPublished - 1 Jan 1999

Publication series

NameAdaptive Computation and Machine Learning
PublisherMIT Press

Fingerprint

Dive into the research topics of 'Latent Variable Models'. Together they form a unique fingerprint.

Cite this