Variational inference engine for probabilistic graphical models

Data processing: structural design – modeling – simulation – and em – Modeling by mathematical expression

Reexamination Certificate

Rate now

  [ 0.00 ] – not rated yet Voters 0   Comments 0

Details

C703S022000, C700S028000, C706S022000

Reexamination Certificate

active

06556960

ABSTRACT:

FIELD OF THE INVENTION
This invention relates generally to data modeling and analysis such as probabilistic graphical models, and more particularly to variational inference engines for such models.
BACKGROUND OF THE INVENTION
Data modeling has become an important tool in solving complex and large real-world computerizable problems. Applications of data modeling include data compression, density estimation and data visualization. A data modeling technique used for these and other applications is probabilistic modeling. It has proven to be a popular technique for data modeling applications such as speech recognition, vision, handwriting recognition, information retrieval and intelligent interfaces. One framework for developing such applications involves the representation of probability distributions as directed acyclic graphs, which are also known as Bayesian networks, belief networks, and probabilistic independence networks, among other terms.
In a probabilistic model, there are a number of variables. Each variable is represented as a node in a directed acyclic graph. An example of such a graph is shown in the diagram of FIG.
2
. The graph
200
includes nodes
202
,
204
,
206
,
208
and
210
, labeled as X
1
, X
2
, X
3
, X
4
and X
5
, respectively. Each node corresponds to a variable, which may correspond to either observed or observable data, or unobserved or unobservable data. For example, if the graph
200
were to correspond to a model for factors involved in starting an automobile, the observed variables might include the starting or otherwise of the engine, while the unobserved variables could include the presence of absence of fuel in the tank and the state of the battery.
The joint distribution over the variables is expressed as the product of a conditional distribution for each node, conditioned on the states of its parents in the graph,
P

(
X
1
,



,
X
M
)
=

i
=
1
M



P

(
X
i

pa
i
&AutoLeftMatch;
)
where pa
i
denotes the parents of X
i
. A specific model is determined by the structure of graph, as well as the choice of the conditional distributions P(X
i
|pa
i
). For example, given the graph
200
of
FIG. 2
, the factorization is
P
(
X
1
, X
2
, X
3
, X
4
, X
5
)=
P
(
X
1
)
P
(
X
2
)
P
(
X
3
|X
1
, X
2
)
P
(X
4
|X
2
)
P
(
X
5
|X
3
, X
4
).
Because some of the variables are unobservable, to effectively use the model represented by the graph, it is necessary to infer the corresponding posterior distribution of at least a subset of the unobservable variables. After this is accomplished, the posterior distribution can then be used to make predictions based on the model. However, exact solutions of probabilistic models are generally intractable for all but the simplest examples. Therefore, approximation schemes are used to approximate the posterior distributions. Such approximation schemes generally fall into one of three classes: (1) Laplace's method and similar semi-analytic approximations; (2) Markov chain Monte Carlo methods, such as Gibbs sampling; and, (3) variational methods.
The last of these approximation schemes, variational methods, generally involve the introduction of a distribution that provides an approximation to the true posterior distribution. However, for each model that is to be approximated, researchers must painstakingly work out the mathematics necessary to apply variational inference, and then develop special-purpose computer code to implement the resulting variational algorithm. This can be costly, from both a time and a monetary perspective, and thus limits the usefulness of variational inference as a manner by which to develop usable probabilistic models. For this and other reasons, there is a need for the present invention.
SUMMARY OF THE INVENTION
The invention relates to a variational inference engine for probabilistic graphical models. The engine allows a user to design, implement and solve broad classes of models without recourse to mathematical analysis or computer coding. A model, for example, can be specified using a scripting language, or by the user drawing a graph of the probability distribution using a graphical user interface. The engine determines the posterior distribution, and thus allows the resulting probabilistic model to be used for prediction purposes.
In one embodiment, a computer-implemented method includes inputting a specification for a model that has observable variables and unobservable variables. The specification includes a functional form for the conditional distributions of the model, and a structure for a graph of model that has nodes for each of the variables. The model is usually such that an exact posterior distribution is intractable. The method determines a distribution for the unobservable variables that approximates the exact posterior distribution, based on the structure for the graph of the model, as well as the functional form for the conditional distributions of the model. This distribution is then output by the method.
As can be appreciated by those of ordinary skill within the art, the approach outlined herein can be extended to include the possibility of combining sampling methods, such as Markov chain Monte Carlo (e.g., Gibbs sampling) and exact methods along with variational methods, so that the engine could employ a combination of two or three different approaches to solve a particular model.
The invention includes computer-implemented methods, machine-readable media, computerized systems, and computers of varying scopes. Other aspects, embodiments and advantages of the invention, beyond those described here, will become apparent by reading the detailed description and with reference to the drawings.


REFERENCES:
patent: 5465321 (1995-11-01), Smyth
patent: 5616504 (1997-04-01), Brown et al.
patent: 5704017 (1997-12-01), Heckerman et al.
patent: 6021403 (2000-02-01), Horvitz et al.
patent: 6058206 (2000-05-01), Kortge
patent: 6212509 (2001-04-01), Pao et al.
patent: 6408290 (2002-06-01), Thiesson et al.
Norsys Corporation, Netica belief network software, http://www.norsys.com/, Aug. 11, 1999.
M. Ramoni, P. Sebastiani, An introduction to the robust Bayesian classifier, Knowledge Media Institute Technical Report, KMi-TR-79, Mar. 9, 1999.
Matlab 5, Bayes Net Toolbox, http://www.cs.berkeley.edu/~murphyk/bnt.html, Jun. 4, 1999.
Dept of Computer Science, Univ. of Helsinki, Bayesian Predictive Discriminant Analysis, http://www.cs.helsinki.fi/research/cosco/projects/None/SW/, Mar. 18, 1998.
D. MacKay, Ensemble learning and evidence maximization, NIPS 1995, May 1, 1995.
Decision Systems Laboratory, GeNIe, http://www2.sis.pitt.edu/~henie/about_Genie.html, Jul. 1, 1998.
S. Srinivas, J. Breese, IDEAL, http://www.rpal.rockwell.com/ideal.html, Jan. 1, 1990.
Hugin, Introducing the Hugin System, http://www.hugin.dk/hugintro/hugin_system_pane.html, Aug. 11, 1999.
BUGS, Introduction to the BUGS project, http://www.mrc-bsu.cam.ac.uk/bugs/intro/www/intro.html, Aug. 11, 1999.
F. Cozman, Introduction to JavaBayes, http://www.ds.cmu.edu/~javabayes/Home
ode2.html, Aug. 13, 1998.
J. Cheng, D. Bell, W. Liu, Learning Bayesian networks from data, Proc. 6th ACM Int'l Conf. Information & Knowledge Management, Jan. 1, 1997.
S.R. Waterhouse, A.J. Robinson, Non-linear prediction of acoustic vectors using hierarchical mixtures of experts, in Neural Information Processing Systems 7, MIT Press, Jan. 1, 1995.
R.M. Neal, G.E. Hinton, A view of the EM algorithm that justifies incremental-sparse and other variants, in Learning in Graphical Models, pp. 355-368 (Klewe), Jan. 1, 1998.

LandOfFree

Say what you really think

Search LandOfFree.com for the USA inventors and patents. Rate them and share your experience with other people.

Rating

Variational inference engine for probabilistic graphical models does not yet have a rating. At this time, there are no reviews or comments for this patent.

If you have personal experience with Variational inference engine for probabilistic graphical models, we encourage you to share that experience with our LandOfFree.com community. Your opinion is very important and Variational inference engine for probabilistic graphical models will most certainly appreciate the feedback.

Rate now

     

Profile ID: LFUS-PAI-O-3072749

  Search
All data on this website is collected from public sources. Our data reflects the most accurate information available at the time of publication.