Title: Dynamic Bayesian Networks
1Dynamic Bayesian Networks
2Table of Contents
- Motivation
- Bayesian Networks
- Bayes Theorem
- Dynamic Bayesian Networks
- Representation
- Basic Inference
- Approximate Inference
- Special Cases
- Applications
- Conclusions
- BLOg
- References
3Motivation
- Bayesian Networks offer a number of
well-documented advantages for the representation
and processing of knowledge and uncertainty. - BN can be understood by humans as well as learned
from observed data - Real time algorithms for reasoning under
uncertainty exist.
4Bayesian Networks
Graphical representations of joint distributions
Static world, each random variable has a single
fixed value.
5Bayes Theorem
Mathematical formula used for calculating
conditional probabilities. Develop by the
mathematician and theologian Thomas Bayes
(published in 1763)
6Dynamic Bayesian Network
- How can we model dynamic situations?
The process of change can de viewed as a set of
slices, each of which describes the state of the
world at a particular time.
Set of unobservable state variables at time t.
Set of observable evidence variables at time t.
The term dynamic means we are modeling a
dynamic system, not that the network structure
changes over time.
7DBN - Representation
- Example
- Target Is it raining today?
next step specify dependencies among the
variables.
8DBN - Representation
- Necessity to specify an unbounded number of
conditional probability table, one for each
variable in each slice, - Each one might involve an unbounded number of
parents.
- Assume that change in the world state are caused
by a stationary process (unmoving process over
time).
is the same for all t
9DBN - Representation
2. Use Markov assumption - The current state
depends on only in a finite history of previous
states. Using the first-order Markov process
Transition Model
In addition to restricting the parents of the
state variable Xt, we must restrict the parents
of the evidence variable Et
Sensor Model
10Dynamic Bayesian Networks
Bayesian network structure corresponding to a
first-order of Markov process with state defined
by the variables Xt.
A second order of Markov process
11Dynamic Bayesian Networks
Bayesian network structure describing the
umbrella world. The transition model is
and
the sensor model is
12Dynamic Bayesian Networks
- There are two possible fixes if the approximation
is too inaccurate - Increasing the order of the Markov process model.
For example, adding as a parent of
, which might give slightly more accurate
predictions. - Increasing the set of state variables. For
example, adding to allow to
incorporate historical records of rainy seasons,
or adding ,
and Pressure to allow to use a
physical model of rainy conditions.
13DBN Basic Inference
- Filtering or monitoring
- Prediction
- Smoothing or hindsight
- Most likely explanation
- The details of how to perform these computations
depend on which model and which algorithm is used.
14DBN Basic Inference
- Filtering or Monitoring
-
- Compute the belief state - the posterior
distribution over the current state, given all
evidence to date.
Filtering is what a rational agent needs to do in
order to keep track of the current state so that
the rational decisions can be made.
15DBN Basic Inference
Given the results of filtering up to time t, one
can easily compute the result for t1 from the
new evidence
(for some function f)
(dividing up the evidence)
(using Bayes Theorem)
(by the Marcov propertyof evidence)
a is a normalizing constant used to make
probabilities sum up to 1.
16DBN Basic Inference
The second term represents a
one-step prediction of the next step, and the
first term updates this with
the new evidence. Now we obtain the one-step
prediction for the next step by conditioning on
the current state Xt
(using the Marcov property)
17DBN Basic Inference
- Illustration for two steps in the Umbrella
example - On day 1, the umbrella appears so U1true. The
prediction from t0 to t1 is
and updating it with the evidence for t1 gives
- On day 2, the umbrella appears so U2true. The
prediction from t1 to t2 is
and updating it with the evidence for t2 gives
18DBN Basic Inference
- Prediction
-
- Compute the posterior distribution over the
future state, given all evidence to date.
for some kgt0
The task of prediction can be seen simply as
filtering without the addition of new evidence.
19DBN Basic Inference
- Smoothing or hindsight
-
- Compute the posterior distribution over the past
state, given all evidence up to the present.
for some k such that 0 k lt t.
Hindsight provides a better estimate of the state
than was available at the time, because it
incorporates more evidence.
20DBN Basic Inference
- Most likely explanation
-
- Compute the sequence of states that is most
likely to have generated a given sequence of
observation.
Algorithms for this task are useful in many
applications, including speech recognition.
21DBN Basic Inference
- Most likely explanation cont.
There exist a recursive relationship between the
most likely paths to each state Xt1 and the most
likely paths to each state Xt. This relationship
can be write as an equation connecting the
probabilities of the paths
22DBN Approximate inference
- Even though we can use DBN to represent very
complex temporal process with many sparsely
connected variables, we cannot reason efficiently
and exact about those process. - Thus, we must fall back on approximate methods.
23DBN Approximate inference
- Particle Filtering Algorithms
Focus the set of samples on the high probability
regions of the state space, throwing away samples
that have very low weight, according to the
observation, while multiplying those that have
high weight. In that way the population of
samples will stay reasonably close to reality.
24Particle Filtering Algorithm
25Dynamic Bayesian Networks
- In addition to these tasks, methods are needed
for learning the transition and sensor models
from observation. - Learning can be done by inference, where
inference provides an estimate of what
transitions actually occurred and of what states
generated the sensor readings. These estimates
can be used to update the models. - The updated models provide new estimates, and the
process iterates to convergence.
26Dynamic Bayesian Networks
- Learning requires the full smoothing inference,
rather than filtering, because it provides better
estimates of the state of the process. - Learning the parameters of a BN is done using
Expectation Maximization (EM) Algorithms.
Iterative optimization method to estimate some
unknowns parameters.
27DBN Special Cases
- Hidden Markov Model (HMMs)
- Temporal probabilistic model in which the state
of the process is described by a single discrete
random variable. (The simplest kind of DBN ) - Kalman Filter Models (KFMs)
- Estimate the state of a physical system from
noisy observations over time. Also known as
linear dynamical systems (LDSs).
28Dynamic Bayesian Networks
- Conclusions
- A DBN is a Bayesian network that represents a
temporal probability model. - In general each slice of a DBN can have any
number of state variables and evidence
variables . - For simplicity it can be assume that the
variables and their links are exactly replicated
from slice to slice and that the DBN represents a
first order Markov process. -
29DBN - Applications
- Areas such as Academics, Biology, Business and
Finance, Computer Games, Computer Vision,
Computer Software, Medicine, Planning,
Psychology, Scheduling, Speech Recognition, etc.
- The most widely used are the ones embedded in
Microsoft's products, including the Answer Wizard
of Office 95, the Office Assistant (the bouncy
paperclip guy) of Office 97, and over 30
Technical Support Troubleshooters.
(research.microsoft.com) - Another interesting fielded application is the
Vista system, developed by Eric Horvitz. The
Vista system is a decision-theoretic system that
has been used at NASA Mission Control Center in
Houston for several years. (research.microsoft.com
/research/dtg/horvitz/vista.htm)
30BLOG Bayesian LOGic
- A formal, compact and intuitive language for
defining probability distribution models over
outcomes with varying sets of objects (worlds
with unknowns objects and identity uncertainty).
31DBN - References
- Russel, S. and Norvig, P. Artificial
Intelligence A Modern Approach. Second Edition.
Prentice Hall. 2003. - Murphy, K. Dynamic Bayesian Networks
Representation, inference and Learning. Phd
Thesis. UC Berkeley, Computer Science Division,
July 2002. - Neopolotan, R. Learning Bayesian Networks.
Prentice Hall. 2003. - http//en.wikipedia.org/wiki/Bayesian_network
- http//www.cs.ubc.ca/murphyk/Bayes/bnintro.htmli
nfer