Title: A Non-Parametric Bayesian Method for Inferring Hidden Causes
1A Non-Parametric Bayesian Method for Inferring
Hidden Causes
- by F. Wood, T. L. Griffiths and Z. Ghahramani
Discussion led by Qi An ECE, Duke University
2Outline
- Introduction
- A generative model with hidden causes
- Inference algorithms
- Experimental results
- Conclusions
3Introduction
- A variety of methods from Bayesian statistics
have been applied to find model structure from a
set of observed variables - Find the dependencies among the set of observed
variables - Introduce some hidden causes and infer their
influence on observed variables
4Introduction
- Learning model structure containing hidden causes
presents a significant challenge - The number of hidden causes is unknown and
potentially unbounded - The relation between hidden causes and observed
variables is unknown - Previous Bayesian approaches assume the number of
hidden causes is finite and fixed.
5A hidden causal structure
- Assume we have T samples of N BINARY variables.
Let be the data and be a dependency
matrix among . - Introduce K BINARY hidden causes with T samples.
Let be hidden causes and be a
dependency matrix between and - K can potentially be infinite.
6A hidden causal structure
Hidden causes (Diseases)
Observed variables (Symptoms)
7A generative model
- Our goal is to estimate the dependency matrix Z
and hidden causes Y. - From Bayes rule, we know
- We start by assuming K is finite, and then
consider the case where K?8
8A generative model
- Assume the entries of X are conditionally
independent given Z and Y, and are generated from
a noise-OR distribution.
where , e is a baseline
probability that , and ? is the
probability with which any of hidden causes is
effective
9A generative model
- The entries of Y are assumed to be drawn from a
Bernoulli distribution - Each column of Z is assumed to be Bernoulli(?k)
distributed. If we further assume a Beta(a/K,1)
hyper-prior and integrate out ?k
where
These assumptions on Z are exactly the same as
the assumption in IBP
10Taking the infinite limit
- If we let K approach infinite, the distribution
on X remains well-defined, and we only need to
concern about rows in Y that the corresponding
mkgt0. - After some math and reordering of Z, the
distribution on Z can be obtained as
11The Indian buffet process is defined in terms of
a sequence of N customers entering a restaurant
and choosing from an infinite array of dishes.
The first customer tries the first Poisson(a)
dishes. The remaining customers then enter one by
one and pick previously sampled dishes with
probability and then tries Poisson(a/i)
new dishes.
12Reversible jump MCMC
13Gibbs sampler for Infinite case
14Experimental results
Synthetic Data
15number of iterations
16Real data
17Conclusions
- A non-parametric Bayesian technique is developed
and demonstrated - Recovers the number of hidden causes correctly
and can be used to obtain reasonably good
estimate of the causal structure - Can be integrated into Bayesian structure
learning both on observed variables and on hidden
causes.