Forgetting Counts : Constant Memory Inference for a Dependent Hierarchical Pitman-Yor Process

1 / 13
About This Presentation
Title:

Forgetting Counts : Constant Memory Inference for a Dependent Hierarchical Pitman-Yor Process

Description:

... A hierarchical Bayesian language model based on Pitman-Yor processes N-gram Markov chain language ... Leaf restaurants The SMC algorithm The ... – PowerPoint PPT presentation

Number of Views:6
Avg rating:3.0/5.0
Slides: 14
Provided by: mc138
Learn more at: http://www.ece.duke.edu

less

Transcript and Presenter's Notes

Title: Forgetting Counts : Constant Memory Inference for a Dependent Hierarchical Pitman-Yor Process


1
Forgetting Counts Constant Memory Inference for
a Dependent Hierarchical Pitman-Yor Process
  • Nicholas Bartlett, David Pfau, Frank Wood
  • Presented by Yingjian Wang
  • Nov. 17, 2010

2
Outline
  • Background
  • The sequential memoizer
  • Forgetting
  • The dependent HPY
  • Experiment results

3
Background
2006,Teh, A hierarchical Bayesian language model
based on Pitman-Yor processes
N-gram Markov chain language model with the HPY
prior.
The Sequential Memoizer (SM) with linear
space/time inference scheme. (lossless)
2009, Wood, A Stochastic Memoizer for Sequence
Data
Combine the SM with an arithmetic coder to
develop a compressor (PLUMP/dePLUMP), see
www.deplump.com.
2010, Gasthaus, Lossless compression based
on the Sequence Memoizer
2010, Bartlett, Forgetting Counts Constant
Memory Inference for a Dependent HPY
Develop a constant memory/space inference for the
SM, by using a dependent HPY. (with loss)
4
SM-Two concepts
  • Memoizer (Donald Michie, 1968) A device which
  • returns former results under the same input
    instead of recalculating in order to save time.
  • Stochastic Memoizer (Wood, 2009) The returned
    results can change since the prediction
    probability is based upon a stochastic process.

5
SM-model and trie
  • model
  • The prefix trie
  • restaurants.

6
SM-the NSP (1)
  • The Normalized Stable Process (Perman, 1990)

Pitman-Yor Process
Concentration parameter c0
Discount parameter d0
Dirichlet Process
A Normalized Stable Process
7
SM-the NSP (2)
  • Collapse the middle restaurants
  • Theorem
  • If
  • Then
  • Prefix tree
  • restaurants
  • (Weiner, 1973
  • Ukkonen, 1995)

8
SM-linear space inference
9
Forgetting
  • Motivation to achieve constant memory inference
    on the basis of SM. How to do? ---
  • Methods Forgetting/delete the restaurants.
  • Restaurants - the basic memory units in the
    context tree
  • How to delete? two deletion schemes random
    deletion greedy deleting.

10
Deletion schemes
  • Random deletion uniformly delete one leaf
    restaurant.
  • Greedy deletion least negatively impacts the
    estimated likelihood of the observed sequence.

Leaf restaurants
11
The SMC algorithm
12
The dependent HPY
  • But wait, what we get after the
    deletion-addition? Will the processes be
    independent? No (Since the seating arrangement
    in the parent restaurant has been changed.)

13
The experiment results
Write a Comment
User Comments (0)