Document Classification using Deep Belief Nets - PowerPoint PPT Presentation

About This Presentation
Title:

Document Classification using Deep Belief Nets

Description:

Corpus: Wikipedia XML Corpus. Single-labeled data each ... Increasing iterations may (partially) make up for learning poor features. Configuration (v,h) ... – PowerPoint PPT presentation

Number of Views:132
Avg rating:3.0/5.0
Slides: 10
Provided by: Ted134
Learn more at: https://nlp.stanford.edu
Category:

less

Transcript and Presenter's Notes

Title: Document Classification using Deep Belief Nets


1
Document Classification using Deep Belief Nets
  • Lawrence McAfee
  • 6/9/08
  • CS224n, Sprint 08

2
Overview
Doc1 Food
Doc1
Doc1
Doc1
Doc2 Brazil
Doc3
Classifier
Doc2
Doc3 Presidents
  • Corpus Wikipedia XML Corpus
  • Single-labeled data each document falls under
    single category
  • Binary Feature Vectors
  • Bag-of-words
  • 1 indicates word occurred one or more times in
    document

3
Background on Deep Belief Nets
Very abstract features
RBM 3
Higher level features
RBM 2
Features/basis vectors for training data
RBM
RBM 1
Training Data
  • Unsupervised, clustering training algorithm

4
Inside an RBM
hidden
j
Energy
i
visible
Input/Training data
  • Goal in training RBM is to minimize energy of
    configurations corresponding to input data
  • Train RBM by repeatedly sampling hidden and
    visible units for a given data input

5
Depth
  • Binary representation does not capture word
    frequency information
  • Inaccurate features learned at each level of DBN

6
Training Iterations
  • Accuracy increases with more training iterations
  • Increasing iterations may (partially) make up for
    learning poor features

Energy
Energy
7
Comparison to SVM, NB
30 categories
  • Binary features do not provide good starting
    point for learning higher level features
  • Binary still useful, as 22 is better than random
  • Time DBN-2h,13m SVM-4sec NB-3sec

8
Lowercasing
  • Supposedly richer vocabulary when lowercasing
  • Overfitting we dont need these extra words
  • Other experiments show only top 500 words relevant

9
Suggestions for Improvement
  • Use appropriate continuous-valued neurons
  • Linear or Gaussian neurons
  • Slower to train
  • Not much documentation on using continuous-valued
    neurons with RBMs
  • Implement backpropagation to fine-tune weights
    and biases
  • Propagate error derivatives from top level RBM
    back to inputs
  • Unsupervised training gives good initial weights,
    while backpropagation slightly modifies
    weights/biases
  • Backpropagation cannot be used alone, as it tends
    to get stuck in local optima
Write a Comment
User Comments (0)
About PowerShow.com