Chapter 6: HIDDEN MARKOV AND MAXIMUM ENTROPY - PowerPoint PPT Presentation

1 / 56
About This Presentation
Title:

Chapter 6: HIDDEN MARKOV AND MAXIMUM ENTROPY

Description:

... 3 3333f3 33 3 3 33 f3 3 3 ff3fff ff f3f33f3ff3 f3f3 ffff3fffff ffff ... 3 f 3 33 3f 3 3 3 f f3 ff f f f 3 f 3 f 3 f ... – PowerPoint PPT presentation

Number of Views:361
Avg rating:3.0/5.0
Slides: 57
Provided by: Inderje9
Category:

less

Transcript and Presenter's Notes

Title: Chapter 6: HIDDEN MARKOV AND MAXIMUM ENTROPY


1
Chapter 6 HIDDEN MARKOV AND MAXIMUM ENTROPY
  • Heshaam Faili
  • hfaili_at_ece.ut.ac.ir
  • University of Tehran

2
Introduction
  • Hidden Markov Model (HMM)
  • Maximum Entropy
  • Maximum Entropy Markov Model (MEMM)
  • machine learning methods
  • A sequence classifier or sequence labeler is a
    model whose job is to assign some label or class
    to each unit in a sequence
  • finite-state transducer is a non-probabilistic
    sequence classifier for transducing from
    sequences of words to sequences of morphemes
  • HMM and MEMM extend this notion by being
    probabilistic sequence classifiers

3
Markov chain
  • Observed Markov model
  • Weighted finite-state automaton
  • Markov Chain a weighted automaton in which the
    input sequence uniquely determines which states
    the automaton will go through
  • cant represent inherently ambiguous problems
  • useful for assigning probabilities to unambiguous
    sequences

4
Markov Chain
5
Formal Description
6
Formal Description
  • First-order Markov Chain the probability of a
    particular state is dependent only on the
    previous state
  • Markov Assumption P(qiq1...qi-1) P(qiqi-1)

7
Markov Chain example
compute the probability of each of the following
sequences hot hot hot hot cold hot cold hot
8
Hidden Markov Model
  • in POS tagging we didnt observe POS tags in the
    world we saw words, and had to infer the correct
    tags from the word sequence. We call the POS tags
    hidden because they are not observed
  • HMM allows us to talk HIDDEN MARKOV about both
    observed MODEL events (like words) and hidden
    events (like POS tags) that we think of as causal
    factors in our probabilistic model

9
Jason Eisner (2002) example
  • Imagine that you are a climatologist in the year
    2799 studying the history of global warming. You
    cannot find any records of the weather in
    Baltimore, Maryland, for the summer of 2007, but
    you do find Jason Eisners diary, which lists how
    many ice creams Jason ate every day that summer.
  • Our goal is to use these observations to estimate
    the temperature every day
  • Given a sequence of observations O, each
    observation an integer corresponding to the
    number of ice creams eaten on a given day, figure
    out the correct hidden sequence Q of weather
    states (H or C) which caused Jason to eat the ice
    cream

10
Formal Description
11
Formal Description
12
HMM Example
13
Fully-connected (Ergodic) Left-to-right (Bakis)
HMM
14
Three fundamental problems
  • Problem 1 (Computing Likelihood) Given an HMM ?
    (A,B) and an observation sequence O, determine
    the likelihood P(O ?)
  • Problem 2 (Decoding) Given an observation
    sequence O and an HMM ? (A,B), discover the
    best hidden state sequence Q
  • Problem 3 (Learning) Given an observation
    sequence O and the set of states in the HMM,
    learn the HMM parameters A and B

15
COMPUTING LIKELIHOOD THE FORWARD ALGORITHM
  • Given an HMM ? (A,B) and an observation
    sequence O, determine the likelihood P(O ?)
  • For a Markov chain we could compute the
    probability of 3 1 3 just by following the states
    labeled 3 1 3 and multiplying the probabilities
    along the arcs
  • We want to determine the probability of an
    ice-cream observation sequence like 3 1 3, but we
    dont know what the hidden state sequence is!
  • Markov chain Suppose we already knew the
    weather, and wanted to predict how much ice cream
    Jason would eat
  • For a given hidden state sequence (e.g. hot hot
    cold) we can easily compute the output likelihood
    of 3 1 3.

16
THE FORWARD ALGORITHM
17
THE FORWARD ALGORITHM
18
THE FORWARD ALGORITHM
  • dynamic programming O(N2T)
  • N hidden states and an observation sequence of T
    observations
  • ??T (j) represents the probability of being in
    state j after seeing the first t observations,
    given the automaton ?
  • qt j means the probability that the tth state
    in the sequence of states is state j

19
(No Transcript)
20
THE FORWARD ALGORITHM
21
THE FORWARD ALGORITHM
22
THE FORWARD ALGORITHM
23
DECODING THE VITERBI
24
DECODING THE VITERBI ALGORITHM
  • vt (j) represents the probability that the HMM is
    in state j after seeing the first t observations
    and passing through the most probable state
    sequence q0,q1,...,qt-1, given the automaton ?

25
TRAINING HMMS THE FORWARD-BACKWARD ALGORITHM
  • Given an observation sequence O and the set of
    possible states in the HMM, learn the HMM
    parameters A and B
  • Ice-Cream task we would start with a sequence of
    observations O 1,3,2, ...,, and the set of
    hidden states H and C.
  • part-of-speech tagging task we would start with
    a sequence of observations O w1,w2,w3 . . .
    and a set of hidden states NN, NNS, VBD, IN,...

26
forward-backward
  • Forward-backward or Baum-Welch algorithm (Baum,
    1972), a special case of the Expectation-Maximizat
    ion (EM algorithm)
  • Start on Markov Model no emission probabilities
    B (alternatively we could view a Markov chain as
    a degenerate Hidden Markov Model where all the b
    probabilities are 1.0 for the observed symbol and
    0 for all other symbols)
  • Only need to train transition probability A

27
forward-backward
  • For Markov Chain only need to compute the state
    transition based on observation and calculate
    matrix A
  • For Hidden Markov Model we can not count this
    transition
  • Baum-Welch algorithm uses two intuitions
  • The first idea is to iteratively estimate the
    counts
  • computing the forward probability for an
    observation and then dividing that probability
    mass among all the different paths that
    contributed to this forward probability

28
backward probability.
29
backward probability.
30
backward probability.
31
forward-backward
32
forward-backward
33
forward-backward
34
forward-backward
  • The probability of being in state j at time t,
    which we will call ?t (j)

35
forward-backward
36
forward-backward
37
(No Transcript)
38
MAXIMUM ENTROPY MODELS
  • Machine learning framework called Maximum Entropy
    modeling, MAXEnt
  • Used for Classification
  • The task of classification is to take a single
    observation, extract some useful features
    describing the observation, and then based on
    these features, to classify the observation into
    one of a set of discrete classes.
  • Probabilistic classifier gives the probability
    of the observation being in that class
  • Non-sequential classification
  • in text classification we might need to decide
    whether a particular email should be classified
    as spam or not
  • In sentiment analysis we have to determine
    whether a particular sentence or document
    expresses a positive or negative opinion.
  • well need to classify a period character (.)
    as either a sentence boundary or not

39
MaxEnt
  • MaxEnt belongs to the family of classifiers known
    as the exponential or log-linear classifiers
  • MaxEnt works by extracting some set of features
    from the input, combining them linearly (meaning
    that we multiply each by a weight and then add
    them up), and then using this sum as an exponent
  • Example tagging
  • A feature for tagging might be this word ends in
    -ing or the previous word was the

40
Linear Regression
  • Two different names for tasks that map some input
    features into some output value regression when
    the output is real-valued, and classification
    when the output is one of a discrete set of
    classes

41
Linear Regression, Example
42
Multiple linear regression
  • pricew0w1 Num Adjectivesw2 Mortgage Ratew3
    Num Unsold Houses

43
Learning in linear regression
  • sum-squared error

44
Logistic regression
  • Classification in which the output y we are
    trying to predict takes on one from a small set
    of discrete values
  • binary classification
  • Odds
  • logit function

45
Logistic regression
46
Logistic regression
47
Logistic regression Classification
hyperplane
48
Learning in logistic regression
conditional maximum likelihood estimation.
49
Learning in logistic regression
  • Convex Optimization

50
MAXIMUM ENTROPY MODELING
  • multinomial logistic regression(MaxEnt)
  • Most of the time, classification problems that
    come up in language processing involve larger
    numbers of classes (part-of-speech classes)
  • y is a value take on C different value
    corresponding to classes C1,,Cn

51
Maximum Entropy Modeling
  • Indicator function A feature that only takes on
    the values 0 and 1

52
Maximum Entropy Modeling
  • Example
  • Secretariat/NNP is/BEZ expected/VBN to/TO race/??
    tomorrow/

53
Maximum Entropy Modeling
54
Why do we call it Maximum Entropy?
  • From of all possible distributions, the
    equiprobable distribution has the maximum entropy

55
Why do we call it Maximum Entropy?
56
Maximum Entropy
  • probability distribution of a multinomial
    logistic regression model whose weights W
    maximize the likelihood of the training data!
    Thus the exponential model
Write a Comment
User Comments (0)
About PowerShow.com