I256: Applied Natural Language Processing - PowerPoint PPT Presentation

About This Presentation
Title:

I256: Applied Natural Language Processing

Description:

I256: Applied Natural Language Processing – PowerPoint PPT presentation

Number of Views:12
Avg rating:3.0/5.0
Slides: 22
Provided by: coursesIs8
Category:

less

Transcript and Presenter's Notes

Title: I256: Applied Natural Language Processing


1
I256 Applied Natural Language Processing
Marti Hearst Sept 27, 2006    
2
Evaluation Measures
3
Evaluation Measures
  • Precision
  • Proportion of those you labeled X that the gold
    standard thinks really is X
  • correctly labeled by alg/ all labels assigned
    by alg
  • True Positive / (True Positive False
    Positive)
  • Recall
  • Proportion of those items that are labeled X in
    the gold standard that you actually label X
  • correctly labeled by alg / all possible correct
    labels
  • True Positive / (True Positive False
    Negative)

4
F-measure
  • Can cheat with precision scores by labeling
    (almost) nothing with X.
  • Can cheat on recall by labeling everything with
    X.
  • The better you do on precision, the worse on
    recall, and vice versa
  • The F-measure is a balance between the two.
  • 2precisionrecall / (recallprecision)

5
Evaluation Measures
  • Accuracy
  • Proportion that you got right
  • (True Positive True Negative) / N
  • N TP TN FP FN
  • Error
  • (False Positive False Negative)/N

6
Prec/Recall vs. Accuracy/Error
  • When to use Precision/Recall?
  • Useful when there are only a few positives and
    many many negatives
  • Also good for ranked ordering
  • Search results ranking
  • When to use Accuracy/Error
  • When every item has to be judged, and its
    important that every item be correct.
  • Error is better when the differences between
    algorithms are very small lets you focus on
    small improvements.
  • Speech recognition

7
Evaluating Partial Parsing
  • How do we evaluate it?

8
Evaluating Partial Parsing
9
Testing our Simple Fule
  • Lets see where we missed

10
Update rules Evaluate Again
11
Evaluate on More Examples
12
Incorrect vs. Missed
  • Add code to print out which were incorrect

13
Missed vs. Incorrect
14
What is a good Chunking Baseline?
15
(No Transcript)
16
The Tree Data Structure
17
Baseline Code (continued)
18
Evaluating the Baseline
19
Cascaded Chunking
20
(No Transcript)
21
Next Time
  • Summarization
Write a Comment
User Comments (0)
About PowerShow.com