2D1431%20Machine%20Learning - PowerPoint PPT Presentation

About This Presentation
Title:

2D1431%20Machine%20Learning

Description:

Bagging & Boosting. Bagging and Boosting aggregate multiple hypotheses ... Bagging and Boosting generate a classifier with a smaller error on the training ... – PowerPoint PPT presentation

Number of Views:28
Avg rating:3.0/5.0
Slides: 12
Provided by: unkn651
Category:

less

Transcript and Presenter's Notes

Title: 2D1431%20Machine%20Learning


1
2D1431 Machine Learning
  • Boosting

2
Classification Problem
  • Assume a set s of N instances xi ? X each
    belonging to one of M classes c1,cM .
  • The training set consists of pairs (xi,ci).
  • A classifier C assigns a classification
  • C(x) ? c1,cM to an instance x.
  • The classifier learned in trial t is denoted Ct
    while C is the composite boosted classifier.

3
Linear Discriminant Analysis
  • Possible classes ,
  • LDA if w1 x1 w2x2 w3 gt 0
  • otherwise

x
o
x
o
x2
w1x1w2x2w30
x
x
x
x
o
o
x
o
o
o
x1
4
Bagging Boosting
  • Bagging and Boosting aggregate multiple
    hypotheses generated by the same learning
    algorithm invoked over different distributions of
    training data Breiman 1996, Freund Schapire
    1996.
  • Bagging and Boosting generate a classifier with a
    smaller error on the training data as it combines
    multiple hypotheses which individually have a
    larger error.

5
Boosting
  • Boosting maintains a weight wi for each instance
    ltxi, cigt in the training set.
  • The higher the weight wi, the more the instance
    xi influences the next hypothesis learned.
  • At each trial, the weights are adjusted to
    reflect the performance of the previously learned
    hypothesis, with the result that the weight of
    correctly classified instances is decreased and
    the weight of incorrectly classified instances is
    increased.

6
Boosting
  • Construct an hypothesis Ct from the current
    distribution of instances described by wt.
  • Adjust the weights according to the
    classification error et of classifier Ct.
  • The strength at of a hypothesis depends on its
    training error et. at ½ ln ((1-et)/et)

learn hypothesis
Set of weighted wit instances xi
hypothesis Ct strength at
adjust weights
7
Boosting
  • The final hypothesis CBO aggregates the
    individual hypotheses Ct by weighted voting.
  • cBO(x) argmax cj?C St1T at d(cj,ct(x))
  • Each hypothesis vote is a function of its
    accuracy.
  • Let wit denote the weight of an instance xi at
    trial t, for every xi, wi1 1/N. The weight wit
    reflects the importance (e.g. probability of
    occurrence) of the instance xi in the sample set
    St.
  • At each trial t 1,,T an hypothesis Ct is
    constructed from the given instances under the
    distribution wt. This requires that the learning
    algorithm can deal with fractional examples.

8
Boosting
  • The error of the hypothesis Ct is measured with
    respect to the weights
  • et Si Ct(xi) ? ci wit / Si wit
  • at ½ ln ((1-et)/et)
  • Update the weights wit of correctly and
    incorrectly classified instances by
  • wit1 wit e-at if Ct(xi) ci
  • wit1 wit eat if Ct(xi) ? ci
  • Afterwards normalize the wit1 such that they
    form a proper distribution Si wit1 1

9
Boosting
  • The classification cBO(x) of the boosted
    hypothesis is obtained by summing the votes of
    the hypotheses C1, C2,, CT where the vote of
    each hypothesis Ct is weights at .

cBO(x) argmax cj?C St1T at d(cj,ct(x))


10
Boosting
  • Given ltx1,c1gt,ltxm,cmgt
  • Initialize wi11/m
  • For t1,,T
  • train weak learner using distribution wit
  • get weak hypothesis Ct X ? C with error
  • et Si Ct(xi) ? ci wit / Si wit
  • choose at ½ ln ((1-et)/et)
  • update
  • wit1 wit e-at if Ct(xi) ci
  • wit1 wit eat if Ct(xi) ? ci
  • Output the final hypothesis
  • cBO(x) argmax cj?C St1T at d(cj,ct(x))

11
Bayes MAP Hypothesis
  • Bayes MAP hypothesis for two classes x and o
  • red incorrect classified instances

12
Boosted Bayes MAP Hypothesis
  • Boosted Bayes MAP hypothesis has more complex
    decision surface than individual hypotheses alone
Write a Comment
User Comments (0)
About PowerShow.com