Instruction Matrix based Genetic Programming for Classification - PowerPoint PPT Presentation

1 / 17
About This Presentation
Title:

Instruction Matrix based Genetic Programming for Classification

Description:

Most learning methods have fixed models, they learn by ... 5-fold. each fold 20 independent runs. G3-P is a grammar-guided GP. Decition tree. Fuzzy rule base ... – PowerPoint PPT presentation

Number of Views:53
Avg rating:3.0/5.0
Slides: 18
Provided by: appsrvCs
Category:

less

Transcript and Presenter's Notes

Title: Instruction Matrix based Genetic Programming for Classification


1
Instruction Matrix based Genetic Programming for
Classification
  • Li Gang
  • May 3, 2007

2
Outline
  • IMGP
  • Classification
  • Gradient descent
  • Program tree complexity
  • Experiment

3
IMGP - representation
  • Program Tree
  • (x-y)(yc)
  • (yc)(xy)
  • Instruction Matrix

4
IMGP-Algorithm
5
Classification
  • Given a set of training data (xi,ti), we need
    to learn a classifier y f(xT), such that the
    following error function is minimized
  • Most learning methods have fixed models, they
    learn by changing the parametersT

6
Gradient Descent
  • It is a numerical method to adjust the arguments
    to minimize the objective function
  • An example in Neural Network

7
GP for Classification
  • The power (or weakness?) of GP is that it
    searches for the optimal structure and parameters
    simultaneously
  • The program tree is a mathematical form with the
    constants as the parameters
  • Can we optimize the constants given a fixed tree
    structure?

8
Gradient Descent for GP
  • Usually gradient descent is used when the
    mathematical form is given, so we derive the
    updating rule offline
  • However, we can calculate the gradient by
    traversing the program tree recursively

9
Computation Cost
  • The computation cost is relatively large
  • We need to calculate the gradient of the internal
    nodes besides their values
  • The constants needs to be updated for a few times
    before it stabilizes
  • Gradient Descent is therefore only applied on the
    current best individual for a few steps

10
Program Tree Complexity
  • How to enhance the generalization?
  • Occams razor or Minimum Description Length
  • A penalty of the program tree complexity is added
    to the fitness (smaller is better)
  • fitness fitness w complexity

11
Tree Size?
  • A naïve way is to use the tree size as the
    complexity of the program tree
  • Linear function is the simplest model of class
    boundary
  • We count only the tree nodes of multiplication
    and division
  • This is the approach I am using at present
  • fitness fitness w ,/

12
Experiment
  • 5-fold
  • each fold 20 independent runs
  • G3-P is a grammar-guided GP
  • Decition tree
  • Fuzzy rule base
  • Neural network
  • Petri-net

13
(No Transcript)
14
(No Transcript)
15
Second Derivative
  • Intuitively, the ruggedness of the function curve
    reflects the function complexity (board)
  • Analytically, the second derivative of the
    function measures the function complexity
  • (axb)0
  • (ax2bxc)2a (constant)
  • (ax3bx2cxd)6ax2b (variable)
  • fitness fitness w sum(f(xi))

16
A simpler way?
  • fitness fitness w sum(f(xi))
  • Is it possible to avoid calculating the sum?
  • Can we infer the complexity from the operators
    directly?
  • (xy)xyxy2xy
  • If xy1, which implies xy2?
  • So (xy)22812? Or max(2,2,8)8?

17
  • Thank You
Write a Comment
User Comments (0)
About PowerShow.com