-Artificial Neural Network- Chapter 5 Back Propagation Network PowerPoint PPT Presentation

presentation player overlay
About This Presentation
Transcript and Presenter's Notes

Title: -Artificial Neural Network- Chapter 5 Back Propagation Network


1
-Artificial Neural Network-Chapter 5 Back
Propagation Network
  • ??????
  • ?????
  • ??? ??

2
Introduction (1)
  • BPN Back Propagation Network
  • BPN is a layered feedforward supervised network.
  • BPN provides an effective means of allowing a
    computer to examine data patterns that may be
    incomplete or noisy.
  • Architecture

?1
?2
Yj
Xn
?h
3
Introduction (2)
  • Input layer X1,X2,.Xn.
  • Hidden layer can have more than one layer.
  • derive net1, net2, neth transfer
    output H1, H2,,Hh,
  • to be used as the input to derive
    the result for output layer
  • Output layer Y1,Yj.
  • Weights Wij.
  • Transfer function Nonlinear ? Sigmoid function
  • () The nodes in the hidden layers organize
    themselves in a way that different nodes learn to
    recognize different features of the total input
    space.

4
Processing Steps (1)
  • Briefly describe the processing Steps as follows.
  • Based on the problem domain, set up the network.
  • Randomly generate weights Wij.
  • Feed a training set, X1,X2,.Xn, into BPN.
  • Compute the weighted sum and apply the transfer
    function on each node in each layer. Feeding the
    transferred data to the next layer until the
    output layer is reached.
  • The output pattern is compared to the desired
    output and an error is computed for each unit.

5
Processing Steps (2)
  • 5. Feedback the error back to each node in the
    hidden layer.
  • Each unit in hidden layer receives only a portion
    of total errors and these errors then feedback to
    the input layer.
  • Go to step 4 until the error is very small.
  • Repeat from step 3 again for another training
    set.

6
Computation Processes(1/10)
  • The detailed computation processes of BPN.
  1. Set up the network according to the input nodes
    and the output nodes required.
  2. Randomly assigned the weights.
  3. Feed the training pattern (set) into the network
    and do the following computation.

7
Computation Processes(2/10)
  • 4. Compute from the Input layer to hidden layer
    for each node.

5. Compute from the hidden layer to output layer
for each node.
8
Computation Processes(3/10)
  • 6. Calculate the total error find the
    difference for correction
  • djYj(1-Yj)( Tj -Yj)
  • dhHh(1- Hh) SjWhj dj
  • 7. ?Whj?dj Hh ?Tj -?dj
  • ?Wih?dh Xi ?Th -?dh
  • 8. update weights
  • WhjWhj?Whj ,WihWih?Wih ,
  • Tj Tj ?Tj, Th Th ?Th
  • 9. Repeat steps 48, until the error is very
    small.
  • 10.Repeat steps 39, until all the training
    patterns are learned.

9
EX Use BPN to solve XOR (1)
  • Use BPN to solve the XOR problem
  • Let W111, W21 -1, W12 -1, W221, W131,
    W231, T11, T21,T31, ?10

X1 X2 T
-1 -1 0
-1 1 1
1 -1 1
1 1 0
10
EX BPN Solve XOR (2)
  • ?W12?d1 X1 (10)(-0.018)(-1)0.18
  • ?W21?d1 X2 (10)(-0.018)(-1)0.18
  • ?T1 -?d1 -(10)(-0.018)0.18
  • ?????????????.

1.18
X1
0.754
0.82
0.82
1.915
0.754
X2
1.18
11
BPN Discussion
  • Number of hidden nodes increase, the convergence
    will get slower. But the error can be minimized
  • The general concept of designing the number of
    hidden node uses
  • of hidden nodes(Input nodes Output
    nodes)/2, or
  • of hidden nodes(Input nodes Output
    nodes)1/2
  • Usually, 12 hidden layer is enough for learning
    a complex problem. Too many layers will cause
    the learning very slow. When the problem is
    hyper-dimension and very complex, then we add an
    extra layer.
  • Learning rate, ?, usually set as 0.5, 1.0, but
    it depends on how fast and how detail the network
    shall learn.

12
The Gradient Steepest Descent Method(SDM) (1)
  • The gradient steepest descent method
  • Recall
  • We want the difference of computed output and
    expected output getting close to 0.
  • Therefore, we want to obtain so that we
    can update weights to improve the network results.

13
The Gradient Steepest Descent Method(SDM) (2)
14
The Gradient Steepest Descent Method(SDM) (3)
15
The Gradient Steepest Descent Method(SDM) (4)
16
The Gradient Steepest Descent Method(SDM) (5)
17
The Gradient Steepest Descent Method(SDM) (6)
  • Learning computation

18
The Gradient Steepest Descent Method(SDM) (7)
Write a Comment
User Comments (0)
About PowerShow.com