Title: Sieci neuronowe
1Sieci neuronowe bezmodelowa analiza danych?
- K. M. Graczyk
- IFT, Uniwersytet Wroclawski
- Poland
2Abstract
- Podczas seminarium opowiem o zastosowaniu
jednokierunkowych sieci neuronowych do analizy
danych eksperymentalnych. W szczególnosci skupie
uwage na podejsciu bayesowskim, które pozwala na
klasyfikacje i wybór najlepszej hipotezy
badawczej. Metoda ta ma w naturalny sposób
wbudowane tzw. kryterium brzytwy Ockhama,
preferujace modele o mniejszym stopniu
zlozonosci. Dodatkowym atutem podejscia jest brak
wymogu uzywania tzw. zbioru testowego do
weryfikacji procesu uczenia. - W drugiej czesci seminarium omówie wlasna
implementacje sieci neuronowej, zawierajaca
metody uczenia bayesowskiego. Na zakonczenie
pokaze moje pierwsze zastosowania w analizie
danych rozproszeniowych.
3Why Neural Networks?
- Look at Electromagnetic Form Factor data
- Simple
- Strightforward
- Then attac more serious problems
- Inspired by C. Giunti (Torino)
- Papers of Forte et al.. (JHEP 0205062,200, JHEP
0503080,2005, JHEP 0703039,2007,
Nucl.Phys.B8091-63,2009). - A kind of model independet way of fitting data
and computing assiosiated uncertienty. - Cooperation with R. Sulej (IPJ, Warszawa) and P.
Plonski (Politechnika Warszawska) - NetMaker
- GrANNet ) my own C library
4Road map
- Artificial Neural Networks (NN) idea
- FeedForward NN
- Bayesian statistics
- Bayesian approach to NN
- PDFs by NN
- GrANNet
- Form Factors by NN
5Inspired by Nature
6Aplications, general list
- Function approximation, or regression analysis,
including time series prediction, fitness
approximation and modeling. - Classification, including pattern and sequence
recognition, novelty detection and sequential
decision making. - Data processing, including filtering, clustering,
blind source separation and compression. - Robotics, including directing manipulators,
Computer numerical control.
7Artificial Neural Network
Output, target
Input layer
Hidden layer
8threshold
9A map from one vector space to another
10Neural Networks
- The universal approximation theorem for neural
networks states that every continuous function
that maps intervals of real numbers to some
output interval of real numbers can be
approximated arbitrarily closely by a multi-layer
perceptron with just one hidden layer. This
result holds only for restricted classes of
activation functions, e.g. for the sigmoidal
functions. (Wikipedia.org)
11Feed-Forward-Network
activation function
- Heavside function q(x)
- ? 0 or 1 signal
- Sigmoid function
- Tanh()
12architecture
- 3 -layers network, two hidden
- 1211
- 221 121 par9
Bias neurons, instead of thresholds
G(Q2)
Q2
Linear Function
Symmetric Sigmoid Function
13Supervised Learning
- Propose the Error Function (Standard Error
Function, chi2, etc, , any continous function
which has a global minimum) - Consider set of the data
- Train given network with data ? marginalize the
error function - Back propagation algorithms
- Iterative procedure which fixes weights
14Learning
- Gradient Algorithms
- Gradient descent
- QuickProp (Fahlman)
- RPROP (Ridmiller Braun)
- Conjugate gradients
- Levenberg-Marquardt (hessian)
- Newtonian method (hessian)
- Monte Carlo algorithms (based on the Marcov chain
algorithm)
15Overfitting
- More complex models describe data in better way,
but lost generalities - bias-variance trade-off
- After fitting one needs to compare with the test
set (must twice larger than original) - Overfitting ? large values of the wigths
- Regularization ? additional penalty term to error
function
16Fitting data with Artificial Neural Networks
- The goal of the network training is not to learn
on exact representation of the training data
itself, but rather to built statistical model for
the process which generates the data - C. Bishop, Neural Networks for Pattern
Recognation
17Parton Distribution Function with NN
18Parton Distributions Functions S. Forte, L.
Garrido, J. I. Latorre and A. Piccione, JHEP 0205
(2002) 062
- A kind of model independent analysis of the data
- Construction of the probability density PG(Q2)
in the space of the structure functions - In practice only one Neural Network architecture
- Probability density in the space of parameters of
one particular NN
But in reality Forte at al.. did
19The idea comes from W. T. Giele and S. Keller
Training Nrep neural networks, one for each set
of Ndat pseudo-data
The Nrep trained neural networks ? provide a
representation of the probability measure in the
space of the structure functions
20uncertainty
correlation
2110, 100 and 1000 replicas
22short
enough long
too long
30 data points, overfitting
23(No Transcript)
24My criticism
- Artificial data, and chi2 error function ?
overestimate error function? - Do not discuss other architectures?
- Problems with overfitting?
25Form Factors with NN, done with FANN library
26How to apply NN to the ep data
- First stage checking if the NN are able to work
on the reasonable level - GE and GM and Ratio separately
- Input Q2 ? output Form Factor
- The standard error function
- GE 200 points
- GM 86 points
- Ratio 152 points
- Combination of the GE, GM, and Ratio
- Input Q2 output GM and GE
- The standard error function a sum of three
functions - GEGMRatio around 260 points
- One needs to constrain the fits by adding some
artificial points with GE(0)GM(0)/mp1
27GMp
28GMp
29GMp
Neural Networks
Fit with TPE (our work)
30GEp
31GEp
32GEp
33Ratio
34GEn
35GEn
36GEn
37GMn
38GMn
39Bayesian Approach
- common sense reduced to calculations
40Bayesian Framework for BackProp NN, MacKay,
Bishop,
- Objective Criteria for comparing alternative
network solutions, in particular with different
architectures - Objective criteria for setting decay rate a
- Objective choice of reularising function Ew
- Comparing with test data is not requiered.
41Notation and Conventions
42Model Classification
- A collection of models, H1, H2, , Hk
- We belive that models are classified by P(H1),
P(H2), , P(Hk) (sum to 1) - After observing data D ? Bayes rule ?
- Usually at the beginning P(H1)P(H2) P(Hk)
43Single Model Statistics
- Assume that model Hi is correct one
- The neural network A with weights w is considered
- Task 1 Assuming some prior probability of w,
construct Posterior after including data
44Hierarchy
45Constructing prior and posterior function
Weight distribution!!!
likelihood
Prior
Posterior probability
w0
46Computing Posterior
hessian
Covariance matrix
47How to fix proper a
- Two ideas
- Evidence Approximation (MacKay)
- Hirerchical
- Find wMP
- Find aMP
- Perform analitically integrals over a
If sharply peaked!!!
48Getting aMP
The effective number of well-determined parameters
Iterative procedure during training
49Bayesian Model Comparison Occam Factor
Occam Factor
- The log of Occam Factor ? amount of
- Information we gain after data have arraived
- Large Occam factor ?? complex models
- larger accesible phase space (larger range of
posterior) - Small Occam factor ?? simple models
- larger accesible phase space (larger range of
posterior)
Best fit likelihood
50Evidence
51What about cross sections
- GE and GM simultaneously,
- Input Q2 and e ? cross sections
- Standard error function
- the chi-2-like function, with the covariance
matrix obtained from the Rosenbluth separation - Possibilities
- The set of Neural Networks becomes a natural
distribution of the differential cross sections - One can produce artificial data in the wide range
of the epsilon and perform the Rosenbluth
separation, searching the nonlinearities of sR in
the epsilon dependence.
52What about TPE?
- Q2, epsilon ? GE, GM and TPE?
- In the perfect case the change of the epsilon
should not affect the GE and GM. - training by the NN by series of the artificial
cross section data with fixed epsilon? - Collecting data in the epsilon bins, and Q2 bins,
then showing network the set of data with
particular epsilon in the wide range of Q2.
53constraining error function
every cycle computed with different epsilon!
54One network!
GM
GE
TPE
Yellow line have the vanishing weights they do
not transfer signal
55(No Transcript)
56GEn
57GEn
58(No Transcript)
59(No Transcript)
60(No Transcript)
61GMn
62(No Transcript)
63(No Transcript)
64(No Transcript)
65GEp
66(No Transcript)
67GMp
68(No Transcript)
69(No Transcript)
70(No Transcript)