Bayesian belief networks 2. PCA and ICA - PowerPoint PPT Presentation

1 / 22
About This Presentation
Title:

Bayesian belief networks 2. PCA and ICA

Description:

PCA 6. ... PCA 7. ... To calculate easily the PCA we can use the iterative algorithm. ... – PowerPoint PPT presentation

Number of Views:45
Avg rating:3.0/5.0
Slides: 23
Provided by: fda31
Category:
Tags: ica | pca | bayesian | belief | networks | pca

less

Transcript and Presenter's Notes

Title: Bayesian belief networks 2. PCA and ICA


1
Bayesian belief networks2. PCA and ICA
Peter Andras andrasp_at_ieee.org
2
Principal component analysisPCA 1.
Idea the high dimensional data might be situated
on a lower dimensional surface.
3
PCA 2.
How to find the lower dimensional surface ?
We look for linear surfaces, i.e., hyperplanes.
We decompose the correlation matrix of data
conform its eigenvectors.
4
PCA 3.
The eigenvectors are called principal component
vectors. The new data vectors are formed by the
projections of the original data vectors onto the
principal component vectors.
5
PCA 4.
are the data vectors
The correlation matrix is
6
PCA 5.
The eigenvectors are determined by the equation
where ? is a real number.
Example with two eigenvectors
7
PCA 6.
In principle we should find d eigenvectors if the
dimensionality of the data vectors is d. If the
data vectors are situated on a lower dimensional
linear surface we find less than d eigenvectors
(i.e., the determinant of the correlation matrix
is zero).
8
PCA 7.
If v1, v2, , vm, mltd, are the eigenvectors of R
then the new, transformed data vectors are
calculated as
9
PCA 8.
How to calculate the eigenvectors of R ?
First method use standard matrix algebra
methods. (it is very laborious)
Second method iterative calculation of the
eigenvectors inspired by artificial neural
networks.
10
PCA 9.
Iterative calculation of the eigenvectors
Let w1 ?Rd a randomly chosen vector, such that
w11
Perform iteratively the calculation
where yiw1Txi and ? is a learning constant.
The algorithm converges to the eigenvector
corresponding to the largest eigenvalue (?).
11
PCA 10.
To calculate the following eigenvectors we modify
the iterative algorithm. Now we use the
calculation formula
where
and ujiwjTxi.
This iterative algorithm converges to wk the k-th
eigenvector.
12
PCA 11.
If the algorithm doesnt converge the situation
can be a. the vector enters in a cycle b. the
values doesnt form any cycle.
If we have a cycle, all the vectors of the cycle
are eigenvectors, and their corresponding
eigenvalues are very close.
If we have no convergence and no cycle, that
means that there is no more eigenvector that can
be determined.
13
PCA 12.
How to use the PCA for dimension reduction ?
Select the important eigenvectors. Many times
all of the eigenvectors can be determined but
only part of them are important. The importance
of the eigenvectors is shown by their associated
eigenvalue.
14
PCA 13.
Selecting the important eigenvectors.
1. Graphical method
15
PCA 14.
Selecting the important eigenvectors.
2. Relative power
3. Cumulative power
16
PCA 15.
Summary
The PCA is used for dimensionality
reduction. The data vectors are projected on the
eigenvectors of their correlation matrix to
obtain the transformed data vectors. To
calculate easily the PCA we can use the iterative
algorithm. To reduce the data dimension we
consider only the important eigenvectors.
17
Independent component analysisICA 1.
The idea if the data vectors are linear
combination of statistically independent data
components, they should be separable in their
components. This is true if the component
vectors have non-Gaussian distribution, with
sharper or flatter peak.
18
ICA 2.
Suppose xiAsi, where xi are the data vectors, si
are the vectors of statistically independent
components (sji) Our goal is to find the matrix A
(more precisely, the rows of it). Example
cocktail-party effect many independent voices
registered together goal separate the
independent voices the recorded mixture is a
linear mixture.
19
ICA 3.
How to find the independent components ?
Optimize
All solution vectors (w) are local minimum
solutions, and they correspond to one of the
independent components, i.e., on the components
of the si vectors.
20
ICA 4.
How to do it practically ?
FastICA algorithm (Hyvarinen and Oja)
Calculates by iterations the w vectors. The
calculation formula is
w converges to one of the vectors corresponding
to one of the independent components.
21
ICA 5.
In practice we have to calculate several w
vectors. To test whether the generated
independent components are really independent we
can use statistical tests. Let us consider
s1iw1Txi and s2iw2Txi. Then we can test the
independence of s1 and s2 by calculating their
correlation and testing their identical origin by
the F-test (they may not be strongly correlated
but at the same time they may have identical
origin). If the testing accepts the independence
of the two series we may accept w2 as a new
vector that corresponds to a separate independent
component.
22
ICA 6.
Remarks
By calculating the independent components we get
a new representation of the data, which has the
property that the components contain minimum
mutual information. We can use the ICA to select
the independent non-Gaussian components, but we
cannot separate the Gaussian mixtures.
Write a Comment
User Comments (0)
About PowerShow.com