Title: Change%20Detection%20in%20Stochastic%20Shape%20Dynamical%20Models%20with%20Application%20to%20Activity%20Recognition
1Change Detection in Stochastic Shape Dynamical
Models with Application to Activity Recognition
- Namrata Vaswani
- Thesis Advisor Prof. Rama Chellappa
2Acknowledgements
- Part of this work is joint work with Dr Amit Roy
Chowdhury and Prof Rama Chellappa.
3The Group Activity Recognition Problem
4Problem Formulation
- The Problem
- Model activities performed by a group of moving
and interacting objects (which can be people or
vehicles or robots or diff. parts of human body).
Use the models for abnormal activity detection
and tracking - Our Approach
- Treat objects as point objects landmarks.
- Changing configuration of objects deforming
shape - Abnormality change from learnt shape dynamics
- Related Approaches for Group Activity
- Co-occurrence statistics, Dynamic Bayes Nets
5The Framework
- Define a Stochastic State-Space Model (a
continuous state HMM) for shape deformations in a
given activity, with shape scaled Euclidean
motion forming the hidden state vector and
configuration of objects forming the observation.
- Use a particle filter to track a given
observation sequence, i.e. estimate the hidden
state given observations. - Define Abnormality as a slow or drastic change in
the shape dynamics with unknown change
parameters. We propose statistics for slow
drastic change detection.
6Overview
- The Group Activity Recognition Problem
- Slow and Drastic Change Detection
- Landmark Shape Dynamical Models
- Applications, Experiments and Results
- Principal Components Null Space Analysis
- Future Directions Summary of Contributions
7A Group of People Abnormal Activity Detection
Normal Activity
Abnormal Activity
8Human Action Tracking
Cyan Observed Green Ground Truth Red SSA
Blue NSSA
9A Group of Robots
10Applications to Speech/Audio Processing?
- Slow change detection
- Shape Dynamical Models Entire framework
applicable to observations coming from acoustic
sensors, only observation model will change - Sensor fusion is easy combine audio/video
- PCNSA an algorithm for high dimensional pattern
classification, applicable here as well
11Overview
- The Group Activity Recognition Problem
- Slow and Drastic Change Detection
- Landmark Shape Dynamical Models
- Applications, Experiments and Results
- Principal Components Null Space Analysis
12The Problem
- General Hidden Markov Model (HMM) Markov state
sequence Xt, Observation sequence Yt - Finite duration change in system model which
causes a permanent change in probability
distribution of state - Change is slow or drastic Tracking Error
Observation Likelihood do not detect slow
changes. Use dist. of Xt - Change parameters unknown use Log-Likelihood(Xt)
- State is partially observed use MMSE estimate of
LL(Xt) given observations, ELL(Xt)Y1t) ELL - Nonlinear dynamics Particle filter to estimate
ELL
13The General HMM
Qt(Xt1Xt)
State, Xt
State, Xt1
Xt2
?t(YtXt)
Observation, Yt1
Observation, Yt
Yt2
14Related Work
- Change Detection using Particle filters, unknown
change parameters - CUSUM on Generalized LRT (Y1t) assumes finite
parameter set - Modified CUSUM statistic on Generalized LRT
(Y1t) - Testing if uj Pr(YtltyjY1t-1) are uniformly
distributed - Tracking Error (TE) error b/w Yt its
prediction based on past - Threshold negative log likelihood of observations
(OL) - All of these approaches use observation
statistics Do not detect slow changes - PF is stable and hence is able to track a slow
change - Average Log Likelihood of i.i.d. observations
used often - But ELL E-LL(Xt)Y1t (MMSE of LL given
observations) in context of general HMMs is new
15Particle Filtering
16Change Detection Statistics
- Slow Change Propose Expected Log Likelihood
(ELL) - ELL Kerridge Inaccuracy b/w pt (posterior) and
pt0 (prior) - ELL(Y1t )E-log pt0 (Xt)Y1tEp-log pt0
(Xt)K(pt pt0) - A sufficient condition for detectable changes
using ELL - EELL(Y1t0) K(pt0pt0)H(pt0),
EELL(Y1tc) K(ptcpt0) - Chebyshev Inequality With false alarm miss
probabilities of 1/9, ELL detects all changes
s.t. - K(ptcpt0) -H(pt0)gt3 vVarELL(Y1tc)
vVarELL(Y1t0) - Drastic Change ELL does not work, use OL or TE
- OL Neg. log of current observation likelihood
given past - OL -log Pr(YtY0t-1,H0) -logltptt-1 ,
?tgt - TE Tracking Error. If white Gaussian observation
noise, TE OL
17ELL OL Slow Drastic Change
- ELL fails to detect drastic changes
- Approximating posterior for changed system
observations using a PF optimal for unchanged
system error large for drastic changes - OL relies on the error introduced due to the
change to detect it - OL fails to detect slow changes
- Particle Filter tracks slow changes correctly
- Assuming change till t-1 was tracked correctly
(error in posterior small), OL only uses change
introduced at t, which is also small - ELL uses total change in posterior till time t
the posterior is approximated correctly for a
slow change so ELL detects a slow change when
its total magnitude becomes detectable - ELL detects change before loss of track, OL
detects after
18A Simulated Example
- Change introduced in system model from t5 to t15
OL
ELL
19Practical Issues
- Defining pt0(x)
- Use part of state vector which has linear
Gaussian dynamics can define pt0(x) in closed
form - OR
- Assume a parametric family for pt0(x), learn
parameters using training data - Declare a change when either ELL or OL exceed
their respective thresholds. - Set ELL threshold to a little above H(pt0)
- Set OL threshold to a little above
EOL0,0H(YtY1t-1) - Single frame estimates of ELL or OL may be noisy
- Average the statistic or average no. of detects
or modify CUSUM
20Change Detection
Yes
Change (Slow)
ELLEp-log pt0(Xt) gt Threshold?
PF
ptt-1N pttN ptN
Yt
Yes
Change (Drastic)
OL -logltptt-1 , ?tgt gt Threshold?
21Approximation Errors
- Total error lt Bounding error Exact filtering
error PF error - Bounding error Stability results hold only for
bounded fns but LL is unbounded. So approximate
LL by min-log pt0(Xt),M - Exact filtering error Error b/w exact filtering
with changed system model with original model.
Evaluating ptc,0 (using Qt0 ) instead of ptc,c
(using Qtc) - PF Error Error b/w exact filtering with original
model particle filtering with original model.
Evaluating ptc,0,N which is a Monte Carlo
estimate of ptc,0
22Stability / Asymptotic Stability
- The ELL approximation error averaged over
observation sequences PF realizations is
eventually monotonically decreasing ( hence
stable), for large enough N if - Change lasts for a finite time
- Unnormalized filter kernels are mixing
- Certain boundedness (or uniform convergence of
bounded approximation) assumptions hold - Asymptotically stable if the kernels are
uniformly mixing - Use stability results of LeGland Oudjane
- Analysis generalizes to errors in MMSE estimate
of any fn of state evaluated using a PF with
system model error
23Unnormalized filter kernel mixing
- Unnormalized filter kernel, Rt, is state
transition kernel,Qt, weighted by likelihood of
observation given state - Mixing measures the rate at which the
transition kernel forgets its initial condition
or eqvtly. how quickly the state sequence becomes
ergodic. Mathematically, - Example LeGland et al State transition, Xt
Xt-1nt is not mixing. But if Yth(Xt)wt, wt
is truncated noise, then Rt is mixing
24Complementary Behavior of ELL OL
- ELL approx. error, etc,0, is upper bounded by an
increasing function of OLkc,0, tclt k lt t - Implication Assume detectable change i.e.
ELLc,c large - OL fails gt OLkc,0,tcltkltt small gt ELL error,
etc,0 smallgt ELLc,0 large gt ELL detects - ELL fails gt ELLc,0 small gtELL error, etc,0
large gt at least one of OLkc,0,tcltkltt large gt
OL detects
25Rate of Change Bound
- The total error in ELL estimation is upper
bounded by increasing functions of the rate of
change (or system model error per time step)
with all increasing derivatives. - OLc,0 is upper bounded by increasing function of
rate of change. - Metric for rate of change (or equivalently
system model error per time step) for a given
observation Yt DQ,t is
26 The Bound
Assume Change for finite time, Unnormalized
filter kernels mixing, Posterior state space
bounded
27Implications
- If change slow, ELL works and OL does not work
- ELL error can blow up very quickly as rate of
change increases (its upper bound blows up) - A small error in both normal changed system
models introduces less total error than a perfect
transition kernel for normal system large error
in changed system - A sequence of small changes will introduce less
total error than one drastic change of same
magnitude
28Possible Applications
- Abnormal activity detection, Detecting motion
disorders in human actions, Activity Segmentation - Neural signal processing detecting changes in
stimuli - Congestion Detection
- Video Shot change or Background model change
detection - System model change detection in target tracking
problems without the tracker loses track
29Approximation Errors
- Total error lt Bounding error Exact filtering
error PF error - Bounding error Stability results hold only for
bounded fns but LL is unbounded.
BEELLtc,c-ELLtc,c,M - Exact filtering error Error b/w exact filtering
with original system model with changed model,
MEELLtc,c,M-ELLtc,0,M - PF Error Error b/w exact filtering with changed
model particle filtering with changed model,
PEELLtc,0,M-ELLtc,0,M,N
30Asymptotic Stability
- If (i) Change lasts for finite time, (ii)
Unnormalized filter kernels are uniformly mixing,
(iii) Bounded posterior state space increase
of Mt (bound on expected value of LL) with t is
polynomial and (iv) E?t lt ? for all t, then - If (i), (ii), (iii) LL unbounded but expected
value of its bounded approx. converges to true
value uniformly in t and (iv), then - Both 1. 2. imply Error avged. over obs.
sequences PF runs is asymptotically stable
31Stability
- If (i), (ii) Unnormalized filter kernels Mixing,
(iii), then - limN?8(Error avged over obs. seq. PF runs) is
stable ?t eventually strictly decreasing - 2. If (i), (ii) Mixing, (iii)Bounded
posterior state space, - limN?8(Error avged over PF runs)/Mt is stable
almost surely for all obs. seq ?t strictly
decreasing.
32We have shown
- Asymptotic stability of errors in ELL estimation
if change lasts for a finite time, unnormalized
filter kernels are uniformly mixing some
boundedness assumptions hold. - Stability for large N if the kernels are only
mixing - ELL error upper bounded by an increasing fn. of
OLc,0 ELL works when OL fails vice versa - ELL error upper bounded by an increasing fn. of
rate of change, with incr. derivatives of all
orders. OLc,0 upper bounded by increasing fn. of
rate of change - Analysis generalizes to errors in MMSE estimate
of any fn. of state evaluated using a PF with
system model error
33More Practical Issues
- Estimates from single frames are noisy and
affected by outliers - Average the no. of detects over past p time
instants - Or average the statistic over past p time
instants - aOL(p) (1/p) -log Pr(Yt-p1tY0t-p,H0)
- Either avg. ELL, aELL, or use joint ELL over
past p states, jELL(p) (1/p)E-log
pt-pt(Xt-pt)Yot - Or, modify CUSUM for unknown change parameters,
i.e. change if max1ltpltt dp gt ?, dp
Statistic(p)Tp,t.
34Overview
- The Group Activity Recognition Problem
- Slow and Drastic Change Detection
- Landmark Shape Dynamical Models
- Applications, Experiments and Results
- Principal Components Null Space Analysis
35A Group of Human Body Parts
36What is Shape?
- Shape is the geometric information that remains
when location, scale and rotation effects are
filtered out Kendall - Shape of k landmarks in 2D
- Represent the X Y coordinates of the k points
as a k-dimensional complex vector Configuration - Translation Normalization Centered Configuration
- Scale Normalization Pre-shape
- Rotation Normalization Shape
37Activities on the Shape Sphere in Ck-1
38Related Work
- Related Approaches for Group Activity
- Co-occurrence Statistics
- Dynamic Bayesian Networks
- Shape for robot formation control
- Shape Analysis/Deformation
- Pairs of Thin plate splines, Principal warps
- Active Shape Models affine deformation in
configuration space - Deformotion scaled Euclidean motion of shape
deformation - Piecewise geodesic models for tracking on
Grassmann manifolds - Particle Filters for Multiple Moving Objects
- JPDAF (Joint Probability Data Association
Filter) for tracking multiple independently
moving objects
39Motivation
- A generic and sensor invariant approach for
activity - Only need to change observation model depending
on the landmark, the landmark extraction method
and the sensor used - Easy to fuse sensors in a Particle filtering
framework - Shape invariant to translation, zoom,
in-plane rotation - Single global framework for modeling and tracking
independent motion interactions of groups of
objects - Co-occurrence statistics Req. individual joint
histograms - JPDAF Cannot model object interactions for
tracking - Active Shape Models good for only approx. rigid
objects - Particle Filter is better than the Extended
Kalman Filter - Able to get back in track after loss of track due
to outliers, - Handle multimodal system or observation process
40The HMM
- Observation, Yt Centered configurations
- State, Xt?t, ct, st, ?t
- Current Shape (zt),
- Shape Velocity (ct) Tangent coordinate w.r.t.
zt-1 - Scale (st),
- Rotation angle (?t)
-
- Use complex vector notation to simplify equations
- Use a particle filter to approximate the optimal
non-linear filter, pt(dx) Pr(Xt?dxY0t)
posterior state distribution conditioned on
observations upto time t, by an N-particle
empirical estimate of pt
41Hidden Markov Shape Model
XtShape(zt), Shape Velocity(ct), Scale(st),
Rotation(?t)
Xt1
State Dynamics
Observation Model Yt h(Xt) wt ztstej?
wt wt i.i.d observation noise
Using complex notation
Observation, Yt Centered Configuration
Shape X Rotation, SO(2) X Scale, R Centered
Config, Ck-1
42State Dynamics
- Shape Dynamics Linear Markov model on shape
velocity - Shape velocity at t in tangent space w.r.t.
shape at t-1, zt-1 - Orthogonal basis of the tangent space, U(zt-1)
- Linear Gauss-Markov model for shape velocity
- Move zt-1 by an amount vt on shape manifold to
get zt - Motion (Scale, Rotation)
- Linear Gauss-Markov dynamics for log st,
unwrapped ?t
43The HMM
Observation Model Shape,Motion?Centered
Configuration
System Model Shape and Motion Dynamics
Motion Dynamics
Shape Dynamics
- Linear Gauss-Markov models for log st and ?t
- Can be stationary or non-stationary
44Three Cases
- Non-Stationary Shape Activity (NSSA)
- Tangent space, U(zt-1), changes at every t
- Most flexible Detect abnormality and also track
it - Stationary Shape Activity (SSA)
- Tangent space, U(µ), is constant (µ is a mean
shape) - Track normal behavior, detect abnormality
- Piecewise Stationary Shape Activity (PSSA)
- Tangent space is piecewise constant, U(µk)
- Change time fixed or decided on the fly using
ELL - PSSA ELL Activity Segmentation
45Stationary, Non-Stationary
Stationary Shape
Non-Stationary Shape
46Stationary Shape Activity
- Mean shape is constant, so set ?t µ (Procrustes
mean), for all t, ?t not part of state vector,
learn mean shape using training data. - Define a single tangent space w.r.t. µ shape
dynamics simplifies to linear Gauss-Markov model
in tangent space - Since shape space is not a vector space, data
mean may not lie in shape space, evaluate
Procrustes mean an intrinsic mean on the shape
manifold.
47What is Procustes Mean?
- Proc mean µ, minimizes sum of squares of Proc
distances of the set of pre-shapes from itself - Proc distance is Euclidean distance between
Proc fit of one pre-shape onto another - Proc fit scale or rotate a pre-shape to
optimally align it with another pre-shape - Optimally minimum Euclidean distance between
the two pre-shapes after alignment
48Learning Procrustes Mean
- Procrustes mean of set of preshapes wi
Dryden,Mardia
49Procrustes DistanceDryden Mardia
- Translation Scale normaln. of config. ?
Pre-Shape - Procrustes fit of pre-shape w onto y
- Procrustes distance
50Learning Stationary Shape Dynamics
Learn Procrustes mean, µ
µ
Learn AR model in tangent space
µ, Sv, A, Sn
51Overview
- The Group Activity Recognition Problem
- Slow and Drastic Change Detection
- Landmark Shape Dynamical Models
- Applications, Experiments and Results
- Principal Components Null Space Analysis
52Abnormal Activity Detection
- Define abnormal activity as
- Slow or drastic change in shape statistics with
change parameters unknown. - System is a nonlinear HMM, tracked using a PF
- This motivated research on slow drastic change
detection in general HMMs - Tracking Error detects drastic changes. We
proposed a statistic called ELL for slow change. - Use a combination of ELL Tracking Error and
declare change if either exceeds its threshold.
53Tracking to obtain observations
- CONDENSATION tracker framework
- State Shape, shape velocity, scale, rotation,
translation, Observation Configuration vector - Measurement model Motion detection locally
around predicted object locations to obtain
observation - Predicted object configuration obtained by
prediction step of Particle filter - Predicted motion information can be used to move
the camera (or any other sensor) - Combine with abnormality detection for drastic
abnormalities will not get observation for a set
of frames, if outlier then only for 1-2 frames
54Activity Segmentation
- Use PSSA model for tracking
- At time t, let current mean shape µk
- Use ELL w.r.t. µk to detect change time, tk1
(segmentation boundary) - At tk1, set current mean shape to posterior
Procrustes mean of current shape, i.e. - µk1largest eigenvector of EpztztSi1N
zt(i) zt(i) - Setting the current mean as above is valid only
if tracking error (or OL) has not exceeded the
threshold (PF still in track)
55A Common Framework for
- Tracking
- Groups of people or vehicles
- Articulated human body tracking
- Abnormal Activity Detection / Activity Id
- Suspicious behavior, Lane change detection
- Abnormal action detection, e.g. motion disorders
- Human Action Classification, Gait recognition
- Activity Sequence Segmentation
- Fusing different sensors
- Video, Audio, Infra-Red, Radar
56Possible Applications
- Modeling group activity to detect suspicious
behavior - Airport example
- Lane change detection in traffic
- Model human actions track a given sequence of
actions, detect abnormal actions (medical
application to detect motion disorders) - Activity sequence segmentation unsupervised
training - Sensor independent Acoustic/Radar/Infra-Red,
Sensor fusion - Low level processing for Dynamic Bayesian
Networks - Medical Image Processing, e.g. Shape deformation
models for heart
57Apply to Gait Verification
- Model diff. parts of human body head, torso,
fore hind arms legs as landmarks - Learn the landmark shape dynamics for diff.
peoples gait - Verification Given a test seq. a possible
match (say, from face recognition stage), verify
if match is correct - Start tracking test seq. using the shape
dynamical model of the possible match - If dynamics does not match at all, PF will lose
track - If dynamics is close but not correct, ELL w.r.t.
the possible match will exceed its threshold
58Gait Recognition
- System Identification approach
- Assuming the test seq. has negligible observation
noise, learn the shape dynamical model parameters
for the test sequence - Find distance of parameters for test seq. from
those for diff people in the database. Similar
idea to Soatto, Ashok - Match time series of shape velocity of probe
gallery - Save the shape velocity sequence for the diff
people in database. - Test sequence estimate the shape velocity seq.,
use DTW Kale to match it against all peoples
gait
59Experiments and Results
60Why Particle Filter?
- N does not increase (much) with incr. state dim.,
Approx. posterior distrib. of state only in high
probability regions (so fixed N works for all t
state space at t Dt) - Better than Extended KF because of asymptotic
stability - Able to track in spite of wrong initial
distribution - Get back in track after losing track due to an
outlier observation - Slowly changing system Able to track it and yet
detect the change using ELL (explained later) - Can handle Multi-Modal prior/posterior, EKF cannot
61Time-Varying No. of Landmarks?
- Ill posed problem Interpolate the curve formed
by joining the landmarks re-sample it to a
fixed no. of landmarks k - Experimented with 2 interpolation/re-sampling
schemes - Uniform Re-samples independently along x y
- Assumes observed landmarks are uniformly sampled
from some continuous function of a dummy variable
s - All observed landmark get equal weight while
re-sampling - Very sensitive to change in of landmarks, but
also able to detect abnormality caused by two
closely spaced points - Arc-length Parameterizes x y coordinates by
the length of the arc upto that landmark - Assumes observed landmarks are non-uniformly
sampled points from a continuous fn. of length x
(l), y (l) - Smoothens out motion of closely spaced points,
thus misses abnormality caused by two closely
spaced points
62Experiments
- Group Activity
- Normal activity Group of people deplaning
walking towards airport terminal used SSA model - Abnormality A person walks away in an un-allowed
direction distorts the normal shape - Simulated walking speeds of 1,2,4,16,32 pixels
per time step (slow to drastic distortion in
shape) - Compared detection delays using TE and ELL
- Plotted ROC curves to compare performance
- Human actions
- Defined NSSA model for tracking a figure skater
- Abnormality abnormal motion of one body part
- Able to detect as well as track slow abnormality
63Abnormality
- Abnormality introduced at t5
- Observation noise variance 9
- OL plot very similar to TE plot (both same to
first order)
Tracking Error (TE)
ELL
64ROC ELL
- Plot of Detection delay against Mean time b/w
False Alarms (MTBFA) for varying detection
thresholds - Plots for increasing observation noise
Drastic Change ELL Fails
Slow Change ELL Works
65ROC Tracking Error(TE)
- ELL Detection delay 7 for slow change ,
Detection delay 60 for drastic - TE Detection delay 29 for slow change,
Detection delay 4 for drastic
Slow Change TE Fails
Drastic Change TE Works
66ROC Combined ELL-TE
- Plots for observation noise variance 81
(maximum) - Detection Delay lt 8 achieved for all rates of
change
67Human Action Tracking
Normal Action SSA better than NSSA
Abnormality NSSA works, SSA fails
Green Observed, Magenta SSA, Blue NSSA
68NSSA Tracks and Detects Abnormality
Abnormality introduced at t20
Tracking Error
ELL
Red SSA, Blue NSSA
69Temporal Abnormality
- Abnormality introduced at t 5, Observation Noise
Variance 81 - Using uniform re-sampling, Not detected using
arc length
70Overview
- The Group Activity Recognition Problem
- Slow and Drastic Change Detection
- Landmark Shape Dynamical Models
- Applications, Experiments and Results
- Principal Components Null Space Analysis
71Typical Data Distributions
Apples from Apples problem All algorithms work
well
Apples from Oranges problem Worst case for
SLDA, PCA
72PCNSA Algorithm
- Subtract common mean µ, Obtain PCA space
- Project all training data into PCA space,
evaluate class mean, covariance in PCA space µi,
Si - Obtain class Approx. Null Space (ANS) for each
class Mi trailing eigenvectors of Si - Valid classification directions in ANS if
distance between class means is significant
WiNSA - Classification Project query Y into PCA space,
XWPCAT(Y- µ), choose Most Likely class, c, as
73Classification Error Probability
- Two class problem. Assumes 1-dim ANS, 1 LDA
direction - Generalizes to M dim ANS and to non-Gaussian but
unimodal symmetric distributions
74Applications
- Image Video retrieval
- Applied to human action retrieval
- Hierarchical image/video retrieval PCNSA
followed by LDA - Activity Classification Abnormal Activity
Detection
75Applications
Face recognition, Large pose variation
Object recognition
Face recognition, Large expression variation
Facial Feature Matching
76Discussion Ideas
- PCNSA test approximates the LRT (optimal Bayes
solution) as condition no. of Si tends to
infinity - Fuse PCNSA and LDA get an algorithm very similar
to Multispace KL - For multiclass problems, use error probability
expressions to decide which of PCNSA or SLDA is
better for a given set of 2 classes - Perform facial feature matching using PCNSA, use
this for face registration followed by warping to
standard geometry
77Ongoing and Future Work
- Change Detection
- Implications of Bound on Errors is increasing fn.
of rate of change - CUSUM on ELL OL
- Quantitative performance analysis of ELL OL
- Find examples of mixing unnormalized filter
kernels - Non-Stationary Piecewise Stationary Shape
Activities - Application to sequences of different kinds of
actions - PSSA ELL for activity segmentation
- Joint tracking and abnormality detection
- Time varying number of Landmarks?
- What is best strategy to get a fixed no. k
of landmarks? - Can we deal with changing dimension of shape
space? - Multiple Simultaneous Activities, Multi-sensor
fusion - 3D Shape, General shape spaces
78Contributions
- ELL for slow change detection, Stability of ELL
approximation error - Complementary behavior of ELL OL, ELL error
proportional to rate of change with all
increasing derivatives - Stochastic dynamical models for landmark shapes
NSSA, SSA, PSSA - Modeling the changing configuration of a group of
moving point objects as a deforming shape shape
activity. - Using ELL PSSA for activity segmentation
- PCNSA its error probability analysis,
application to action retrieval, abnormal
activity detection
79Contributions
- Slow and drastic change detection in general HMMs
using particle filters. We have shown - Asymptotic stability / stability of errors in ELL
approximation - Complementary behavior of ELL OL for slow
drastic changes - Upper bound on ELL error is an increasing
function of rate of change, with all increasing
derivatives - Stochastic state space models (HMMs) for
simultaneously moving and deforming shapes. - Stationary, non-stationary p.w. stationary
cases - Group activity human actions modeling,
detecting abnormality - NSSA for tracking slow abnormality, ELL for
detecting it - PSSA ELL Apply to activity segmentation
80Other Contributions
- A linear subspace algorithm for pattern
classification motivated by PCA - Approximates the optimal Bayes classifier for
Gaussian pdfs with unequal covariance matrices. - Useful for apples from oranges type problems.
- Derived tight upper bound on its classification
error probability - Compared performance with Subspace LDA both
analytically experimentally - Applied to object recognition, face recognition
under large pose variation, action retrieval. - Fast algorithms for infra-red image compression
81Special Cases
- For i.i.d. observation sequence, Yth(Xt)wt
- ELL(Y0t)E-log pt(Xt)Y0tE-log pt(Xt)Yt
- -log pt(h-1(Yt)-Eh-1(wt))- log
pt(h-1(Yt)) - OL(Yt)const. if Eh-1(wt)0
- For zero (negligible) observation noise case,
Yth(Xt) - ELL(Y0t) E-log pt(Xt)Y0t -log
pt(h-1(Yt))OL(Yt)const.
82Particle Filtering Algorithm
- At t0, generate N Monte Carlo samples from
initial state distribution, p0p0 - For all t,
- Prediction Given posterior at t-1 as an
empirical distr., pt-1, sample from the state
transition kernel Qt (xt-1,dxt) to generate
samples from ptt-1 - Update/Correction
- Weight each sample of ptt-1 by probability of
the observation given that sample, ?t(Ytx) pt - Use multinomial sampling to resample from these
weighted particles to generate particles
distributed according to pt
83Classification and Tracking Algorithms Using
Landmark Shape Analysis and their Application to
Face and Gait
- Namrata Vaswani
- Dept. of Electrical Computer Engineering
- University of Maryland, College Park
- http//www.cfar.umd.edu/namrata
84Principal Component Null Space Analysis (PCNSA)
for Face Recognition
85Related Work
- PCA uses projection directions with maximum
inter-class variance but do not minimize the
intra-class variance - LDA uses directions that maximize the ratio of
inter-class variance and intra-class variance - Subspace LDA for large dimensional data, use PCA
for dim. reduction followed by LDA - Multi-space KL (similar to PCNSA)
- Other work ICA, Kernel PCA LDA, Neural nets
86Motivation
- Example PCA or SLDA good for face recognition
under small pose variation, PCNSA proposed for
larger pose variation - PCNSA addresses such Apples from Oranges type
classification problems - PCA assumes classes are well separated along all
directions in PCA space Si Ss2I - SLDA assumes all classes have similar directions
of min. max. variance Si S, for all i - If minimum variance direction for one class is
maximum variance for the other a worst case for
SLDA or PCA
87Assumptions Extensions
- Assumptions required For each class,
- (i) an approx. null space exists,
- (ii) valid classification directions exist
- Progressive-PCNSA
- Defines a heuristic for choosing dimension of ANS
when (ii) is not satisfied. - Also defines a heuristic for new (untrained)
class detection
88Experimental Results
- PCNSA misclassified least, followed by SLDA and
then PCA - New class detection ability of PCNSA was better
- PCNSA most sensitive to training data size, PCA
most robust