Title: Dynamic%20Integration%20of%20Virtual%20Predictors
1Dynamic Integration of Virtual Predictors
- Vagan Terziyan
- Information Technology Research Institute,
University of Jyvaskyla, FINLAND - e-mail vagan_at_it.jyu.fi
- URL http//www.cs.jyu.fi/ai
- CIMA2001/AIDA2001
- Bangor, June 19-22, 2001
2Acknowledgements
Academy of Finland Project (1999) Dynamic
Integration of Classification Algorithms
Information Technology Research Institute
(University of Jyvaskyla) Customer-oriented
research and development in Information
Technology http//www.titu.jyu.fi/eindex.html
Multimeetmobile (MMM) Project (2000-2001) Locati
on-Based Service System and Transaction
Management in Mobile Electronic Commerce
http//www.cs.jyu.fi/mmm
3Acknowledgements also to a Team (external to MMM
project)
Department of Computer Science and Information
Systems, University of Jyvaskyla http//www.cs.jyu
.fi
Alexey Tsymbal
Irina Skrypnik
Seppo Puuronen
4Contents
- The problem
- Virtual Predictor
- Classification Team
- Team Direction
- Dynamic Selection of Classification Team
- Implementation for Mobile e-Commerce
- Conclusion
5Inductive learning with integration of predictors
Learning Environment
Sample Instances
Predictors/Classifiers
P1
P2
...
Pn
yt
6Virtual Classifier
Virtual Classifier is a group of seven
cooperative agents
TC - Team Collector TM - Training Manager TP -
Team Predictor TI - Team Integrator
FS - Feature Selector DE - Distance Evaluator CL
- Classification Processor
7Classification Team Feature Selector
FS - Feature Selector
8Feature Selection
- Feature selection methods try to pick a subset of
features that are relevant to the target concept - Each of these methods has its strengths and
weaknesses based on data types and domain
characteristics - The choice of a feature selection method depends
on various data set characteristics (i) data
types, (ii) data size, and (iii) noise.
9Classification of feature selection methods Dash
and Liu, 1997
10Feature Selector to find the minimally sized
feature subset that is sufficient for correct
classification of the instance
Sample Instances
Sample Instances
Feature Selector
11Classification Team Distance Evaluator
DE - Distance Evaluator
12Use of distance evaluation
- Distance between instances is useful to recognize
nearest neighborhood of any classified instance - Distance between classes is useful to define the
misclassification error - Distance between classifiers is useful to
evaluate weights of every classifier for their
further integration.
13Well known distance functions Wilson Martinez
1997
14Distance Evaluator to measure distance between
instances based on their numerical or nominal
attribute values
Distance Evaluator
15Classification Team Classification Processor
CL - Classification Processor
16Classification Processor to predict class for a
new instance based on its selected features and
its location relatively to sample instances
Feature Selector
Sample Instances
Classification Processor
Distance Evaluator
17Team InstructorsTeam Collector
TC - Team Collector completes Classification
Teams for training
18Team Collector - completes classification teams
for future training
Distance Evaluation functions
Classification rules
Feature Selection methods
Team Collector
FSi
DEj
CLk
19Team InstructorsTraining Manager
TM - Training Manager trains all completed teams
on sample instances
20Training Manager - trains all completed teams on
sample instances
Training Manager
CLk1
FSi1
DEj1
Sample Instances
Sample Metadata
CLk2
FSi2
DEj2
CLkn
FSin
DEjn
Classification Teams
21Team InstructorsTeam Predictor
TP - Team Predictor predicts weights for every
classification team in certain location
22Team Predictor - predicts weights for every
classification team in certain location
Predicted weights of classification teams
Location
Training Manager e.g. WNN algorithm
Sample Metadata
23Team Predictor - predicts weights for every
classification team in certain location
Sample metadata
NN2
ltw21, w22,, w2ngt
ltw31, w32,, w3ngt
NN3
d2
NN1
ltw11, w12,, w1ngt
d3
d1
Pi
dmax
ltwi1, wi2,, wingt
NN4
ltw41, w42,, w4ngt
wij F(w1j, d1, w2j, d2, w3j, d3, dmax)
24Team Prediction Locality assumption
Each team has certain subdomains in the space of
instance attributes, where it is more reliable
than the others This assumption is supported by
the experiences, that classifiers usually work
well not only in certain points of the domain
space, but in certain subareas of the domain
space Quinlan, 1993 If a team does not work
well with the instances near a new instance, then
it is quite probable that it will not work well
with this new instance also.
25Team InstructorsTeam Integrator
TI - Team Integrator produces classification resul
t for a new instance by integrating appropriate
outcomes of learned teams
26Team integrator - produces classification result
for a new instance by integrating appropriate
outcomes of learned teams
Weights of classification teams in the location
of a new instance
New instance
CLk1
yt1
FSi1
DEj1
CLk2
yt2
FSi2
DEj2
yt
Team Integrator
CLkn
yt1
FSin
DEjn
Classification teams
27Simple case static or dynamic selection of a
classification team from two
Assume that we have two different classification
teams and they have been learned on a same sample
set with n instances. Let the first team
classifies correctly m1, and the second one m2
sample instances respectively. We consider two
possible cases to select the best team for
further classification a static selection case
and a dynamic selection case.
28Static Selection
- Static selection means that we try all teams on a
sample set and for further classification select
one, which achieved the best classification
accuracy among others for the whole sample set.
Thus we select a team only once and then use it
to classify all new domain instances.
29Dynamic Selection
- Dynamic selection means that the team is being
selected for every new instance separately
depending on where this instance is located. If
it has been predicted that certain team can
better classify this new instance than other
teams, then this team is used to classify this
new instance. In such case we say that the new
instance belongs to the competence area of that
classification team.
30Theorem
- The average classification accuracy in the case
of (dynamic) selection of a classification team
for every instance is expected to be not worse
than the one in the case of (static) selection
for the whole domain. - The accuracy of these two cases can be equal if
and only if
where k is amount of instances correctly
classified by both teams
31Competence areas of classification teams in
dynamic selection
n instances
m2 instances
m1 instances
k instances
32M-Commerce LBS systemhttp//www.cs.jyu.fi/mmm
In the framework of the Multi Meet Mobile (MMM)
project at the University of Jyväskylä, a LBS
pilot system, MMM Location-based Service system
(MLS), has been developed. MLS is a general LBS
system for mobile users, offering map and
navigation across multiple geographically
distributed services accompanied with access to
location-based information through the map on
terminals screen. MLS is based on Java, XML and
uses dynamic selection of services for customers
based on their profile and location.
33Architecture of LBS system
34Sample from the location-based services access
history
Mobile customer description
Ordered service
35Adaptive interface for MLS client
Only predicted services, for the customer with
known profile and location, will be delivered
from MLS and displayed at the mobile terminal
screen as clickable points of interest
36Conclusion
- Knowledge discovery with an ensemble of
classifiers is known to be more accurate than
with any classifier alone e.g. Dietterich,
1997. - If a classifier somehow consists of certain
feature selection algorithm, distance evaluation
function and classification rule, then why not to
consider these parts also as ensembles making a
classifier itself more flexible? - We expect that classification teams completed
from different feature selection, distance
evaluation, and classification methods will be
more accurate than any ensemble of known
classifiers alone, and we focus our research and
implementation on this assumption.