Title: Learning
1Learning
- Leslie Pack Kaelbling
- Computer Science and Artificial Intelligence
Laboratory - Massachusetts Institute of Technology
2What is learning?
- Learning denotes changes in the system that are
adaptive in the sense that they enable the system
to do the task or tasks drawn from the same
population more efficiently and more effectively
the next time. -- Herb Simon
3Learning is crucial
- Any system that is connected to a real external
environment needs to learn - Humans, even if expert, are inarticulate
- Single system should be adaptable to multiple
environments - Environments change over time
4Some machine learning successes
- assessing loan credit risk
- detecting cell phone fraud
- improved manufacturing electric pumps, steel
rolling, rotogravure, separation of gas and oil - cataloging astronomical images
- helping NBA coaches analyze performance
- personalizing news and web searches
- steering an autonomous car across the US
5Supervised learning
- Given data (training set)
- Goal find a hypothesis h in hypothesis class H
that does a good job of mapping x to y
input
output
Classification discrete Y Regression continuous
Y
6Spam filtering
- Represent document as a bag of words
- How many times has each word occurred in the
document? - x lt 0, 0, 0, 1, 0, 0, 2, 0, 0, 4, 0, 0, 0,
1, 0, 1, gt - Label is it spam or not?
- Classifier given a new document, predict whether
its spam
7Attribute vectors
- Is this an image of a face?
- ltpixel1, pixel2, pixel3, gt
- Will John be late to the meeting?
- lt time, location, topic, preceding appt, gt
- Will my proposal be funded?
- x lt dollar-amount, num-pis, topic, length,
font, gt
8Memory
yes
no
No
9Statistics
yes
no
Yes
10Generalization
?
?
11Hypothesis
yes
blue?
oval?
yes
no
big?
no
no
yes
12Decision Tree
blue?
oval?
yes
big?
no
?
no
yes
13Decision Tree
blue?
oval?
yes
big?
no
?
no
yes
14Whats the right hypothesis?
15Whats the right hypothesis?
16Now, whats the right hypothesis?
17Now, whats the right hypothesis?
18How about now?
19How about now? Answer 1
20How about now? Answer 2
21How about now?
22How about now?
23No Free Lunch
- Unless you know something about the distribution
of problems your learning algorithm will
encounter, any hypothesis that agrees with all
your data is as good as any other. - You cant learn anything unless you already know
something.
24Supervised learning is reasonably well solved
- Learning methods differ in terms of
- the form of the hypothesis
- the way the computer finds a hypothesis given the
data - Variety of methods
- nearest neighbor
- decision trees
- neural networks
- support vector machines
25Support vector machines
- Based on two important technical ideas
- some linear separators are better than others
- complex separators in the original space are
linear separators in a complex space - Similar performance to multi-layer neural nets,
but no local optima (much easier to apply) - Computationally expensive, but manageable
26SVMs Maximize the margin
27SVMs Maximize the margin
28SVMs Maximize the margin
29SVMs Maximize the margin
30SVMs Kernel space
- Data are not separable in one dimension
x
31SVMs Kernel space
- Data are separable in ltx, x2gt space
x2
x
32SVMs Kernel space
- Data are separable in ltx, x2gt space
x2
x
33SVMs Kernel space
- Clever mathematical kernel trick allows us to
find - maximum margin separator
- in many high or infinite-dimensional spaces
- in time (empirically) polynomial in the number of
data points - no local optima!
34Framing a Problem
- Clever human needs to
- decide what information about the examples is
important to the prediction problem - choose an encoding of that information
- choose a class of hypotheses that is likely to
contain a reasonably good solution - gather training data
- set parameters in the learning algorithm
- finally, run the learning algorithm on the data
to get a prediction rule
35So is all of learning well solved?
36Personal assistant that learns
- You have a meeting with Bob at 2 today. Are you
going to talk about the budget or the system
architecture? - Architecture. Could you get me the architecture
slide Alice showed at last weeks design meeting?
- Here it is. By the way, you might want to
consider taking the metro to the meeting it
looks like traffic is badly backed up downtown. - I never take the metro. But maybe you can
suggest the best driving directions given the
current traffic. - It looks like going around the beltway is the
best option
37PAL has to learn
- Your world meetings, projects, participants,
rooms, locations, topics, and how they are all
related - How to perceptually recognize people, locations,
activities people are engaging in - How to extract information from documents about
your world - Your practices and preferences how you commute,
what the standard office procedures are
38What makes learning in PAL hard?
- Huge sets of labeled training examples not
available - Large variety of possible tasks
- Learning problems not framed explicitly in
advance by humans
39Four technological problems
- learning with much less labeled data
- using richer representations of situations and
hypotheses - learning to behave in complex environments
- life-long and life-sized learning
40Four technological problems
- learning with much less labeled data
- using richer representations of situations and
hypotheses - learning to behave in complex environments
- life-long and life-sized learning
41Using unlabeled data
- Most of generalization depends on a notion of
distance similar objects should have similar
properties - Labeled data is expensive unlabeled data is
cheap - Use unlabeled data to learn underlying properties
of the data space
42Unlabeled data
43Underlying manifold
44A few labeled examples
45Generalization
46Four technological problems
- learning with much less labeled data
- using richer representations of situations and
hypotheses - learning to behave in complex environments
- life-long and life-sized learning
47Learning richer knowledge
- Standard learning methods do not represent or
learn relational knowledge - Susan is Joes boss
- Susan works on the CALO project
- A purchase order must be signed by the
purchasers boss - Its okay to schedule a meeting of a project if
all but one of its members can attend
48Learning logical representations
- Vector-space representation is effective but
limiting - Inductive logic programming learns logical rules
- alpha0(A,B) -
- position(A,E,O), not_aromatic(O),
- small_or_polar(O), position(A,B,C),
very_hydrophobic(C), - can_sign_proposal(A, B) -
- higher_in_org_chart(A, B),
- no_conflict_of_interest(A, B)
49Learning probabilistic representations
- Bayesian networks represent probabilistic
relations among state variables - gene regulatory networks
- observations, locations in mapping
- diseases and symptoms, computer diagnosis,
50Probabilistic relational models
- Learn uncertain relations between properties of
individuals, independent of particular
individuals
51Instantiating the model
- boss(John) Jane
- boss(Mary) Jane
- boss(Jane) Pat
John is rich
John is happy
Jane is rich
Jane is happy
Mary is happy
Pat is happy
Mary is rich
52Broad range of applications
- disease transmission
- paper citations
- vehicle tracking
- gene expression
53Dynamics of complex worlds
- Agents actions change relations among objects
- call(Person, Message)
- urgent(Message),inMeeting(Person) ?
0.7 knows(Person, Message), happy(Person) 0.2
knows(Person, Message), angry(Person) 0.05 angry(P
erson) 0.05 nothing changes
54Four technological problems
- learning with much less labeled data
- using richer representations of situations and
hypotheses - learning to behave in complex environments
- life-long and life-sized learning
55Reinforcement learning
- given a connection to the environment
- find a behavior that maximizes long-run
reinforcement
State
Reinforcement
Observation
Action
56Why reinforcement learning?
- Supervision signal is rarely available to agents
- Reward is easier than behavior for humans to
specify - for transmitting message
- - for emailing
- - - for phoning in office
- - - - for phoning during meeting
57Reinforcement learning is hard
- less information per training instance
- requires active exploration
- actions have long-term consequences
- on-line performance important
58RL successes
- backgammon player ties human world chamption
- elevator scheduling
- cell-phone channel allocation
- network routing
- All in simulation
- need too much data for online use in real domains
59Learning by watching
- Initial exploration phases intolerably long
- Get help from humans
- built-in reflexes
- demonstration
- leading the robot by the hand
- declarative advice
60Learning a world model
- Let your hypotheses die in your stead. Popper
- learn a model
- mentally simulateexperience
- no bad consequences
- transfer to other tasks
61Four technological problems
- learning with much less labeled data
- using richer representations of situations and
hypotheses - learning to behave in complex environments
- life-long and life-sized learning
62Lifelong learning
- Knowledge learned today provides inductive
leverage for learning tomorrow - Once agent can recognize people, it can analyze
social structure of meetings - Once agent learns the boss relation, it can
generalize rule for who can sign travel claims
63Learning tasks interconnect
- Natural language techniques to extract relational
information from text - Prior information about relations helps
disambiguate text - Perceptual recognition should be primed by
expectations from relational knowledge (relative
positions of objects, attendees of meetings)
64Learning to live
- You are simultaneously trying to
- remain nourished
- retain your job
- have a good time
- not fall asleep in this talk
- You know about
- dancing
- differential equations
- donuts
- dinosaurs
65Learning to live
- You are simultaneously trying to
- remain nourished
- retain your job
- have a good time
- not fall asleep in this talk
- You know about
- dancing
- differential equations
- donuts
- dinosaurs
How can you possibly decide what to do next?
66Performance curves
computers
performance
humans
life
domain complexity
67Dynamic problem reformulation
tractable sub-problem
perception
action
68Multiple-resolution plans
Fine view of near-term high-probability
events Coarse view of distant low-probability
events
69Learning applications
70Learning to perceive
- Virtually every perceptual system built today
relies on learning - vision person and face recognition,
segmentation, activity recognition - speech recognition of phonemes, words, sentence
structures - language grammatical models, word-sense
disambiguation, named-entity extraction - None of these systems could have been built by
hand.
71Learning in computer systems
- active document retrieval
- repetitive editing by example
- scheduling machine instructions to optimize
program execution time - adaptive routing
- system policies
- when to spin hard disk up /down
- when to turn wireless transmitter on/off
- where to cache data in distributed system
72Learning is the future
- Learning techniques will be a basis for every
application that involves a connection to a real
world - Basic learning algorithms are ready for use in
limited applications today - Prospects for broader future application make for
exciting fundamental research and development
opportunities
73(No Transcript)