Human-level AI - PowerPoint PPT Presentation

About This Presentation
Title:

Human-level AI

Description:

Representation, learning, and prior knowledge. Vision and NLP coming soon. Decision architectures and ... Humans are more intelligent than pigeons ... – PowerPoint PPT presentation

Number of Views:22
Avg rating:3.0/5.0
Slides: 23
Provided by: russ158
Learn more at: https://cs.stanford.edu
Category:
Tags: human | level

less

Transcript and Presenter's Notes

Title: Human-level AI


1
Human-level AI
  • Stuart Russell
  • Computer Science Division
  • UC Berkeley

2
  • Everything Josh said

3
  • Almost everything Yann said

4
Outline
  • Representation, learning, and prior knowledge
  • Vision and NLP coming soon
  • Decision architectures and metareasoning
  • Structure in behavior
  • Agent architectures
  • Tasks and platforms
  • Open problems
  • What if we succeed?

5
Acerbic remarks
  • Humans are more intelligent than pigeons
  • Because progress has been made on logical
    reasoning, chess, etc., does not mean they are
    easy. They have benefited from 2.5 millenia of
    focused thought from Aristotle to von Neumann,
    Turing, Wiener, Shannon, McCarthy. Such abstract
    tasks probably do capture an important level of
    cognition (but not a purely deterministic or
    disconnected one). And theyre not done yet.

6
Representation
  • Expressive language gt concise models gt fast
    learning, sometimes fast inference
  • E.g., rules of chess 1 page in first-order
    logic, 100,000 pages in propositional logic
  • E.g., DBN vs HMM inference
  • Significant progress occurring, expanding contact
    layer between AI systems and real-world data
    (both relational data and rawdata requiring
    relational modelling)

7
Expressiveness
17th C
20th C
21st C
probability
5th C B.C.
19th C
logic
atomic
propositional
first-order/relational
8
Learning
Learning
knowledge
data
9
Learning
prior knowledge
Learning
knowledge
data
10
Learning
prior knowledge
Learning
knowledge
data
11
Learning
prior knowledge
Learning
knowledge
data
Learned knowledge in a form usable as prior
knowledge, not always one step down in
abstraction Self-reinforcing accumulation of
knowledge and representational
abstractions
12
Vision
  • Learning and probabilistic modelling making huge
    inroads -- dominant paradigm?
  • Expressive generative models should help,
    particularly unknown-worlds models
  • MCMC on generative models with discriminatively
    learned proposals is becoming a bit of a cliché
    in vision (E.S.)
  • Vision will be mostly working within 10 years

13
Language
  • Surface statistical models maxing out
  • Grammar and semantics, formerly well-unified in
    logical approaches, are returning in
    probabilistic clothes to aid disambiguation,
    compensate for inaccurate models
  • Probabilistic semantics w/ text as evidence
  • Choose meaning most likely to be true
  • The money is in the bank

14
Language
  • The money is not in the bank
  • Generative speaker plan world semantics
    syntax
  • (Certainly not n-grams I said word n because I
    said words n-1 and n-2)
  • Hard to reach human-level NLP, but useful
    knowledge creation systems will emerge in 5-10
    years
  • Must include real integrated multi-concept
    learning

15
Simple decision architectures
  • Core state estimation maintains belief state
  • Lots of progress extend to open worlds
  • Reflex p(s)
  • Action-value argmaxa Q(s,a)
  • Goal-based a such that G(result(a,s))
  • Utility-based argmaxa E(U(result(a,s)))
  • Would be nice to understand when one is better
    (e.g., more learnable) than another
  • Any or all can be applied anywhere in the
    architecture that decisions are made

16
Metareasoning
  • Computations are actions too
  • Controlling them is essential, esp. for
    model-based architectures that can do lookahead
    and for approximate inference on intractable
    models
  • Effective control based on expected value of
    computation
  • Methods for learning this must be built-in
    brains unlikely to have fixed, highly engineered
    algorithms that will correctly dictate trillions
    of computational actions over agents lifetime

17
Structure in behaviour
  • One billion seconds, one trillion (parallel)
    actions
  • Unlikely to be generated from a flat solution to
    the unknown POMDP of life
  • Hierarchical structuring of behaviour
  • enunciating this syllable
  • saying this word
  • saying this sentence
  • explaining structure in
    behaviour
  • giving a talk about
    human-level AI
  • .

18
Subroutines and value
  • Key point decisions within subroutines are
    independent of almost all state variables
  • E.g., say(word,prosody) not
  • say(word,prosody,NYSEprices,NASDAQ)
  • Value functions decompose into additive factors,
    both temporally and functionally
  • Structure ultimately reflects properties of
    domain transition model and reward function
  • Partial programming languages -gt declarative
    procedural knowledge (what (not) to do),
    expressed in some extension of temporal logic
    learning within these constraints
  • (Applies to computational behaviour too.)

19
Agent architecture
  • My boxes and arrows vs your boxes and arrows?
  • Well-designed/evolved architecture solves what
    optimization problem? What forces drive design
    choices?
  • Generate optimal actions
  • Generate them quickly
  • Learn to do this from few experiences
  • Each by itself leads to less interesting
    solutions (e.g., omitting learnability favours
    the all-compiled solution)
  • Bounded-optimal solutions have interesting
    architectures Russell Subramanian 95, Livnat
    and Pippenger 05, but architectures are unlikely
    to emerge from a sea of gates/neurons

20
(No Transcript)
21
Challenge problems should involve
  • Continued existence
  • Behavioral structure at several time scales (not
    just repetition of small task)
  • Finding good decisions should sometimes require
    extended deliberation
  • Environment with many, varied objects, nontrivial
    perception (other agents, language optional)
  • Examples cook, house cleaner, secretary, courier
  • Wanted a human-scale dextrous 4-legged robot (or
    a sufficiently rich simulated environment)

22
Open problems
  • Learning better representations need a new
    understanding of reification/generalization
  • Learning new behavioural structures
  • Generating new goals from utility soup
  • Do neuroscience and cognitive science have
    anything to tell us?
  • What if we succeed? Can we design probably
    approximately safe agents?
Write a Comment
User Comments (0)
About PowerShow.com