Oliver Schulte - PowerPoint PPT Presentation

1 / 34
About This Presentation
Title:

Oliver Schulte

Description:

Intelligent Agents CHAPTER 2 Oliver Schulte Summer2011 Learning agents(Taxi driver) Performance element How it currently drives Taxi driver Makes quick left turn ... – PowerPoint PPT presentation

Number of Views:36
Avg rating:3.0/5.0
Slides: 35
Provided by: csSfuCac7
Category:
Tags: oliver | schulte

less

Transcript and Presenter's Notes

Title: Oliver Schulte


1
Intelligent Agents
  • CHAPTER 2
  • Oliver Schulte
  • Summer2011

2
Outline
  • Agents and environments
  • Rationality
  • PEAS (Performance measure, Environment,
    Actuators, Sensors)
  • Environment types
  • Agent types

3
Agents
  • An agent is anything that can be viewed as
    perceiving its environment through sensors and
    acting upon that environment through actuators
  • Human agent
  • eyes, ears, and other organs for sensors
  • hands, legs, mouth, and other body parts for
    actuators
  • Robotic agent
  • cameras and infrared range finders for sensors
  • various motors for actuators

4
Agents and environments
  • The agent function maps from percept histories to
    actions
  • f P ? A
  • The agent program runs on the physical
    architecture to produce f
  • agent architecture program

5
Vacuum-cleaner world
Demo http//www.ai.sri.com/oreilly/aima3ejava/ai
ma3ejavademos.html
  • Percepts location and contents, e.g., A,Dirty
  • Actions Left, Right, Suck, NoOp
  • Agents function ? look-up table
  • For many agents this is a very large table

6
Rational agents
  • Rationality
  • Performance measuring success
  • Agents prior knowledge of environment
  • Actions that agent can perform
  • Agents percept sequence to date
  • Rational Agent For each possible percept
    sequence, a rational agent should select an
    action that is expected to maximize its
    performance measure, given the evidence provided
    by the percept sequence and whatever built-in
    knowledge the agent has.

7
Examples of Rational Choice
  • See File intro-choice.doc

8
Rationality
  • Rational is different from omniscience
  • Percepts may not supply all relevant information
  • E.g., in card game, dont know cards of others.
  • Rational is different from being perfect
  • Rationality maximizes expected outcome while
    perfection maximizes actual outcome.

9
Autonomy in Agents
The autonomy of an agent is the extent to which
its behaviour is determined by its own
experience, rather than knowledge of designer.
  • Extremes
  • No autonomy ignores environment/data
  • Complete autonomy must act randomly/no program
  • Example baby learning to crawl
  • Ideal design agents to have some autonomy
  • Possibly become more autonomous with experience

10
PEAS
  • PEAS Performance measure, Environment,
    Actuators, Sensors
  • Must first specify the setting for intelligent
    agent design
  • Consider, e.g., the task of designing an
    automated taxi driver
  • Performance measure Safe, fast, legal,
    comfortable trip, maximize profits
  • Environment Roads, other traffic, pedestrians,
    customers
  • Actuators Steering wheel, accelerator, brake,
    signal, horn
  • Sensors Cameras, sonar, speedometer, GPS,
    odometer, engine sensors, keyboard

11
PEAS
  • Agent Part-picking robot
  • Performance measure Percentage of parts in
    correct bins
  • Environment Conveyor belt with parts, bins
  • Actuators Jointed arm and hand
  • Sensors Camera, joint angle sensors

12
PEAS
  • Agent Interactive English tutor
  • Performance measure Maximize student's score on
    test
  • Environment Set of students
  • Actuators Screen display (exercises,
    suggestions, corrections)
  • Sensors Keyboard

13
Environment types
  • Fully observable (vs. partially observable)
  • Deterministic (vs. stochastic)
  • Episodic (vs. sequential)
  • Static (vs. dynamic)
  • Discrete (vs. continuous)
  • Single agent (vs. multiagent)

14
Fully observable (vs. partially observable)
  • Is everything an agent requires to choose its
    actions available to it via its sensors? Perfect
    or Full information.
  • If so, the environment is fully accessible
  • If not, parts of the environment are inaccessible
  • Agent must make informed guesses about world.
  • In decision theory perfect information vs.
    imperfect information.

15
Deterministic (vs. stochastic)
  • Does the change in world state
  • Depend only on current state and agents action?
  • Non-deterministic environments
  • Have aspects beyond the control of the agent
  • Utility functions have to guess at changes in
    world

Cross Word
Backgammon
Taxi driver
Part picking robot
Poker
Image analysis
Cross Word
Backgammon
Taxi driver
Part
Poker
Image analysis
Deterministic
Deterministic
Stochastic
Stochastic
Stochastic
Stochastic
16
Episodic (vs. sequential)
  • Is the choice of current action
  • Dependent on previous actions?
  • If not, then the environment is episodic
  • In non-episodic environments
  • Agent has to plan ahead
  • Current choice will affect future actions

Cross Word
Backgammon
Taxi driver
Part picking robot
Poker
Image analysis
Sequential
Sequential
Sequential
Sequential
Episodic
Episodic
17
Static (vs. dynamic)
  • Static environments dont change
  • While the agent is deliberating over what to do
  • Dynamic environments do change
  • So agent should/could consult the world when
    choosing actions
  • Alternatively anticipate the change during
    deliberation OR make decision very fast
  • Semidynamic If the environment itself does not
    change with the passage of time but the agent's
    performance score does.

Cross Word
Backgammon
Taxi driver
Part picking robot
Poker
Image analysis
Dynamic
Semi
Static
Static
Static
Dynamic
Another example off-line route planning vs.
on-board navigation system
18
Discrete (vs. continuous)
  • A limited number of distinct, clearly defined
    percepts and actions vs. a range of values
    (continuous)

Cross Word
Backgammon
Taxi driver
Part picking robot
Poker
Image analysis
Conti
Discrete
Discrete
Discrete
Conti
Conti
19
Single agent (vs. multiagent)
  • An agent operating by itself in an environment or
    there are many agents working together

Cross Word
Backgammon
Taxi driver
Part picking robot
Poker
Image analysis
Single
Single
Single
Multi
Multi
Multi
20
Summary.
Observable
Deterministic
Static
Episodic
Agents
Discrete
Cross Word
Deterministic
Sequential
Static
Discrete
Single
Fully
Fully
Poker
Stochastic
Sequential
Static
Discrete
Multi
Partially
Backgammon
Stochastic
Sequential
Static
Discrete
Multi
Partially
Taxi driver
Multi
Stochastic
Sequential
Dynamic
Conti
Single
Part picking robot
Stochastic
Episodic
Dynamic
Conti
Partially
Single
Image analysis
Deterministic
Episodic
Semi
Conti
Fully
21
Choice under (Un)certainty
Fully Observable
yes
no
Deterministic
no
yes
Certainty Search
Uncertainty
22
Agent types
  • Four basic types in order of increasing
    generality
  • Simple reflex agents
  • Reflex agents with state/model
  • Goal-based agents
  • Utility-based agents
  • All these can be turned into learning agents
  • http//www.ai.sri.com/oreilly/aima3ejava/aima3eja
    vademos.html

23
Simple reflex agents
24
Simple reflex agents
  • Simple but very limited intelligence.
  • Action does not depend on percept history, only
    on current percept.
  • Therefore no memory requirements.
  • Infinite loops
  • Suppose vacuum cleaner does not observe location.
    What do you do given location clean? Left of A
    or right on B -gt infinite loop.
  • Fly buzzing around window or light.
  • Possible Solution Randomize action.
  • Thermostat.
  • Chess openings, endings
  • Lookup table (not a good idea in general)
  • 35100 entries required for the entire game

25
States Beyond Reflexes
  • Recall the agent function that maps from percept
    histories to actions
  • f P ? A
  • An agent program can implement an agent function
    by maintaining an internal state.
  • The internal state can contain information about
    the state of the external environment.
  • The state depends on the history of percepts and
    on the history of actions taken
  • f P, A? S ?A where S is the set of states.
  • If each internal state includes all information
    relevant to information making, the state space
    is Markovian.

26
States and Memory Game Theory
  • If each state includes the information about the
    percepts and actions that led to it, the state
    space has perfect recall.
  • Perfect Information Perfect Recall Full
    Observability.

27
Goal-based agents
  • knowing state and environment? Enough?
  • Taxi can go left, right, straight
  • Have a goal
  • A destination to get to
  • Uses knowledge about a goal to guide its actions
  • E.g., Search, planning

28
Goal-based agents
  • Reflex agent breaks when it sees brake lights.
    Goal based agent reasons
  • Brake light -gt car in front is stopping -gt I
    should stop -gt I should use brake

29
Model-based reflex agents
  • Know how world evolves
  • Overtaking car gets closer from behind
  • How agents actions affect the world
  • Wheel turned clockwise takes you right
  • Model base agents update their state

30
Utility-based agents
  • Goals are not always enough
  • Many action sequences get taxi to destination
  • Consider other things. How fast, how safe..
  • A utility function maps a state onto a real
    number which describes the associated degree of
    happiness, goodness, success.
  • Where does the utility measure come from?
  • Economics money.
  • Biology number of offspring.
  • Your life?

31
Utility-based agents
32
Learning agents
  • Performance element is what was previously the
    whole agent
  • Input sensor
  • Output action
  • Learning element
  • Modifies performance element.

33
Learning agents
  • Critic how the agent is doing
  • Input checkmate?
  • Fixed
  • Problem generator
  • Tries to solve the problem differently instead of
    optimizing.
  • Suggests exploring new actions -gt new problems.

34
Learning agents(Taxi driver)
  • Performance element
  • How it currently drives
  • Taxi driver Makes quick left turn across 3 lanes
  • Critics observe shocking language by passenger
    and other drivers and informs bad action
  • Learning element tries to modify performance
    elements for future
  • Problem generator suggests experiment out
    something called Brakes on different Road
    conditions
  • Exploration vs. Exploitation
  • Learning experience can be costly in the short
    run
  • shocking language from other drivers
  • Less tip
  • Fewer passengers
Write a Comment
User Comments (0)
About PowerShow.com