AI-Chapter 2 Part1
AI-Chapter 2 Part1
CHAPTER 2
INTELLIGENT AGENTS
Objectives:
Understand the agent’s environment and its behavior
Differentiate rationality from omniscience, understand how to specify task environment of an
agent
Explain the properties of task environment
Describe the structure of an agent and the working principle of Simple-Reflex agent
Describe the working principle of Model-based reflex agent and Goal-based agent.
Describe the working principle of Utility-based agent and Learning agent.
Partial tabulation of simple agent function for the vacuum –cleaner world:
Rational agent: is one that does the right thing conceptually, every entry in the table for the agent function
is filled out correctly.
Obviously, there is not one fixed performance measure for all tasks and agents; typically, a designer will devise
one appropriate to the circumstances.
As a general rule, it is better to design performance measures according to what one actually wants in the
environment, rather than according to how one thinks the agent should behave.
We claim that under these circumstances the agent is indeed rational; the same agent would be irrational under
different circumstances.
An omniscient agent knows the actual outcome of its actions and can act accordingly; but omniscience is
impossible in reality.
Example: A person is walking along the road and there is no traffic nearby and he is being rational, he start to
cross the street, meanwhile, at 33,000 feet, a cargo door falls off a passing airliner, and before he make it to the
other side of the street he flattened.
The point is that if we expect an agent to do what turns out to be the best action after the fact, it will be
impossible to design an agent to fulfill this specification.
Mizan-Tepi University Page 3
Department of Information Systems School of Computing and Informatics
Our definition of rationality does not require omniscience, then, because the rational choice depends only on the
percept sequence to date. Information gathering and exploration, is an important part of rationality – Doing
actions in order to modify future percepts.
Rational ≠ omniscient
Rational=> exploration ,learning and autonomy.
A Rational agent should not only to gather information, but also to learn as much as possible from what
it perceives.
A Rational agent should be autonomous-it should learn what it can to compensate for partial or incorrect
prior knowledge.
To the extent that an agent relies on the prior knowledge of its designer rather than on its own percepts, we say
that the agent lacks autonomy.
Example: A vacuum-cleaning agent that learns to foresee where and when additional dirt will appear will do
better than one that does not.
Taskenvironments: which are essentially the “problems” to which rational agents are the “solutions”.
Specifying the task environment:
In designing an agent, the first step must always be to specify the task environment as fully as possible.
We call this the PEAS (Performance, Environment, Actuators, and Sensors) description.
If an agent’s sensors give it access to the complete state of the environment ateach point in time, then we say
that the task environment is fully observable.
An environment might be partially observable because of noisy and inaccurate sensors or because parts of the
state are simply missing from the sensor data.
Unobservable: if the agent has no sensors at all then the environment is unobservable.
If the next state of the environment is completely determined by the current state and the action executed
by the agent, then we say the environment is deterministic, otherwise, it is stochastic.
In principle, an agent need not worry about uncertainty in a fully observable, deterministic environment.
Example: A game can be deterministic even though each agent may be unable to predict the actions of
others.If the environment is partially observable, then it could appear to be stochastic.
Example: Taxi driving is stochastic, because one can never predict the behavior of traffic exactly.
Example: vacuum-cleaner world as we described it is deterministic, but variations can include stochastic
elements such as randomly appearing dir and unreliable suction mechanism.
Uncertain environment:
An environment is uncertain if it is not fully observable or not deterministic.
Note:
Our use of the word “Stochastic” generally implies that uncertainty about outcomes is quantified in
terms of probabilities.
Non deterministic environment is one in which actions are characterized by their possible outcomes, but
no probabilities are attached to them.
Episodic: In this environment, the agent’s experience is divided into atomic episodes. In each episode the agent
receives a percept and then performs a single action. The next episode does not depend on the actions taken in
previous episodes.
Example:An agent that has to spot defective parts on an assembly line bases each decision in the
current part, regardless of previous decisions; moreover the current decision doesn’t affect whether the
next part is defective.
Sequential: In this environment, the current decision could affect all future decisions.
Example: Chess and taxi driving – in both cases, short term actions can have long-term consequences.
Episodic environments are much simpler than sequential environments because the agent does not need to think
ahead.
If the environment can change while an agent is deliberating, then it is dynamic for that agent; otherwise
it is static.
Static environments: are easy to deal with because the agent need not keep looking at the work while it is
deciding on an action, nor need it worry about the passage of time.
Dynamic environments: are continuously asking the agent what it wants to do; if it hasn’t decided yet, that
counts as deciding to do nothing.
Semidynamic environment: If the environment itself does not change with the passage of time but the agent’s
performance sure does, then we say that it is semidynamic environment.
Example:
Taxi driving – is dynamic
Chess – is semidynamic, when played with a clock.
Crossword puzzles – are static
These discrete/continuous distinction applies to the state of the environment, to the way time is handled, and to
the percepts and actions of the agent.
Taxi driving - is a continuous-state and continuous time problem: the speed and location of the taxi and
of the other vehicles sweep through a range of continuous values and so smoothly over time.
Taxi driving actions – are also continuous (steering angles….)
This distinction refers not to the environment itself but to the agent’s (or designer’s) state of knowledge about
the “laws of physics” of the environment.
Known environment:
The outcomes (or outcome probabilities of the environment is stochastic) for all actions are given.
Unknown environment:
Here the agent will have to learn how it works in order to make good decisions.
Note: The distinction between known and unknown environments is not the same as the one between fully and
partially observable environments.