0% found this document useful (0 votes)
30 views

Intelliget Agent 2 How To Design

The document discusses designing intelligent agents including defining their environment and task. An intelligent agent interacts with its environment through sensors and actuators. The environment can have different properties like being observable, deterministic, episodic, static, discrete, or single/multi-agent. The document also discusses representing the environment and agent using state transformer functions and defining a system as the combination of an agent and its environment.

Uploaded by

thejaka aloka
Copyright
© © All Rights Reserved
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
30 views

Intelliget Agent 2 How To Design

The document discusses designing intelligent agents including defining their environment and task. An intelligent agent interacts with its environment through sensors and actuators. The environment can have different properties like being observable, deterministic, episodic, static, discrete, or single/multi-agent. The document also discusses representing the environment and agent using state transformer functions and defining a system as the combination of an agent and its environment.

Uploaded by

thejaka aloka
Copyright
© © All Rights Reserved
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 24

How to Design an

Intelligent Agent

1
About Intelligent Agents
• Agent comprises with
– Environment
– Agent
– Percepts
– Action
– Sensors

2
Environment
• Determine to a large degree the interaction
between the “outside world” and the agent
– the “outside world” is not necessarily the “real
world” as we perceive it
– it may be a real or virtual environment the agent
lives in
• in many cases, environments are
implemented within computers
– they may or may not have a close correspondence
to the “real world

3
Environment types
• Fully observable (vs. partially observable): An agent's
sensors give it access to the complete state of the
environment at each point in time.

• Deterministic (vs. stochastic): The next state of the


environment is completely determined by the current
state and the action executed by the agent. (If the
environment is deterministic except for the actions of
other agents, then the environment is strategic)

• Episodic (vs. sequential): An agent’s action is divided


into atomic episodes. Decisions do not depend on
previous decisions/actions.
Environment types
• Static (vs. dynamic): The environment is unchanged while an
agent is deliberating. (The environment is semidynamic if the
environment itself does not change with the passage of time
but the agent's performance score does)

• Discrete (vs. continuous): A limited number of distinct, clearly


defined percepts and actions.
How do we represent or abstract or model the world?

• Single agent (vs. multi-agent): An agent operating by itself in


an environment. Does the other agent interfere with my
performance measure?
Environment Properties
• Fully observable vs. partially observable
– sensors capture all relevant information from the
environment
• deterministic vs. stochastic (non-deterministic)
– changes in the environment are predictable
• episodic vs. sequential (non-episodic)
– independent perceiving-acting episodes
• static vs. dynamic
– no changes while the agent is “thinking”
• discrete vs. continuous
– limited number of distinct percepts/actions
• single vs. multiple agents
– interaction and collaboration among agents
– competitive, cooperative
6
Environment Programs
• Environment simulators for experiments with
agents
– gives a percept to an agent
– receives an action
– updates the environment
• Often divided into environment classes for
related tasks or types of agents
• The environment frequently provides
mechanisms for measuring the performance
of agents

7
Task Environment
• Before we design an intelligent agent, we
must specify its “task environment”:
• PEAS:
– Performance measure
– Environment
– Actuators
– Sensors

8
Agents with State
• We now consider agents that maintain
state:
Agent

see action

next state

Environment

9
Example
Vacuum cleaner robot

10
Solution

11
PEAS

• Example: Agent = taxi driver


– Performance measure: Safe, fast, legal, comfortable
trip, maximize profits

– Environment: Roads, other traffic, pedestrians,


customers

– Actuators: Steering wheel, accelerator, brake, signal,


horn

– Sensors: Cameras, sonar, speedometer, GPS,


odometer, engine sensors, keyboard
PEAS
• Example: Agent = Medical diagnosis system

Performance measure: Healthy patient, minimize costs,


lawsuits

Environment: Patient, hospital, staff

Actuators: Screen display (questions, tests, diagnoses,


treatments, referrals)

Sensors: Keyboard (entry of symptoms, findings, patient's


answers)
PEAS
• Example: Agent = Part-picking robot

• Performance measure: Percentage of parts in correct


bins

• Environment: Conveyor belt with parts, bins

• Actuators: Jointed arm and hand

• Sensors: Camera, joint angle sensors


task env. observable determ./ episodic/ static/ discrete/ agents
stochastic sequential dynamic continuous

crossword fully determ. sequential static discrete single


puzzle
chess with fully strategic sequential semi discrete multi
clock
poker partial stochastic sequential static discrete multi

back fully stochastic sequential static discrete multi


gammon
taxi partial stochastic sequential dynamic continuous multi
driving
medical partial stochastic sequential dynamic continuous single
diagnosis
image fully determ. episodic semi continuous single
analysis
partpicking partial stochastic episodic dynamic continuous single
robot

refinery partial stochastic sequential dynamic continuous single


controller
interact. partial stochastic sequential dynamic discrete multi
Eng. tutor
Abstract Architecture for Agents
• Assume the environment may be in any of a finite
set E of discrete, instantaneous states:

• Agents are assumed to have a repertoire of


possible actions available to them, which transform
the state of the environment:

• A run, r, of an agent in an environment is a


sequence of interleaved environment states and
actions:

2-16
Abstract Architecture for Agents
• Let:
– R be the set of all such possible finite
sequences (over E and Ac)
– RAc be the subset of these that end with an
action
– RE be the subset of these that end with an
environment state

2-17
State Transformer Functions
• A state transformer function represents behavior
of the environment:

• Note that environments are…


– history dependent
– non-deterministic
• If (r)=, then there are no possible successor
states to r. In this case, we say that the system
has ended its run
• Formally, we say an environment Env is a triple
Env =E,e0, where: E is a set of environment
states, e0 E is the initial state, and  is a state
transformer function
2-18
Agents
• Agent is a function which maps runs to
actions:

An agent makes a decision about what


action to perform based on the history of
the system that it has witnessed to date.
Let AG be the set of all agents

2-19
Systems
• A system is a pair containing an agent and
an environment
• Any system will have associated with it a
set of possible runs; we denote the set of
runs of agent Ag in environment Env by
R(Ag, Env)
• (We assume R(Ag, Env) contains only
terminated runs)

2-20
Systems
• Formally, a sequence

represents a run of an agent Ag in


environment Env =E,e0, if:
1. e0 is the initial state of Env
2. 0 = Ag(e0); and
3. For u > 0,

2-21
Purely Reactive Agents
• Some agents decide what to do without
reference to their history — they base their
decision making entirely on the present,
with no reference at all to the past
• We call such agents purely reactive:

• A thermostat is a purely reactive agent

2-22
Software agent- Design
• Why Agent than the Objects
• Get Idea about Environment
• Input
• Output
• Processes
• Communication/ Action
• How to provide Intelligent capabilities

23
Example

24

You might also like