AI: Intelligent Agents. Chapter 2

Similar documents
Intelligent Agents. Outline. Agents. Agents and environments

Intelligent Agents. Chapter 2

22c:145 Artificial Intelligence

Dr. Mustafa Jarrar. Chapter 2 Intelligent Agents. Sina Institute, University of Birzeit

Chapter 2: Intelligent Agents

Intelligent Agents. Chapter 2 ICS 171, Fall 2009

Artificial Intelligence Intelligent agents

Intelligent Agents. Russell and Norvig: Chapter 2

CS 771 Artificial Intelligence. Intelligent Agents

Intelligent Agents. CmpE 540 Principles of Artificial Intelligence

Intelligent Autonomous Agents. Ralf Möller, Rainer Marrone Hamburg University of Technology

Artificial Intelligence. Intelligent Agents

CS 331: Artificial Intelligence Intelligent Agents

CS 331: Artificial Intelligence Intelligent Agents

Web-Mining Agents Cooperating Agents for Information Retrieval

Intelligent Agents. Soleymani. Artificial Intelligence: A Modern Approach, Chapter 2

Foundations of Artificial Intelligence

Princess Nora University Faculty of Computer & Information Systems ARTIFICIAL INTELLIGENCE (CS 370D) Computer Science Department

CS 331: Artificial Intelligence Intelligent Agents. Agent-Related Terms. Question du Jour. Rationality. General Properties of AI Systems

Contents. Foundations of Artificial Intelligence. Agents. Rational Agents

Agents. This course is about designing intelligent agents Agents and environments. Rationality. The vacuum-cleaner world

Web-Mining Agents Cooperating Agents for Information Retrieval

Intelligent Agents. BBM 405 Fundamentals of Artificial Intelligence Pinar Duygulu Hacettepe University. Slides are mostly adapted from AIMA

Agents & Environments Chapter 2. Mausam (Based on slides of Dan Weld, Dieter Fox, Stuart Russell)

Artificial Intelligence Agents and Environments 1

Agents & Environments Chapter 2. Mausam (Based on slides of Dan Weld, Dieter Fox, Stuart Russell)

Outline for Chapter 2. Agents. Agents. Agents and environments. Vacuum- cleaner world. A vacuum- cleaner agent 8/27/15

Outline. Chapter 2 Agents & Environments. Agents. Types of Agents: Immobots

Intelligent Agents. Instructor: Tsung-Che Chiang

Introduction to Artificial Intelligence 2 nd semester 2016/2017. Chapter 2: Intelligent Agents

Artificial Intelligence CS 6364

Intelligent Agents. Instructor: Tsung-Che Chiang

Rational Agents (Chapter 2)

Agents. Formalizing Task Environments. What s an agent? The agent and the environment. Environments. Example: the automated taxi driver (PEAS)

Agents and Environments

Vorlesung Grundlagen der Künstlichen Intelligenz

Agents. Environments Multi-agent systems. January 18th, Agents

Intelligent Agents. Philipp Koehn. 16 February 2017

KECERDASAN BUATAN 3. By Sirait. Hasanuddin Sirait, MT

Lecture 2 Agents & Environments (Chap. 2) Outline

How do you design an intelligent agent?

Agents and Environments. Stephen G. Ware CSCI 4525 / 5525

Agents and State Spaces. CSCI 446: Artificial Intelligence

Artificial Intelligence

What is AI? The science of making machines that: Think rationally. Think like people. Act like people. Act rationally

CS324-Artificial Intelligence

Artificial Intelligence Lecture 7

Artificial Intelligence

Ar#ficial Intelligence

Module 1. Introduction. Version 1 CSE IIT, Kharagpur

Agents and Environments

Solutions for Chapter 2 Intelligent Agents

AI Programming CS F-04 Agent Oriented Programming

Overview. What is an agent?

Agents and State Spaces. CSCI 446: Ar*ficial Intelligence Keith Vertanen

CS343: Artificial Intelligence


Rational Agents (Ch. 2)

Silvia Rossi. Agent as Intentional Systems. Lezione n. Corso di Laurea: Informatica. Insegnamento: Sistemi multi-agente.

Agents. Robert Platt Northeastern University. Some material used from: 1. Russell/Norvig, AIMA 2. Stacy Marsella, CS Seif El-Nasr, CS4100

COMP329 Robotics and Autonomous Systems Lecture 15: Agents and Intentions. Dr Terry R. Payne Department of Computer Science

Rational Agents (Ch. 2)

AGENT-BASED SYSTEMS. What is an agent? ROBOTICS AND AUTONOMOUS SYSTEMS. Today. that environment in order to meet its delegated objectives.

1 What is an Agent? CHAPTER 2: INTELLIGENT AGENTS

Robotics Summary. Made by: Iskaj Janssen

GRUNDZÜGER DER ARTIFICIAL INTELLIGENCE

PCT 101. A Perceptual Control Theory Primer. Fred Nickols 8/27/2012

Definitions. The science of making machines that: This slide deck courtesy of Dan Klein at UC Berkeley

A PCT Primer. Fred Nickols 6/30/2011

A Computational Framework for Concept Formation for a Situated Design Agent

Introduction to Arti Intelligence

Sequential Decision Making

What is AI? The science of making machines that:

Behavior Architectures

Appendix G: Methodology checklist: the QUADAS tool for studies of diagnostic test accuracy 1

Bill Drake August 2, 2016 FIRO - B

Mindset For Optimal Performance: Essential Mental Skills DR. RICK MCGUIRE DIRECTOR OF SPORT PSYCHOLOGY ANNE SHADLE M.ED.

Embodiment in GLAIR: A Grounded Layered Architecture. with Integrated Reasoning for Autonomous Agents. Henry Hexmoor. Johan Lammens.

Robot Behavior Genghis, MIT Callisto, GATech

HearIntelligence by HANSATON. Intelligent hearing means natural hearing.

EEL-5840 Elements of {Artificial} Machine Intelligence

Introduction to Deep Reinforcement Learning and Control

Problem Solving Agents

The Role of Action in Object Categorization

Biologically-Inspired Human Motion Detection

Artificial Intelligence. Outline

STPA Applied to Automotive Automated Parking Assist

Using Diverse Cognitive Mechanisms for Action Modeling

Semiotics and Intelligent Control

Deliberating on Ontologies: The Present Situation. Simon Milton Department of Information Systems, The University of Melbourne

Test Driven Development (TDD)

Katsunari Shibata and Tomohiko Kawano

Inferring Actions and Observations from Interactions

Reinforcement Learning

Unmanned autonomous vehicles in air land and sea

John E. Laird University of Michigan 1101 Beal Ave. Ann Arbor, MI

CS148 - Building Intelligent Robots Lecture 5: Autonomus Control Architectures. Instructor: Chad Jenkins (cjenkins)

Pavlovian, Skinner and other behaviourists contribution to AI

5.8 Departure from cognitivism: dynamical systems

Proposal for a Multiagent Architecture for Self-Organizing Systems (MA-SOS)

Transcription:

AI: Intelligent Agents Chapter 2

Outline <>Agents and environments <> Rationality <> PEAS (Performance measure, Environment, Actuators, Sensors) <> Environment types <> Agent types

Agents An agent is anything that can be viewed as perceiving its environment through sensors and acting upon that environment through actuators. Percept refers to the agent s perceptual inputs at a given time instant; an agent s perceptual sequence is the complete history of everything the agent has ever perceived. In general, an agent s choice of action at any given instant can depend on the entire precept sequence observed to date, but not on anything it hasn t perceived.

Agents and environments Agents include humans, robots, softbots, thermostats, etc. The agent function maps from percept histories to actions: f : P A The agent program runs on the physical architecture to produce f.

Vacuum-cleaner world Percepts: location and contents, e.g., [A,Dirty] Actions: Left, Right, Suck, NoOp A vacuum-cleaner agent Percept sequence Action [A,Clean] Right [A,Dirty] Suck [B,Clean] Left [B,Dirty] Suck [A,Clean], [A,Clean] Right [A,Clean], [A,Dirty] Suck...... What is the right function? Can it be implemented in a small agent program?

Good Behavior: Rationality A rational agent is one that does the right thing, i.e. the table for the agent function is filled out correctly. But what does it mean to do the right thing? We use a performance measure to evaluate any given sequence of environment states. Importantly, we emphasize that the performance is assessed in terms of environment states and not agent states; self-assessment is often susceptible to self-delusion. Here is a relevant rule of thumb: It is advisable to design performance measures according to what one actually wants in the environment, as opposed to how one believes that agent should behave.

Rationality What is rational at any given time depends on (at least) four things: (1) The performance measure (2) The agent s prior knowledge (3) The actions the agents can perform (4) The agent s percept sequence to date. Definition of a rational agent: For each possible precept sequence, a rational agent should select an action that is expected to maximize its performance measure, given the evidence provided by the percept sequence and whatever built-in knowledge the agent possesses.

Rationality Note that rationality is not the same as omniscience; an omniscient agenknows the actual outcome of its actions and can act accordingly. Percepts may not supply all relevant information. Similarly, rationality is not the same thing as clairvoyance (action outcomes may be unexpected) nor perfection (we maximize expected performance, not actual performance). Performing actions in order to modify future percepts (i.e. information gathering) is a crucial part of rationality and is closely aligned with exploration.

An intelligent agent should not only gather information, but also learn. The agent s initial configuration could reflect some prior knowledge of the environment, but as the agent gains experience, this may be modified and augmented (an extreme case is when the environment is known a priori). Generally speaking, a rational agent should be autonomous, in the sense that it learns what it can to compensate for partial or incorrect prior knowledge. After sufficient experience of its environment, the behavior of a rational agent can become effectively independent of its prior knowledge. Ideally, the incorporation of learning allows for the design of a single rational agent that will succeed in a variety of different environments and for a variety of tasks (the goal of AGI).

PEAS To design a rational agent, we must specify the task environment. Consider, e.g., the task of designing an automated taxi: Performance measure?? Environment?? Actuators?? Sensors??

PEAS To design a rational agent, we must specify the task environment Consider, e.g., the task of designing an automated taxi: Performance measure safety, destination, profits, legality, comfort,... Environment US streets/freeways, traffic, pedestrians, weather,... Actuators steering, accelerator, brake, horn, speaker/display,... Sensors video, accelerometers, gauges, engine sensors, keyboard, GPS,...

PEAS Agent: Medical diagnosis system Performance measure: Healthy patient, minimize costs, lawsuits Environment: Patient, hospital, staff Actuators: Screen display (questions, tests, diagnoses, treatments, referrals) Sensors: Keyboard (entry of symptoms, findings, patient's answers)

Agent: Part-picking robot PEAS Performance measure: Percentage of parts in correct bins Environment: Conveyor belt with parts, bins Actuators: Jointed arm and hand Sensors: Camera, joint angle sensors

PEAS Agent: Interactive English tutor Performance measure: Maximize student's score on test Environment: Set of students Actuators: Screen display (exercises, suggestions, corrections) Sensors: Keyboard

Environment types Fully observable (vs. partially observable): An agent's sensors give it access to the complete state of the environment at each point in time. Deterministic (vs. stochastic): The next state of the environment is completely determined by the current state and the action executed by the agent. (If the environment is deterministic except for the actions of other agents, then the environment is strategic). Episodic (vs. sequential): The agent's experience is divided into atomic "episodes" (each episode consists of the agent perceiving and then performing a single action), and the choice of action in each episode depends only on the episode itself.

Environment types Static (vs. dynamic): The environment is unchanged while an agent is deliberating. (The environment is semidynamic if the environment itself does not change with the passage of time but the agent's performance score does) Discrete (vs. continuous): A limited number of distinct, clearly defined percepts and actions. Single agent (vs. multiagent): An agent operating by itself in an environment.

Environment types Chess with Chess without Taxi driving a clock a clock Fully observable Yes Yes No Deterministic Strategic Strategic No Episodic No No No Static Semi Yes No Discrete Yes Yes No Single agent No No No The environment type largely determines the agent design The real world is (of course) partially observable, stochastic, sequential, dynamic, continuous, multi-agent

Solitaire Backgammon Internet shopping Taxi Observable?? Yes Yes No No Deterministic?? Yes No Partly No Episodic?? No No No No Static?? Yes Semi Semi No Discrete?? Yes Yes Yes No Single-agent?? Yes No Yes (except auctions) No The environment type largely determines the agent design

Agent types Four basic types in order of increasing generality: simple reflex agents reflex agents with state goal-based agents utility-based agents All these can be turned into learning agents

Simple Reflex Agents The simplest kind of agent. These agents select actions on the basis of the current percept, ignoring the rest of the percept history. (An example for the vacuum world is below). function Reflex-Vacuum-Agent([location,status]) returns an action if status = Dirty then return Suck else if location = A then return Right else if location = B then return Left Notice that the vacuum agent program is very small compared to a look-up table; the chief reduction comes from ignoring the percept history (reduction in rule set from 4^T to just 4). Simple reflex agents are, naturally, simple, but they turn out to be of limited intelligence. The agent will only work if the correct decision can be made on the basis of only the current percept (so only if the environment is fully observable).

Function MODEL-BASED-REFLEX-AGENT (percept) returns an action Persistent: state, the agent s current conception of the world state model, a description of how the next state depends on the current state and action rules, a set of condition-action rules action, the most recent action, initially none state<- UPDATE-STATE (state, action percept, model) rule <- RULE-MATCH(state, rules) action <- rule.action return action

Goal-based Reflex Agents Often, to make the correct decision, the agent needs some sort of goal information that describes situations that are desirable. Occasionally, goal-based action selection is straightforward (e.g. follow the action that leads directly to the goal); at other times, however, the agent must consider also search and planning. Decision making of this latter kind involves consideration of the future. Goal-based agents are commonly more flexible than reflex agents.

Utility-based Reflex Agents Goals alone are not enough to generate high-quality behavior in most environments. An agent s utility function is essentially an international of the performance measure. If the internal utility function and the external performance measure are in agreement, then an agent that chooses actions to maximize its utility will be rational according to the external performance measure. A utility-based agent has many advantages in terms of flexibility and learning (e.g. in the case of conflicting goals and cases when there exist several goals).

Learning Agents A learning agent is comprised of (4) components: (1) the learning element, which is responsible for making improvements; (2) the performance element, which is responsible for selecting external actions; (3) the critic, which gives feedback to the agent, and determines how the performance should be modified; (4) the problem generator is responsible for suggesting actions that will lead to new and informative experiences. Consider the taxi example: Performance element: whatever collection of knowledge and procedures the taxi has for selecting its driving actions. Learning element: Formulates a rule based on experience Critic: Adds new rules, based on feedback. Problem Generator: Identify certain areas of behavior in need of improvement and suggest experiments.

Summary Agents interact with environments through actuators and sensors The agent function describes what the agent does in all circumstances The performance measure evaluates the environment sequence A perfectly rational agent maximizes expected performance Agent programs implement (some) agent functions PEAS descriptions define task environments Environments are categorized along several dimensions: observable? deterministic? episodic? static? discrete? single-agent? Several basic agent architectures exist: reflex, reflex with state, goal-based, utility-based