Rational Agents (Chapter 2)

Similar documents
Intelligent Agents. Outline. Agents. Agents and environments

Intelligent Agents. Chapter 2 ICS 171, Fall 2009

Web-Mining Agents Cooperating Agents for Information Retrieval

Intelligent Autonomous Agents. Ralf Möller, Rainer Marrone Hamburg University of Technology

Agents & Environments Chapter 2. Mausam (Based on slides of Dan Weld, Dieter Fox, Stuart Russell)

Dr. Mustafa Jarrar. Chapter 2 Intelligent Agents. Sina Institute, University of Birzeit

Agents & Environments Chapter 2. Mausam (Based on slides of Dan Weld, Dieter Fox, Stuart Russell)

Chapter 2: Intelligent Agents

Web-Mining Agents Cooperating Agents for Information Retrieval

CS 331: Artificial Intelligence Intelligent Agents

AI: Intelligent Agents. Chapter 2

Outline. Chapter 2 Agents & Environments. Agents. Types of Agents: Immobots

CS 771 Artificial Intelligence. Intelligent Agents

CS 331: Artificial Intelligence Intelligent Agents

Intelligent Agents. BBM 405 Fundamentals of Artificial Intelligence Pinar Duygulu Hacettepe University. Slides are mostly adapted from AIMA

Intelligent Agents. Chapter 2

Intelligent Agents. Soleymani. Artificial Intelligence: A Modern Approach, Chapter 2

Agents. Formalizing Task Environments. What s an agent? The agent and the environment. Environments. Example: the automated taxi driver (PEAS)

CS 331: Artificial Intelligence Intelligent Agents. Agent-Related Terms. Question du Jour. Rationality. General Properties of AI Systems

Lecture 2 Agents & Environments (Chap. 2) Outline

Artificial Intelligence Agents and Environments 1

Artificial Intelligence. Intelligent Agents

Outline for Chapter 2. Agents. Agents. Agents and environments. Vacuum- cleaner world. A vacuum- cleaner agent 8/27/15

Artificial Intelligence CS 6364

Agents. Environments Multi-agent systems. January 18th, Agents

What is AI? The science of making machines that: Think rationally. Think like people. Act like people. Act rationally

Intelligent Agents. Instructor: Tsung-Che Chiang

22c:145 Artificial Intelligence

Intelligent Agents. Instructor: Tsung-Che Chiang

Princess Nora University Faculty of Computer & Information Systems ARTIFICIAL INTELLIGENCE (CS 370D) Computer Science Department

Vorlesung Grundlagen der Künstlichen Intelligenz

Intelligent Agents. CmpE 540 Principles of Artificial Intelligence

Agents. This course is about designing intelligent agents Agents and environments. Rationality. The vacuum-cleaner world

Agents and State Spaces. CSCI 446: Artificial Intelligence

Agents and Environments

Intelligent Agents. Russell and Norvig: Chapter 2

Introduction to Artificial Intelligence 2 nd semester 2016/2017. Chapter 2: Intelligent Agents

Agents and Environments. Stephen G. Ware CSCI 4525 / 5525

Artificial Intelligence Intelligent agents

Foundations of Artificial Intelligence

How do you design an intelligent agent?

Contents. Foundations of Artificial Intelligence. Agents. Rational Agents

Artificial Intelligence

KECERDASAN BUATAN 3. By Sirait. Hasanuddin Sirait, MT

Intelligent Agents. Philipp Koehn. 16 February 2017

Agents and Environments

Overview. What is an agent?

Artificial Intelligence Lecture 7

Ar#ficial Intelligence

AI Programming CS F-04 Agent Oriented Programming

Agents and State Spaces. CSCI 446: Ar*ficial Intelligence Keith Vertanen

CS343: Artificial Intelligence

Solutions for Chapter 2 Intelligent Agents

Module 1. Introduction. Version 1 CSE IIT, Kharagpur

Rational Agents (Ch. 2)

Rational Agents (Ch. 2)

CS324-Artificial Intelligence

Silvia Rossi. Agent as Intentional Systems. Lezione n. Corso di Laurea: Informatica. Insegnamento: Sistemi multi-agente.

Artificial Intelligence

Agents. Robert Platt Northeastern University. Some material used from: 1. Russell/Norvig, AIMA 2. Stacy Marsella, CS Seif El-Nasr, CS4100

Sequential Decision Making

AGENT-BASED SYSTEMS. What is an agent? ROBOTICS AND AUTONOMOUS SYSTEMS. Today. that environment in order to meet its delegated objectives.

Lecture 13: Finding optimal treatment policies

COMP329 Robotics and Autonomous Systems Lecture 15: Agents and Intentions. Dr Terry R. Payne Department of Computer Science


Learning to Identify Irrelevant State Variables

Reinforcement Learning

CS 4365: Artificial Intelligence Recap. Vibhav Gogate

GRUNDZÜGER DER ARTIFICIAL INTELLIGENCE

Lecture 5- Hybrid Agents 2015/2016

Lecture 2: Learning and Equilibrium Extensive-Form Games

1 What is an Agent? CHAPTER 2: INTELLIGENT AGENTS

Bayesian Perception & Decision for Intelligent Mobility

Programming with Goals (3)

CS148 - Building Intelligent Robots Lecture 5: Autonomus Control Architectures. Instructor: Chad Jenkins (cjenkins)

EEL-5840 Elements of {Artificial} Machine Intelligence

Probabilistic Graphical Models: Applications in Biomedicine

What is AI? The science of making machines that:

POND-Hindsight: Applying Hindsight Optimization to POMDPs

Organizational. Architectures of Cognition Lecture 1. What cognitive phenomena will we examine? Goals of this course. Practical Assignments.

Persuasive Technology

Introduction to Arti Intelligence

Consumer Behavior, Ninth Edition. Schiffman & Kanuk

Towards Learning to Ignore Irrelevant State Variables

Study Unit 3 -Part 2. Consumer Learning SIM University. All rights reserved. Introduction. In this presentation, you will learn:

Chapter 7 Consumer Learning. Consumer Behavior, Ninth Edition. Schiffman & Kanuk. Chapter Outline

Modeling Agents as Qualitative Decision Makers

IE 5203 Decision Analysis Lab I Probabilistic Modeling, Inference and Decision Making with Netica

Computational Models for Belief Revision, Group Decisions and Cultural Shifts

Intelligent Machines That Act Rationally. Hang Li Bytedance AI Lab

Behavior Architectures

CS 2310 Software Engineering Final Project Report. Arm Motion Assessment for Remote Physical Therapy Using Kinect

Problem Solving Agents

A Framework for Sequential Planning in Multi-Agent Settings

Artificial Intelligence Programming Probability

Robotics Summary. Made by: Iskaj Janssen

COMMITMENT. &SOLUTIONS Act like someone s life depends on what we do. UNPARALLELED

Introduction to Game Theory Autonomous Agents and MultiAgent Systems 2015/2016

The optimism bias may support rational action

Katsunari Shibata and Tomohiko Kawano

Transcription:

Rational Agents (Chapter 2)

Agents An agent is anything that can be viewed as perceiving its environment through sensors and acting upon that environment through actuators

Example: Vacuum-Agent Percepts: Location and status, e.g., [A,Dirty] Actions: Left, Right, Suck, NoOp function Vacuum-Agent([location,status]) returns an action if status = Dirty then return Suck else if location = A then return Right else if location = B then return Left

Rational agents For each possible percept sequence, a rational agent should select an action that is expected to maximize its performance measure, given the evidence provided by the percept sequence and the agent s built-in knowledge Performance measure (utility function): An objective criterion for success of an agent's behavior Expected utility: EU(action) = outcomes P(outcome action)u(outcome) Can a rational agent make mistakes?

Back to Vacuum-Agent Percepts: Location and status, e.g., [A,Dirty] Actions: Left, Right, Suck, NoOp function Vacuum-Agent([location,status]) returns an action if status = Dirty then return Suck else if location = A then return Right else if location = B then return Left Is this agent rational? Depends on performance measure, environment properties

Specifying the task environment PEAS: Performance measure, Environment, Actuators, Sensors P: a function the agent is maximizing (or minimizing) Assumed given In practice, needs to be computed somewhere E: a formal representation for world states For concreteness, a tuple (var 1 =val 1, var 2 =val 2,,var n =val n ) A: actions that change the state according to a transition model Given a state and action, what is the successor state (or distribution over successor states)? S: observations that allow the agent to infer the world state Often come in very different form than the state itself E.g., in tracking, observations may be pixels and state variables 3D coordinates

PEAS Example: Autonomous taxi Performance measure Safe, fast, legal, comfortable trip, maximize profits Environment Roads, other traffic, pedestrians, customers Actuators Steering wheel, accelerator, brake, signal, horn Sensors Cameras, LIDAR, speedometer, GPS, odometer, engine sensors, keyboard

Another PEAS example: Spam filter Performance measure Minimizing false positives, false negatives Environment A user s email account, email server Actuators Mark as spam, delete, etc. Sensors Incoming messages, other information about user s account

Environment types Fully observable vs. partially observable Deterministic vs. stochastic Episodic vs. sequential Static vs. dynamic Discrete vs. continuous Single agent vs. multi-agent Known vs. unknown

Fully observable vs. partially observable Do the agent's sensors give it access to the complete state of the environment? For any given world state, are the values of all the variables known to the agent? vs. Source: L. Zettlemoyer

Deterministic vs. stochastic Is the next state of the environment completely determined by the current state and the agent s action? Is the transition model deterministic (unique successor state given current state and action) or stochastic (distribution over successor states given current state and action)? Strategic: the environment is deterministic except for the actions of other agents vs.

Episodic vs. sequential Is the agent s experience divided into unconnected single decisions/actions, or is it a coherent sequence of observations and actions in which the world evolves according to the transition model? vs.

Static vs. dynamic Is the world changing while the agent is thinking? Semidynamic: the environment does not change with the passage of time, but the agent's performance score does vs.

Discrete vs. continuous Does the environment provide a fixed number of distinct percepts, actions, and environment states? Are the values of the state variables discrete or continuous? Time can also evolve in a discrete or continuous fashion vs.

Single-agent vs. multiagent Is an agent operating by itself in the environment? vs.

Known vs. unknown Are the rules of the environment (transition model and rewards associated with states) known to the agent? Strictly speaking, not a property of the environment, but of the agent s state of knowledge vs.

Examples of different environments Word jumble solver Chess with a clock Scrabble Autonomous driving Observable Deterministic Episodic Fully Deterministic Episodic Fully Partially Partially Strategic Stochastic Stochastic Sequential Sequential Sequential Static Static Semidynamic Static Dynamic Discrete Single agent Discrete Single Discrete Discrete Continuous Multi Multi Multi

Preview of the course Deterministic environments: search, constraint satisfaction, classical planning Can be sequential or episodic Multi-agent, strategic environments: minimax search, games Can also be stochastic, partially observable Stochastic environments Episodic: Bayesian networks, pattern classifiers Sequential, known: Markov decision processes Sequential, unknown: reinforcement learning

Review: PEAS

Review: PEAS P: Performance measure Function the agent is maximizing (or minimizing) E: Environment A formal representation for world states For concreteness, a tuple (var 1 =val 1, var 2 =val 2,,var n =val n ) A: Actions Transition model: Given a state and action, what is the successor state (or distribution over successor states)? S: Sensors Observations that allow the agent to infer the world state Often come in very different form than the state itself

Review: Environment types Fully observable vs. partially observable Deterministic vs. stochastic (vs. strategic) Episodic vs. sequential Static vs. dynamic (vs. semidynamic) Discrete vs. continuous Single agent vs. multi-agent Known vs. unknown