Artificial Intelligence Agents and Environments 1

Similar documents
Intelligent Agents. Outline. Agents. Agents and environments

Dr. Mustafa Jarrar. Chapter 2 Intelligent Agents. Sina Institute, University of Birzeit

Agents & Environments Chapter 2. Mausam (Based on slides of Dan Weld, Dieter Fox, Stuart Russell)

Agents & Environments Chapter 2. Mausam (Based on slides of Dan Weld, Dieter Fox, Stuart Russell)

Intelligent Agents. Chapter 2 ICS 171, Fall 2009

Outline. Chapter 2 Agents & Environments. Agents. Types of Agents: Immobots

Intelligent Autonomous Agents. Ralf Möller, Rainer Marrone Hamburg University of Technology

Chapter 2: Intelligent Agents

Artificial Intelligence. Intelligent Agents

Web-Mining Agents Cooperating Agents for Information Retrieval

Intelligent Agents. BBM 405 Fundamentals of Artificial Intelligence Pinar Duygulu Hacettepe University. Slides are mostly adapted from AIMA

Web-Mining Agents Cooperating Agents for Information Retrieval

CS 771 Artificial Intelligence. Intelligent Agents

Outline for Chapter 2. Agents. Agents. Agents and environments. Vacuum- cleaner world. A vacuum- cleaner agent 8/27/15

Princess Nora University Faculty of Computer & Information Systems ARTIFICIAL INTELLIGENCE (CS 370D) Computer Science Department

Intelligent Agents. Soleymani. Artificial Intelligence: A Modern Approach, Chapter 2

Artificial Intelligence CS 6364

Intelligent Agents. CmpE 540 Principles of Artificial Intelligence

Lecture 2 Agents & Environments (Chap. 2) Outline

CS 331: Artificial Intelligence Intelligent Agents

CS 331: Artificial Intelligence Intelligent Agents

AI: Intelligent Agents. Chapter 2

Introduction to Artificial Intelligence 2 nd semester 2016/2017. Chapter 2: Intelligent Agents

Intelligent Agents. Instructor: Tsung-Che Chiang

CS 331: Artificial Intelligence Intelligent Agents. Agent-Related Terms. Question du Jour. Rationality. General Properties of AI Systems

Intelligent Agents. Instructor: Tsung-Che Chiang

Agents and Environments

22c:145 Artificial Intelligence

Agents. Environments Multi-agent systems. January 18th, Agents

Agents. Formalizing Task Environments. What s an agent? The agent and the environment. Environments. Example: the automated taxi driver (PEAS)

Rational Agents (Chapter 2)

Intelligent Agents. Chapter 2

Vorlesung Grundlagen der Künstlichen Intelligenz

Intelligent Agents. Russell and Norvig: Chapter 2

Agents. This course is about designing intelligent agents Agents and environments. Rationality. The vacuum-cleaner world

Artificial Intelligence Intelligent agents

Foundations of Artificial Intelligence

Agents and State Spaces. CSCI 446: Artificial Intelligence

Contents. Foundations of Artificial Intelligence. Agents. Rational Agents

What is AI? The science of making machines that: Think rationally. Think like people. Act like people. Act rationally

Artificial Intelligence Lecture 7

Intelligent Agents. Philipp Koehn. 16 February 2017

Artificial Intelligence

How do you design an intelligent agent?

Artificial Intelligence

KECERDASAN BUATAN 3. By Sirait. Hasanuddin Sirait, MT

Agents and Environments. Stephen G. Ware CSCI 4525 / 5525

Ar#ficial Intelligence

Agents and Environments

Overview. What is an agent?

Rational Agents (Ch. 2)

CS324-Artificial Intelligence

Agents and State Spaces. CSCI 446: Ar*ficial Intelligence Keith Vertanen

Module 1. Introduction. Version 1 CSE IIT, Kharagpur

Rational Agents (Ch. 2)

AI Programming CS F-04 Agent Oriented Programming

CS343: Artificial Intelligence

Solutions for Chapter 2 Intelligent Agents


Agents. Robert Platt Northeastern University. Some material used from: 1. Russell/Norvig, AIMA 2. Stacy Marsella, CS Seif El-Nasr, CS4100

CS148 - Building Intelligent Robots Lecture 5: Autonomus Control Architectures. Instructor: Chad Jenkins (cjenkins)

Silvia Rossi. Agent as Intentional Systems. Lezione n. Corso di Laurea: Informatica. Insegnamento: Sistemi multi-agente.

GRUNDZÜGER DER ARTIFICIAL INTELLIGENCE

Four Example Application Domains

COMP329 Robotics and Autonomous Systems Lecture 15: Agents and Intentions. Dr Terry R. Payne Department of Computer Science

AGENT-BASED SYSTEMS. What is an agent? ROBOTICS AND AUTONOMOUS SYSTEMS. Today. that environment in order to meet its delegated objectives.

Unmanned autonomous vehicles in air land and sea

Robotics Summary. Made by: Iskaj Janssen

Problem Solving Agents

Reinforcement Learning

Pavlovian, Skinner and other behaviourists contribution to AI

Definitions. The science of making machines that: This slide deck courtesy of Dan Klein at UC Berkeley

Classifying Cognitive Load and Driving Situation with Machine Learning

Modeling Emotion and Temperament on Cognitive Mobile Robots

Sequential Decision Making

Embodiment in GLAIR: A Grounded Layered Architecture. with Integrated Reasoning for Autonomous Agents. Henry Hexmoor. Johan Lammens.

Lecture 5- Hybrid Agents 2015/2016

Behavior Architectures

The Misjudgment of Risk due to Inattention

ICS 606. Intelligent Autonomous Agents 1. Intelligent Autonomous Agents ICS 606 / EE 606 Fall Reactive Architectures

OPTIC FLOW IN DRIVING SIMULATORS

1 What is an Agent? CHAPTER 2: INTELLIGENT AGENTS

Bayesian Perception & Decision for Intelligent Mobility


Re: ENSC 370 Project Gerbil Functional Specifications

Lecture 13: Finding optimal treatment policies

Using Probabilistic Reasoning to Develop Automatically Adapting Assistive

Introduction to Arti Intelligence

Multimodal Driver Displays: Potential and Limitations. Ioannis Politis

Artificial Intelligence. Outline

Biologically-Inspired Human Motion Detection

STPA Applied to Automotive Automated Parking Assist

Robot Learning Letter of Intent

On Three Layer Architectures (Erann Gat) Matt Loper / Brown University Presented for CS296-3

Part I Part 1 Robotic Paradigms and Control Architectures

Bundles of Synergy A Dynamical View of Mental Function

What is AI? The science of making machines that:

HearIntelligence by HANSATON. Intelligent hearing means natural hearing.

Autonomous Mobile Robotics

Semiotics and Intelligent Control

Transcription:

Artificial Intelligence and Environments 1 Instructor: Dr. B. John Oommen Chancellor s Professor Fellow: IEEE; Fellow: IAPR School of Computer Science, Carleton University, Canada. 1 The primary source of these notes are the slides of Professor Hwee Tou Ng from Singapore. I sincerely thank him for this. 1/24

2/24 Intelligent, Evaluating Agent Anything that can be viewed as perceiving its environment Perception done through sensors Acting upon that environment through actuators Human agent Eyes, ears, and other organs for sensors Hands, legs, mouth, and other body parts for actuators Robotic agent Cameras and infrared range finders for sensors Various motors for actuators

2/24 Intelligent, Evaluating Agent Anything that can be viewed as perceiving its environment Perception done through sensors Acting upon that environment through actuators Human agent Eyes, ears, and other organs for sensors Hands, legs, mouth, and other body parts for actuators Robotic agent Cameras and infrared range finders for sensors Various motors for actuators

2/24 Intelligent, Evaluating Agent Anything that can be viewed as perceiving its environment Perception done through sensors Acting upon that environment through actuators Human agent Eyes, ears, and other organs for sensors Hands, legs, mouth, and other body parts for actuators Robotic agent Cameras and infrared range finders for sensors Various motors for actuators

3/24... Evaluating Percepts Agent Environment? Actions Agent: Mapping: Percept Sequences Actions

4/24... Evaluating Agent Function Maps from percept histories to actions: [F : P A] Agent Program Runs on the physical architecture to produce F Agent = Architecture + Program Vacuum Cleaner Agent Percepts: Location and Contents: {[LocA, Dirty],... } Actions: Left, Right, Suck, VacuumOn, VacuumOff Agent: Function(PerceptHistory, Vacuum-agent-function-table)

4/24... Evaluating Agent Function Maps from percept histories to actions: [F : P A] Agent Program Runs on the physical architecture to produce F Agent = Architecture + Program Vacuum Cleaner Agent Percepts: Location and Contents: {[LocA, Dirty],... } Actions: Left, Right, Suck, VacuumOn, VacuumOff Agent: Function(PerceptHistory, Vacuum-agent-function-table)

4/24... Evaluating Agent Function Maps from percept histories to actions: [F : P A] Agent Program Runs on the physical architecture to produce F Agent = Architecture + Program Vacuum Cleaner Agent Percepts: Location and Contents: {[LocA, Dirty],... } Actions: Left, Right, Suck, VacuumOn, VacuumOff Agent: Function(PerceptHistory, Vacuum-agent-function-table)

4/24... Evaluating Agent Function Maps from percept histories to actions: [F : P A] Agent Program Runs on the physical architecture to produce F Agent = Architecture + Program Vacuum Cleaner Agent Percepts: Location and Contents: {[LocA, Dirty],... } Actions: Left, Right, Suck, VacuumOn, VacuumOff Agent: Function(PerceptHistory, Vacuum-agent-function-table)

5/24... Evaluating Agent should strive to do the right thing : Based on what it can perceive and actions it can do The right action : One that will cause the agent to be most successful Performance measure: Objective criterion for success of an agent s behavior Performance of a vacuum-cleaner agent could be: Amount of dirt cleaned up Amount of time taken Amount of electricity consumed Amount of noise generated, etc.

5/24... Evaluating Agent should strive to do the right thing : Based on what it can perceive and actions it can do The right action : One that will cause the agent to be most successful Performance measure: Objective criterion for success of an agent s behavior Performance of a vacuum-cleaner agent could be: Amount of dirt cleaned up Amount of time taken Amount of electricity consumed Amount of noise generated, etc.

5/24... Evaluating Agent should strive to do the right thing : Based on what it can perceive and actions it can do The right action : One that will cause the agent to be most successful Performance measure: Objective criterion for success of an agent s behavior Performance of a vacuum-cleaner agent could be: Amount of dirt cleaned up Amount of time taken Amount of electricity consumed Amount of noise generated, etc.

5/24... Evaluating Agent should strive to do the right thing : Based on what it can perceive and actions it can do The right action : One that will cause the agent to be most successful Performance measure: Objective criterion for success of an agent s behavior Performance of a vacuum-cleaner agent could be: Amount of dirt cleaned up Amount of time taken Amount of electricity consumed Amount of noise generated, etc.

5/24... Evaluating Agent should strive to do the right thing : Based on what it can perceive and actions it can do The right action : One that will cause the agent to be most successful Performance measure: Objective criterion for success of an agent s behavior Performance of a vacuum-cleaner agent could be: Amount of dirt cleaned up Amount of time taken Amount of electricity consumed Amount of noise generated, etc.

6/24 Rational? Evaluating There is a: Performance measure Percept sequence Agent s knowledge about the Environment Agent s action repertoire Rational Agent: For each percept sequence Acts so as to maximize expected performance measure Given percept sequence and its built-in knowledge

6/24 Rational? Evaluating There is a: Performance measure Percept sequence Agent s knowledge about the Environment Agent s action repertoire Rational Agent: For each percept sequence Acts so as to maximize expected performance measure Given percept sequence and its built-in knowledge

7/24 Rational? Evaluating Rationality is distinct from omniscience All-knowing with infinite knowledge can perform actions to modify future percepts Use this to obtain useful information Information gathering, Exploration An Autonomous Agent: Behavior is determined by its own experience With ability to learn and adapt

7/24 Rational? Evaluating Rationality is distinct from omniscience All-knowing with infinite knowledge can perform actions to modify future percepts Use this to obtain useful information Information gathering, Exploration An Autonomous Agent: Behavior is determined by its own experience With ability to learn and adapt

7/24 Rational? Evaluating Rationality is distinct from omniscience All-knowing with infinite knowledge can perform actions to modify future percepts Use this to obtain useful information Information gathering, Exploration An Autonomous Agent: Behavior is determined by its own experience With ability to learn and adapt

7/24 Rational? Evaluating Rationality is distinct from omniscience All-knowing with infinite knowledge can perform actions to modify future percepts Use this to obtain useful information Information gathering, Exploration An Autonomous Agent: Behavior is determined by its own experience With ability to learn and adapt

8/24 Rational :PEAS Evaluating PEAS: Performance measure, Environment, Actuators, Sensors Must first specify the setting for intelligent agent design Example: Task of designing an Automated Taxi Driver Performance: Safe, fast, legal, comfort, maximize profits Environment: Roads, other traffic, pedestrians, customers Actuators: Steering wheel, accelerator, brake, signal, horn Sensors: Cameras, sonar, speedometer, GPS, odometer, engine sensors, keyboard

8/24 Rational :PEAS Evaluating PEAS: Performance measure, Environment, Actuators, Sensors Must first specify the setting for intelligent agent design Example: Task of designing an Automated Taxi Driver Performance: Safe, fast, legal, comfort, maximize profits Environment: Roads, other traffic, pedestrians, customers Actuators: Steering wheel, accelerator, brake, signal, horn Sensors: Cameras, sonar, speedometer, GPS, odometer, engine sensors, keyboard

8/24 Rational :PEAS Evaluating PEAS: Performance measure, Environment, Actuators, Sensors Must first specify the setting for intelligent agent design Example: Task of designing an Automated Taxi Driver Performance: Safe, fast, legal, comfort, maximize profits Environment: Roads, other traffic, pedestrians, customers Actuators: Steering wheel, accelerator, brake, signal, horn Sensors: Cameras, sonar, speedometer, GPS, odometer, engine sensors, keyboard

9/24 Rational :PEAS Evaluating PEAS: Agent: Medical Diagnosis System Performance: Healthy patient, minimize costs, lawsuits Environment: Patient, hospital, staff Actuators: Screen (questions, tests, diagnoses, treatments, referrals) Sensors: Keyboard (entry of symptoms, findings, patient s answers)

10/24 Rational :PEAS Evaluating PEAS: Agent: Part-picking Robot Performance measure: Percentage of parts in correct bins Environment: Conveyor belt with parts, bins Actuators: Jointed arm and hand Sensors: Camera, joint angle sensors

11/24 Rational :PEAS Evaluating PEAS: Agent: Interactive English Tutor Performance measure: Maximize student s score on test Environment: Set of students Actuators: Screen (exercises, suggestions, corrections) Sensors: Keyboard

12/24 Rational :PEAS Evaluating Four basic types in order of increasing generality Simple reflex agents Model-based reflex agents Goal-based agents Utility-based (not just that we reach the goal) agents We consider (3) and (4) together.

12/24 Rational :PEAS Evaluating Four basic types in order of increasing generality Simple reflex agents Model-based reflex agents Goal-based agents Utility-based (not just that we reach the goal) agents We consider (3) and (4) together.

13/24 Evaluating agent sensors what the world is like now condaction rules what action should I do now? effectors e n v i r o n m e n t

Evaluating function Agent (percept) returns action static: memory memory UpdateMemory(memory, percept) ; Agent stores percept sequences in memory ; Only one input percept per invocation action ChooseBestAction(memory) memory UpdateMemory(memory, action) ; Performance measure: Evaluated externally return action Issues to be considered: Model-based Reflex agents Keeping track of the world agents Goal-based agents Utility-based agents... 14/24

15/24 Evaluating agent state how the world changes what my actions do condaction rules sensors what the world is like now what it will be like if I do A what action should I do now? effectors e n v i r o n m e n t

16/24 Evaluating Works only if a correct decision can be made on basis of current percept (à la subsumption architecture) function Agent (percept) returns action static: rules state InterpretInput(percept) ;Description of world s state from percept rule RuleMatch(state, rules) ;Returns a rule matching state description action RuleAction(rule) return action NEXT: What to do when world is partially observable

Evaluating function Agent (percept) returns action static: rules state ;World state state InterpretInput(percept) ;Description of world state from percept state UpdateState(state, percept) ;Hard! Presupposes knowledge about how: ;(1) World changes independently of agent ;(2) Agent s actions effect the world rule RuleMatch(state, rules) ;Returns a rule matching state description action RuleAction(rule) state UpdateState(state, action) ;Hard! Record unsensed parts of World ;Hard! Record effects of agent s actions return action 17/24

18/24 Evaluating Actions depends on current state and goal... Often: Goal satisfaction requires sequences of actions What will happen if I do this? Credit assignment Goals are not enough Some goal-achieving sequences are cheaper, faster, etc. Utility: states reals Tradeoffs on goal...

18/24 Evaluating Actions depends on current state and goal... Often: Goal satisfaction requires sequences of actions What will happen if I do this? Credit assignment Goals are not enough Some goal-achieving sequences are cheaper, faster, etc. Utility: states reals Tradeoffs on goal...

19/24 Evaluating agent state how the world changes what my actions do utility goals rules sensors what the world is like now what it will be like if I do A how happy will I be in such a state? what action should I do now? effectors e n v i r o n m e n t

Evaluating function RunEvalEnvironment (state, UpdateFn, agents,termination,performfn) ;Have multiple agents; Returns scores ;State, UpdateFn: Simulate Environment; ;These are unseen by agents! ;Agent s states: Constructed from percepts ; have no access to PerformFn! repeat for each agent in agents do Percept[agent] GetPercept(agent, state) for each agent in agents do Action[agent] Program[agent](Percept[agent]) state UpdateFn(actions, agents, state) scores PerformFn(scores, agents, state) until termination return scores 20/24

21/24 Types of Environments Source of Actions Types of Environments Fully Observable/Accessible vs. Partially Observable: Agent s sensors: Access environment s complete state Deterministic (or not) i.e., Stochastic Next state completely determined by current state & action If the environment is deterministic except for the actions of other agents, then the environment is strategic Episodic (or not) The agent s experience is divided into atomic episodes Each episode consists of the agent perceiving and then performing a single action The choice of action in each episode depends only on the episode itself

21/24 Types of Environments Source of Actions Types of Environments Fully Observable/Accessible vs. Partially Observable: Agent s sensors: Access environment s complete state Deterministic (or not) i.e., Stochastic Next state completely determined by current state & action If the environment is deterministic except for the actions of other agents, then the environment is strategic Episodic (or not) The agent s experience is divided into atomic episodes Each episode consists of the agent perceiving and then performing a single action The choice of action in each episode depends only on the episode itself

21/24 Types of Environments Source of Actions Types of Environments Fully Observable/Accessible vs. Partially Observable: Agent s sensors: Access environment s complete state Deterministic (or not) i.e., Stochastic Next state completely determined by current state & action If the environment is deterministic except for the actions of other agents, then the environment is strategic Episodic (or not) The agent s experience is divided into atomic episodes Each episode consists of the agent perceiving and then performing a single action The choice of action in each episode depends only on the episode itself

22/24 Types of Environments Source of Actions Types of Environments Static (or not) Environment does not change while the agent deliberates Discrete (or not) Fixed number of well-defined percepts and actions Single agent (vs. Multiagent) An agent operating by itself in an environment The Real World Of course: partially observable, stochastic, sequential, dynamic, continuous, multi-agent Chess: Accessible, Deterministic, Episodic, Static, Discrete Diagnosis: Access., Determin., Episodic, Static, Discrete

22/24 Types of Environments Source of Actions Types of Environments Static (or not) Environment does not change while the agent deliberates Discrete (or not) Fixed number of well-defined percepts and actions Single agent (vs. Multiagent) An agent operating by itself in an environment The Real World Of course: partially observable, stochastic, sequential, dynamic, continuous, multi-agent Chess: Accessible, Deterministic, Episodic, Static, Discrete Diagnosis: Access., Determin., Episodic, Static, Discrete

22/24 Types of Environments Source of Actions Types of Environments Static (or not) Environment does not change while the agent deliberates Discrete (or not) Fixed number of well-defined percepts and actions Single agent (vs. Multiagent) An agent operating by itself in an environment The Real World Of course: partially observable, stochastic, sequential, dynamic, continuous, multi-agent Chess: Accessible, Deterministic, Episodic, Static, Discrete Diagnosis: Access., Determin., Episodic, Static, Discrete

22/24 Types of Environments Source of Actions Types of Environments Static (or not) Environment does not change while the agent deliberates Discrete (or not) Fixed number of well-defined percepts and actions Single agent (vs. Multiagent) An agent operating by itself in an environment The Real World Of course: partially observable, stochastic, sequential, dynamic, continuous, multi-agent Chess: Accessible, Deterministic, Episodic, Static, Discrete Diagnosis: Access., Determin., Episodic, Static, Discrete

22/24 Types of Environments Source of Actions Types of Environments Static (or not) Environment does not change while the agent deliberates Discrete (or not) Fixed number of well-defined percepts and actions Single agent (vs. Multiagent) An agent operating by itself in an environment The Real World Of course: partially observable, stochastic, sequential, dynamic, continuous, multi-agent Chess: Accessible, Deterministic, Episodic, Static, Discrete Diagnosis: Access., Determin., Episodic, Static, Discrete

23/24 Types of Environments Source of Actions Source of Actions Selected by the Agent Performance Element Agent program to select actions Learning Element Improves PE and makes agent s behavior robust In initially unknown environments Problem Generator Suggests actions May lead to new, informative experiences Exploitation vs Exploration

24/24 Types of Environments Source of Actions Source of Actions Selected by the Agent Performance Standard feedback Agent learning goals Critic Learning Element Problem Generator changes Knowledge Sensors Performance Element Actuators E n v i r o n m e n t