Intelligent Agents. Soleymani. Artificial Intelligence: A Modern Approach, Chapter 2

Similar documents
Intelligent Agents. Outline. Agents. Agents and environments

Artificial Intelligence. Intelligent Agents

Intelligent Agents. Chapter 2 ICS 171, Fall 2009

Intelligent Agents. BBM 405 Fundamentals of Artificial Intelligence Pinar Duygulu Hacettepe University. Slides are mostly adapted from AIMA

Dr. Mustafa Jarrar. Chapter 2 Intelligent Agents. Sina Institute, University of Birzeit

Princess Nora University Faculty of Computer & Information Systems ARTIFICIAL INTELLIGENCE (CS 370D) Computer Science Department

Chapter 2: Intelligent Agents

CS 771 Artificial Intelligence. Intelligent Agents

Intelligent Agents. Instructor: Tsung-Che Chiang

Intelligent Autonomous Agents. Ralf Möller, Rainer Marrone Hamburg University of Technology

Web-Mining Agents Cooperating Agents for Information Retrieval

Intelligent Agents. Instructor: Tsung-Che Chiang

Agents and Environments

Artificial Intelligence CS 6364

Introduction to Artificial Intelligence 2 nd semester 2016/2017. Chapter 2: Intelligent Agents

Web-Mining Agents Cooperating Agents for Information Retrieval

CS 331: Artificial Intelligence Intelligent Agents

22c:145 Artificial Intelligence

Agents & Environments Chapter 2. Mausam (Based on slides of Dan Weld, Dieter Fox, Stuart Russell)

CS 331: Artificial Intelligence Intelligent Agents

Outline for Chapter 2. Agents. Agents. Agents and environments. Vacuum- cleaner world. A vacuum- cleaner agent 8/27/15

AI: Intelligent Agents. Chapter 2

Outline. Chapter 2 Agents & Environments. Agents. Types of Agents: Immobots

Agents & Environments Chapter 2. Mausam (Based on slides of Dan Weld, Dieter Fox, Stuart Russell)

Artificial Intelligence Agents and Environments 1

CS 331: Artificial Intelligence Intelligent Agents. Agent-Related Terms. Question du Jour. Rationality. General Properties of AI Systems

Intelligent Agents. CmpE 540 Principles of Artificial Intelligence

Foundations of Artificial Intelligence

Lecture 2 Agents & Environments (Chap. 2) Outline

Contents. Foundations of Artificial Intelligence. Agents. Rational Agents

Artificial Intelligence

Agents. This course is about designing intelligent agents Agents and environments. Rationality. The vacuum-cleaner world

Artificial Intelligence Lecture 7

Rational Agents (Chapter 2)

Intelligent Agents. Russell and Norvig: Chapter 2

Agents. Environments Multi-agent systems. January 18th, Agents

Artificial Intelligence Intelligent agents

Intelligent Agents. Chapter 2

What is AI? The science of making machines that: Think rationally. Think like people. Act like people. Act rationally

Agents and State Spaces. CSCI 446: Artificial Intelligence

Agents. Formalizing Task Environments. What s an agent? The agent and the environment. Environments. Example: the automated taxi driver (PEAS)

Vorlesung Grundlagen der Künstlichen Intelligenz

Intelligent Agents. Philipp Koehn. 16 February 2017

How do you design an intelligent agent?

Agents and Environments. Stephen G. Ware CSCI 4525 / 5525

KECERDASAN BUATAN 3. By Sirait. Hasanuddin Sirait, MT

CS324-Artificial Intelligence

Artificial Intelligence

Ar#ficial Intelligence

Agents and Environments

Overview. What is an agent?

Agents and State Spaces. CSCI 446: Ar*ficial Intelligence Keith Vertanen

AI Programming CS F-04 Agent Oriented Programming

CS343: Artificial Intelligence

Solutions for Chapter 2 Intelligent Agents

Agents. Robert Platt Northeastern University. Some material used from: 1. Russell/Norvig, AIMA 2. Stacy Marsella, CS Seif El-Nasr, CS4100

Rational Agents (Ch. 2)

Module 1. Introduction. Version 1 CSE IIT, Kharagpur

Rational Agents (Ch. 2)

Silvia Rossi. Agent as Intentional Systems. Lezione n. Corso di Laurea: Informatica. Insegnamento: Sistemi multi-agente.


AGENT-BASED SYSTEMS. What is an agent? ROBOTICS AND AUTONOMOUS SYSTEMS. Today. that environment in order to meet its delegated objectives.

COMP329 Robotics and Autonomous Systems Lecture 15: Agents and Intentions. Dr Terry R. Payne Department of Computer Science

What is AI? The science of making machines that:

Artificial Intelligence. Admin.

GRUNDZÜGER DER ARTIFICIAL INTELLIGENCE

Definitions. The science of making machines that: This slide deck courtesy of Dan Klein at UC Berkeley

Unmanned autonomous vehicles in air land and sea

1 What is an Agent? CHAPTER 2: INTELLIGENT AGENTS

Behavior Architectures

Katsunari Shibata and Tomohiko Kawano

Four Example Application Domains

Artificial Intelligence. Outline

A Scoring Policy for Simulated Soccer Agents Using Reinforcement Learning

OPTIC FLOW IN DRIVING SIMULATORS

CS148 - Building Intelligent Robots Lecture 5: Autonomus Control Architectures. Instructor: Chad Jenkins (cjenkins)

GENERAL STATEMENT OF JOB. Specific Duties and Responsibilities

Sequential Decision Making

Problem Solving Agents

Robotics Summary. Made by: Iskaj Janssen

Re: ENSC 370 Project Gerbil Functional Specifications

Robot Behavior Genghis, MIT Callisto, GATech

Bayesian Perception & Decision for Intelligent Mobility

Effects of Sequential Context on Judgments and Decisions in the Prisoner s Dilemma Game

About Varibel. Varibel strives to improve the quality of life for everyone who comes in touch with our products.

Discover Simple Neuroscience Body Hacks that Easily Increase Personal Performance, Accelerate Physical Results and Relieve Pain

AVR Based Gesture Vocalizer Using Speech Synthesizer IC

Introduction to Arti Intelligence

Speak Out! Sam Trychin, Ph.D. Copyright 1990, Revised Edition, Another Book in the Living With Hearing Loss series

Affective Preference From Physiology in Videogames: a Lesson Learned from the TORCS Experiment

Embodiment in GLAIR: A Grounded Layered Architecture. with Integrated Reasoning for Autonomous Agents. Henry Hexmoor. Johan Lammens.

Introduction to Decision Making

Solving problems by searching

Ophthalmology. Chairs That Optimize Eye Exams and Eye Surgeries. Designed and manufactured in Denmark since 1935

Protecting Workers with Smart E-Vests

Perceptual Anchoring with Indefinite Descriptions

Introduction to Deep Reinforcement Learning and Control

Measuring the Impossible : Quantifying the Subjective (A method for developing a Risk Perception Scale applied to driving)

Chapter 5 Car driving

Achievement: Approach versus Avoidance Motivation

Transcription:

Intelligent Agents CE417: Introduction to Artificial Intelligence Sharif University of Technology Spring 2016 Soleymani Artificial Intelligence: A Modern Approach, Chapter 2

Outline Agents and environments Rationality PEAS (Performance measure, Environment, Actuators, Sensors) Environment types Agent types 2

Agents An agent is anything that can be viewed as Sensors: perceive environment Actuators: act upon environment Samples of agents 3 Human agent Sensors: eyes, ears, and other organs for sensors Actuators: hands, legs, vocal tract, and other movable or changeable body parts Robotic agent Sensors: cameras and infrared range finders Actuators: various motors Software agents Sensors: keystrokes, file contents, received network packages Actuators: displays on the screen, files, sent network packets

Agents & environments Agent behavior can be described as an agent function that maps entire perception histories to actions: f: P A Percept sequence to date Action set The agent program runs on the physical architecture to produce f Program is a concrete implementation of agent function Architecture includes sensors, actuators, computing device agent = architecture + program 4

Vacuum-cleaner world Percepts: location and dirt/clean status of its location e.g., [A,Dirty] Actions: Left, Right, Suck, NoOp 5

A vacuum-cleaner agent Tabulation of the agent function Percept Sequence [A, Clean] [A, Dirty] [B, Clean] Action Right Suck Left [ B, Dirty] Suck [A, Clean], [A, Clean] [A, Clean], [A, Dirty] Right Suck One simple rule implementing the agent function: If the current square is dirty then suck, otherwise move to the other square 6 [A, Clean], [A, Clean], [A, Clean] [A, Clean], [A, Clean], [A, Dirty] Right Suck

Rational agents "do the right thing" based on the perception history and the actions it can perform. Rational Agent: For each possible percept sequence, a rational agent should select an action that is expected to maximize its performance measure, given the evidence provided by the percept sequence and whatever built-in knowledge the agent has. 7

Performance measure Evaluates the sequence of environment states Vacuum-cleaner agent: samples of performance measure Amount of dirt cleaned up One point award for each clean square at each time step Penalty for electricity consumption & generated noise Mediocre job or periods of high and low activation? 8

Rational agents (vacuum cleaner example) Is this rational? If dirty then suck, otherwise move to the other square Depends on Performance measure, e.g., Penalty for energy consumption? Environment, e.g., New dirt can appear? Actuators, e.g., No-op action? Sensors, e.g., Only sense dirt in its location? 9

Rationality vs. Omniscience Rationality is distinct from omniscience (all-knowing with infinite knowledge, impossible in reality) Doing actions in order to modify future percepts to obtain useful information information gathering or exploration (important for rationality) e.g., eyeballs and/or neck movement in human to see different directions 10

Autonomy An agent is autonomous if its behavior is determined by its own experience (with ability to learn and adapt) Not just relies only on prior knowledge of designer Learns to compensate for partial or incorrect prior knowledge Benefit: changing environment Starts by acting randomly or base on designer knowledge and then learns form experience Rational agent should be autonomous Example: vacuum-cleaner agent If dirty then suck, otherwise move to the other square Does it yield an autonomous agent? learning to foresee occurrence of dirt in squares 11

Task Environment (PEAS) Performance measure Environment Actuators Sensors 12

PEAS Samples Agent:Automated taxi driver Performance measure: Safe, fast, legal, comfortable trip, maximize profits, Environment: Roads, other traffic, pedestrians, customers, Actuators: Steering wheel, accelerator, brake, signal, horn, display Sensors: Cameras, sonar, speedometer, GPS, odometer, accelerometer, engine sensors, keyboard 13

PEAS Samples Agent: Medical diagnosis system Performance measure: Healthy patient, minimize costs Environment: Patient, hospital, staff Actuators: Screen display (questions, tests, diagnoses, treatments, referrals) Sensors: Keyboard (entry of symptoms, findings, patient's answers) 14

PEAS Samples Satellite image analysis system Performance measure: Correct image categorization Environment: Downlink from orbiting satellite Actuators: Display of scene categorization Sensors: Color pixel array 15

PEAS Samples Agent: Part picking robot Performance measure: Percentage of parts in correct bins Environment: Conveyor belt with parts, bins Actuators: Jointed arm and hand Sensors: Camera, joint angle sensors 16

PEAS Samples Agent: Interactive English tutor Performance measure: Maximize student's score on test Environment: Set of students Actuators: Screen display (exercises, suggestions, corrections) Sensors: Keyboard 17

PEAS Samples Agent: Pacman Performance measure: Score, lives Environment: Maze containing white dots, four ghosts, power pills, occasionally appearing fruit Actuators:Arrow keys Sensors: Game screen 18

Environment types Fully observable (vs. partially observable): Sensors give access to the complete state of the environment at each time Sensors detect all aspects relevant to the choice of action Convenient (need not any internal state) Noisy and inaccurate sensors or missing parts of the state from sensors cause partially observability 19

Environment types Deterministic (vs. stochastic): Next state can be completely determined by the current state and the executed action If the environment is deterministic except for the actions of other agents, then the environment is strategic (we ignore this uncertainty) Partially observable environment could appear to be stochastic. Environment is uncertain if it is not fully observable or not deterministic 20

Environment types Single agent (vs. multi-agent): Crossword puzzle is a single-agent game (chess is a multi-agent one) Is B an agent or just an object in the environment? B is an agent when its behavior can be described as maximizing a performance measure whose value depends on A s behavior. Multi-agent: competitive, cooperative Randomized behavior and communication can be rational Discrete (vs. continuous): A limited number of distinct, clearly defined states, percepts and actions, time steps Chess has finite number of discrete states, and discrete set of percepts and actions while Taxi driving has continuous states, and actions 21

Environment types Episodic (vs. sequential): The agent's experience is divided into atomic "episodes where the choice of action in each episode depends only on the episode itself. E.g., spotting defective parts on an assembly line (independency) In sequential environments, short-term actions can have long-term consequences Episodic environment can be much simpler Static (vs. dynamic): The environment is unchanged while an agent is deliberating. Semi-dynamic: if the environment itself does not change with the passage of time but the agent's performance score does. Static (cross-word puzzles), dynamic (taxi driver), semi-dynamic (clock chess) 22

Environment types Known (vs. unknown): the outcomes or (outcomes probabilities for all actions are given. It is not strictly a property of the environment Related to agent s or designer s state of knowledge about laws of physics of the environment The real world is partially observable, multi-agent, stochastic, sequential, dynamic, continuous, (and unknown) Hardest type of environment The environment type largely determines the agent design 23

Pacman game Fully observable? Single-agent? Deterministic? Discrete? Episodic? Static? Known? 24

Environment types 25

Environment types 26

Structure of agents An agent is completely specified by the agent function (that maps percept sequences to actions) One agent function or small equivalent class is rational Agent program implements agent function (focus of our course) Agent program takes just the current percept as input Agent needs to remember the whole percept sequence, if requiring it (internal state) 27

Agent Program Types Lookup table Basic types of agent program in order of increasing generality: Simple reflexive Model-based reflex agents Goal-based agents We mainly focus on this type in our course Utility-based agents Learning-based agents 28

Look Up Table Agents Benefits: Easy to implement Drawbacks: T space ( t=1 P t ; P : set of possible percepts, T: lifetime) For chess at least 10 150 entries while less than 10 80 atoms in the observable universe the designer have not time to create table no agent could ever learn the right table entries from its experience how to fill in table entries? 29

Agent program Mapping is not necessarily using a table. AI intends to find programs producing rational behavior (to the extent possible) from a smallish program instead of a vast table Can AI do for general intelligent behavior what Newton did for square roots? Huge tables of square roots before 1970s 1.0 1.00000 1.1 1.04898 1.2 1.09565 1.3 1.14056... function SQRT( double X ) { double r = 1.0 ; while ( fabs( r * r - x ) > 10-8 ) r = r - ( r * r - x ) / 2r ; return r ; } 30

Simple Reflex Agents Agent Program 31

Simple Reflex Agents Select actions on the basis of the current percept ignoring the rest of the percept history If car-in-front-is-braking then initiate-braking Blinking reflex 32

Simple Reflex Agents Simple, but very limited intelligence Works only if the correct decision can be made on the basis of the current percept (fully observability) Infinite loops in partially observable environment 33

34 Model-based reflex agents

Model-based reflex agents Partial observability Internal state (based on percept history) reflects some unobserved aspects of the current state Updating the internal state information requires two kinds of knowledge Information about how the world evolves (independent of agent) Information about how the agent's own actions affects the world Only determine the best guess for the current state of a partially observable environment 35

Goal-based agents 36

Goal-based agents Knowing about the current state is not always enough to decide what to do Situations that are desirable must be specified (goal) Usually requires search and planning to find action sequences achieving goal 37

Goal-based agents vs. reflex-based agents Consideration of future Goal-based agents may be less efficient but are more flexible Knowledge is represented explicitly and can be changed easily Example: going to a new destination Goal-based agent: specifying that destination as the goal Reflexive agent: agent's rules for when to turn and when to go straight must be rewritten 38

Utility-based agents 39

Utility-based agents More general performance measure than goals How happy would each world state make the agent? Utility function is an internalization of performance measure Advantages Like goal-based agents show flexibility and learning advantages Can trade-off conflicting goals (e.g. speed and safety) Where none of several goals can be achieved with certainty likelihood of success can be weighted by importance of goals Rational utility-based agent chooses the action that maximizes the expected utility of action outcomes Many chapters of AIMA book is about this Handling uncertainty in partially observable and stochastic environments 40

Learning Agents 41

Learning Agents Create state-of-the-art systems in many areas of AI Four conceptual components Performance element: selects actions based on percepts (considered as entire agent before) Learning element: makes improvements by modifying knowledge (performance element) based on critic feedback Critic: feedbacks on how the agents is doing Problem generator: suggests actions leading to new and informative experiences Performance standard Fixed, out of the agent Percepts themselves do not provide indication of success Distinguishes part of the incoming percept as a reward or penalty 42

Learning Agents Learning element is based on performance element (i.e., agent design) The learning element can make changes to any of the "knowledge" components of previous agent diagrams To bring components into closer agreement with feedback (yielding better overall performance) 43