Agents and Environments. Stephen G. Ware CSCI 4525 / 5525

Similar documents
Introduction to Artificial Intelligence 2 nd semester 2016/2017. Chapter 2: Intelligent Agents

Chapter 2: Intelligent Agents

CS 331: Artificial Intelligence Intelligent Agents

CS 331: Artificial Intelligence Intelligent Agents

Artificial Intelligence. Intelligent Agents

CS 331: Artificial Intelligence Intelligent Agents. Agent-Related Terms. Question du Jour. Rationality. General Properties of AI Systems

Intelligent Agents. Outline. Agents. Agents and environments

Foundations of Artificial Intelligence

Contents. Foundations of Artificial Intelligence. Agents. Rational Agents

Artificial Intelligence

Agents and Environments

Princess Nora University Faculty of Computer & Information Systems ARTIFICIAL INTELLIGENCE (CS 370D) Computer Science Department

Dr. Mustafa Jarrar. Chapter 2 Intelligent Agents. Sina Institute, University of Birzeit

Intelligent Agents. Chapter 2 ICS 171, Fall 2009

Agents & Environments Chapter 2. Mausam (Based on slides of Dan Weld, Dieter Fox, Stuart Russell)

Intelligent Autonomous Agents. Ralf Möller, Rainer Marrone Hamburg University of Technology

CS 771 Artificial Intelligence. Intelligent Agents

AI: Intelligent Agents. Chapter 2

Agents & Environments Chapter 2. Mausam (Based on slides of Dan Weld, Dieter Fox, Stuart Russell)

Intelligent Agents. Soleymani. Artificial Intelligence: A Modern Approach, Chapter 2

Web-Mining Agents Cooperating Agents for Information Retrieval

Artificial Intelligence CS 6364

Intelligent Agents. Chapter 2

Intelligent Agents. Instructor: Tsung-Che Chiang

Agents and State Spaces. CSCI 446: Artificial Intelligence

Agents. This course is about designing intelligent agents Agents and environments. Rationality. The vacuum-cleaner world

Intelligent Agents. Instructor: Tsung-Che Chiang

Web-Mining Agents Cooperating Agents for Information Retrieval

AI Programming CS F-04 Agent Oriented Programming

Agents. Environments Multi-agent systems. January 18th, Agents

Intelligent Agents. BBM 405 Fundamentals of Artificial Intelligence Pinar Duygulu Hacettepe University. Slides are mostly adapted from AIMA

Outline. Chapter 2 Agents & Environments. Agents. Types of Agents: Immobots

Artificial Intelligence Lecture 7

Agents and Environments

Artificial Intelligence Agents and Environments 1

22c:145 Artificial Intelligence

Rational Agents (Chapter 2)

Agents. Formalizing Task Environments. What s an agent? The agent and the environment. Environments. Example: the automated taxi driver (PEAS)

Intelligent Agents. Russell and Norvig: Chapter 2

Overview. What is an agent?

Rational Agents (Ch. 2)

Artificial Intelligence Intelligent agents

Rational Agents (Ch. 2)

CS324-Artificial Intelligence

Vorlesung Grundlagen der Künstlichen Intelligenz

Outline for Chapter 2. Agents. Agents. Agents and environments. Vacuum- cleaner world. A vacuum- cleaner agent 8/27/15

Intelligent Agents. CmpE 540 Principles of Artificial Intelligence

Ar#ficial Intelligence

Agents. Robert Platt Northeastern University. Some material used from: 1. Russell/Norvig, AIMA 2. Stacy Marsella, CS Seif El-Nasr, CS4100

Lecture 2 Agents & Environments (Chap. 2) Outline

Intelligent Agents. Philipp Koehn. 16 February 2017

How do you design an intelligent agent?

Module 1. Introduction. Version 1 CSE IIT, Kharagpur

Solutions for Chapter 2 Intelligent Agents

KECERDASAN BUATAN 3. By Sirait. Hasanuddin Sirait, MT

What is AI? The science of making machines that: Think rationally. Think like people. Act like people. Act rationally

Agents and State Spaces. CSCI 446: Ar*ficial Intelligence Keith Vertanen

Artificial Intelligence

CS343: Artificial Intelligence

Silvia Rossi. Agent as Intentional Systems. Lezione n. Corso di Laurea: Informatica. Insegnamento: Sistemi multi-agente.


Behavior Architectures

Artificial Intelligence. Outline

COMP329 Robotics and Autonomous Systems Lecture 15: Agents and Intentions. Dr Terry R. Payne Department of Computer Science

What is AI? The science of making machines that:

Definitions. The science of making machines that: This slide deck courtesy of Dan Klein at UC Berkeley

Sequential Decision Making

AGENT-BASED SYSTEMS. What is an agent? ROBOTICS AND AUTONOMOUS SYSTEMS. Today. that environment in order to meet its delegated objectives.

GRUNDZÜGER DER ARTIFICIAL INTELLIGENCE

Introduction to Arti Intelligence

1 What is an Agent? CHAPTER 2: INTELLIGENT AGENTS

What is Artificial Intelligence? A definition of Artificial Intelligence. Systems that act like humans. Notes

Category-Based Intrinsic Motivation

Problem Solving Agents

Robot Behavior Genghis, MIT Callisto, GATech

Solving problems by searching

Robotics Summary. Made by: Iskaj Janssen

Robot Learning Letter of Intent

LAURIE PARMA, UNIVERSITY OF CAMBRIDGE ENGAGEMENT, CREATIVITY, PASSION AND FLOW

Go Suck a Lemon: Strategies for Improving Your Emotional Intelligence. Track 1 Session 4

Artificial Intelligence Programming Probability

Reinforcement Learning

Dynamic Control Models as State Abstractions

Cognitive and Computational Neuroscience of Aging

Beyond Skill Acquisition;

Lecture 13: Finding optimal treatment policies

Neurophysiology and Information: Theory of Brain Function

COGNITIVE DEVELOPMENT AP PSYCHOLOGY

A Theory of Complexity and its Relevance to Better Decision Making

Applying Appraisal Theories to Goal Directed Autonomy

The Vine Assessment System by LifeCubby

Multidimensional Approach

Introduction to Computational Neuroscience

HearIntelligence by HANSATON. Intelligent hearing means natural hearing.

Bayesian Perception & Decision for Intelligent Mobility

Emotional Intelligence

CS148 - Building Intelligent Robots Lecture 5: Autonomus Control Architectures. Instructor: Chad Jenkins (cjenkins)

ISSN (PRINT): , (ONLINE): , VOLUME-5, ISSUE-5,

Affective Game Engines: Motivation & Requirements

Content Area: Physical Education Grade Level Expectations: Fifth Grade Standard: 1. Movement Competence & Understanding in Physical Education

Transcription:

Agents and Environments Stephen G. Ware CSCI 4525 / 5525

Agents An agent (software or hardware) has: Sensors that perceive its environment Actuators that change its environment Environment Sensors Actuators

Vacuum World Environment: Room A Room B Dirt Vacuum Agent: Room Sensor Dirt Sensor Motor Vacuum A B

Agent Function An agent function maps any given percept sequence to an action. Agent functions are generally implemented in software as agent programs.

Vacuum Agent Function Percept sequence: 1. You are in Room A. 2. Room A is dirty. Action: suck Percept sequence: 1. You are in Room A. 2. Room A is clean. Action: go right A A B B

Vacuum Agent Function Percept sequence: 1. You are in Room A. 2. Room A is clean. 3. You are in Room B. 4. Room B is dirty. Action: suck A B

Performance Measures An agent s success is defined according to some performance measure, which must be defined for each task the agent is expected to perform.

Vacuum Performance Example: Score 1 point for every pile of dirt sucked up in an 8 hour shift. Agent sucks dirt, then dumps it out, then sucks it up again, then dumps it out again, etc Excuse me, do you know what time it is? Yes, I do.

Vacuum Performance Better: Score 1 point for each clean room every minute. Lesson: Design performance measures according to what you want in the environment, not how you think the agent should behave.

Rational Agent Things to consider: The performance measure The agent s prior knowledge The percept sequence to date The agent s available actions

Rational Agent For each possible percept sequence, a rational agent should select an action that is expected to maximize its performance measure, given the evidence provided by the percept sequence and whatever built-in knowledge the agent has.

Omniscience Rational agents maximize expected performance, not necessarily actual performance. Agents are not expected to be omniscient.

Coping with Limited Knowledge Some agents know what they do not know and act to improve their knowledge via information gathering, such as sensing and exploration. Some agents can update their own agent programs based on observations, called learning. Agents which gather information and learn are autonomous, whereas agent which rely only on prior knowledge lack autonomy and are fragile.

A Bug s Life Dung Beetle Sphex Wasp

Environments Single Agent vs. Multi-Agent Episodic vs. Sequential Static vs. Dynamic Fully Observable vs. Partially Observable Deterministic vs. Stochastic Known vs. Unknown Discrete vs. Continuous

Single Agent vs. Multi-Agent Single Agent: Everything else in the environment can be treated as an object, rather than an agent. Multi-Agent: The actions of other objects in the environment will depend on the agent s actions, and should therefore be viewed as other agents. Competitive: Maximizing my performance may hurt another s performance. Cooperative: Through coordinated action, we can both achieve better performance.

Episodic vs. Sequential Episodic: An agent s current choice of actions does not depend on previous choices. Sequential: An agent s current choice of actions will effect which actions are available in the future, and thus some kind of planning is required.

Static vs. Dynamic Static: The agents are the only cause of change in the environment. Time is not a factor. Dynamic: The environment may change independently of the agents. Time is a factor. Semidynamic: The agents are the only source of change, but time is a factor.

Fully vs. Partially Observable Fully Observable: Every relevant piece of information is available on demand. The agent does not need to track the environment s state. Partially Observable: Some relevant information is missing and must be assumed or gathered. Information is unavailable due to the agent s state. Appropriate sensors are missing or malfunctioning. Sensory data is noisy.

Deterministic vs. Uncertain Deterministic: Given the current state, we can know exactly the outcome of an action. Uncertain: The next state cannot be known with full confidence because the environment is not fully observable or action outcomes unpredictable. Stochastic: Outcome probabilities are known. Nondeterministic: Probabilities not known.

Known vs. Unknown Known: The rules (even if they are stochastic) of the environment are known in advance. Unknown: The rules of the environment are not known and must be learned.

Discrete vs. Continuous Discrete: The environment s state and time and the agent s actions can be easily divided into clearly labeled classes. Continuous: State and/or time and/or actions cannot be clearly divided into distinct classes.

Environments Single Agent vs. Multi-Agent Episodic vs. Sequential Static vs. Dynamic Fully Observable vs. Partially Observable Deterministic vs. Stochastic Known vs. Unknown Discrete vs. Continuous

The Easiest Environment Single Agent vs. Multi-Agent Episodic vs. Sequential Static vs. Dynamic Fully Observable vs. Partially Observable Deterministic vs. Stochastic Known vs. Unknown Discrete vs. Continuous

The Real World Single Agent vs. Multi-Agent Episodic vs. Sequential Static vs. Dynamic Fully Observable vs. Partially Observable Deterministic vs. Stochastic Known vs. Unknown Discrete vs. Continuous

Vacuum World A B Observable: Agents: Deterministic: Episodic: Static: Discrete: Partially Single Deterministic Episodic Static Discrete

Crossword Puzzle Observable: Agents: Deterministic: Episodic: Static: Discrete: Fully Single Deterministic Sequential Static Discrete

Facial Recognition Observable: Agents: Deterministic: Episodic: Static: Discrete: Fully Single Deterministic Episodic Static Continuous

Chess Observable: Agents: Deterministic: Episodic: Static: Discrete: Fully Multi Deterministic Sequential Static Discrete

Chess with a Clock Observable: Agents: Deterministic: Episodic: Static: Discrete: Fully Multi Deterministic Sequential Semidynamic Discrete

Poker Observable: Agents: Deterministic: Episodic: Static: Discrete: Partially Multi Stochastic Sequential Static Discrete

Intelligent Tutoring System Observable: Agents: Deterministic: Episodic: Static: Discrete: Partially Multi Stochastic Sequential Dynamic Discrete

Driving a Taxi Observable: Agents: Deterministic: Episodic: Static: Discrete: Partially Multi Nondeterministic Sequential Dynamic Continuous

Developing AI Solutions Solving real world problems in AI: 1. Make as many simplifying assumptions as possible. 2. Solve the easy problem. 3. Relax one of the simplifying assumptions. 4. Solve the harder problem. 5. Repeat steps 3 and 4 until you have solved the real world problem. A B