Goal Regression Planning, Constraint Automata and Causal Graphs
|
|
- Arthur McKenzie
- 5 years ago
- Views:
Transcription
1 Goal Regression Planning, Constraint Automata and Causal Graphs Contributions: Dan Weld Seung Chung Erez Karpas David Wang Prof Brian Williams MIT CSAIL courtesy of JPL New Text J/6.834J Cognitive Robotics Leap Day - February 29 th, 2016 This image is in the public domain. 1
2 After lecture you will know how to Generate plans by working backwards from goals. Generate least commitment, partial order plans. Encode actions with indirect effects as concurrent automata with constraints (CCA). Analyze action dependence using causal graphs. Generate plans with out search, for CCA that have tree structured causal graphs. Use causal graph planning as a heuristic to HFS. 2/29/ J / 6.834J Goal Regression & Causal Graph Planning 2
3 Assignments Today: D. Weld, An introduction to least commitment planning, AI Magazine 15(4):27-61, B. Williams and P. Nayak, "A reactive planner for a model-based executive, IJCAI, Next: D. Wang and B. Williams, tburton: A Divide and Conquer Temporal Planner, AAAI, Homework: PSet #3 PDDL Modeling, out today, due Wed, March 9 th. Advanced Lecture Pset #1, out today, due Fri, March 4 th. 2/29/ J / 6.834J Goal Regression & Causal Graph Planning 3
4 Outline Review: programs on state Planning as goal regression (SNLP) Goal regression planning with causal graphs (Burton) Appendix: HFS planning with the causal graph heuristic (Fast Downward) 4
5 A single cognitive system language and executive. Enterprise Programs that: Uhura User Goals & models in RMPL Collaboratively resolves goal failures Kirk Make choices Sketches mission and assigns tasks Burton Pike Plans actions Adjust timing Control Commands Coordinates and monitors tasks Sulu Bones Plans paths Diagnoses likely failures Monitor conditions 2/29/2016 source unknown. All rights reserved. This content is excluded from our Creative Commons license. For more information, see Goal Regression and Causal Graph Planning 5
6 In a State-Space Program, Activity Planning Maps Desired States to Actions no-fly zone lake2 fire2 no-fly zone fire1 base2 no-fly zone lake1 uav1 base1 class Main{ UAV uav1; Lake lake1; Lake lake2; Fire fire1; Fire fire2;... Main (){ uav1 = new UAV(); uav1.location= base_1_location; uav1.flying = no; uav1.loaded = no; lake1 = new Lake(); lake1.location = lake_1_location; lake2 = new Lake(); lake2.location = lake_2_location; class UAV { Roadmap location; Boolean flying; Boolean loaded; primitive method takeoff() flying == no => flying == yes; primitive method land() flying == yes => flying == no; primitive method load_water(lake lakespot) ((flying == yes) && (loaded == no) && (lakespot.location == location)) => loaded == yes; method run() { sequence{ }}} (fire1 == out); (fire2 == out); (uav1.flying == no && uav1.location == base_1_location); } fire1 = new Fire(); fire1.location = fire_1_location; fire1 = high; fire2 = new Fire(); fire2.location = fire_2_location; fire2 = high; } primitive method drop_water_high_altiture(fire firespot) ((flying == yes) && (loaded == yes) && (firespot.location == location) && (firespot == high)) => ((loaded == no) && (firespot == medium)); primitive method drop_water_low_altiture(fire firespot) ((flying == yes) && (loaded == yes) && (firespot.location == location) && (firespot == medium)) => ((loaded == no) && (firespot == out)); #MOTION_PRIMITIVES(location, fly, flying==yes) 6
7 Planning maps desired states to actions no-fly zone fire2 no-fly zone base2 no-fly zone lake1 lake2 fire1 uav1 base1 Roadmap Path Planning Classical Action Planning 7
8 A single cognitive system language and executive. Enterprise State Space Programs that: Uhura User Goals & models in RMPL Collaboratively resolves goal failures Kirk Sketches mission and assigns tasks Burton Plan to achieve discrete states. Plans actions Control Commands Pike Coordinates and monitors tasks Sulu Plan to achieve continuous states. Bones Plans paths Diagnoses likely failures Monitor continuous states. source unknown. All rights reserved. This content is excluded from our Creative Commons license. For more information, see 2/29/2016 Goal Regression and Causal Graph Planning 8
9 Last Week: Classic Activity Planning Representation (STRIPS/PDDL) Action Model: Objects (things): Predicates (used to create true or false statements about the world, facts ): (On A, B ) (Clear A ) Actions (used to change truth of predicates): (Clear ) A (Clear ) B Stack (Clear B ) (On A, B ) Initial: Goal: preconditions (Clear ) (Clear ) (Clear ) (On, ) (On, ) effects 9
10 Last Week: Forward Search action3 s 2 0 action1 s 1 0 Initial State (s o ) action2 s 1 0 action4 goal action2 s 2 1 action3 s 1 0 action3 s 2 2 Time 10
11 This Week: Goal-Regression Search action3 S t 0 action1 Initial State (s o ) action2 action2 S t 1 action4 goal action3 S t 2 action2 Time 11
12 This Week: Planning to Control Complex Devices no-fly zone fire2 no-fly zone base2 no-fly zone lake1 lake2 fire1 uav1 base1 Roadmap Path Planning Classical Action Planning Thrust Goals Delta_V(direction=b, magnitude=200) Power This image is in the public domain. Attitude Point(a) Turn(a,b) Point(b) Turn(b,a) Device Reconfiguration & Repair Engine Off Warm Up Thrust (b, 200) O ff Time-line Planning 12
13 This Week: Automata Representation Action Model: Concurrent Constraint Automata (CCA) (like a state-machine): Initial: Goal: Note: This is a very simple example, there are usually many automata, and guards on the transitions. We will use CCA to support indirect effects, concurrency and metric time. Algorithms exist to map between the two representations. 13
14 Outline Review: programs on state Planning as goal regression (SNLP/UCPOP) Partial Order Planning Overview Partial Order Planning Problem Partial Order Plan Generation Goal regression planning with causal graphs (Burton) Appendix: HFS planning with the causal graph heuristic (Fast Downward) 14
15 Classical Planning Problem Propositions: P i Initial Conditions: P 1 P 2 P 3 P 4 Operators: pre 1 pre 2 pre 3 Op eff 1 eff 2 Goals: Goal 1 Goal 2 Goal 3 2/29/ J / 6.834J Goal Regression & Causal Graph Planning 15
16 Simple Spacecraft Problem Observation-1 target instruments Observation-2 Observation-3 Observation-4 pointing calibrated This image is in the public domain. Propositions: Operators: Target Pointed To, Camera Calibrated?, Has Image? Calibrate, Turn to Y, and Take Image. 2/29/ J / 6.834J Goal Regression & Causal Graph Planning 16
17 Example Init Actions Goal p C p C C c I A p y p x T y p x c Im I x p x Propositions: Operators: Target Pointed To, Camera Calibrated?, Has Image? Calibrate, Turn to Y, and Take Image. 2/29/ J / 6.834J Goal Regression & Causal Graph Planning 17
18 Actions in the Planning Domain Description Language (PDDL) (:action TakeImage :parameters (?target,?instr) :precondition (and (Status?instr Calibrated) (Pointing?target)) :effect (Image?target)) By convention, parameters start with?, as in?var. (:action Calibrate :parameters (?instrument) :precondition (and (Status?instr On) (Calibration-Target?target), (Pointing?target) :effect (and (not (Status?inst On)) (Status?instr Calibrated))) (:action Turn :parameters (?target) :precondition (and (Pointing?direction)?direction?target :effect (and (not (Pointing?direction) (Pointing?target))) 2/29/ J / 6.834J Goal Regression & Causal Graph Planning 18
19 Partial Order Plan p C C c S p A Im I A F p C T A p C 2/29/ J / 6.834J Goal Regression & Causal Graph Planning 19
20 Planning from Goals: Partial Order Causal Link Planning (SNLP, UCPOP) I A F 1. Select an open condition; 2. Choose an op that can achieve it: Link to an existing instance or Add a new instance; 3. Resolve threats. p C c I Im A F p A C c p A Im I A F S p C p C C c p A T A Im p C Alternatives: Forward Search, CSP I A S F S p C C c p A p A p C T A 2/29/ J / 6.834J Goal Regression & Causal Graph Planning p C Im I A p C C c Im I A F F 20
21 Planning as Goal Regression Partial Planning Overview Partial Order Planning Problem Problem Encoding Partial Order Plans Plan Correctness Partial Order Plan Generation 2/29/ J / 6.834J Goal Regression & Causal Graph Planning 21
22 Example Problem Initial State: At(Home) Sells(HWS,Drill) Sells(SM,Milk) Sells(SM,Ban.) Goal: Have(Milk) At(Home) Have(Ban.) Have(Drill) Operators: At(HWS) Go(SM) At(SM) At(SM), Sells(SM,Ban.) Buy(Ban.) Have(Ban) At(Home) Go(HWS) At(HWS) At(SM) Go(Home) At(Home) At(HWS) Sells(HWS,Drill) Buy(Drill) Have(Drill) At(SM), Sells(SM,Milk) Buy(Milk) Have(Milk) 2/29/ J / 6.834J Goal Regression & Causal Graph Planning 22
23 Initial and Goal States Encoded as Operators Start At(Home) Sells(HWS,Drill) Sells(SM,Milk) Sells(SM,Ban.) Why encode as operators? Don t need to introduce (partial) states as separate objects. Keeps theory minimal. Have(Milk) At(Home) Have(Ban.) Have(Drill) Finish 2/29/ J / 6.834J Goal Regression & Causal Graph Planning 23
24 Partial Order Plan <Actions,Orderings,Links> Start At(Home) Sells(HWS,Drill) Sells(SM,Milk) Sells(SM,Ban.) At(Home) Go(HWS) At(HWS) Sells(HWS,Drill) Buy(Drill) At(HWS) Go(SM) At(SM), Sells(SM,Milk) Buy(Milk) At(SM), Sells(SM,Ban.) Buy(Ban.) At(SM) Go(Home) Have(Milk) At(Home) Have(Ban.) Have(Drill) Finish 2/29/ J / 6.834J Goal Regression & Causal Graph Planning 24
25 Partial Order Plan <Actions,Orderings,Links> Start At(Home) Sells(HWS,Drill) Sells(SM,Milk) Sells(SM,Ban.) At(Home) Go(HWS) At(HWS) Sells(HWS,Drill) Buy(Drill) At(HWS) Go(SM) At(SM), Sells(SM,Milk) Buy(Milk) At(SM), Sells(SM,Ban.) Buy(Ban.) At(SM) Go(Home) Have(Milk) At(Home) Have(Ban.) Have(Drill) Finish 2/29/ J / 6.834J Goal Regression & Causal Graph Planning 25
26 Partial Order Plan <Actions,Orderings,Links> Start At(Home) Sells(HWS,Drill) Sells(SM,Milk) Sells(SM,Ban.) At(Home) Go(HWS) At(HWS) Sells(HWS,Drill) Buy(Drill) At(HWS) Go(SM) At(SM), Sells(SM,Milk) Buy(Milk) At(SM), Sells(SM,Ban.) Buy(Ban.) At(SM) Go(Home) Have(Milk) At(Home) Have(Ban.) Have(Drill) Finish 2/29/ J / 6.834J Goal Regression & Causal Graph Planning 26
27 Partial Order Plan <Actions,Orderings,Links> Start At(Home) Sells(HWS,Drill) Sells(SM,Milk) Sells(SM,Ban.) At(Home) Go(HWS) At(HWS) Sells(HWS,Drill) Buy(Drill) At(HWS) Go(SM) At(SM), Sells(SM,Milk) Buy(Milk) At(SM), Sells(SM,Ban.) Buy(Ban.) At(SM) Go(Home) Have(Milk) At(Home) Have(Ban.) Have(Drill) Finish 2/29/ J / 6.834J Goal Regression & Causal Graph Planning 27
28 Planning as Goal Regression Partial Planning Overview Partial Order Planning Problem Problem Encoding Partial Order Plans Plan Correctness Partial Order Plan Generation 2/29/ J / 6.834J Goal Regression & Causal Graph Planning 28
29 What Constitutes a Correct Plan? Complete Plan Achieves all Goals Achieves all preconditions No actions intervene to undo a needed precondition Consistent Plan There exists an execution sequence that is consistent with the ordering At(SM), Sells(SM,Milk) Buy(Milk) Start At(Home) Sells(HWS,Drill) Sells(SM,Milk) Sells(SM,Ban.) At(Home) Go(HWS) At(HWS) Sells(HWS,Drill) Buy(Drill) At(HWS) Go(SM) At(SM) Go(Home) Finish At(SM), Sells(SM,Ban.) Buy(Ban.) Have(Milk) At(Home) Have(Ban.) Have(Drill) 2/29/ J / 6.834J Goal Regression & Causal Graph Planning 29
30 Why is an ordering needed? At(HWS) Go(SM) At(SM), not At(HWS) At(SM) Buy(Milk) Have(Milk) At(SM) Go(Home) At(Home), not At(SM) 2/29/ J / 6.834J Goal Regression & Causal Graph Planning 30
31 Why is an ordering needed? Suppose the other order is allowed, what happens? At(HWS) Go(SM) At(SM), not At(HWS) Threatened Precondition At(SM) Buy(Milk) At(SM) Go(Home) At(Home), not At(SM) 2/29/ J / 6.834J Goal Regression & Causal Graph Planning 31
32 Why is an ordering needed? Link indicates protected time interval for the precondition. Suppose the other order is allowed, what happens? At(SM) Buy(Milk) At(HWS) Go(SM) At(SM), not At(HWS) At(SM) Go(Home) At(Home), not At(SM) 2/29/ J / 6.834J Goal Regression & Causal Graph Planning 32
33 Why is an ordering needed? The ordering resolves the threat. At(SM) Buy(Milk) At(HWS) Go(SM) At(SM), not At(HWS) At(SM) Go(Home) At(Home), not At(SM) 2/29/ J / 6.834J Goal Regression & Causal Graph Planning 33
34 Solution: A Complete and Consistent Plan Complete Plan IFF every precondition of every step is achieved. A step s precondition is achieved iff its the effect of some preceding step, no intervening step can undo it. Consistent Plan At(SM), Sells(SM,Milk) Buy(Milk) IFF there is no contradiction in the ordering constraint. i.e., never s i < s j and s j < s i, or the causal links + orderings are loop free. Start At(Home) Sells(HWS,Drill) Sells(SM,Milk) Sells(SM,Ban.) At(Home) Go(HWS) At(HWS) Sells(HWS,Drill) Buy(Drill) At(HWS) Go(SM) At(SM) Go(Home) Finish At(SM), Sells(SM,Ban.) Buy(Ban.) Have(Milk) At(Home) Have(Ban.) Have(Drill) 2/29/ J / 6.834J Goal Regression & Causal Graph Planning 34
35 Planning as Goal Regression Partial Planning Overview Partial Order Planning Problem Partial Order Plan Generation Derivation from Completeness and Consistency Backward Chaining Threat Resolution The POP algorithm 2/29/ J / 6.834J Goal Regression & Causal Graph Planning 35
36 Partial Order Planning Algorithm The algorithm falls out of Consistency and Completeness Completeness: Must achieve all preconditions Backward chain from goals to initial state, by inserting actions and causal links. Must avoid intervening actions that threaten After each action is inserted, find any action that threatens its effects, and impose ordering to resolve. Consistent: Ordering must be consistent After each causal link and ordering is inserted, check for loops. 2/29/ J / 6.834J Goal Regression & Causal Graph Planning 36
37 Planning as Goal Regression Partial Planning Overview Partial Order Planning Problem Partial Order Plan Generation Derivation from Completeness and Consistency Backward Chaining Threat Resolution The POP algorithm 2/29/ J / 6.834J Goal Regression & Causal Graph Planning 37
38 Start At(Home) Sells(HWS,Drill) Sells(SM,Milk) Sells(SM,Ban.) Have(Drill) Have(Milk) Have(Ban.) at(home) Finish 2/29/ J / 6.834J Goal Regression & Causal Graph Planning 38
39 Start At(Home) Sells(HWS,Drill) Sells(SM,Milk) Sells(SM,Ban.) At(HWS) Sells(HWS,Drill) Buy(Drill) At(SM), Sells(SM,Milk) Buy(Milk) At(SM), Sells(SM,Ban.) Buy(Ban.) Have(Drill) Have(Milk) Have(Ban.) at(home) Finish 2/29/ J / 6.834J Goal Regression & Causal Graph Planning 39
40 Start At(Home) Sells(HWS,Drill) Sells(SM,Milk) Sells(SM,Ban.) At(HWS) Sells(HWS,Drill) At(SM), Sells(SM,Milk) At(SM), Sells(SM,Ban.) Buy(Drill) Buy(Milk) Buy(Ban.) Have(Drill) Have(Milk) Have(Ban.) at(home) Finish 2/29/ J / 6.834J Goal Regression & Causal Graph Planning 40
41 Start At(Home) Sells(HWS,Drill) Sells(SM,Milk) Sells(SM,Ban.) At(?x) Go(HWS) At(?x) Go(SM) At(HWS) Sells(HWS,Drill) At(SM), Sells(SM,Milk) At(SM), Sells(SM,Ban.) Buy(Drill) Buy(Milk) Buy(Ban.) Have(Drill) Have(Milk) Have(Ban.) at(home) Finish 2/29/ J / 6.834J Goal Regression & Causal Graph Planning 41
42 Planning as Goal Regression Partial Planning Overview Partial Order Planning Problem Partial Order Plan Generation Derivation from Completeness and Consistency Backward Chaining Threat Resolution The POP algorithm 2/29/ J / 6.834J Goal Regression & Causal Graph Planning 42
43 After adding a causal link/action, find threat with any existing action/link At(HWS) Go(SM) At(SM), not At(HWS) At(SM) Buy(Milk) At(SM) Go(Home) At(Home), not At(SM) 2/29/ J / 6.834J Goal Regression & Causal Graph Planning 43
44 To remove threats At(HWS) Go(SM) At(SM), not At(HWS) At(SM) Buy(Milk) At(SM) Go(Home) At(Home), not At(SM) 2/29/ J / 6.834J Goal Regression & Causal Graph Planning 44
45 To remove threats promote the threat or At(SM) Buy(Milk) At(HWS) Go(SM) At(SM), not At(HWS) At(SM) Go(Home) At(Home), not At(SM) 2/29/ J / 6.834J Goal Regression & Causal Graph Planning 45
46 To remove threats promote the threat or demote the threat At(HWS) Go(SM) At(SM), not At(HWS) At(SM) Go(Home) At(Home), not At(SM) At(SM) Buy(Milk) But only allow demotion/promotion if schedulable consistent = loop free no action precedes initial state 2/29/ J / 6.834J Goal Regression & Causal Graph Planning 46
47 Go(SM) threatens At(HWS) Promote Start At(Home) Sells(HWS,Drill) Sells(SM,Milk) Sells(SM,Ban.) At(Home) Go(HWS) At(Home) Go(SM) At(HWS) Sells(HWS,Drill) At(SM), Sells(SM,Milk) At(SM), Sells(SM,Ban.) Buy(Drill) Buy(Milk) Buy(Ban.) Have(Drill) Have(Milk) Have(Ban.) at(home) Finish 2/29/ J / 6.834J Goal Regression & Causal Graph Planning 47
48 Go(HWS) threatens At(Home) Try Promote Start Inconsistent! At(Home) Sells(HWS,Drill) Sells(SM,Milk) Sells(SM,Ban.) At(Home) Go(HWS) At(Home) Go(SM) At(HWS) Sells(HWS,Drill) At(SM), Sells(SM,Milk) At(SM), Sells(SM,Ban.) Buy(Drill) Buy(Milk) Buy(Ban.) Have(Drill) Have(Milk) Have(Ban.) at(home) Finish 2/29/ J / 6.834J Goal Regression & Causal Graph Planning 48
49 Use Go(HWS) to produce At(?x) Finish by Go(Home). Start At(Home) Sells(HWS,Drill) Sells(SM,Milk) Sells(SM,Ban.) At(Home) Go(HWS) At(?x) Go(SM) HWS At(HWS) Sells(HWS,Drill) At(SM), Sells(SM,Milk) At(SM), Sells(SM,Ban.) Buy(Drill) Buy(Milk) Buy(Ban.) Have(Drill) Have(Milk) Have(Ban.) at(home) Finish At(SM) Go(Home) 2/29/ J / 6.834J Goal Regression & Causal Graph Planning 49
50 POP(<A,O,L>, agenda, actions) 1. Termination: If agenda is empty, return plan <A,O,L>. 2. Goal Selection: Select and remove open condition <p, a need > from agenda. 3. Action Selection: Choose new or existing action a add that can precede a need and whose effects include p. Link and order actions. 4. Update Agenda: If a add is new, add its preconditions to agenda. 5. Threat Detection: For every action a threat that might threaten some causal link from a produce to a consume, choose a consistent ordering: a) Demote: Add a threat < a produce b) Promote: Add a consume < a threat 6. Recurse: on modified plan and agenda Choose is nondeterministic 2/29/ J / 6.834J Goal Regression & Causal Graph Planning Select is deterministic 50
51 Lets Start with Goal-Regression Search action3 S t 0 action1 Initial State (s o ) action2 action2 S t 1 action4 goal Time action3 S t 2 action2 Why can Goal Regression be slow? Multiple actions can achieve goals. Many possible (sub-)goal orderings. Dead-ends can be discovered late. We try a real-world example next! 51
52 What assumptions are implied by the STRIPS representation? TakeImage (?target,?instr): Pre: Status(?instr, Calibrated), Pointing(?target) Eff: Image(?target) Calibrate (?instrument): Pre: Status(?instr, On), Calibration-Target(?target), Pointing(?target) Eff: Status(?inst, On), Status(?instr, Calibrated) Turn (?target): Pre: Pointing(?direction),?direction?target Eff: Pointing(?direction), Pointing(?target) STRIPS Assumptions: Atomic time, Agent is omniscient (no sensing necessary), Agent is sole cause of change, Actions have deterministic effects, and No indirect effects. One action at a time. No metric time. No goals over time. 2/29/ J / 6.834J Goal Regression & Causal Graph Planning 52
53 Outline Review: programs on state Planning as goal regression (SNLP/UCPOP) Goal regression planning with causal graphs (Burton) Appendix: HFS planning with the causal graph heuristic (Fast Downward) 53
54 DS1 Revisited: Planning to Control Interacting Devices over Time Thrust Goals Delta_V(direction=b, magnitude=200) Power Time-line Planning (tburton) Attitude Point(a) Turn(a,b) Point(b) Turn(b,a) Engine Off Warm Up Thrust (b, 200) O ff This image is in the public domain. Reactive Device Reconfiguration and Repair (Burton) 54
55 Domain Independent Planning in the Real-World classical off-line Speed Heuristic-based Searches & Symbolic Search (not presented) real-time Burton (real-time planning) Expressivity tburton (planning with time and indirect effects) Both Burtons use Causal Graphs to analyze weak and tight couplings and to Decompose. 55
56 Planning with Indirect Effects: DS 1 Attitude Control System Example Flight Computer Data Commands PDE Propulsion Drive Electronics SRU Stellar Reference Unit PDU Power Distribution Unit z facing thrusters He N 2 H 4 x facing thrusters Bus Controller GDE Gimbal Drive Electronics PASM Power Activation & Switching Module DSEU Diagnostics Sensors Electronics Unit Dynamics 1553 bus ACE Attitude Control Electronics control mode xatt error yatt error zatt error 56
57 Why is Controlling an Engineered Device Easier than a Puzzle? Propulsion Drive Electronics Stellar Reference Unit Power Distribution Unit Gimbal Drive Electronics Power Activation & Switching Module Diagnostics Sensors Electronics Unit Attitude Control Electronics Flight Computer PDE He Data Commands SRU PDU z facing thrusters N 2 H 4 x facing thrusters Bus Controller GDE 3 5 PASM DSEU Dynamics 1553 bus ACE control mode xatt erro r yatt erro r zatt erro r 1. Actions are reversible. 2. Devices hold state. 3. Causes and effects form a tree. 1. Actions are reversible. 2. Devices hold state. 3. Causes and effects form tight cycles. 57
58 Outline Review: programs on state Planning as goal regression (SNLP/UCPOP) Goal regression planning with causal graphs (Burton) Constraint automata planning problem Causal graphs Using causal graphs to order goals Computing policies for selecting actions Appendix: Planning for cyclic causal graphs Appendix: HFS planning with the causal graph heuristic (Fast Downward) 58
59 Concurrent Constraint Automata Planning Problem flow in dcmd out = vcmd in dcmd in Variables and Domains: flow out State: driver in {on, off, resettable, failed}, valve in {open, closed, stuck-open, stuck-closed}. Control: dcmd in in {idle, on, off, reset, open, close} Dependent: flow in, flow out in {pos, neg, zero} dcmd out, vcmd in in Domain{dcmd in } Initial assignment: {driver = on, valve = closed} Goal assignment: {driver = off, valve = open} 59
60 Concurrent Constraint Automata Planning Problem (cont.) flow in dcmd out = vcmd in dcmd in Constraint automata (one per state variable): flow out driver dcmd in = dcmd out valve flowin in = flow out On dcmd in = reset Resettable Open Stuck Open dcmd in = on dcmd in = off dcmd in = off vcmd in = open vcmd in = close Off Failed Closed Stuck Closed Assume: transitions independently controlled, each location can idle. State constraints: {dcmd out = vcmd in } 60
61 Outline Review: programs on state Planning as goal regression (SNLP/UCPOP) Goal regression planning with causal graphs (Burton) Constraint automata planning problem Causal graphs Using causal graphs to order goals Computing policies for selecting actions Appendix: Planning for cyclic causal graphs Appendix: HFS planning with the causal graph heuristic (Fast Downward) 61
62 Observation: Engineered systems are largely loop free. Computer Bus Control Remote Terminal Remote Terminal Driver Driver Valve Valve 62
63 Causal Graph G of concurrent automata S: Vertices are control and state variables of automata. Edge from v i to v j if v j s transition is conditioned on v i. Driver On dcmd in = dcmd out dcmd in = reset Resettable dcmd in Driver Valve dcmd in = on dcmd in = off Off dcmd in = off Failed Valve Open flowin in = flow out Stuck Open dcmd in vcmd in = open vcmd in = close dcmd out = vcmd in Closed Stuck Closed 63
64 Outline Review: programs on state Planning as goal regression (SNLP/UCPOP) Goal regression planning with causal graphs (Burton) Constraint automata planning problem Causal graphs Using causal graphs to order goals Computing policies for selecting actions Appendix: Planning for cyclic causal graphs Appendix: HFS planning with the causal graph heuristic (Fast Downward) 64
65 Idea: use causal graph analysis to eliminate ALL forms of search (Burton) POP(<A,O,L>, agenda, actions): 1. Termination: If agenda is empty, return plan <A,O,L>. 2. Goal Selection: Select and remove open condition <p, a need > from agenda. 3. Action Selection: Choose new or existing action a add that can precede a need and whose effects include p. Link and order actions. 4. Update Agenda: If a add is new, add its preconditions to agenda. 5. Threat Detection: For every action a threat that might threaten some causal link from a produce to a consume, choose a consistent ordering: a) Demote: Add a threat < a produce b) Promote: Add a consume < a threat 6. Recurse: on modified plan and agenda Burton [Williams & Nayak, IJCAI 1997] 2/29/ J / 6.834J Goal Regression & Causal Graph Planning 65
66 Why do goal orderings matter? 1. An achieved goal can be clobbered by a subsequent goal. command Example: Current State: Goal State: driver = on, valve = closed. driver = off, valve = open. Achieving (driver = off), followed by (valve = open) clobbers (driver = off). Achieve valve goal before driver goal. Effect Cause 66
67 Goal Ordering for Causal Graph Planning Require: The CCA causal graph to be acyclic. Causal Graph Driver dcmd in Valve Idea: Achieve conjunctive goals upstream within the causal graph, from effects to causes (i.e., children to parents). 67
68 Property: Safe to achieve goals in an upstream order The only variables used to set some variable v 7 is its ancestors. Variable v 7 can be changed without affecting its descendants. Affected Unaffected Exploits: Each transitions independently controlled, each location can idle. Simple check: 1. Number the causal graph depth-first from leaves. Child has lower number than parents 2. Achieve goals in the order of increasing depth-first number. 68
69 Why do goal orderings matter? 2. Two goals can compete for the same variable associated with their sub-goals. data Switc h Latch 1 Latch 2 Example: Latch Data at Latches 1 and 2 If Latch1 and Latch2 goals achieved at same time, Latch1 and Latch2 compete for Switch position. 1 2 Solve one goal completely before the other (serially). 69
70 Property: Safe to achieve one goal before starting next sibling (serialization). Sibling goals v 7 and v 4 may both need shared ancestors. Not Shared Shared Unaffected Competition gone once sibling goal v 7 is satisfied. Not Shared Unaffected 70
71 Goal regression, CU causal graph planning example Driver 1 Driver 2 Valve 1 Valve 2 Burton s goal ordering rules: 1. Achieve effect goals before their causes, with control actiions last. 2. Achieve goals serially (like Hacker[Sussman PhD]). Valve 1 = open Valve 2 = open Driver 1 = off Driver 2 = off Open Open Turn-Off Turn-Off Driver 1 = on Driver 2 = on CU = on CU = on Turn-On CU = on CU = on CU = on CU = on Turn-On No Search! First action generated first! Turn-On Worst Case per action: Depth * Sub-goal branch factor. CU = off Average Cost per action: Sub-goal branch factor. 71
72 To select actions reactively, convert constraint automata to lookup policies Driver 2 Valve 1 dcmd out = vcmd in dcmd in = dcmd out inflow = outflow on dcmd in = reset resettable open stuck open cmd in = on dcmd in = off dcmd in = off vcmd in = open vcmd in = close off failed closed stuck closed 72
73 To select actions reactively, convert constraint automata to lookup policies Algorithm: Instance of APSP Driver 2 Valve 1 dcmd out = vcmd in dcmd in = dcmd out dcmd on in = reset resettable Goal Current On Off On idle cmd = off cmd in = on dcmd in = off dcmd in = off Off cmd = on idle off failed Resettable cmd = reset cmd = off 73
74 To select actions reactively, convert constraint automata to lookup policies Driver 2 Valve 1 dcmd out = vcmd in Goal inflow = outflow Current Open Open idle Closed driver = on cmd = close open stuck open Closed driver = on cmd = open idle vcmd in = open vcmd in = close Stuck fail fail closed stuck closed 74
75 Plan by passing sub-goals up causal graph Goal: Driver = off, Valve = closed Current: Driver = off, Valve = open Driver 2 Algorithm: see [williams and nayak, IJCAI97] Valve 1 Goal Current On Off Goal Current Open Closed On idle cmd = off Open idle driver = on cmd = close Off cmd = on idle Closed driver = on cmd = open idle Resettable cmd = reset cmd = off Stuck fail fail 75
76 Plan by passing sub-goals up causal graph Goal: Driver = off, Valve = closed Current: Driver = off, Valve = open Driver 2 Algorithm: see [williams and nayak, IJCAI97] Valve 1 Goal Current On Off Goal Current Open Closed On idle cmd = off Open idle driver = on cmd = close Off cmd = on idle Closed driver = on cmd = open idle Resettable cmd = reset cmd = off Stuck fail fail 76
77 Plan by passing sub-goals up causal graph Goal: Driver = off, Valve = closed Current: Driver = off, Valve = open Send: cmd = on Driver 2 Algorithm: see [williams and nayak, IJCAI97] Valve 1 Goal Current On Off Goal Current Open Closed On idle cmd = off Open idle driver = on cmd = close Off cmd = on idle Closed driver = on cmd = open idle Resettable cmd = reset cmd = off Stuck fail fail 77
78 Plan by passing sub-goals up causal graph Goal: Driver = off, Valve = closed Current: Driver = resettable, Valve = open Driver Failed Resettable 2 Algorithm: see [williams and nayak, IJCAI97] Valve 1 Goal Current On Off Goal Current Open Closed On idle cmd = off Open idle driver = on cmd = close Off cmd = on idle Closed driver = on cmd = open idle Resettable cmd = reset cmd = off Stuck fail fail 78
79 Plan by passing sub-goals up causal graph Goal: Driver = off, Valve = closed Current: Driver = resettable, Valve = open Driver 2 Algorithm: see [williams and nayak, IJCAI97] Valve 1 Goal Current On Off Goal Current Open Closed On idle cmd = off Open idle driver = on cmd = close Off cmd = on idle Closed driver = on cmd = open idle Resettable cmd = reset cmd = off Stuck fail fail 79
80 Plan by passing sub-goals up causal graph Goal: Driver = off, Valve = closed Current: Driver = resettable, Valve = open Send cmd = reset Driver 2 Algorithm: see [williams and nayak, IJCAI97] Valve 1 Goal Current On Off Goal Current Open Closed On idle cmd = off Open idle driver = on cmd = close Off cmd = on idle Closed driver = on cmd = open idle Resettable cmd = reset cmd = off Stuck fail fail 80
81 What if the causal graph G contains cycles? Computer Bus Control K-band Transmitter Amplifier Antenna K-band Transmitter Problem: Plan is no longer serializable. Solution: Amplifier Antenna Isolate the cyclic components (compute Strongly Connected Components). compose each cycle into a single component. New causal graph G is acyclic. Goals of G are serializable. 81
82 Action Policy for Composed Components Transmitter Amplifier on on on T on A cmd A = off cmd A = on on T off A A = off cmd T = on A = off cmd T = off T = on cmd A = on cmd A = off cmd T = on cmd T = off off off off T on A cmd A = off off T off A 82
83 Action Policy for Composed Components Problem: Composition grows exponential in space usage. Solution: Use BDD encoding. [Chung and Williams, Self-adaptive SW 03] on T on A cmd A = off cmd A = on on T off A cmd T = on cmd T = off Current Goal On T, On A On T, Off A Off T, Off A Off T, On A off T on A cmd A = off off T off A On T, On A idle cmd A = off cmd A = off fail On T, Off A cmd A = on idle cmd T = off fail Off T, Off A cmd T = on cmd T = on idle fail Off T, On A fail fail cmd A = off idle 83
84 Outline Review: programs on state Planning as goal regression (SNLP/UCPOP) Goal regression planning with causal graphs (Burton) Appendix: HFS planning with the causal graph heuristic (Fast Downward) 84
85 Causal Graph Heuristic for PDDL Recall: The Fast Forward (FF) Heuristic is computed over a Relaxed Planning Graph. Likewise: The Causal Graph (CG) Heuristic is computed over a Causal Graph. Map PDDL to concurrent automata, and extract causal graph (called domain transition graph (DTG). 2/29/2016 Goal Regression and Causal Graph Planning 85
86 Problem Reformulation Original Representation: STRIPS or PDDL Init Goal TruckIn(t, c 2 ) BoxIn(b,c 1 ) BoxIn(b, Paris) Operators, e.g. Drive(t, c 1, c 2 ) Pre: TruckIn(c 1 ) Add: TruckIn(c 2 ) Del: TruckIn(c 1 ) New Representation: Multi-valued Planning Task Variables: truck := {c 1, c 2, Paris} box := {ontruck, c 1, c 2, Paris} Init truck = c 2 box = c 1 Goal box = Paris Operators, e.g. Drive(t, c 1, c 2 ) Pre: truck = c 1 Post: truck = c 2 86
87 Domain Transition Graphs (DTG) One DTG per variable. Edges represent possible transitions (actions) and are guarded by preconditions truck c 1 Causal Edge Drive Drive Drive Drive Drive c 2 Drive Paris A causal edge between the DTG represents that the box DTG has preconditions that depend on the truck DTG. box Load truck=c 1 Unload truck=paris c 1 Unload truck=c 1 Load truck=c 2 Paris t c 2 Load truck=paris Unload truck=c 2 87
88 Calculating the Heuristic Initial: BoxIn(b, c 1 ) TruckIn(t, c 2 ) Goal: BoxIn(b, Paris) truck c 1 Drive Drive Drive Drive Paris Drive Drive c 2 box c 1 Load truck=c 1 Unload truck=paris Unload truck=c 1 Load truck=c 2 Paris t c 2 Load truck=paris Unload truck=c 2 88
89 Calculating the Heuristic Initial: BoxIn(b, c 1 ) TruckIn(t, c 2 ) Goal: BoxIn(b, Paris) truck c 1 Drive Drive Drive Drive Paris # of transitions to get the box to Paris: 2 Drive c 2 Drive box c 1 Load truck=c 1 Unload truck=paris Unload truck=c 1 Load truck=c 2 Paris t c 2 Load truck=paris Unload truck=c 2 89
90 Calculating the Heuristic Initial: BoxIn(b, c 1 ) TruckIn(t, c 2 ) Goal: BoxIn(b, Paris) truck c 1 Drive Drive Drive Drive Paris # of transitions to get the box to Paris: 2 Drive c 2 Drive # of transitions to get the truck to c 1 : 1 box c 1 Load truck=c 1 Unload truck=paris Unload truck=c 1 Load truck=c 2 Paris t c 2 Load truck=paris Unload truck=c 2 90
91 Calculating the Heuristic Initial: BoxIn(b, c 1 ) TruckIn(t, c 2 ) Goal: BoxIn(b, Paris) truck c 1 Drive Drive Drive Drive Paris # of transitions to get the box to Paris: 2 Drive c 2 Drive # of transitions to get the truck to c 1 : 1 # of transitions to get the truck to Paris: 1 box c 1 Load truck=c 1 Unload truck=paris Unload truck=c 1 Load truck=c 2 Paris t c 2 Load truck=paris Unload truck=c 2 91
92 Calculating the Heuristic Initial: BoxIn(b, c 1 ) TruckIn(t, c 2 ) Goal: BoxIn(b, Paris) truck c 1 Drive Drive Drive Drive Paris # of transitions to get the box to Paris: 2 Drive c 2 Drive # of transitions to get the truck to c 1 : 1 # of transitions to get the truck to Paris: 1 box c 1 Sum the transitions to get the heuristic value for the initial state h CG = 4 Load truck=c 1 Unload truck=paris Unload truck=c 1 Load truck=c 2 Paris t c 2 Load truck=paris Unload truck=c 2 92
93 Causal Graph Heuristic h CG ( s) v dom( s cost * ) v ( s( v), s * ( v)) The cost of getting from the current state, s, to the goal state, s* Idea: Sum the domain transition costs to get to the goal. 1. Identify each variable involved in the goal. 2. Recurse from child to parent in the causal graph. For each variable, sum the cost of along the shortest path to change the current value to the goal value. If changing that variable s value has preconditions, also add the cost of changing its parent variable. 93
94 Causal Graph Heuristic Notes Can not handle cyclic causal graphs Relax some links until the graph is acyclic Calculation performed differently in practice Modified Dijkstra algorithm Each cost calculation can over estimate. Not admissible Assumes no helpful interactions between subgoals 94
95 Breaking Cycles Break action with multiple effects into multiple unary effect actions If there are still cycles, ignore some preconditions of actions Pickup(d1, cream, w1,5) Pre: d1_location=w1, w1_cream=5 Eff: w1_cream=4, d1_payload=cream Pickup1(d1, cream, w1,5) Pre: d1_location=w1, w1_cream=5 Eff: d1_payload=cream Pickup2(d1, cream, w1,5) Pre: d1_location=w1, w1_cream=5 Eff: w1_cream=4 95
96 Breaking Cycles d1_loca tion d1_loca tion Actions split into unary effects d1_payl oad w1_inven tory d1_payl oad w1_inven tory c1_wa nt c1_wa nt 96
97 Causal Graph Heuristic Summary Graph capture dependence between variables Requires re-achieving conditions Multi-valued formulation reveals structure h CEA heuristic extends to cyclic graphs Performs extremely well on classical problems 97
98 Heuristic Families There are many other heuristics, we ve only covered two Type Relaxation Heuristics Estimates Relaxed planning graph Ignore deletes h max, h add, h ff, h cs, h m, ADHG h+ (cost of relaxed plan) Causal Graph Limited interactions h CG, h cea h+ (cost of relaxed plan) Projection Abstractions Projection h PDB, h STAN, h FORK h (cost of optimal) 98
99 Key Ideas Heuristic forward search is one of the fastest current planning techniques Domain independent heuristic design still problematic Fast Forward Heuristic Solve a relaxed problem in a Planning Graph Causal Graph Heuristic Consider problem s structure in a Causal Graph 99
100 MIT OpenCourseWare J / 6.834J Cognitive Robotics Spring 2016 For information about citing these materials or our Terms of Use, visit:
CS 730/730W/830: Intro AI
CS 730/730W/830: Intro AI 1 handout: slides Wheeler Ruml (UNH) Lecture 18, CS 730 1 / 15 Comparison Extensions Setting Break Wheeler Ruml (UNH) Lecture 18, CS 730 2 / 15 Comparison Comparison Extensions
More informationPlan Recognition through Goal Graph Analysis
Plan Recognition through Goal Graph Analysis Jun Hong 1 Abstract. We present a novel approach to plan recognition based on a two-stage paradigm of graph construction and analysis. First, a graph structure
More informationHoare Logic and Model Checking. LTL and CTL: a perspective. Learning outcomes. Model Checking Lecture 12: Loose ends
Learning outcomes Hoare Logic and Model Checking Model Checking Lecture 12: Loose ends Dominic Mulligan Based on previous slides by Alan Mycroft and Mike Gordon Programming, Logic, and Semantics Group
More informationArtificial Intelligence Lecture 7
Artificial Intelligence Lecture 7 Lecture plan AI in general (ch. 1) Search based AI (ch. 4) search, games, planning, optimization Agents (ch. 8) applied AI techniques in robots, software agents,... Knowledge
More informationComputational Tree Logic and Model Checking A simple introduction. F. Raimondi
Computational Tree Logic and Model Checking A simple introduction F. Raimondi I am Franco Raimondi, f.raimondi@cs.ucl.ac.uk Slides, coursework, coursework solutions can be found online: http://www.cs.ucl.ac.uk/staff/f.raimondi/
More informationPlan Recognition through Goal Graph Analysis
Plan Recognition through Goal Graph Analysis Jun Hong 1 Abstract. We present a novel approach to plan recognition based on a two-stage paradigm of graph construction and analysis. First, a graph structure
More informationChapter 2. Knowledge Representation: Reasoning, Issues, and Acquisition. Teaching Notes
Chapter 2 Knowledge Representation: Reasoning, Issues, and Acquisition Teaching Notes This chapter explains how knowledge is represented in artificial intelligence. The topic may be launched by introducing
More informationCS148 - Building Intelligent Robots Lecture 5: Autonomus Control Architectures. Instructor: Chad Jenkins (cjenkins)
Lecture 5 Control Architectures Slide 1 CS148 - Building Intelligent Robots Lecture 5: Autonomus Control Architectures Instructor: Chad Jenkins (cjenkins) Lecture 5 Control Architectures Slide 2 Administrivia
More informationCourse Syllabus. Operating Systems, Spring 2016, Meni Adler, Danny Hendler and Amnon Meisels 1 3/14/2016
Course Syllabus. Introduction - History; Views; Concepts; Structure 2. Process Management - Processes; State + Resources; Threads; Unix implementation of Processes 3. Scheduling Paradigms; Unix; Modeling
More informationPART - A 1. Define Artificial Intelligence formulated by Haugeland. The exciting new effort to make computers think machines with minds in the full and literal sense. 2. Define Artificial Intelligence
More informationAgents and Environments
Agents and Environments Berlin Chen 2004 Reference: 1. S. Russell and P. Norvig. Artificial Intelligence: A Modern Approach. Chapter 2 AI 2004 Berlin Chen 1 What is an Agent An agent interacts with its
More informationEvolutionary Programming
Evolutionary Programming Searching Problem Spaces William Power April 24, 2016 1 Evolutionary Programming Can we solve problems by mi:micing the evolutionary process? Evolutionary programming is a methodology
More informationKnowledge Based Systems
Knowledge Based Systems Human Expert A human expert is a specialist for a specific differentiated application field who creates solutions to customer problems in this respective field and supports them
More informationCS 4365: Artificial Intelligence Recap. Vibhav Gogate
CS 4365: Artificial Intelligence Recap Vibhav Gogate Exam Topics Search BFS, DFS, UCS, A* (tree and graph) Completeness and Optimality Heuristics: admissibility and consistency CSPs Constraint graphs,
More informationProgramming with Goals (3)
Programming with Goals (3) M. Birna van Riemsdijk, TU Delft, The Netherlands GOAL slides adapted from MAS course slides by Hindriks 4/24/11 Delft University of Technology Challenge the future Outline GOAL:
More informationBayesian Networks Representation. Machine Learning 10701/15781 Carlos Guestrin Carnegie Mellon University
Bayesian Networks Representation Machine Learning 10701/15781 Carlos Guestrin Carnegie Mellon University March 16 th, 2005 Handwriting recognition Character recognition, e.g., kernel SVMs r r r a c r rr
More informationHuman Activities: Handling Uncertainties Using Fuzzy Time Intervals
The 19th International Conference on Pattern Recognition (ICPR), Tampa, FL, 2009 Human Activities: Handling Uncertainties Using Fuzzy Time Intervals M. S. Ryoo 1,2 and J. K. Aggarwal 1 1 Computer & Vision
More informationAn Evolutionary Approach to Tower of Hanoi Problem
An Evolutionary Approach to Tower of Hanoi Problem Jie Li, Rui Shen Department of Computer Engineering Shandong Aerospace Electro-Technology Institute Yantai, China jie_yi_ehw@163.com Abstract:The Tower
More informationDynamic Rule-based Agent
International Journal of Engineering Research and Technology. ISSN 0974-3154 Volume 11, Number 4 (2018), pp. 605-613 International Research Publication House http://www.irphouse.com Dynamic Rule-based
More informationArtificial Intelligence Programming Probability
Artificial Intelligence Programming Probability Chris Brooks Department of Computer Science University of San Francisco Department of Computer Science University of San Francisco p.1/25 17-0: Uncertainty
More informationIntelligent Agents. CmpE 540 Principles of Artificial Intelligence
CmpE 540 Principles of Artificial Intelligence Intelligent Agents Pınar Yolum pinar.yolum@boun.edu.tr Department of Computer Engineering Boğaziçi University 1 Chapter 2 (Based mostly on the course slides
More informationUsing Bayesian Networks to Analyze Expression Data. Xu Siwei, s Muhammad Ali Faisal, s Tejal Joshi, s
Using Bayesian Networks to Analyze Expression Data Xu Siwei, s0789023 Muhammad Ali Faisal, s0677834 Tejal Joshi, s0677858 Outline Introduction Bayesian Networks Equivalence Classes Applying to Expression
More informationINFERENCING STRATEGIES
INFERENCING STRATEGIES Table of Content Introduction Categories of Reasoning Inference Techniques Control Strategies Comparative Summary of Backward and Forward Chaining Conflict Resolution Goal Agenda
More informationSTIN2103. Knowledge. engineering expert systems. Wan Hussain Wan Ishak. SOC 2079 Ext.: Url:
& Knowledge STIN2103 engineering expert systems Wan Hussain Wan Ishak SOC 2079 Ext.: 4786 Email: hussain@uum.edu.my Url: http://www.wanhussain.com Outline Knowledge Representation Types of knowledge Knowledge
More informationIntroduction to Artificial Intelligence 2 nd semester 2016/2017. Chapter 2: Intelligent Agents
Introduction to Artificial Intelligence 2 nd semester 2016/2017 Chapter 2: Intelligent Agents Mohamed B. Abubaker Palestine Technical College Deir El-Balah 1 Agents and Environments An agent is anything
More informationERA: Architectures for Inference
ERA: Architectures for Inference Dan Hammerstrom Electrical And Computer Engineering 7/28/09 1 Intelligent Computing In spite of the transistor bounty of Moore s law, there is a large class of problems
More informationImproving the Accuracy of Neuro-Symbolic Rules with Case-Based Reasoning
Improving the Accuracy of Neuro-Symbolic Rules with Case-Based Reasoning Jim Prentzas 1, Ioannis Hatzilygeroudis 2 and Othon Michail 2 Abstract. In this paper, we present an improved approach integrating
More informationCS 771 Artificial Intelligence. Intelligent Agents
CS 771 Artificial Intelligence Intelligent Agents What is AI? Views of AI fall into four categories 1. Thinking humanly 2. Acting humanly 3. Thinking rationally 4. Acting rationally Acting/Thinking Humanly/Rationally
More informationRobust Execution of Contingent, Temporally Flexible Plans
Robust xecution of Contingent, Temporally Flexible Plans tephen. Block, ndreas F. Wehowsky and Brian C. Williams Computer cience and rtificial Intelligence Laboratory, Massachusetts Institute of Technology,
More informationLiving with Newton's Laws
Task #1 - Newton s 1 st Law - This is a pain in the neck Let's suppose you are in your car, waiting at a stop light. Like any good driver, you have your seat belt buckled. (It's the law.) Suddenly, a car
More informationSeminar Thesis: Efficient Planning under Uncertainty with Macro-actions
Seminar Thesis: Efficient Planning under Uncertainty with Macro-actions Ragnar Mogk Department of Computer Science Technische Universität Darmstadt ragnar.mogk@stud.tu-darmstadt.de 1 Introduction This
More informationInferencing in Artificial Intelligence and Computational Linguistics
Inferencing in Artificial Intelligence and Computational Linguistics (http://www.dfki.de/~horacek/infer-ai-cl.html) no classes on 28.5., 18.6., 25.6. 2-3 extra lectures will be scheduled Helmut Horacek
More informationSolving problems by searching
Solving problems by searching Chapter 3 14 Jan 2004 CS 3243 - Blind Search 1 Outline Problem-solving agents Problem types Problem formulation Example problems Basic search algorithms 14 Jan 2004 CS 3243
More informationPrincess Nora University Faculty of Computer & Information Systems ARTIFICIAL INTELLIGENCE (CS 370D) Computer Science Department
Princess Nora University Faculty of Computer & Information Systems 1 ARTIFICIAL INTELLIGENCE (CS 370D) Computer Science Department (CHAPTER-3) INTELLIGENT AGENTS (Course coordinator) CHAPTER OUTLINE What
More informationPolicy Gradients. CS : Deep Reinforcement Learning Sergey Levine
Policy Gradients CS 294-112: Deep Reinforcement Learning Sergey Levine Class Notes 1. Homework 1 milestone due today (11:59 pm)! Don t be late! 2. Remember to start forming final project groups Today s
More informationIntro to Theory of Computation
Intro to Theory of Computation LECTURE 13 Last time: Turing Machines and Variants Today Turing Machine Variants Church-Turing Thesis Sofya Raskhodnikova 2/23/2016 Sofya Raskhodnikova; based on slides by
More informationChapter 2: Intelligent Agents
Chapter 2: Intelligent Agents Outline Last class, introduced AI and rational agent Today s class, focus on intelligent agents Agent and environments Nature of environments influences agent design Basic
More informationLecture 2: Foundations of Concept Learning
Lecture 2: Foundations of Concept Learning Cognitive Systems - Machine Learning Part I: Basic Approaches to Concept Learning Version Space, Candidate Elimination, Inductive Bias last change October 18,
More informationIntro, Graph and Search
GAI Questions Intro 1. How would you describe AI (generally), to not us? 2. Game AI is really about The I of I. Which is what? Supporting the P E which is all about making the game more enjoyable Doing
More informationIntelligent Agents. Philipp Koehn. 16 February 2017
Intelligent Agents Philipp Koehn 16 February 2017 Agents and Environments 1 Agents include humans, robots, softbots, thermostats, etc. The agent function maps from percept histories to actions: f : P A
More informationProgress in Risk Science and Causality
Progress in Risk Science and Causality Tony Cox, tcoxdenver@aol.com AAPCA March 27, 2017 1 Vision for causal analytics Represent understanding of how the world works by an explicit causal model. Learn,
More informationHigh-level Vision. Bernd Neumann Slides for the course in WS 2004/05. Faculty of Informatics Hamburg University Germany
High-level Vision Bernd Neumann Slides for the course in WS 2004/05 Faculty of Informatics Hamburg University Germany neumann@informatik.uni-hamburg.de http://kogs-www.informatik.uni-hamburg.de 1 Contents
More informationArtificial Intelligence
Artificial Intelligence Intelligent Agents Chapter 2 & 27 What is an Agent? An intelligent agent perceives its environment with sensors and acts upon that environment through actuators 2 Examples of Agents
More informationHow do you design an intelligent agent?
Intelligent Agents How do you design an intelligent agent? Definition: An intelligent agent perceives its environment via sensors and acts rationally upon that environment with its effectors. A discrete
More informationBayesian (Belief) Network Models,
Bayesian (Belief) Network Models, 2/10/03 & 2/12/03 Outline of This Lecture 1. Overview of the model 2. Bayes Probability and Rules of Inference Conditional Probabilities Priors and posteriors Joint distributions
More informationIncorporating Action into Diagnostic Problem Solving (An Abridged Report)
Incorporating Action into Diagnostic Problem Solving (An Abridged Report) Sheila A. McIlraith Department of Computer Science, University of Toronto Toronto, ON M5S 1A4 Canada e-mail: mcilrait@cs.toronto.edu
More informationChapter 3 Software Packages to Install How to Set Up Python Eclipse How to Set Up Eclipse... 42
Table of Contents Preface..... 21 About the Authors... 23 Acknowledgments... 24 How This Book is Organized... 24 Who Should Buy This Book?... 24 Where to Find Answers to Review Questions and Exercises...
More informationLecture 13: Finding optimal treatment policies
MACHINE LEARNING FOR HEALTHCARE 6.S897, HST.S53 Lecture 13: Finding optimal treatment policies Prof. David Sontag MIT EECS, CSAIL, IMES (Thanks to Peter Bodik for slides on reinforcement learning) Outline
More informationThe Semantics of Intention Maintenance for Rational Agents
The Semantics of Intention Maintenance for Rational Agents Michael P. Georgeffand Anand S. Rao Australian Artificial Intelligence Institute Level 6, 171 La Trobe Street, Melbourne Victoria 3000, Australia
More informationResource Access Protocols. LS 12, TU Dortmund
Resource Access Protocols Prof. Dr. Jian-Jia Chen LS 12, TU Dortmund 30, Oct. 2018 Prof. Dr. Jian-Jia Chen (LS 12, TU Dortmund) 1 / 27 Why do We Have to Worry about Resource Sharing? Shared Resources:
More informationYou can use this app to build a causal Bayesian network and experiment with inferences. We hope you ll find it interesting and helpful.
icausalbayes USER MANUAL INTRODUCTION You can use this app to build a causal Bayesian network and experiment with inferences. We hope you ll find it interesting and helpful. We expect most of our users
More informationCS324-Artificial Intelligence
CS324-Artificial Intelligence Lecture 3: Intelligent Agents Waheed Noor Computer Science and Information Technology, University of Balochistan, Quetta, Pakistan Waheed Noor (CS&IT, UoB, Quetta) CS324-Artificial
More informationCPSC 121 Some Sample Questions for the Final Exam
CPSC 121 Some Sample Questions for the Final Exam [0] 1. Tautologies and Contradictions: Determine whether the following statements are tautologies (definitely true), contradictions (definitely false),
More informationCHAPTER 4 CONTENT LECTURE 1 November :28 AM
CHAPTER 4 By Radu Muresan University of Guelph Page 1 CHAPTER 4 CONTENT LECTURE 1 November 07 12 10:28 AM UNIPROCESSOR SCHEDULING Real Time Task Model Concepts Types of Real Time Tasks and Their Characteristics
More informationUnmanned autonomous vehicles in air land and sea
based on Gianni A. Di Caro lecture on ROBOT CONTROL RCHITECTURES SINGLE AND MULTI-ROBOT SYSTEMS: A CASE STUDY IN SWARM ROBOTICS Unmanned autonomous vehicles in air land and sea Robots and Unmanned Vehicles
More informationFoundations of AI. 10. Knowledge Representation: Modeling with Logic. Concepts, Actions, Time, & All the Rest
Foundations of AI 10. Knowledge Representation: Modeling with Logic Concepts, Actions, Time, & All the Rest Wolfram Burgard, Andreas Karwath, Bernhard Nebel, and Martin Riedmiller 10/1 Contents Knowledge
More informationLearning to Use Episodic Memory
Learning to Use Episodic Memory Nicholas A. Gorski (ngorski@umich.edu) John E. Laird (laird@umich.edu) Computer Science & Engineering, University of Michigan 2260 Hayward St., Ann Arbor, MI 48109 USA Abstract
More informationStatistical analysis DIANA SAPLACAN 2017 * SLIDES ADAPTED BASED ON LECTURE NOTES BY ALMA LEORA CULEN
Statistical analysis DIANA SAPLACAN 2017 * SLIDES ADAPTED BASED ON LECTURE NOTES BY ALMA LEORA CULEN Vs. 2 Background 3 There are different types of research methods to study behaviour: Descriptive: observations,
More informationRational Agents (Ch. 2)
Rational Agents (Ch. 2) Extra credit! Occasionally we will have in-class activities for extra credit (+3%) You do not need to have a full or correct answer to get credit, but you do need to attempt the
More informationOperant matching. Sebastian Seung 9.29 Lecture 6: February 24, 2004
MIT Department of Brain and Cognitive Sciences 9.29J, Spring 2004 - Introduction to Computational Neuroscience Instructor: Professor Sebastian Seung Operant matching Sebastian Seung 9.29 Lecture 6: February
More informationApplying STPA to the Artificial Pancreas for People with Type 1 Diabetes
Applying STPA to the Artificial Pancreas for People with Type 1 Diabetes Lane Desborough Product Strategist Medtronic Diabetes Northridge, California Type 1 Diabetes Artificial Pancreas Challenges Applying
More informationVorlesung Grundlagen der Künstlichen Intelligenz
Vorlesung Grundlagen der Künstlichen Intelligenz Reinhard Lafrenz / Prof. A. Knoll Robotics and Embedded Systems Department of Informatics I6 Technische Universität München www6.in.tum.de lafrenz@in.tum.de
More informationOverview. cis32-spring2003-parsons-lect15 2
EXPERT SYSTEMS Overview The last lecture looked at rules as a technique for knowledge representation. In a while we will go on to look at logic as a means of knowledge representation. First, however, we
More informationA Scoring Policy for Simulated Soccer Agents Using Reinforcement Learning
A Scoring Policy for Simulated Soccer Agents Using Reinforcement Learning Azam Rabiee Computer Science and Engineering Isfahan University, Isfahan, Iran azamrabiei@yahoo.com Nasser Ghasem-Aghaee Computer
More informationAgents and State Spaces. CSCI 446: Artificial Intelligence
Agents and State Spaces CSCI 446: Artificial Intelligence Overview Agents and environments Rationality Agent types Specifying the task environment Performance measure Environment Actuators Sensors Search
More informationOverview EXPERT SYSTEMS. What is an expert system?
EXPERT SYSTEMS Overview The last lecture looked at rules as a technique for knowledge representation. In a while we will go on to look at logic as a means of knowledge representation. First, however, we
More informationFoundations of Natural Language Processing Lecture 13 Heads, Dependency parsing
Foundations of Natural Language Processing Lecture 13 Heads, Dependency parsing Alex Lascarides (slides from Alex Lascarides, Henry Thompson, Nathan Schneider and Sharon Goldwater) 6 March 2018 Alex Lascarides
More informationCS343: Artificial Intelligence
CS343: Artificial Intelligence Introduction: Part 2 Prof. Scott Niekum University of Texas at Austin [Based on slides created by Dan Klein and Pieter Abbeel for CS188 Intro to AI at UC Berkeley. All materials
More informationAn Intent-Driven Planner for Multi-Agent Story Generation
An Intent-Driven Planner for Multi-Agent Story Generation Mark Owen Riedl Liquid Narrative Group Department of Computer Science North Carolina State University Raleigh, NC 27695 moriedl@ncsu.edu Abstract
More informationWays of counting... Review of Basic Counting Principles. Spring Ways of counting... Spring / 11
Ways of counting... Review of Basic Counting Principles Spring 2013 Ways of counting... Spring 2013 1 / 11 1 Introduction 2 Basic counting principles a first view 3 Counting with functions 4 Using functions
More informationAgent-Based Systems. Agent-Based Systems. Michael Rovatsos. Lecture 5 Reactive and Hybrid Agent Architectures 1 / 19
Agent-Based Systems Michael Rovatsos mrovatso@inf.ed.ac.uk Lecture 5 Reactive and Hybrid Agent Architectures 1 / 19 Where are we? Last time... Practical reasoning agents The BDI architecture Intentions
More informationSymbolic CTL Model Checking
Symbolic CTL Model Checking Crystal Chang Din Precise Modeling and Analysis group (PMA), University of Oslo INF9140 - Specification and Verification of Parallel Systems Crystal Chang Din @ UiO Symbolic
More informationBehavior Architectures
Behavior Architectures 5 min reflection You ve read about two very different behavior architectures. What are the most significant functional/design differences between the two approaches? Are they compatible
More information(2) In each graph above, calculate the velocity in feet per second that is represented.
Calculus Week 1 CHALLENGE 1-Velocity Exercise 1: Examine the two graphs below and answer the questions. Suppose each graph represents the position of Abby, our art teacher. (1) For both graphs above, come
More informationRe: ENSC 370 Project Gerbil Functional Specifications
Simon Fraser University Burnaby, BC V5A 1S6 trac-tech@sfu.ca February, 16, 1999 Dr. Andrew Rawicz School of Engineering Science Simon Fraser University Burnaby, BC V5A 1S6 Re: ENSC 370 Project Gerbil Functional
More informationEEL-5840 Elements of {Artificial} Machine Intelligence
Menu Introduction Syllabus Grading: Last 2 Yrs Class Average 3.55; {3.7 Fall 2012 w/24 students & 3.45 Fall 2013} General Comments Copyright Dr. A. Antonio Arroyo Page 2 vs. Artificial Intelligence? DEF:
More informationPolicy Gradients. CS : Deep Reinforcement Learning Sergey Levine
Policy Gradients CS 294-112: Deep Reinforcement Learning Sergey Levine Class Notes 1. Homework 1 due today (11:59 pm)! Don t be late! 2. Remember to start forming final project groups Today s Lecture 1.
More informationAUTOMATED DEBUGGING. Session Leader: Zongxu Mu
AUTOMATED DEBUGGING Session Leader: Zongxu Mu THE FIRST COMPUTER BUG Source: https://plus.google.com/107048744275438760860/posts/twfceqadsld 10/22/2013 Automated Debugging - Zongxu Mu 2 A RANDOM COMPUTER
More information(c) KSIS Politechnika Poznanska
Fundamentals of Autonomous Systems Control architectures in robotics Dariusz Pazderski 1 1 Katedra Sterowania i In»ynierii Systemów, Politechnika Pozna«ska 9th March 2016 Introduction Robotic paradigms
More informationn Outline final paper, add to outline as research progresses n Update literature review periodically (check citeseer)
Project Dilemmas How do I know when I m done? How do I know what I ve accomplished? clearly define focus/goal from beginning design a search method that handles plateaus improve some ML method s robustness
More informationA Guide to Algorithm Design: Paradigms, Methods, and Complexity Analysis
A Guide to Algorithm Design: Paradigms, Methods, and Complexity Analysis Anne Benoit, Yves Robert, Frédéric Vivien To cite this version: Anne Benoit, Yves Robert, Frédéric Vivien. A Guide to Algorithm
More informationArtificial Intelligence. Intelligent Agents
Artificial Intelligence Intelligent Agents Agent Agent is anything that perceives its environment through sensors and acts upon that environment through effectors. Another definition later (Minsky) Humans
More informationArtificial Intelligence CS 6364
Artificial Intelligence CS 6364 Professor Dan Moldovan Section 2 Intelligent Agents Intelligent Agents An agent is a thing (e.g. program, or system) that can be viewed as perceiving its environment and
More informationExpert Systems. Artificial Intelligence. Lecture 4 Karim Bouzoubaa
Expert Systems Artificial Intelligence Lecture 4 Karim Bouzoubaa Artificial Intelligence Copyright Karim Bouzoubaa 2 Introduction ES: Capture, represent, store and apply human K using a machine Practical
More informationSequential Decision Making
Sequential Decision Making Sequential decisions Many (most) real world problems cannot be solved with a single action. Need a longer horizon Ex: Sequential decision problems We start at START and want
More informationRobot Learning Letter of Intent
Research Proposal: Robot Learning Letter of Intent BY ERIK BILLING billing@cs.umu.se 2006-04-11 SUMMARY The proposed project s aim is to further develop the learning aspects in Behavior Based Control (BBC)
More informationPLANNING HEURISTICS FOR CONJUNCTIVE AND DISJUNCTIVE GOALS. A Preliminary Report. Submitted to the Faculty. Purdue University.
PLANNING HEURISTICS FOR CONJUNCTIVE AND DISJUNCTIVE GOALS A Preliminary Report Submitted to the Faculty of Purdue University by Lin Zhu In Partial Fulfillment of the Requirements for the Degree of Doctor
More informationHRM Triax 100 is engineered to provide quick, one-touch access to a variety of critical information during training and in competition.
1 Information is power. HRM Triax 100 is engineered to provide quick, one-touch access to a variety of critical information during training and in competition. Used properly this heart rate monitor will
More informationOverview. What is an agent?
Artificial Intelligence Programming s and s Chris Brooks Overview What makes an agent? Defining an environment Overview What makes an agent? Defining an environment Department of Computer Science University
More information2 Psychological Processes : An Introduction
2 Psychological Processes : An Introduction 2.1 Introduction In our everyday life we try to achieve various goals through different activities, receive information from our environment, learn about many
More informationStepwise Knowledge Acquisition in a Fuzzy Knowledge Representation Framework
Stepwise Knowledge Acquisition in a Fuzzy Knowledge Representation Framework Thomas E. Rothenfluh 1, Karl Bögl 2, and Klaus-Peter Adlassnig 2 1 Department of Psychology University of Zurich, Zürichbergstraße
More informationAutomated Conflict Detection Between Medical Care Pathways
Automated Conflict Detection Between Medical Care Pathways Philip Weber, Bosco Filho, Mark Lee, Ian Litchfield, Ruth Backman University of Birmingham, UK School of Computer Science Institute of Applied
More informationUnifying Data-Directed and Goal-Directed Control: An Example and Experiments
Unifying Data-Directed and Goal-Directed Control: An Example and Experiments Daniel D. Corkill, Victor R. Lesser, and Eva Hudlická Department of Computer and Information Science University of Massachusetts
More informationReactivity and Deliberation in Decision-Making Systems
11 Reactivity and Deliberation in Decision-Making Systems Carle Côté 11.1 Introduction 11.2 Let s Begin at the Beginning 11.3 Common Pitfall #1 : One Decision Model to Rule Them All! 11.4 Common Pitfall
More informationHarvard-MIT Division of Health Sciences and Technology HST.952: Computing for Biomedical Scientists. Data and Knowledge Representation Lecture 6
Harvard-MIT Division of Health Sciences and Technology HST.952: Computing for Biomedical Scientists Data and Knowledge Representation Lecture 6 Last Time We Talked About Medical Coding Systems UMLS Today
More informationMore Examples and Applications on AVL Tree
CSCI2100 Tutorial 11 Jianwen Zhao Department of Computer Science and Engineering The Chinese University of Hong Kong Adapted from the slides of the previous offerings of the course Recall in lectures we
More informationHRM Triax 15 is engineered to provide quick, one-touch access to a variety of critical information during training and in competition.
1 Information is power. HRM Triax 15 is engineered to provide quick, one-touch access to a variety of critical information during training and in competition. Used properly this heart rate monitor will
More informationA Reinforcement Learning Approach Involving a Shortest Path Finding Algorithm
Proceedings of the 003 IEEE/RSJ Intl. Conference on Intelligent Robots and Systems Proceedings Las Vegas, of Nevada the 003 October IEEE/RSJ 003 Intl. Conference on Intelligent Robots and Systems Las Vegas,
More informationToday we will... Foundations of Natural Language Processing Lecture 13 Heads, Dependency parsing. Evaluating parse accuracy. Evaluating parse accuracy
Today we will... Foundations of Natural Language Processing Lecture 13 Heads, Dependency parsing Alex Lascarides (slides from Alex Lascarides, Henry Thompson, Nathan chneider and haron Goldwater) 6 March
More information