Goal Regression Planning, Constraint Automata and Causal Graphs

Size: px
Start display at page:

Download "Goal Regression Planning, Constraint Automata and Causal Graphs"

Transcription

1 Goal Regression Planning, Constraint Automata and Causal Graphs Contributions: Dan Weld Seung Chung Erez Karpas David Wang Prof Brian Williams MIT CSAIL courtesy of JPL New Text J/6.834J Cognitive Robotics Leap Day - February 29 th, 2016 This image is in the public domain. 1

2 After lecture you will know how to Generate plans by working backwards from goals. Generate least commitment, partial order plans. Encode actions with indirect effects as concurrent automata with constraints (CCA). Analyze action dependence using causal graphs. Generate plans with out search, for CCA that have tree structured causal graphs. Use causal graph planning as a heuristic to HFS. 2/29/ J / 6.834J Goal Regression & Causal Graph Planning 2

3 Assignments Today: D. Weld, An introduction to least commitment planning, AI Magazine 15(4):27-61, B. Williams and P. Nayak, "A reactive planner for a model-based executive, IJCAI, Next: D. Wang and B. Williams, tburton: A Divide and Conquer Temporal Planner, AAAI, Homework: PSet #3 PDDL Modeling, out today, due Wed, March 9 th. Advanced Lecture Pset #1, out today, due Fri, March 4 th. 2/29/ J / 6.834J Goal Regression & Causal Graph Planning 3

4 Outline Review: programs on state Planning as goal regression (SNLP) Goal regression planning with causal graphs (Burton) Appendix: HFS planning with the causal graph heuristic (Fast Downward) 4

5 A single cognitive system language and executive. Enterprise Programs that: Uhura User Goals & models in RMPL Collaboratively resolves goal failures Kirk Make choices Sketches mission and assigns tasks Burton Pike Plans actions Adjust timing Control Commands Coordinates and monitors tasks Sulu Bones Plans paths Diagnoses likely failures Monitor conditions 2/29/2016 source unknown. All rights reserved. This content is excluded from our Creative Commons license. For more information, see Goal Regression and Causal Graph Planning 5

6 In a State-Space Program, Activity Planning Maps Desired States to Actions no-fly zone lake2 fire2 no-fly zone fire1 base2 no-fly zone lake1 uav1 base1 class Main{ UAV uav1; Lake lake1; Lake lake2; Fire fire1; Fire fire2;... Main (){ uav1 = new UAV(); uav1.location= base_1_location; uav1.flying = no; uav1.loaded = no; lake1 = new Lake(); lake1.location = lake_1_location; lake2 = new Lake(); lake2.location = lake_2_location; class UAV { Roadmap location; Boolean flying; Boolean loaded; primitive method takeoff() flying == no => flying == yes; primitive method land() flying == yes => flying == no; primitive method load_water(lake lakespot) ((flying == yes) && (loaded == no) && (lakespot.location == location)) => loaded == yes; method run() { sequence{ }}} (fire1 == out); (fire2 == out); (uav1.flying == no && uav1.location == base_1_location); } fire1 = new Fire(); fire1.location = fire_1_location; fire1 = high; fire2 = new Fire(); fire2.location = fire_2_location; fire2 = high; } primitive method drop_water_high_altiture(fire firespot) ((flying == yes) && (loaded == yes) && (firespot.location == location) && (firespot == high)) => ((loaded == no) && (firespot == medium)); primitive method drop_water_low_altiture(fire firespot) ((flying == yes) && (loaded == yes) && (firespot.location == location) && (firespot == medium)) => ((loaded == no) && (firespot == out)); #MOTION_PRIMITIVES(location, fly, flying==yes) 6

7 Planning maps desired states to actions no-fly zone fire2 no-fly zone base2 no-fly zone lake1 lake2 fire1 uav1 base1 Roadmap Path Planning Classical Action Planning 7

8 A single cognitive system language and executive. Enterprise State Space Programs that: Uhura User Goals & models in RMPL Collaboratively resolves goal failures Kirk Sketches mission and assigns tasks Burton Plan to achieve discrete states. Plans actions Control Commands Pike Coordinates and monitors tasks Sulu Plan to achieve continuous states. Bones Plans paths Diagnoses likely failures Monitor continuous states. source unknown. All rights reserved. This content is excluded from our Creative Commons license. For more information, see 2/29/2016 Goal Regression and Causal Graph Planning 8

9 Last Week: Classic Activity Planning Representation (STRIPS/PDDL) Action Model: Objects (things): Predicates (used to create true or false statements about the world, facts ): (On A, B ) (Clear A ) Actions (used to change truth of predicates): (Clear ) A (Clear ) B Stack (Clear B ) (On A, B ) Initial: Goal: preconditions (Clear ) (Clear ) (Clear ) (On, ) (On, ) effects 9

10 Last Week: Forward Search action3 s 2 0 action1 s 1 0 Initial State (s o ) action2 s 1 0 action4 goal action2 s 2 1 action3 s 1 0 action3 s 2 2 Time 10

11 This Week: Goal-Regression Search action3 S t 0 action1 Initial State (s o ) action2 action2 S t 1 action4 goal action3 S t 2 action2 Time 11

12 This Week: Planning to Control Complex Devices no-fly zone fire2 no-fly zone base2 no-fly zone lake1 lake2 fire1 uav1 base1 Roadmap Path Planning Classical Action Planning Thrust Goals Delta_V(direction=b, magnitude=200) Power This image is in the public domain. Attitude Point(a) Turn(a,b) Point(b) Turn(b,a) Device Reconfiguration & Repair Engine Off Warm Up Thrust (b, 200) O ff Time-line Planning 12

13 This Week: Automata Representation Action Model: Concurrent Constraint Automata (CCA) (like a state-machine): Initial: Goal: Note: This is a very simple example, there are usually many automata, and guards on the transitions. We will use CCA to support indirect effects, concurrency and metric time. Algorithms exist to map between the two representations. 13

14 Outline Review: programs on state Planning as goal regression (SNLP/UCPOP) Partial Order Planning Overview Partial Order Planning Problem Partial Order Plan Generation Goal regression planning with causal graphs (Burton) Appendix: HFS planning with the causal graph heuristic (Fast Downward) 14

15 Classical Planning Problem Propositions: P i Initial Conditions: P 1 P 2 P 3 P 4 Operators: pre 1 pre 2 pre 3 Op eff 1 eff 2 Goals: Goal 1 Goal 2 Goal 3 2/29/ J / 6.834J Goal Regression & Causal Graph Planning 15

16 Simple Spacecraft Problem Observation-1 target instruments Observation-2 Observation-3 Observation-4 pointing calibrated This image is in the public domain. Propositions: Operators: Target Pointed To, Camera Calibrated?, Has Image? Calibrate, Turn to Y, and Take Image. 2/29/ J / 6.834J Goal Regression & Causal Graph Planning 16

17 Example Init Actions Goal p C p C C c I A p y p x T y p x c Im I x p x Propositions: Operators: Target Pointed To, Camera Calibrated?, Has Image? Calibrate, Turn to Y, and Take Image. 2/29/ J / 6.834J Goal Regression & Causal Graph Planning 17

18 Actions in the Planning Domain Description Language (PDDL) (:action TakeImage :parameters (?target,?instr) :precondition (and (Status?instr Calibrated) (Pointing?target)) :effect (Image?target)) By convention, parameters start with?, as in?var. (:action Calibrate :parameters (?instrument) :precondition (and (Status?instr On) (Calibration-Target?target), (Pointing?target) :effect (and (not (Status?inst On)) (Status?instr Calibrated))) (:action Turn :parameters (?target) :precondition (and (Pointing?direction)?direction?target :effect (and (not (Pointing?direction) (Pointing?target))) 2/29/ J / 6.834J Goal Regression & Causal Graph Planning 18

19 Partial Order Plan p C C c S p A Im I A F p C T A p C 2/29/ J / 6.834J Goal Regression & Causal Graph Planning 19

20 Planning from Goals: Partial Order Causal Link Planning (SNLP, UCPOP) I A F 1. Select an open condition; 2. Choose an op that can achieve it: Link to an existing instance or Add a new instance; 3. Resolve threats. p C c I Im A F p A C c p A Im I A F S p C p C C c p A T A Im p C Alternatives: Forward Search, CSP I A S F S p C C c p A p A p C T A 2/29/ J / 6.834J Goal Regression & Causal Graph Planning p C Im I A p C C c Im I A F F 20

21 Planning as Goal Regression Partial Planning Overview Partial Order Planning Problem Problem Encoding Partial Order Plans Plan Correctness Partial Order Plan Generation 2/29/ J / 6.834J Goal Regression & Causal Graph Planning 21

22 Example Problem Initial State: At(Home) Sells(HWS,Drill) Sells(SM,Milk) Sells(SM,Ban.) Goal: Have(Milk) At(Home) Have(Ban.) Have(Drill) Operators: At(HWS) Go(SM) At(SM) At(SM), Sells(SM,Ban.) Buy(Ban.) Have(Ban) At(Home) Go(HWS) At(HWS) At(SM) Go(Home) At(Home) At(HWS) Sells(HWS,Drill) Buy(Drill) Have(Drill) At(SM), Sells(SM,Milk) Buy(Milk) Have(Milk) 2/29/ J / 6.834J Goal Regression & Causal Graph Planning 22

23 Initial and Goal States Encoded as Operators Start At(Home) Sells(HWS,Drill) Sells(SM,Milk) Sells(SM,Ban.) Why encode as operators? Don t need to introduce (partial) states as separate objects. Keeps theory minimal. Have(Milk) At(Home) Have(Ban.) Have(Drill) Finish 2/29/ J / 6.834J Goal Regression & Causal Graph Planning 23

24 Partial Order Plan <Actions,Orderings,Links> Start At(Home) Sells(HWS,Drill) Sells(SM,Milk) Sells(SM,Ban.) At(Home) Go(HWS) At(HWS) Sells(HWS,Drill) Buy(Drill) At(HWS) Go(SM) At(SM), Sells(SM,Milk) Buy(Milk) At(SM), Sells(SM,Ban.) Buy(Ban.) At(SM) Go(Home) Have(Milk) At(Home) Have(Ban.) Have(Drill) Finish 2/29/ J / 6.834J Goal Regression & Causal Graph Planning 24

25 Partial Order Plan <Actions,Orderings,Links> Start At(Home) Sells(HWS,Drill) Sells(SM,Milk) Sells(SM,Ban.) At(Home) Go(HWS) At(HWS) Sells(HWS,Drill) Buy(Drill) At(HWS) Go(SM) At(SM), Sells(SM,Milk) Buy(Milk) At(SM), Sells(SM,Ban.) Buy(Ban.) At(SM) Go(Home) Have(Milk) At(Home) Have(Ban.) Have(Drill) Finish 2/29/ J / 6.834J Goal Regression & Causal Graph Planning 25

26 Partial Order Plan <Actions,Orderings,Links> Start At(Home) Sells(HWS,Drill) Sells(SM,Milk) Sells(SM,Ban.) At(Home) Go(HWS) At(HWS) Sells(HWS,Drill) Buy(Drill) At(HWS) Go(SM) At(SM), Sells(SM,Milk) Buy(Milk) At(SM), Sells(SM,Ban.) Buy(Ban.) At(SM) Go(Home) Have(Milk) At(Home) Have(Ban.) Have(Drill) Finish 2/29/ J / 6.834J Goal Regression & Causal Graph Planning 26

27 Partial Order Plan <Actions,Orderings,Links> Start At(Home) Sells(HWS,Drill) Sells(SM,Milk) Sells(SM,Ban.) At(Home) Go(HWS) At(HWS) Sells(HWS,Drill) Buy(Drill) At(HWS) Go(SM) At(SM), Sells(SM,Milk) Buy(Milk) At(SM), Sells(SM,Ban.) Buy(Ban.) At(SM) Go(Home) Have(Milk) At(Home) Have(Ban.) Have(Drill) Finish 2/29/ J / 6.834J Goal Regression & Causal Graph Planning 27

28 Planning as Goal Regression Partial Planning Overview Partial Order Planning Problem Problem Encoding Partial Order Plans Plan Correctness Partial Order Plan Generation 2/29/ J / 6.834J Goal Regression & Causal Graph Planning 28

29 What Constitutes a Correct Plan? Complete Plan Achieves all Goals Achieves all preconditions No actions intervene to undo a needed precondition Consistent Plan There exists an execution sequence that is consistent with the ordering At(SM), Sells(SM,Milk) Buy(Milk) Start At(Home) Sells(HWS,Drill) Sells(SM,Milk) Sells(SM,Ban.) At(Home) Go(HWS) At(HWS) Sells(HWS,Drill) Buy(Drill) At(HWS) Go(SM) At(SM) Go(Home) Finish At(SM), Sells(SM,Ban.) Buy(Ban.) Have(Milk) At(Home) Have(Ban.) Have(Drill) 2/29/ J / 6.834J Goal Regression & Causal Graph Planning 29

30 Why is an ordering needed? At(HWS) Go(SM) At(SM), not At(HWS) At(SM) Buy(Milk) Have(Milk) At(SM) Go(Home) At(Home), not At(SM) 2/29/ J / 6.834J Goal Regression & Causal Graph Planning 30

31 Why is an ordering needed? Suppose the other order is allowed, what happens? At(HWS) Go(SM) At(SM), not At(HWS) Threatened Precondition At(SM) Buy(Milk) At(SM) Go(Home) At(Home), not At(SM) 2/29/ J / 6.834J Goal Regression & Causal Graph Planning 31

32 Why is an ordering needed? Link indicates protected time interval for the precondition. Suppose the other order is allowed, what happens? At(SM) Buy(Milk) At(HWS) Go(SM) At(SM), not At(HWS) At(SM) Go(Home) At(Home), not At(SM) 2/29/ J / 6.834J Goal Regression & Causal Graph Planning 32

33 Why is an ordering needed? The ordering resolves the threat. At(SM) Buy(Milk) At(HWS) Go(SM) At(SM), not At(HWS) At(SM) Go(Home) At(Home), not At(SM) 2/29/ J / 6.834J Goal Regression & Causal Graph Planning 33

34 Solution: A Complete and Consistent Plan Complete Plan IFF every precondition of every step is achieved. A step s precondition is achieved iff its the effect of some preceding step, no intervening step can undo it. Consistent Plan At(SM), Sells(SM,Milk) Buy(Milk) IFF there is no contradiction in the ordering constraint. i.e., never s i < s j and s j < s i, or the causal links + orderings are loop free. Start At(Home) Sells(HWS,Drill) Sells(SM,Milk) Sells(SM,Ban.) At(Home) Go(HWS) At(HWS) Sells(HWS,Drill) Buy(Drill) At(HWS) Go(SM) At(SM) Go(Home) Finish At(SM), Sells(SM,Ban.) Buy(Ban.) Have(Milk) At(Home) Have(Ban.) Have(Drill) 2/29/ J / 6.834J Goal Regression & Causal Graph Planning 34

35 Planning as Goal Regression Partial Planning Overview Partial Order Planning Problem Partial Order Plan Generation Derivation from Completeness and Consistency Backward Chaining Threat Resolution The POP algorithm 2/29/ J / 6.834J Goal Regression & Causal Graph Planning 35

36 Partial Order Planning Algorithm The algorithm falls out of Consistency and Completeness Completeness: Must achieve all preconditions Backward chain from goals to initial state, by inserting actions and causal links. Must avoid intervening actions that threaten After each action is inserted, find any action that threatens its effects, and impose ordering to resolve. Consistent: Ordering must be consistent After each causal link and ordering is inserted, check for loops. 2/29/ J / 6.834J Goal Regression & Causal Graph Planning 36

37 Planning as Goal Regression Partial Planning Overview Partial Order Planning Problem Partial Order Plan Generation Derivation from Completeness and Consistency Backward Chaining Threat Resolution The POP algorithm 2/29/ J / 6.834J Goal Regression & Causal Graph Planning 37

38 Start At(Home) Sells(HWS,Drill) Sells(SM,Milk) Sells(SM,Ban.) Have(Drill) Have(Milk) Have(Ban.) at(home) Finish 2/29/ J / 6.834J Goal Regression & Causal Graph Planning 38

39 Start At(Home) Sells(HWS,Drill) Sells(SM,Milk) Sells(SM,Ban.) At(HWS) Sells(HWS,Drill) Buy(Drill) At(SM), Sells(SM,Milk) Buy(Milk) At(SM), Sells(SM,Ban.) Buy(Ban.) Have(Drill) Have(Milk) Have(Ban.) at(home) Finish 2/29/ J / 6.834J Goal Regression & Causal Graph Planning 39

40 Start At(Home) Sells(HWS,Drill) Sells(SM,Milk) Sells(SM,Ban.) At(HWS) Sells(HWS,Drill) At(SM), Sells(SM,Milk) At(SM), Sells(SM,Ban.) Buy(Drill) Buy(Milk) Buy(Ban.) Have(Drill) Have(Milk) Have(Ban.) at(home) Finish 2/29/ J / 6.834J Goal Regression & Causal Graph Planning 40

41 Start At(Home) Sells(HWS,Drill) Sells(SM,Milk) Sells(SM,Ban.) At(?x) Go(HWS) At(?x) Go(SM) At(HWS) Sells(HWS,Drill) At(SM), Sells(SM,Milk) At(SM), Sells(SM,Ban.) Buy(Drill) Buy(Milk) Buy(Ban.) Have(Drill) Have(Milk) Have(Ban.) at(home) Finish 2/29/ J / 6.834J Goal Regression & Causal Graph Planning 41

42 Planning as Goal Regression Partial Planning Overview Partial Order Planning Problem Partial Order Plan Generation Derivation from Completeness and Consistency Backward Chaining Threat Resolution The POP algorithm 2/29/ J / 6.834J Goal Regression & Causal Graph Planning 42

43 After adding a causal link/action, find threat with any existing action/link At(HWS) Go(SM) At(SM), not At(HWS) At(SM) Buy(Milk) At(SM) Go(Home) At(Home), not At(SM) 2/29/ J / 6.834J Goal Regression & Causal Graph Planning 43

44 To remove threats At(HWS) Go(SM) At(SM), not At(HWS) At(SM) Buy(Milk) At(SM) Go(Home) At(Home), not At(SM) 2/29/ J / 6.834J Goal Regression & Causal Graph Planning 44

45 To remove threats promote the threat or At(SM) Buy(Milk) At(HWS) Go(SM) At(SM), not At(HWS) At(SM) Go(Home) At(Home), not At(SM) 2/29/ J / 6.834J Goal Regression & Causal Graph Planning 45

46 To remove threats promote the threat or demote the threat At(HWS) Go(SM) At(SM), not At(HWS) At(SM) Go(Home) At(Home), not At(SM) At(SM) Buy(Milk) But only allow demotion/promotion if schedulable consistent = loop free no action precedes initial state 2/29/ J / 6.834J Goal Regression & Causal Graph Planning 46

47 Go(SM) threatens At(HWS) Promote Start At(Home) Sells(HWS,Drill) Sells(SM,Milk) Sells(SM,Ban.) At(Home) Go(HWS) At(Home) Go(SM) At(HWS) Sells(HWS,Drill) At(SM), Sells(SM,Milk) At(SM), Sells(SM,Ban.) Buy(Drill) Buy(Milk) Buy(Ban.) Have(Drill) Have(Milk) Have(Ban.) at(home) Finish 2/29/ J / 6.834J Goal Regression & Causal Graph Planning 47

48 Go(HWS) threatens At(Home) Try Promote Start Inconsistent! At(Home) Sells(HWS,Drill) Sells(SM,Milk) Sells(SM,Ban.) At(Home) Go(HWS) At(Home) Go(SM) At(HWS) Sells(HWS,Drill) At(SM), Sells(SM,Milk) At(SM), Sells(SM,Ban.) Buy(Drill) Buy(Milk) Buy(Ban.) Have(Drill) Have(Milk) Have(Ban.) at(home) Finish 2/29/ J / 6.834J Goal Regression & Causal Graph Planning 48

49 Use Go(HWS) to produce At(?x) Finish by Go(Home). Start At(Home) Sells(HWS,Drill) Sells(SM,Milk) Sells(SM,Ban.) At(Home) Go(HWS) At(?x) Go(SM) HWS At(HWS) Sells(HWS,Drill) At(SM), Sells(SM,Milk) At(SM), Sells(SM,Ban.) Buy(Drill) Buy(Milk) Buy(Ban.) Have(Drill) Have(Milk) Have(Ban.) at(home) Finish At(SM) Go(Home) 2/29/ J / 6.834J Goal Regression & Causal Graph Planning 49

50 POP(<A,O,L>, agenda, actions) 1. Termination: If agenda is empty, return plan <A,O,L>. 2. Goal Selection: Select and remove open condition <p, a need > from agenda. 3. Action Selection: Choose new or existing action a add that can precede a need and whose effects include p. Link and order actions. 4. Update Agenda: If a add is new, add its preconditions to agenda. 5. Threat Detection: For every action a threat that might threaten some causal link from a produce to a consume, choose a consistent ordering: a) Demote: Add a threat < a produce b) Promote: Add a consume < a threat 6. Recurse: on modified plan and agenda Choose is nondeterministic 2/29/ J / 6.834J Goal Regression & Causal Graph Planning Select is deterministic 50

51 Lets Start with Goal-Regression Search action3 S t 0 action1 Initial State (s o ) action2 action2 S t 1 action4 goal Time action3 S t 2 action2 Why can Goal Regression be slow? Multiple actions can achieve goals. Many possible (sub-)goal orderings. Dead-ends can be discovered late. We try a real-world example next! 51

52 What assumptions are implied by the STRIPS representation? TakeImage (?target,?instr): Pre: Status(?instr, Calibrated), Pointing(?target) Eff: Image(?target) Calibrate (?instrument): Pre: Status(?instr, On), Calibration-Target(?target), Pointing(?target) Eff: Status(?inst, On), Status(?instr, Calibrated) Turn (?target): Pre: Pointing(?direction),?direction?target Eff: Pointing(?direction), Pointing(?target) STRIPS Assumptions: Atomic time, Agent is omniscient (no sensing necessary), Agent is sole cause of change, Actions have deterministic effects, and No indirect effects. One action at a time. No metric time. No goals over time. 2/29/ J / 6.834J Goal Regression & Causal Graph Planning 52

53 Outline Review: programs on state Planning as goal regression (SNLP/UCPOP) Goal regression planning with causal graphs (Burton) Appendix: HFS planning with the causal graph heuristic (Fast Downward) 53

54 DS1 Revisited: Planning to Control Interacting Devices over Time Thrust Goals Delta_V(direction=b, magnitude=200) Power Time-line Planning (tburton) Attitude Point(a) Turn(a,b) Point(b) Turn(b,a) Engine Off Warm Up Thrust (b, 200) O ff This image is in the public domain. Reactive Device Reconfiguration and Repair (Burton) 54

55 Domain Independent Planning in the Real-World classical off-line Speed Heuristic-based Searches & Symbolic Search (not presented) real-time Burton (real-time planning) Expressivity tburton (planning with time and indirect effects) Both Burtons use Causal Graphs to analyze weak and tight couplings and to Decompose. 55

56 Planning with Indirect Effects: DS 1 Attitude Control System Example Flight Computer Data Commands PDE Propulsion Drive Electronics SRU Stellar Reference Unit PDU Power Distribution Unit z facing thrusters He N 2 H 4 x facing thrusters Bus Controller GDE Gimbal Drive Electronics PASM Power Activation & Switching Module DSEU Diagnostics Sensors Electronics Unit Dynamics 1553 bus ACE Attitude Control Electronics control mode xatt error yatt error zatt error 56

57 Why is Controlling an Engineered Device Easier than a Puzzle? Propulsion Drive Electronics Stellar Reference Unit Power Distribution Unit Gimbal Drive Electronics Power Activation & Switching Module Diagnostics Sensors Electronics Unit Attitude Control Electronics Flight Computer PDE He Data Commands SRU PDU z facing thrusters N 2 H 4 x facing thrusters Bus Controller GDE 3 5 PASM DSEU Dynamics 1553 bus ACE control mode xatt erro r yatt erro r zatt erro r 1. Actions are reversible. 2. Devices hold state. 3. Causes and effects form a tree. 1. Actions are reversible. 2. Devices hold state. 3. Causes and effects form tight cycles. 57

58 Outline Review: programs on state Planning as goal regression (SNLP/UCPOP) Goal regression planning with causal graphs (Burton) Constraint automata planning problem Causal graphs Using causal graphs to order goals Computing policies for selecting actions Appendix: Planning for cyclic causal graphs Appendix: HFS planning with the causal graph heuristic (Fast Downward) 58

59 Concurrent Constraint Automata Planning Problem flow in dcmd out = vcmd in dcmd in Variables and Domains: flow out State: driver in {on, off, resettable, failed}, valve in {open, closed, stuck-open, stuck-closed}. Control: dcmd in in {idle, on, off, reset, open, close} Dependent: flow in, flow out in {pos, neg, zero} dcmd out, vcmd in in Domain{dcmd in } Initial assignment: {driver = on, valve = closed} Goal assignment: {driver = off, valve = open} 59

60 Concurrent Constraint Automata Planning Problem (cont.) flow in dcmd out = vcmd in dcmd in Constraint automata (one per state variable): flow out driver dcmd in = dcmd out valve flowin in = flow out On dcmd in = reset Resettable Open Stuck Open dcmd in = on dcmd in = off dcmd in = off vcmd in = open vcmd in = close Off Failed Closed Stuck Closed Assume: transitions independently controlled, each location can idle. State constraints: {dcmd out = vcmd in } 60

61 Outline Review: programs on state Planning as goal regression (SNLP/UCPOP) Goal regression planning with causal graphs (Burton) Constraint automata planning problem Causal graphs Using causal graphs to order goals Computing policies for selecting actions Appendix: Planning for cyclic causal graphs Appendix: HFS planning with the causal graph heuristic (Fast Downward) 61

62 Observation: Engineered systems are largely loop free. Computer Bus Control Remote Terminal Remote Terminal Driver Driver Valve Valve 62

63 Causal Graph G of concurrent automata S: Vertices are control and state variables of automata. Edge from v i to v j if v j s transition is conditioned on v i. Driver On dcmd in = dcmd out dcmd in = reset Resettable dcmd in Driver Valve dcmd in = on dcmd in = off Off dcmd in = off Failed Valve Open flowin in = flow out Stuck Open dcmd in vcmd in = open vcmd in = close dcmd out = vcmd in Closed Stuck Closed 63

64 Outline Review: programs on state Planning as goal regression (SNLP/UCPOP) Goal regression planning with causal graphs (Burton) Constraint automata planning problem Causal graphs Using causal graphs to order goals Computing policies for selecting actions Appendix: Planning for cyclic causal graphs Appendix: HFS planning with the causal graph heuristic (Fast Downward) 64

65 Idea: use causal graph analysis to eliminate ALL forms of search (Burton) POP(<A,O,L>, agenda, actions): 1. Termination: If agenda is empty, return plan <A,O,L>. 2. Goal Selection: Select and remove open condition <p, a need > from agenda. 3. Action Selection: Choose new or existing action a add that can precede a need and whose effects include p. Link and order actions. 4. Update Agenda: If a add is new, add its preconditions to agenda. 5. Threat Detection: For every action a threat that might threaten some causal link from a produce to a consume, choose a consistent ordering: a) Demote: Add a threat < a produce b) Promote: Add a consume < a threat 6. Recurse: on modified plan and agenda Burton [Williams & Nayak, IJCAI 1997] 2/29/ J / 6.834J Goal Regression & Causal Graph Planning 65

66 Why do goal orderings matter? 1. An achieved goal can be clobbered by a subsequent goal. command Example: Current State: Goal State: driver = on, valve = closed. driver = off, valve = open. Achieving (driver = off), followed by (valve = open) clobbers (driver = off). Achieve valve goal before driver goal. Effect Cause 66

67 Goal Ordering for Causal Graph Planning Require: The CCA causal graph to be acyclic. Causal Graph Driver dcmd in Valve Idea: Achieve conjunctive goals upstream within the causal graph, from effects to causes (i.e., children to parents). 67

68 Property: Safe to achieve goals in an upstream order The only variables used to set some variable v 7 is its ancestors. Variable v 7 can be changed without affecting its descendants. Affected Unaffected Exploits: Each transitions independently controlled, each location can idle. Simple check: 1. Number the causal graph depth-first from leaves. Child has lower number than parents 2. Achieve goals in the order of increasing depth-first number. 68

69 Why do goal orderings matter? 2. Two goals can compete for the same variable associated with their sub-goals. data Switc h Latch 1 Latch 2 Example: Latch Data at Latches 1 and 2 If Latch1 and Latch2 goals achieved at same time, Latch1 and Latch2 compete for Switch position. 1 2 Solve one goal completely before the other (serially). 69

70 Property: Safe to achieve one goal before starting next sibling (serialization). Sibling goals v 7 and v 4 may both need shared ancestors. Not Shared Shared Unaffected Competition gone once sibling goal v 7 is satisfied. Not Shared Unaffected 70

71 Goal regression, CU causal graph planning example Driver 1 Driver 2 Valve 1 Valve 2 Burton s goal ordering rules: 1. Achieve effect goals before their causes, with control actiions last. 2. Achieve goals serially (like Hacker[Sussman PhD]). Valve 1 = open Valve 2 = open Driver 1 = off Driver 2 = off Open Open Turn-Off Turn-Off Driver 1 = on Driver 2 = on CU = on CU = on Turn-On CU = on CU = on CU = on CU = on Turn-On No Search! First action generated first! Turn-On Worst Case per action: Depth * Sub-goal branch factor. CU = off Average Cost per action: Sub-goal branch factor. 71

72 To select actions reactively, convert constraint automata to lookup policies Driver 2 Valve 1 dcmd out = vcmd in dcmd in = dcmd out inflow = outflow on dcmd in = reset resettable open stuck open cmd in = on dcmd in = off dcmd in = off vcmd in = open vcmd in = close off failed closed stuck closed 72

73 To select actions reactively, convert constraint automata to lookup policies Algorithm: Instance of APSP Driver 2 Valve 1 dcmd out = vcmd in dcmd in = dcmd out dcmd on in = reset resettable Goal Current On Off On idle cmd = off cmd in = on dcmd in = off dcmd in = off Off cmd = on idle off failed Resettable cmd = reset cmd = off 73

74 To select actions reactively, convert constraint automata to lookup policies Driver 2 Valve 1 dcmd out = vcmd in Goal inflow = outflow Current Open Open idle Closed driver = on cmd = close open stuck open Closed driver = on cmd = open idle vcmd in = open vcmd in = close Stuck fail fail closed stuck closed 74

75 Plan by passing sub-goals up causal graph Goal: Driver = off, Valve = closed Current: Driver = off, Valve = open Driver 2 Algorithm: see [williams and nayak, IJCAI97] Valve 1 Goal Current On Off Goal Current Open Closed On idle cmd = off Open idle driver = on cmd = close Off cmd = on idle Closed driver = on cmd = open idle Resettable cmd = reset cmd = off Stuck fail fail 75

76 Plan by passing sub-goals up causal graph Goal: Driver = off, Valve = closed Current: Driver = off, Valve = open Driver 2 Algorithm: see [williams and nayak, IJCAI97] Valve 1 Goal Current On Off Goal Current Open Closed On idle cmd = off Open idle driver = on cmd = close Off cmd = on idle Closed driver = on cmd = open idle Resettable cmd = reset cmd = off Stuck fail fail 76

77 Plan by passing sub-goals up causal graph Goal: Driver = off, Valve = closed Current: Driver = off, Valve = open Send: cmd = on Driver 2 Algorithm: see [williams and nayak, IJCAI97] Valve 1 Goal Current On Off Goal Current Open Closed On idle cmd = off Open idle driver = on cmd = close Off cmd = on idle Closed driver = on cmd = open idle Resettable cmd = reset cmd = off Stuck fail fail 77

78 Plan by passing sub-goals up causal graph Goal: Driver = off, Valve = closed Current: Driver = resettable, Valve = open Driver Failed Resettable 2 Algorithm: see [williams and nayak, IJCAI97] Valve 1 Goal Current On Off Goal Current Open Closed On idle cmd = off Open idle driver = on cmd = close Off cmd = on idle Closed driver = on cmd = open idle Resettable cmd = reset cmd = off Stuck fail fail 78

79 Plan by passing sub-goals up causal graph Goal: Driver = off, Valve = closed Current: Driver = resettable, Valve = open Driver 2 Algorithm: see [williams and nayak, IJCAI97] Valve 1 Goal Current On Off Goal Current Open Closed On idle cmd = off Open idle driver = on cmd = close Off cmd = on idle Closed driver = on cmd = open idle Resettable cmd = reset cmd = off Stuck fail fail 79

80 Plan by passing sub-goals up causal graph Goal: Driver = off, Valve = closed Current: Driver = resettable, Valve = open Send cmd = reset Driver 2 Algorithm: see [williams and nayak, IJCAI97] Valve 1 Goal Current On Off Goal Current Open Closed On idle cmd = off Open idle driver = on cmd = close Off cmd = on idle Closed driver = on cmd = open idle Resettable cmd = reset cmd = off Stuck fail fail 80

81 What if the causal graph G contains cycles? Computer Bus Control K-band Transmitter Amplifier Antenna K-band Transmitter Problem: Plan is no longer serializable. Solution: Amplifier Antenna Isolate the cyclic components (compute Strongly Connected Components). compose each cycle into a single component. New causal graph G is acyclic. Goals of G are serializable. 81

82 Action Policy for Composed Components Transmitter Amplifier on on on T on A cmd A = off cmd A = on on T off A A = off cmd T = on A = off cmd T = off T = on cmd A = on cmd A = off cmd T = on cmd T = off off off off T on A cmd A = off off T off A 82

83 Action Policy for Composed Components Problem: Composition grows exponential in space usage. Solution: Use BDD encoding. [Chung and Williams, Self-adaptive SW 03] on T on A cmd A = off cmd A = on on T off A cmd T = on cmd T = off Current Goal On T, On A On T, Off A Off T, Off A Off T, On A off T on A cmd A = off off T off A On T, On A idle cmd A = off cmd A = off fail On T, Off A cmd A = on idle cmd T = off fail Off T, Off A cmd T = on cmd T = on idle fail Off T, On A fail fail cmd A = off idle 83

84 Outline Review: programs on state Planning as goal regression (SNLP/UCPOP) Goal regression planning with causal graphs (Burton) Appendix: HFS planning with the causal graph heuristic (Fast Downward) 84

85 Causal Graph Heuristic for PDDL Recall: The Fast Forward (FF) Heuristic is computed over a Relaxed Planning Graph. Likewise: The Causal Graph (CG) Heuristic is computed over a Causal Graph. Map PDDL to concurrent automata, and extract causal graph (called domain transition graph (DTG). 2/29/2016 Goal Regression and Causal Graph Planning 85

86 Problem Reformulation Original Representation: STRIPS or PDDL Init Goal TruckIn(t, c 2 ) BoxIn(b,c 1 ) BoxIn(b, Paris) Operators, e.g. Drive(t, c 1, c 2 ) Pre: TruckIn(c 1 ) Add: TruckIn(c 2 ) Del: TruckIn(c 1 ) New Representation: Multi-valued Planning Task Variables: truck := {c 1, c 2, Paris} box := {ontruck, c 1, c 2, Paris} Init truck = c 2 box = c 1 Goal box = Paris Operators, e.g. Drive(t, c 1, c 2 ) Pre: truck = c 1 Post: truck = c 2 86

87 Domain Transition Graphs (DTG) One DTG per variable. Edges represent possible transitions (actions) and are guarded by preconditions truck c 1 Causal Edge Drive Drive Drive Drive Drive c 2 Drive Paris A causal edge between the DTG represents that the box DTG has preconditions that depend on the truck DTG. box Load truck=c 1 Unload truck=paris c 1 Unload truck=c 1 Load truck=c 2 Paris t c 2 Load truck=paris Unload truck=c 2 87

88 Calculating the Heuristic Initial: BoxIn(b, c 1 ) TruckIn(t, c 2 ) Goal: BoxIn(b, Paris) truck c 1 Drive Drive Drive Drive Paris Drive Drive c 2 box c 1 Load truck=c 1 Unload truck=paris Unload truck=c 1 Load truck=c 2 Paris t c 2 Load truck=paris Unload truck=c 2 88

89 Calculating the Heuristic Initial: BoxIn(b, c 1 ) TruckIn(t, c 2 ) Goal: BoxIn(b, Paris) truck c 1 Drive Drive Drive Drive Paris # of transitions to get the box to Paris: 2 Drive c 2 Drive box c 1 Load truck=c 1 Unload truck=paris Unload truck=c 1 Load truck=c 2 Paris t c 2 Load truck=paris Unload truck=c 2 89

90 Calculating the Heuristic Initial: BoxIn(b, c 1 ) TruckIn(t, c 2 ) Goal: BoxIn(b, Paris) truck c 1 Drive Drive Drive Drive Paris # of transitions to get the box to Paris: 2 Drive c 2 Drive # of transitions to get the truck to c 1 : 1 box c 1 Load truck=c 1 Unload truck=paris Unload truck=c 1 Load truck=c 2 Paris t c 2 Load truck=paris Unload truck=c 2 90

91 Calculating the Heuristic Initial: BoxIn(b, c 1 ) TruckIn(t, c 2 ) Goal: BoxIn(b, Paris) truck c 1 Drive Drive Drive Drive Paris # of transitions to get the box to Paris: 2 Drive c 2 Drive # of transitions to get the truck to c 1 : 1 # of transitions to get the truck to Paris: 1 box c 1 Load truck=c 1 Unload truck=paris Unload truck=c 1 Load truck=c 2 Paris t c 2 Load truck=paris Unload truck=c 2 91

92 Calculating the Heuristic Initial: BoxIn(b, c 1 ) TruckIn(t, c 2 ) Goal: BoxIn(b, Paris) truck c 1 Drive Drive Drive Drive Paris # of transitions to get the box to Paris: 2 Drive c 2 Drive # of transitions to get the truck to c 1 : 1 # of transitions to get the truck to Paris: 1 box c 1 Sum the transitions to get the heuristic value for the initial state h CG = 4 Load truck=c 1 Unload truck=paris Unload truck=c 1 Load truck=c 2 Paris t c 2 Load truck=paris Unload truck=c 2 92

93 Causal Graph Heuristic h CG ( s) v dom( s cost * ) v ( s( v), s * ( v)) The cost of getting from the current state, s, to the goal state, s* Idea: Sum the domain transition costs to get to the goal. 1. Identify each variable involved in the goal. 2. Recurse from child to parent in the causal graph. For each variable, sum the cost of along the shortest path to change the current value to the goal value. If changing that variable s value has preconditions, also add the cost of changing its parent variable. 93

94 Causal Graph Heuristic Notes Can not handle cyclic causal graphs Relax some links until the graph is acyclic Calculation performed differently in practice Modified Dijkstra algorithm Each cost calculation can over estimate. Not admissible Assumes no helpful interactions between subgoals 94

95 Breaking Cycles Break action with multiple effects into multiple unary effect actions If there are still cycles, ignore some preconditions of actions Pickup(d1, cream, w1,5) Pre: d1_location=w1, w1_cream=5 Eff: w1_cream=4, d1_payload=cream Pickup1(d1, cream, w1,5) Pre: d1_location=w1, w1_cream=5 Eff: d1_payload=cream Pickup2(d1, cream, w1,5) Pre: d1_location=w1, w1_cream=5 Eff: w1_cream=4 95

96 Breaking Cycles d1_loca tion d1_loca tion Actions split into unary effects d1_payl oad w1_inven tory d1_payl oad w1_inven tory c1_wa nt c1_wa nt 96

97 Causal Graph Heuristic Summary Graph capture dependence between variables Requires re-achieving conditions Multi-valued formulation reveals structure h CEA heuristic extends to cyclic graphs Performs extremely well on classical problems 97

98 Heuristic Families There are many other heuristics, we ve only covered two Type Relaxation Heuristics Estimates Relaxed planning graph Ignore deletes h max, h add, h ff, h cs, h m, ADHG h+ (cost of relaxed plan) Causal Graph Limited interactions h CG, h cea h+ (cost of relaxed plan) Projection Abstractions Projection h PDB, h STAN, h FORK h (cost of optimal) 98

99 Key Ideas Heuristic forward search is one of the fastest current planning techniques Domain independent heuristic design still problematic Fast Forward Heuristic Solve a relaxed problem in a Planning Graph Causal Graph Heuristic Consider problem s structure in a Causal Graph 99

100 MIT OpenCourseWare J / 6.834J Cognitive Robotics Spring 2016 For information about citing these materials or our Terms of Use, visit:

CS 730/730W/830: Intro AI

CS 730/730W/830: Intro AI CS 730/730W/830: Intro AI 1 handout: slides Wheeler Ruml (UNH) Lecture 18, CS 730 1 / 15 Comparison Extensions Setting Break Wheeler Ruml (UNH) Lecture 18, CS 730 2 / 15 Comparison Comparison Extensions

More information

Plan Recognition through Goal Graph Analysis

Plan Recognition through Goal Graph Analysis Plan Recognition through Goal Graph Analysis Jun Hong 1 Abstract. We present a novel approach to plan recognition based on a two-stage paradigm of graph construction and analysis. First, a graph structure

More information

Hoare Logic and Model Checking. LTL and CTL: a perspective. Learning outcomes. Model Checking Lecture 12: Loose ends

Hoare Logic and Model Checking. LTL and CTL: a perspective. Learning outcomes. Model Checking Lecture 12: Loose ends Learning outcomes Hoare Logic and Model Checking Model Checking Lecture 12: Loose ends Dominic Mulligan Based on previous slides by Alan Mycroft and Mike Gordon Programming, Logic, and Semantics Group

More information

Artificial Intelligence Lecture 7

Artificial Intelligence Lecture 7 Artificial Intelligence Lecture 7 Lecture plan AI in general (ch. 1) Search based AI (ch. 4) search, games, planning, optimization Agents (ch. 8) applied AI techniques in robots, software agents,... Knowledge

More information

Computational Tree Logic and Model Checking A simple introduction. F. Raimondi

Computational Tree Logic and Model Checking A simple introduction. F. Raimondi Computational Tree Logic and Model Checking A simple introduction F. Raimondi I am Franco Raimondi, f.raimondi@cs.ucl.ac.uk Slides, coursework, coursework solutions can be found online: http://www.cs.ucl.ac.uk/staff/f.raimondi/

More information

Plan Recognition through Goal Graph Analysis

Plan Recognition through Goal Graph Analysis Plan Recognition through Goal Graph Analysis Jun Hong 1 Abstract. We present a novel approach to plan recognition based on a two-stage paradigm of graph construction and analysis. First, a graph structure

More information

Chapter 2. Knowledge Representation: Reasoning, Issues, and Acquisition. Teaching Notes

Chapter 2. Knowledge Representation: Reasoning, Issues, and Acquisition. Teaching Notes Chapter 2 Knowledge Representation: Reasoning, Issues, and Acquisition Teaching Notes This chapter explains how knowledge is represented in artificial intelligence. The topic may be launched by introducing

More information

CS148 - Building Intelligent Robots Lecture 5: Autonomus Control Architectures. Instructor: Chad Jenkins (cjenkins)

CS148 - Building Intelligent Robots Lecture 5: Autonomus Control Architectures. Instructor: Chad Jenkins (cjenkins) Lecture 5 Control Architectures Slide 1 CS148 - Building Intelligent Robots Lecture 5: Autonomus Control Architectures Instructor: Chad Jenkins (cjenkins) Lecture 5 Control Architectures Slide 2 Administrivia

More information

Course Syllabus. Operating Systems, Spring 2016, Meni Adler, Danny Hendler and Amnon Meisels 1 3/14/2016

Course Syllabus. Operating Systems, Spring 2016, Meni Adler, Danny Hendler and Amnon Meisels 1 3/14/2016 Course Syllabus. Introduction - History; Views; Concepts; Structure 2. Process Management - Processes; State + Resources; Threads; Unix implementation of Processes 3. Scheduling Paradigms; Unix; Modeling

More information

PART - A 1. Define Artificial Intelligence formulated by Haugeland. The exciting new effort to make computers think machines with minds in the full and literal sense. 2. Define Artificial Intelligence

More information

Agents and Environments

Agents and Environments Agents and Environments Berlin Chen 2004 Reference: 1. S. Russell and P. Norvig. Artificial Intelligence: A Modern Approach. Chapter 2 AI 2004 Berlin Chen 1 What is an Agent An agent interacts with its

More information

Evolutionary Programming

Evolutionary Programming Evolutionary Programming Searching Problem Spaces William Power April 24, 2016 1 Evolutionary Programming Can we solve problems by mi:micing the evolutionary process? Evolutionary programming is a methodology

More information

Knowledge Based Systems

Knowledge Based Systems Knowledge Based Systems Human Expert A human expert is a specialist for a specific differentiated application field who creates solutions to customer problems in this respective field and supports them

More information

CS 4365: Artificial Intelligence Recap. Vibhav Gogate

CS 4365: Artificial Intelligence Recap. Vibhav Gogate CS 4365: Artificial Intelligence Recap Vibhav Gogate Exam Topics Search BFS, DFS, UCS, A* (tree and graph) Completeness and Optimality Heuristics: admissibility and consistency CSPs Constraint graphs,

More information

Programming with Goals (3)

Programming with Goals (3) Programming with Goals (3) M. Birna van Riemsdijk, TU Delft, The Netherlands GOAL slides adapted from MAS course slides by Hindriks 4/24/11 Delft University of Technology Challenge the future Outline GOAL:

More information

Bayesian Networks Representation. Machine Learning 10701/15781 Carlos Guestrin Carnegie Mellon University

Bayesian Networks Representation. Machine Learning 10701/15781 Carlos Guestrin Carnegie Mellon University Bayesian Networks Representation Machine Learning 10701/15781 Carlos Guestrin Carnegie Mellon University March 16 th, 2005 Handwriting recognition Character recognition, e.g., kernel SVMs r r r a c r rr

More information

Human Activities: Handling Uncertainties Using Fuzzy Time Intervals

Human Activities: Handling Uncertainties Using Fuzzy Time Intervals The 19th International Conference on Pattern Recognition (ICPR), Tampa, FL, 2009 Human Activities: Handling Uncertainties Using Fuzzy Time Intervals M. S. Ryoo 1,2 and J. K. Aggarwal 1 1 Computer & Vision

More information

An Evolutionary Approach to Tower of Hanoi Problem

An Evolutionary Approach to Tower of Hanoi Problem An Evolutionary Approach to Tower of Hanoi Problem Jie Li, Rui Shen Department of Computer Engineering Shandong Aerospace Electro-Technology Institute Yantai, China jie_yi_ehw@163.com Abstract:The Tower

More information

Dynamic Rule-based Agent

Dynamic Rule-based Agent International Journal of Engineering Research and Technology. ISSN 0974-3154 Volume 11, Number 4 (2018), pp. 605-613 International Research Publication House http://www.irphouse.com Dynamic Rule-based

More information

Artificial Intelligence Programming Probability

Artificial Intelligence Programming Probability Artificial Intelligence Programming Probability Chris Brooks Department of Computer Science University of San Francisco Department of Computer Science University of San Francisco p.1/25 17-0: Uncertainty

More information

Intelligent Agents. CmpE 540 Principles of Artificial Intelligence

Intelligent Agents. CmpE 540 Principles of Artificial Intelligence CmpE 540 Principles of Artificial Intelligence Intelligent Agents Pınar Yolum pinar.yolum@boun.edu.tr Department of Computer Engineering Boğaziçi University 1 Chapter 2 (Based mostly on the course slides

More information

Using Bayesian Networks to Analyze Expression Data. Xu Siwei, s Muhammad Ali Faisal, s Tejal Joshi, s

Using Bayesian Networks to Analyze Expression Data. Xu Siwei, s Muhammad Ali Faisal, s Tejal Joshi, s Using Bayesian Networks to Analyze Expression Data Xu Siwei, s0789023 Muhammad Ali Faisal, s0677834 Tejal Joshi, s0677858 Outline Introduction Bayesian Networks Equivalence Classes Applying to Expression

More information

INFERENCING STRATEGIES

INFERENCING STRATEGIES INFERENCING STRATEGIES Table of Content Introduction Categories of Reasoning Inference Techniques Control Strategies Comparative Summary of Backward and Forward Chaining Conflict Resolution Goal Agenda

More information

STIN2103. Knowledge. engineering expert systems. Wan Hussain Wan Ishak. SOC 2079 Ext.: Url:

STIN2103. Knowledge. engineering expert systems. Wan Hussain Wan Ishak. SOC 2079 Ext.: Url: & Knowledge STIN2103 engineering expert systems Wan Hussain Wan Ishak SOC 2079 Ext.: 4786 Email: hussain@uum.edu.my Url: http://www.wanhussain.com Outline Knowledge Representation Types of knowledge Knowledge

More information

Introduction to Artificial Intelligence 2 nd semester 2016/2017. Chapter 2: Intelligent Agents

Introduction to Artificial Intelligence 2 nd semester 2016/2017. Chapter 2: Intelligent Agents Introduction to Artificial Intelligence 2 nd semester 2016/2017 Chapter 2: Intelligent Agents Mohamed B. Abubaker Palestine Technical College Deir El-Balah 1 Agents and Environments An agent is anything

More information

ERA: Architectures for Inference

ERA: Architectures for Inference ERA: Architectures for Inference Dan Hammerstrom Electrical And Computer Engineering 7/28/09 1 Intelligent Computing In spite of the transistor bounty of Moore s law, there is a large class of problems

More information

Improving the Accuracy of Neuro-Symbolic Rules with Case-Based Reasoning

Improving the Accuracy of Neuro-Symbolic Rules with Case-Based Reasoning Improving the Accuracy of Neuro-Symbolic Rules with Case-Based Reasoning Jim Prentzas 1, Ioannis Hatzilygeroudis 2 and Othon Michail 2 Abstract. In this paper, we present an improved approach integrating

More information

CS 771 Artificial Intelligence. Intelligent Agents

CS 771 Artificial Intelligence. Intelligent Agents CS 771 Artificial Intelligence Intelligent Agents What is AI? Views of AI fall into four categories 1. Thinking humanly 2. Acting humanly 3. Thinking rationally 4. Acting rationally Acting/Thinking Humanly/Rationally

More information

Robust Execution of Contingent, Temporally Flexible Plans

Robust Execution of Contingent, Temporally Flexible Plans Robust xecution of Contingent, Temporally Flexible Plans tephen. Block, ndreas F. Wehowsky and Brian C. Williams Computer cience and rtificial Intelligence Laboratory, Massachusetts Institute of Technology,

More information

Living with Newton's Laws

Living with Newton's Laws Task #1 - Newton s 1 st Law - This is a pain in the neck Let's suppose you are in your car, waiting at a stop light. Like any good driver, you have your seat belt buckled. (It's the law.) Suddenly, a car

More information

Seminar Thesis: Efficient Planning under Uncertainty with Macro-actions

Seminar Thesis: Efficient Planning under Uncertainty with Macro-actions Seminar Thesis: Efficient Planning under Uncertainty with Macro-actions Ragnar Mogk Department of Computer Science Technische Universität Darmstadt ragnar.mogk@stud.tu-darmstadt.de 1 Introduction This

More information

Inferencing in Artificial Intelligence and Computational Linguistics

Inferencing in Artificial Intelligence and Computational Linguistics Inferencing in Artificial Intelligence and Computational Linguistics (http://www.dfki.de/~horacek/infer-ai-cl.html) no classes on 28.5., 18.6., 25.6. 2-3 extra lectures will be scheduled Helmut Horacek

More information

Solving problems by searching

Solving problems by searching Solving problems by searching Chapter 3 14 Jan 2004 CS 3243 - Blind Search 1 Outline Problem-solving agents Problem types Problem formulation Example problems Basic search algorithms 14 Jan 2004 CS 3243

More information

Princess Nora University Faculty of Computer & Information Systems ARTIFICIAL INTELLIGENCE (CS 370D) Computer Science Department

Princess Nora University Faculty of Computer & Information Systems ARTIFICIAL INTELLIGENCE (CS 370D) Computer Science Department Princess Nora University Faculty of Computer & Information Systems 1 ARTIFICIAL INTELLIGENCE (CS 370D) Computer Science Department (CHAPTER-3) INTELLIGENT AGENTS (Course coordinator) CHAPTER OUTLINE What

More information

Policy Gradients. CS : Deep Reinforcement Learning Sergey Levine

Policy Gradients. CS : Deep Reinforcement Learning Sergey Levine Policy Gradients CS 294-112: Deep Reinforcement Learning Sergey Levine Class Notes 1. Homework 1 milestone due today (11:59 pm)! Don t be late! 2. Remember to start forming final project groups Today s

More information

Intro to Theory of Computation

Intro to Theory of Computation Intro to Theory of Computation LECTURE 13 Last time: Turing Machines and Variants Today Turing Machine Variants Church-Turing Thesis Sofya Raskhodnikova 2/23/2016 Sofya Raskhodnikova; based on slides by

More information

Chapter 2: Intelligent Agents

Chapter 2: Intelligent Agents Chapter 2: Intelligent Agents Outline Last class, introduced AI and rational agent Today s class, focus on intelligent agents Agent and environments Nature of environments influences agent design Basic

More information

Lecture 2: Foundations of Concept Learning

Lecture 2: Foundations of Concept Learning Lecture 2: Foundations of Concept Learning Cognitive Systems - Machine Learning Part I: Basic Approaches to Concept Learning Version Space, Candidate Elimination, Inductive Bias last change October 18,

More information

Intro, Graph and Search

Intro, Graph and Search GAI Questions Intro 1. How would you describe AI (generally), to not us? 2. Game AI is really about The I of I. Which is what? Supporting the P E which is all about making the game more enjoyable Doing

More information

Intelligent Agents. Philipp Koehn. 16 February 2017

Intelligent Agents. Philipp Koehn. 16 February 2017 Intelligent Agents Philipp Koehn 16 February 2017 Agents and Environments 1 Agents include humans, robots, softbots, thermostats, etc. The agent function maps from percept histories to actions: f : P A

More information

Progress in Risk Science and Causality

Progress in Risk Science and Causality Progress in Risk Science and Causality Tony Cox, tcoxdenver@aol.com AAPCA March 27, 2017 1 Vision for causal analytics Represent understanding of how the world works by an explicit causal model. Learn,

More information

High-level Vision. Bernd Neumann Slides for the course in WS 2004/05. Faculty of Informatics Hamburg University Germany

High-level Vision. Bernd Neumann Slides for the course in WS 2004/05. Faculty of Informatics Hamburg University Germany High-level Vision Bernd Neumann Slides for the course in WS 2004/05 Faculty of Informatics Hamburg University Germany neumann@informatik.uni-hamburg.de http://kogs-www.informatik.uni-hamburg.de 1 Contents

More information

Artificial Intelligence

Artificial Intelligence Artificial Intelligence Intelligent Agents Chapter 2 & 27 What is an Agent? An intelligent agent perceives its environment with sensors and acts upon that environment through actuators 2 Examples of Agents

More information

How do you design an intelligent agent?

How do you design an intelligent agent? Intelligent Agents How do you design an intelligent agent? Definition: An intelligent agent perceives its environment via sensors and acts rationally upon that environment with its effectors. A discrete

More information

Bayesian (Belief) Network Models,

Bayesian (Belief) Network Models, Bayesian (Belief) Network Models, 2/10/03 & 2/12/03 Outline of This Lecture 1. Overview of the model 2. Bayes Probability and Rules of Inference Conditional Probabilities Priors and posteriors Joint distributions

More information

Incorporating Action into Diagnostic Problem Solving (An Abridged Report)

Incorporating Action into Diagnostic Problem Solving (An Abridged Report) Incorporating Action into Diagnostic Problem Solving (An Abridged Report) Sheila A. McIlraith Department of Computer Science, University of Toronto Toronto, ON M5S 1A4 Canada e-mail: mcilrait@cs.toronto.edu

More information

Chapter 3 Software Packages to Install How to Set Up Python Eclipse How to Set Up Eclipse... 42

Chapter 3 Software Packages to Install How to Set Up Python Eclipse How to Set Up Eclipse... 42 Table of Contents Preface..... 21 About the Authors... 23 Acknowledgments... 24 How This Book is Organized... 24 Who Should Buy This Book?... 24 Where to Find Answers to Review Questions and Exercises...

More information

Lecture 13: Finding optimal treatment policies

Lecture 13: Finding optimal treatment policies MACHINE LEARNING FOR HEALTHCARE 6.S897, HST.S53 Lecture 13: Finding optimal treatment policies Prof. David Sontag MIT EECS, CSAIL, IMES (Thanks to Peter Bodik for slides on reinforcement learning) Outline

More information

The Semantics of Intention Maintenance for Rational Agents

The Semantics of Intention Maintenance for Rational Agents The Semantics of Intention Maintenance for Rational Agents Michael P. Georgeffand Anand S. Rao Australian Artificial Intelligence Institute Level 6, 171 La Trobe Street, Melbourne Victoria 3000, Australia

More information

Resource Access Protocols. LS 12, TU Dortmund

Resource Access Protocols. LS 12, TU Dortmund Resource Access Protocols Prof. Dr. Jian-Jia Chen LS 12, TU Dortmund 30, Oct. 2018 Prof. Dr. Jian-Jia Chen (LS 12, TU Dortmund) 1 / 27 Why do We Have to Worry about Resource Sharing? Shared Resources:

More information

You can use this app to build a causal Bayesian network and experiment with inferences. We hope you ll find it interesting and helpful.

You can use this app to build a causal Bayesian network and experiment with inferences. We hope you ll find it interesting and helpful. icausalbayes USER MANUAL INTRODUCTION You can use this app to build a causal Bayesian network and experiment with inferences. We hope you ll find it interesting and helpful. We expect most of our users

More information

CS324-Artificial Intelligence

CS324-Artificial Intelligence CS324-Artificial Intelligence Lecture 3: Intelligent Agents Waheed Noor Computer Science and Information Technology, University of Balochistan, Quetta, Pakistan Waheed Noor (CS&IT, UoB, Quetta) CS324-Artificial

More information

CPSC 121 Some Sample Questions for the Final Exam

CPSC 121 Some Sample Questions for the Final Exam CPSC 121 Some Sample Questions for the Final Exam [0] 1. Tautologies and Contradictions: Determine whether the following statements are tautologies (definitely true), contradictions (definitely false),

More information

CHAPTER 4 CONTENT LECTURE 1 November :28 AM

CHAPTER 4 CONTENT LECTURE 1 November :28 AM CHAPTER 4 By Radu Muresan University of Guelph Page 1 CHAPTER 4 CONTENT LECTURE 1 November 07 12 10:28 AM UNIPROCESSOR SCHEDULING Real Time Task Model Concepts Types of Real Time Tasks and Their Characteristics

More information

Unmanned autonomous vehicles in air land and sea

Unmanned autonomous vehicles in air land and sea based on Gianni A. Di Caro lecture on ROBOT CONTROL RCHITECTURES SINGLE AND MULTI-ROBOT SYSTEMS: A CASE STUDY IN SWARM ROBOTICS Unmanned autonomous vehicles in air land and sea Robots and Unmanned Vehicles

More information

Foundations of AI. 10. Knowledge Representation: Modeling with Logic. Concepts, Actions, Time, & All the Rest

Foundations of AI. 10. Knowledge Representation: Modeling with Logic. Concepts, Actions, Time, & All the Rest Foundations of AI 10. Knowledge Representation: Modeling with Logic Concepts, Actions, Time, & All the Rest Wolfram Burgard, Andreas Karwath, Bernhard Nebel, and Martin Riedmiller 10/1 Contents Knowledge

More information

Learning to Use Episodic Memory

Learning to Use Episodic Memory Learning to Use Episodic Memory Nicholas A. Gorski (ngorski@umich.edu) John E. Laird (laird@umich.edu) Computer Science & Engineering, University of Michigan 2260 Hayward St., Ann Arbor, MI 48109 USA Abstract

More information

Statistical analysis DIANA SAPLACAN 2017 * SLIDES ADAPTED BASED ON LECTURE NOTES BY ALMA LEORA CULEN

Statistical analysis DIANA SAPLACAN 2017 * SLIDES ADAPTED BASED ON LECTURE NOTES BY ALMA LEORA CULEN Statistical analysis DIANA SAPLACAN 2017 * SLIDES ADAPTED BASED ON LECTURE NOTES BY ALMA LEORA CULEN Vs. 2 Background 3 There are different types of research methods to study behaviour: Descriptive: observations,

More information

Rational Agents (Ch. 2)

Rational Agents (Ch. 2) Rational Agents (Ch. 2) Extra credit! Occasionally we will have in-class activities for extra credit (+3%) You do not need to have a full or correct answer to get credit, but you do need to attempt the

More information

Operant matching. Sebastian Seung 9.29 Lecture 6: February 24, 2004

Operant matching. Sebastian Seung 9.29 Lecture 6: February 24, 2004 MIT Department of Brain and Cognitive Sciences 9.29J, Spring 2004 - Introduction to Computational Neuroscience Instructor: Professor Sebastian Seung Operant matching Sebastian Seung 9.29 Lecture 6: February

More information

Applying STPA to the Artificial Pancreas for People with Type 1 Diabetes

Applying STPA to the Artificial Pancreas for People with Type 1 Diabetes Applying STPA to the Artificial Pancreas for People with Type 1 Diabetes Lane Desborough Product Strategist Medtronic Diabetes Northridge, California Type 1 Diabetes Artificial Pancreas Challenges Applying

More information

Vorlesung Grundlagen der Künstlichen Intelligenz

Vorlesung Grundlagen der Künstlichen Intelligenz Vorlesung Grundlagen der Künstlichen Intelligenz Reinhard Lafrenz / Prof. A. Knoll Robotics and Embedded Systems Department of Informatics I6 Technische Universität München www6.in.tum.de lafrenz@in.tum.de

More information

Overview. cis32-spring2003-parsons-lect15 2

Overview. cis32-spring2003-parsons-lect15 2 EXPERT SYSTEMS Overview The last lecture looked at rules as a technique for knowledge representation. In a while we will go on to look at logic as a means of knowledge representation. First, however, we

More information

A Scoring Policy for Simulated Soccer Agents Using Reinforcement Learning

A Scoring Policy for Simulated Soccer Agents Using Reinforcement Learning A Scoring Policy for Simulated Soccer Agents Using Reinforcement Learning Azam Rabiee Computer Science and Engineering Isfahan University, Isfahan, Iran azamrabiei@yahoo.com Nasser Ghasem-Aghaee Computer

More information

Agents and State Spaces. CSCI 446: Artificial Intelligence

Agents and State Spaces. CSCI 446: Artificial Intelligence Agents and State Spaces CSCI 446: Artificial Intelligence Overview Agents and environments Rationality Agent types Specifying the task environment Performance measure Environment Actuators Sensors Search

More information

Overview EXPERT SYSTEMS. What is an expert system?

Overview EXPERT SYSTEMS. What is an expert system? EXPERT SYSTEMS Overview The last lecture looked at rules as a technique for knowledge representation. In a while we will go on to look at logic as a means of knowledge representation. First, however, we

More information

Foundations of Natural Language Processing Lecture 13 Heads, Dependency parsing

Foundations of Natural Language Processing Lecture 13 Heads, Dependency parsing Foundations of Natural Language Processing Lecture 13 Heads, Dependency parsing Alex Lascarides (slides from Alex Lascarides, Henry Thompson, Nathan Schneider and Sharon Goldwater) 6 March 2018 Alex Lascarides

More information

CS343: Artificial Intelligence

CS343: Artificial Intelligence CS343: Artificial Intelligence Introduction: Part 2 Prof. Scott Niekum University of Texas at Austin [Based on slides created by Dan Klein and Pieter Abbeel for CS188 Intro to AI at UC Berkeley. All materials

More information

An Intent-Driven Planner for Multi-Agent Story Generation

An Intent-Driven Planner for Multi-Agent Story Generation An Intent-Driven Planner for Multi-Agent Story Generation Mark Owen Riedl Liquid Narrative Group Department of Computer Science North Carolina State University Raleigh, NC 27695 moriedl@ncsu.edu Abstract

More information

Ways of counting... Review of Basic Counting Principles. Spring Ways of counting... Spring / 11

Ways of counting... Review of Basic Counting Principles. Spring Ways of counting... Spring / 11 Ways of counting... Review of Basic Counting Principles Spring 2013 Ways of counting... Spring 2013 1 / 11 1 Introduction 2 Basic counting principles a first view 3 Counting with functions 4 Using functions

More information

Agent-Based Systems. Agent-Based Systems. Michael Rovatsos. Lecture 5 Reactive and Hybrid Agent Architectures 1 / 19

Agent-Based Systems. Agent-Based Systems. Michael Rovatsos. Lecture 5 Reactive and Hybrid Agent Architectures 1 / 19 Agent-Based Systems Michael Rovatsos mrovatso@inf.ed.ac.uk Lecture 5 Reactive and Hybrid Agent Architectures 1 / 19 Where are we? Last time... Practical reasoning agents The BDI architecture Intentions

More information

Symbolic CTL Model Checking

Symbolic CTL Model Checking Symbolic CTL Model Checking Crystal Chang Din Precise Modeling and Analysis group (PMA), University of Oslo INF9140 - Specification and Verification of Parallel Systems Crystal Chang Din @ UiO Symbolic

More information

Behavior Architectures

Behavior Architectures Behavior Architectures 5 min reflection You ve read about two very different behavior architectures. What are the most significant functional/design differences between the two approaches? Are they compatible

More information

(2) In each graph above, calculate the velocity in feet per second that is represented.

(2) In each graph above, calculate the velocity in feet per second that is represented. Calculus Week 1 CHALLENGE 1-Velocity Exercise 1: Examine the two graphs below and answer the questions. Suppose each graph represents the position of Abby, our art teacher. (1) For both graphs above, come

More information

Re: ENSC 370 Project Gerbil Functional Specifications

Re: ENSC 370 Project Gerbil Functional Specifications Simon Fraser University Burnaby, BC V5A 1S6 trac-tech@sfu.ca February, 16, 1999 Dr. Andrew Rawicz School of Engineering Science Simon Fraser University Burnaby, BC V5A 1S6 Re: ENSC 370 Project Gerbil Functional

More information

EEL-5840 Elements of {Artificial} Machine Intelligence

EEL-5840 Elements of {Artificial} Machine Intelligence Menu Introduction Syllabus Grading: Last 2 Yrs Class Average 3.55; {3.7 Fall 2012 w/24 students & 3.45 Fall 2013} General Comments Copyright Dr. A. Antonio Arroyo Page 2 vs. Artificial Intelligence? DEF:

More information

Policy Gradients. CS : Deep Reinforcement Learning Sergey Levine

Policy Gradients. CS : Deep Reinforcement Learning Sergey Levine Policy Gradients CS 294-112: Deep Reinforcement Learning Sergey Levine Class Notes 1. Homework 1 due today (11:59 pm)! Don t be late! 2. Remember to start forming final project groups Today s Lecture 1.

More information

AUTOMATED DEBUGGING. Session Leader: Zongxu Mu

AUTOMATED DEBUGGING. Session Leader: Zongxu Mu AUTOMATED DEBUGGING Session Leader: Zongxu Mu THE FIRST COMPUTER BUG Source: https://plus.google.com/107048744275438760860/posts/twfceqadsld 10/22/2013 Automated Debugging - Zongxu Mu 2 A RANDOM COMPUTER

More information

(c) KSIS Politechnika Poznanska

(c) KSIS Politechnika Poznanska Fundamentals of Autonomous Systems Control architectures in robotics Dariusz Pazderski 1 1 Katedra Sterowania i In»ynierii Systemów, Politechnika Pozna«ska 9th March 2016 Introduction Robotic paradigms

More information

n Outline final paper, add to outline as research progresses n Update literature review periodically (check citeseer)

n Outline final paper, add to outline as research progresses n Update literature review periodically (check citeseer) Project Dilemmas How do I know when I m done? How do I know what I ve accomplished? clearly define focus/goal from beginning design a search method that handles plateaus improve some ML method s robustness

More information

A Guide to Algorithm Design: Paradigms, Methods, and Complexity Analysis

A Guide to Algorithm Design: Paradigms, Methods, and Complexity Analysis A Guide to Algorithm Design: Paradigms, Methods, and Complexity Analysis Anne Benoit, Yves Robert, Frédéric Vivien To cite this version: Anne Benoit, Yves Robert, Frédéric Vivien. A Guide to Algorithm

More information

Artificial Intelligence. Intelligent Agents

Artificial Intelligence. Intelligent Agents Artificial Intelligence Intelligent Agents Agent Agent is anything that perceives its environment through sensors and acts upon that environment through effectors. Another definition later (Minsky) Humans

More information

Artificial Intelligence CS 6364

Artificial Intelligence CS 6364 Artificial Intelligence CS 6364 Professor Dan Moldovan Section 2 Intelligent Agents Intelligent Agents An agent is a thing (e.g. program, or system) that can be viewed as perceiving its environment and

More information

Expert Systems. Artificial Intelligence. Lecture 4 Karim Bouzoubaa

Expert Systems. Artificial Intelligence. Lecture 4 Karim Bouzoubaa Expert Systems Artificial Intelligence Lecture 4 Karim Bouzoubaa Artificial Intelligence Copyright Karim Bouzoubaa 2 Introduction ES: Capture, represent, store and apply human K using a machine Practical

More information

Sequential Decision Making

Sequential Decision Making Sequential Decision Making Sequential decisions Many (most) real world problems cannot be solved with a single action. Need a longer horizon Ex: Sequential decision problems We start at START and want

More information

Robot Learning Letter of Intent

Robot Learning Letter of Intent Research Proposal: Robot Learning Letter of Intent BY ERIK BILLING billing@cs.umu.se 2006-04-11 SUMMARY The proposed project s aim is to further develop the learning aspects in Behavior Based Control (BBC)

More information

PLANNING HEURISTICS FOR CONJUNCTIVE AND DISJUNCTIVE GOALS. A Preliminary Report. Submitted to the Faculty. Purdue University.

PLANNING HEURISTICS FOR CONJUNCTIVE AND DISJUNCTIVE GOALS. A Preliminary Report. Submitted to the Faculty. Purdue University. PLANNING HEURISTICS FOR CONJUNCTIVE AND DISJUNCTIVE GOALS A Preliminary Report Submitted to the Faculty of Purdue University by Lin Zhu In Partial Fulfillment of the Requirements for the Degree of Doctor

More information

HRM Triax 100 is engineered to provide quick, one-touch access to a variety of critical information during training and in competition.

HRM Triax 100 is engineered to provide quick, one-touch access to a variety of critical information during training and in competition. 1 Information is power. HRM Triax 100 is engineered to provide quick, one-touch access to a variety of critical information during training and in competition. Used properly this heart rate monitor will

More information

Overview. What is an agent?

Overview. What is an agent? Artificial Intelligence Programming s and s Chris Brooks Overview What makes an agent? Defining an environment Overview What makes an agent? Defining an environment Department of Computer Science University

More information

2 Psychological Processes : An Introduction

2 Psychological Processes : An Introduction 2 Psychological Processes : An Introduction 2.1 Introduction In our everyday life we try to achieve various goals through different activities, receive information from our environment, learn about many

More information

Stepwise Knowledge Acquisition in a Fuzzy Knowledge Representation Framework

Stepwise Knowledge Acquisition in a Fuzzy Knowledge Representation Framework Stepwise Knowledge Acquisition in a Fuzzy Knowledge Representation Framework Thomas E. Rothenfluh 1, Karl Bögl 2, and Klaus-Peter Adlassnig 2 1 Department of Psychology University of Zurich, Zürichbergstraße

More information

Automated Conflict Detection Between Medical Care Pathways

Automated Conflict Detection Between Medical Care Pathways Automated Conflict Detection Between Medical Care Pathways Philip Weber, Bosco Filho, Mark Lee, Ian Litchfield, Ruth Backman University of Birmingham, UK School of Computer Science Institute of Applied

More information

Unifying Data-Directed and Goal-Directed Control: An Example and Experiments

Unifying Data-Directed and Goal-Directed Control: An Example and Experiments Unifying Data-Directed and Goal-Directed Control: An Example and Experiments Daniel D. Corkill, Victor R. Lesser, and Eva Hudlická Department of Computer and Information Science University of Massachusetts

More information

Reactivity and Deliberation in Decision-Making Systems

Reactivity and Deliberation in Decision-Making Systems 11 Reactivity and Deliberation in Decision-Making Systems Carle Côté 11.1 Introduction 11.2 Let s Begin at the Beginning 11.3 Common Pitfall #1 : One Decision Model to Rule Them All! 11.4 Common Pitfall

More information

Harvard-MIT Division of Health Sciences and Technology HST.952: Computing for Biomedical Scientists. Data and Knowledge Representation Lecture 6

Harvard-MIT Division of Health Sciences and Technology HST.952: Computing for Biomedical Scientists. Data and Knowledge Representation Lecture 6 Harvard-MIT Division of Health Sciences and Technology HST.952: Computing for Biomedical Scientists Data and Knowledge Representation Lecture 6 Last Time We Talked About Medical Coding Systems UMLS Today

More information

More Examples and Applications on AVL Tree

More Examples and Applications on AVL Tree CSCI2100 Tutorial 11 Jianwen Zhao Department of Computer Science and Engineering The Chinese University of Hong Kong Adapted from the slides of the previous offerings of the course Recall in lectures we

More information

HRM Triax 15 is engineered to provide quick, one-touch access to a variety of critical information during training and in competition.

HRM Triax 15 is engineered to provide quick, one-touch access to a variety of critical information during training and in competition. 1 Information is power. HRM Triax 15 is engineered to provide quick, one-touch access to a variety of critical information during training and in competition. Used properly this heart rate monitor will

More information

A Reinforcement Learning Approach Involving a Shortest Path Finding Algorithm

A Reinforcement Learning Approach Involving a Shortest Path Finding Algorithm Proceedings of the 003 IEEE/RSJ Intl. Conference on Intelligent Robots and Systems Proceedings Las Vegas, of Nevada the 003 October IEEE/RSJ 003 Intl. Conference on Intelligent Robots and Systems Las Vegas,

More information

Today we will... Foundations of Natural Language Processing Lecture 13 Heads, Dependency parsing. Evaluating parse accuracy. Evaluating parse accuracy

Today we will... Foundations of Natural Language Processing Lecture 13 Heads, Dependency parsing. Evaluating parse accuracy. Evaluating parse accuracy Today we will... Foundations of Natural Language Processing Lecture 13 Heads, Dependency parsing Alex Lascarides (slides from Alex Lascarides, Henry Thompson, Nathan chneider and haron Goldwater) 6 March

More information