270 likes | 401 Views
Probabilistic Planning. Jim Blythe November 6th. A slide from August 30th: Assumptions (until October..). Atomic time All effects are immediate Deterministic effects Omniscience Sole agent of change Goals of attainment. Assumptions (until October..). Atomic time
E N D
Probabilistic Planning Jim Blythe November 6th
A slide from August 30th:Assumptions (until October..) • Atomic time • All effects are immediate • Deterministic effects • Omniscience • Sole agent of change • Goals of attainment
Assumptions (until October..) • Atomic time • All effects are immediate • Deterministic effects • Omniscience • Sole agent of change • Goals of attainment
? ? Sources of uncertainty When we try to execute plans, uncertainty from several different sources can affect success. Firstly, we might have uncertainty about the state of the world
Sources of uncertainty Actions we take might have uncertain effects even when we know the world state ? ?
Me Me X X Sources of uncertainty External agents might be changing the world while we execute our plan.
Dealing with uncertainty: re-planning • Make a plan assuming nothing bad will happen • Monitor for problems during execution • Build a new plan if a problem is found • Either re-plan to the goal state • Or try to patch the existing plan
Dealing with uncertainty: Conditional planning • Deal with contingencies (bad outcomes) at planning time before they occur • Reactive planning might be viewed as conditional planning where every possible contingency is covered (somehow) in the policy.
Tradeoffs in strategies for uncertainty • My re-planner housemate: “Why are you taking an umbrella? It’s not raining!” • Can’t find plans that require steps taken before the contingency is discovered • My conditional planner housemate: “Why are you leaving the house? Class may be cancelled. It might rain. You might have won the lottery. Was that an earthquake?….” • Impossible to plan for every contingency. Need a representation that captures tradeoffs.
Probabilistic planning lets us explore the middle ground • Different contingencies have different probabilities of occurring. • Plan ahead for likely contingencies that may need steps taken before they occur. • Use probability theory to judge plans that address some contingencies: seek a plan that is above some minimum probability of success.
Some issues to think about • How do we figure out the probability of a plan succeeding? Is it expensive to do? • How do we know what the most likely contingencies are? • Can we distinguish bad outcomes (not holding the cup) from really bad outcomes (broken the cup, spilled the anthrax agent..)?
Q Ac Ap Reminder: POP algorithm POP((A, O, L), agenda, PossibleActions): • If agenda is empty, return (A, O, L) • Pick (Q, An) from agenda • Ad = choose an action that adds Q. • If no such action exists, fail. • Add the link Ad Ac to L and the ordering Ad < Ac to O • If Ad is new, add it to A. • Remove (Q, An) from agenda. If Ad is new, for each of its preconditions P add (P, Ad) to agenda. • For every action At that threatens any link • Choose to add At < Ap or Ac < At to O. • If neither choice is consistent, fail. • POP((A, O, L), agenda, PossibleActions) Q
Buridan (an SNLP-based planner) • An SNLP-based planner might come up with this plan for a deterministic action representation:
Modifications to the UCPOP algorithm • Allow more than one causal link for each condition in the plan. • Confront a threat by decreasing the probability that it will happen. (By adding conditions negating the trigger of the threat). • Terminate when sufficient probability reached (may still have threats).
Computing probability of plan success1: forward projection • Simulate the plan, keep track of possible states and their probabilities, finally sum the probabilities of states that satisfy the goal. • Here, the china is packed in the initial state with probability 0.5 (and is not packed with probability 0.5) What is the worst-case time complexity of this algorithm?
Computing the probability of success 2: Bayes nets Time-stamped literal node Action outcome node What is the worst-case time complexity of this algorithm?
Tradeoffs in computing probability of success • Belief net approach is often faster because it ignores irrelevant differences in the state. • Neither approach is guaranteed to be faster. • Often, the time to compute the probability of success dominates the planning time.
Conditional planning in this framework: CNLP and C-Buridan • Tricky to represent conditional branches in partially-ordered plans. • Actions can produce “observation labels” as well as effects, e.g. “the weather is good”. • After introducing an action with observation labels, the possible values can be used as “context labels” assigned to actions ordered after the observation step.
DRIPS:(Decision-theoretic Refinement Planner) • Considers plan utility, taking into account action costs, benefits of different states. • Searches for a plan with Maximum Expected Utility (MEU), not just above a threshold. • A skeletal planner, makes use of ranges of utility of abstract plans in order to search efficiently. • Prune abstract plans whose utility range is completely below the range of some alternative (dominated plans)
MAXPLAN • Inspired by SATPLAN. Compile planning problem to an instance of E-MAJSAT • E-MAJSAT: given a boolean formula with variables that are either choice variables or chance variables, find an assignment to the choice variables that maximises the probability that the formula is true. • Choice variables: we can control them • e.g. which action to use • Chance variables: we cannot control them • e.g. the weather, the outcome of each action, .. • Then use standard algorithm to compute and maximise probability of success
Thinking about MAXPLAN • As it stands, does MAXPLAN build conditional plans? • How could we make MAXPLAN build conditional plans?
Other approaches that have been used • Graphplan (pointers to Weld and Smith’s work in paper) • Prodigy (more in next class) • HTN planning (Cypress) • Markov decision problems (more in the class after next)
With all approaches, we must consider the same issues • Tractability • Plans can have many possible outcomes • How to reason about when to add sensing • Plan utility • Is probability of success enough? • What measures of cost and benefit can be used tractably? • Can operator costs be summed? What difference do time-based utilities like deadlines make? • Observability and conditional planning • Classical planning is “open-loop” with no sensing • A “policy” assumes we can observer everything • Can we model limited observability, noisy sensors, bias..?