Another strand of work models conformant and conditional planning as
a search in the space of belief states. This started with
[15], who concentrated on formulating a set of
admissible pruning conditions for controlling search. There were no
heuristics for choosing among unpruned nodes. GPT
[6] extended this idea to consider a simple form
of reachability heuristic. Specifically, in computing the estimated
cost of a belief state, GPT assumes that the initial state is fully
observable. The cost estimate itself is done in terms of
reachability (with dynamic programming rather than planning graphs).
GPT's reachability heuristic is similar to our
heuristic because they both estimate the cost of the farthest
(maximum distance) state by looking at a deterministic relaxation of
the problem. In comparison to GPT,
and
can be
seen as using heuristics that do a better job of considering the
cost of the belief state across the various possible worlds.
Another family of planners that search in belief states is the
MBP-family of planners--MBP [3], and KACMBP
[1]. In contrast to
but similar to
, the MBP-family of planners all represent belief states in
terms of binary decision diagrams. Action application is modeled as
modifications to the BDDs. MBP supports both progression and
regression in the space of belief states, while KACMBP is a pure
progression planner. Before computing heuristic estimates, KACMBP
pro-actively reduces the uncertainty in the belief state by
preferring uncertainty reducing actions. The motivation for this
approach is that applying cardinality heuristics to belief states
containing multiple states may not give accurate enough direction to
the search. While reducing the uncertainty seems to be an effective
idea, we note that (a) not all domains may contain actions that
reduce belief state uncertainty and (b) the need for uncertainty
reduction may be reduced when we have heuristics that effectively
reason about the multiple worlds (viz., our multiple planning graph
heuristics). Nevertheless, it could be very fruitful to integrate
knowledge goal ideas of KACMBP and the reachability heuristics of
and
to handle domains that contain both high
uncertainty and costly goals.
In contrast to these domain-independent approaches that only require models of the domain physics, PKSPlan [26] is a forward-chaining knowledge-based planner that requires richer domain knowledge. The planner makes use of several knowledge bases, as opposed to a single knowledge base taking the form of a belief state. The knowledge bases separate binary and multi-valued variables, and planning and execution time knowledge.
YKA [28] is a regression conditional planner using BDDs that uses a cardinality heuristic. Recently Rintanen has also developed related reachability heuristics that consider distances for groups of states, which do not rely on planning graphs [29].
More recently, there has been closely related work on heuristics for
constructing conformant plans within the CFF planner [17].
The planner represents belief states implicitly through a set of
known facts, the action history (leading to the belief state), and
the initial belief state. CFF builds a planning graph forward from
the set of known literals to the goal literals and backwards to the
initial belief state. In the planning graph, conditional effects
are restricted to single literals in their antecedent to enable
tractable 2-cnf reasoning. From this planning graph, CFF extracts a
relaxed plan that represents supporting the goal belief state from
all states in the initial belief state. The biggest differences
between the
and the CFF technique are that the
reasons
only forward from the source belief state (assuming an explicit,
albeit symbolic, belief state), and the
does not restrict the
number of literals in antecedents. As a result, the
does not
lose the causal information nor perform backward reasoning to the
initial belief state.
Our handling of uncertainty through labels and label propagation is reminiscent of and related to de Kleer's assumption based truth maintenance system (ATMS) [14]. Where an ATMS uses labels to identify the assumptions (contexts) where a particular statement holds, a traditional truth maintenance system requires extensive backtracking and consistency enforcement to identify other contexts. Similarly, where we can reason about multiple possible worlds (contexts) with the LUG simultaneously, the MG approach requires, not backtracking, but reproduction of planning graphs for other possible worlds.
Finally,
and
are also related to, and an
adaptation of the work on reachability heuristics for classical
planning, including
[23], FF
[18] and HSP-r [5].
is the conformant extension to
that uses
regression search (similar to HSP-r) guided by planning graph
heuristics.
is similar to FF in that it uses progression
search with planning graph heuristics.