Asymptotically-Optimal Motion Planning using Lower Bounds on Cost Oren Salzman and Dan Halperin∗ Abstract— Many path-finding algorithms on graphs such as A* are sped up by using a heuristic function that gives lower bounds on the cost to reach the goal. Aiming to apply similar techniques to speed up sampling-based motion- planning algorithms, we use effective lower bounds on the cost between configurations to tightly estimate the cost-to-go. We then use these estimates in an anytime asymptotically-optimal algorithm which we call Motion Planning using Lower Bounds (MPLB). MPLB is based on the Fast Marching Trees (FMT*) algorithm [1] recently presented by Janson and Pavone. An advantage of our approach is that in many cases (especially as the number of samples grows) the weight of collision detection in the computation is almost negligible with respect to nearest-neighbor calls. We prove that MPLB performs no more collision-detection calls than an anytime version of FMT*. Additionally, we demonstrate in simulations that for certain scenarios, the algorithmic tools presented here enable efficiently producing low-cost paths while spending only a small fraction of the running time on collision detection. I. INTRODUCTION Motion-planning algorithms aim to find a collision-free path for a robot moving amidst obstacles. The most prevalent approach in practice is to use sampling-based techniques [2]. These algorithms sample points in the robot’s configuration- space (C-space) and connect close-by configurations to con- struct a graph called a roadmap. Often, a low-cost path is desired, where cost can be measured in terms of, for example, path length or energy consumption along the path. Sampling-based algorithms rely on two central primitive operations: Collision Detection (CD) and Nearest Neighbors (NN) search. CD determines whether a configuration is collision-free or not and is often used to assess if a path connecting close-by configurations is collision-free. This latter operation is referred to as Local Planning (LP). An NN data structure preprocesses a set of points to efficiently answer queries such as “which are the points within ra- dius r of a given query point?” In practice, the cost of CD, primarily due to LP calls, often dominates the running time of sampling-based algorithms, and is typically regarded as the computational bottleneck for such algorithms. For a summary of the computational complexity of NN, CD and LP in sampling-based motion-planning algorithms see [3]. In their influential work, Karaman and Frazzoli [4] ana- lyzed existing sampling-based algorithms (namely PRM [5] and RRT [6]) and introduced the notion of asymptotic optimality (AO); an algorithm is said to be AO if the cost ∗Blavatnik School of Computer Science, Tel-Aviv University, Israel This work has been supported in part by the Israel Science Foundation (grant no. 1102/11), by the German-Israeli Foundation (grant no. 1150- 82.6/2011), and by the Hermann Minkowski–Minerva Center for Geometry at Tel Aviv University. of the solution produced by it converges to the cost of the optimal solution if the algorithm is run for sufficiently long time. They proposed AO variants of PRM and RRT called PRM* and RRT*, respectively. However, the AO of PRM* and RRT* comes at the cost of increased running time and memory consumption when compared to their non-optimal counterparts. To reduce this cost, several improvements were proposed which modify the sampling scheme [7], [8], the CD [3], or relax the optimality to asymptotic near-optimality (ANO) [9], [10], [11]. An algorithm is said to be ANO if, given an approximation factor ε, the cost of a solution returned by the algorithm is guaranteed to converge to within a factor of 1 + ε of the cost of the optimal solution. Following the introduction of PRM* and RRT*, other AO algorithms have been suggested. RRT# [12], extends its roadmap in a similar fashion to RRT*. However, in contrast to RRT* which only performs local rewiring, RRT# efficiently propagates changes to all the relevant parts of the roadmap. Another AO algorithm, proposed by Janson and Pavone, is the Fast Marching Trees (FMT*) [1] algorithm. FMT*, reviewed in detail in Section II, was shown to converge to an optimal solution faster than PRM* or RRT*. Contribution and paper organization We show how by looking at the roadmap induced by a set of samples, we can compute effective lower bounds on the cost-to-go of nodes. This is done without performing expensive LP calls and allows to efficiently guide the search performed by the algorithm. We call our scheme Motion Planning using Lower Bounds or MPLB for short. An interesting and useful implication of our approach is that the weight of CD and LP becomes negligible when compared to that of the NN calls1. Bialkowski et al. [3] introduced a technique which replaces CD calls by NN calls. Their scheme relies on additional data produced by the CD algorithm used, namely, a bound on the clearance of a configuration (if it is collision free) or on its penetration depth (if it is not collision free). Alas, such a bound is not trivial to compute for some prevalent C-spaces. In contrast, our algorithmic framework can be paired with existing off- the-shelf NN, CD and LP procedures. Thus, our results, which are more general (as they are applicable to general C-spaces), strengthen the conjecture of Bialkowski et al. [3] that NN computation and not CD may be the bottleneck of sampling-based motion-planning algorithms. This work continues and expands our recent work [11] 1Throughout this paper, when we say an NN call we mean a call to find all the nodes within radius r(n) of a given node. r(n) is a radius depending on the number of samples used and will be formally defined in Section II. arXiv:1403.7714v2 [cs.RO] 30 Sep 2014 where we relaxed the AO of RRT* to ANO using lower bounds. The novel component here is that multiple nodes are processed simultaneously while in RRT* the nodes are processed one at a time. This allows to efficiently compute for all nodes an estimation of the cost-to-go which in turn can be used to speed up motion-planning algorithms. Our framework is demonstrated for the case where dis- tance is the cost function via the FMT* algorithm which is reviewed in Section II. As we wish to work in an anytime setting, we introduce in Section III a straightforward adaptation of FMT* for anytime planning which we call aFMT. We then proceed to present MPLB in Section IV. We analyze aFMT* and MPLB with respect to the amount of calls to the NN and LP procedures in Section V and report on experimental results in Section VI. Specifically, we demonstrate in simulations that for certain scenarios, MPLB produces lower-cost paths faster (by a factor of between two and three) than aFMT*. We conclude with a discussion and suggestions for future work in Section VII. II. TERMINOLOGY AND ALGORITHMIC BACKGROUND We begin by formally stating the motion-planning problem and introducing several procedures used by the algorithms we present. We continue by reviewing the FMT* algorithm. A. Problem definition and terminology Let X, Xfree denote the Euclidean2 C-space and free space, respectively, and d the dimension of the C-space. Let (Xfree, xinit, Xgoal) be the motion-planning problem where: xinit ∈Xfree is the initial free configuration of the robot and Xgoal ⊆ Xfree is the goal region. We will make use of the following procedures: sample free(n), a procedure returning n random configurations from Xfree; nearest neighbors(x, V, r) is a procedure that returns all neighbors of x with distance smaller than r within the set V ; collision free(x, y) tests whether the straight- line segment connecting x and y is contained in Xfree; cost(x, y) returns the cost of the straight-line path con- necting x and y, namely, in our case, the distance. We consider weighted graphs G = (V, E), where the weight of an edge (x, y) ∈ E is cost(x, y). Given such a graph G, we denote by costG(x, y) the cost of the weighted shortest path from x to y. Let cost-to-comeG(x) be costG(xinit, x) and cost-to-goG(x) be the minimal costG(x, xgoal) for xgoal ∈Xgoal. Namely for every node x, cost-to-comeG(x) is the minimal cost to reach x from xinit and cost-to-goG(x) is the minimal cost to reach Xgoal from x. Additionally, let BG(xinit, r), BG(Xgoal, r) be the set of all nodes whose cost-to-come (respectively, cost-to-go) value on G is smaller than r. Finally, we denote by Dijkstra(G, x, c) an implementation of Dijkstra’s al- gorithm3 running on the graph G from x until a maximal 2 Although we describe the algorithm for Euclidean spaces, by standard techniques (see, e.g. [2, Section 3.5, Section 7.1.2] or [13]) the algorithm can be applied to non-Euclidean spaces such as SE3. However, the AO proof of FMT*, presented in [1], is shown only for Euclidean spaces. 3Any other algorithm that computes the shortest path from a single source to all nodes in a graph may be used. Algorithm 1 FMT* (xinit, Xgoal, n) 1: V ←{xinit} ∪sample free(n); E ←∅; T ←(V, E) 2: PATH ←search (T , Xgoal, 0, ∞) // See Alg. 2 3: return PATH cost of c has been reached. The algorithm’s implementation updates the cost to reach each node from x and outputs the set of nodes traversed. Given a set of samples V , and a radius r, we denote by G(V, r) the disk graph,4 which is the graph whose set of vertices is V and two vertices x, y ∈V are connected by an edge if the distance between x and y is less than r. B. Fast Marching Trees (FMT*) FMT*, outlined in Alg. 1, performs a “lazy” dynamic programming recursion on a set of sampled configurations to grow a tree rooted at xinit [14]. The algorithm samples n collision-free nodes V (line 1). It searches for a path to Xgoal by building a minimum-cost spanning tree growing in cost- to-come space (line 2 and detailed in Alg. 2). As we explain in Section IV, the algorithm may benefit from using a heuris- tic function estimating the cost-to-go of a node and a bound on the maximal length of the path that should be found. As these are not part of the original formulation of FMT*, we describe the search procedure of FMT* using a cost-to-go estimation of zero for each node and an unbounded maximal path length (marked in red in Alg. 1 and 2). This will allow us to use the same pseudo-code of Alg. 2 to explain the MPLB algorithm in Section IV. The search-tree is built by maintaining two sets of nodes H, W such that H is the set of nodes added to the tree that may be expanded and W is the set of nodes that have not yet been added to the tree (Alg. 2, line 1). It then computes for each node the set of nearest neighbors5 of radius r(n) (line 3). The algorithm repeats the following process: the node z with the lowest cost-to-come value is chosen from H (line 4 and 16). For each neighbor x of z that is not already in H, the algorithm finds its neighbor y ∈H such that the cost-to-come of y added to the distance between y and x is minimal (lines 7-9). If the local path between y and x is free, x is added to H with y as its parent (lines 10-12). At the end of each iteration z is removed from H (line 13). The algorithm runs until a solution is found or there are no more nodes to process. To ensure AO, the radius r(n) used by the algorithm is r(n) = (1 + η) · 2 1 d  1 d µ(Xfree) ζd  1 d log n n  1 d , (1) where η > 0 is some small constant, µ(·) denotes the d- dimensional Lebesgue measure and ζd is the volume of the unit ball in the d-dimensional Euclidean space. 4The disk graph is sometimes referred to as the the neighborhood graph. 5The nearest-neighbor computation can be delayed and performed only when a node is processed but we present the batched mode of computation to simplify the exposition. Algorithm 2 search (T , Xgoal, cost to go, cmax) 1: W ←V \ {xinit}; H ←{xinit} 2: for all v ∈V do 3: Nv ←nearest neighbors(V \ {v}, v, r(n)) 4: z ←xinit 5: while z /∈XGoal do 6: Hnew ←∅; Xnear ←W ∩Nz 7: for x ∈Xnear do 8: Ynear ←H ∩Nx 9: ymin ←arg miny∈Ynear{costT (y) + dist(y, x)} 10: if collision free(ymin, x) then 11: T .parent(x) ←ymin 12: Hnew ←Hnew ∪{x}; W ←W \ {x} 13: H ←(H ∪Hnew) \ {z} 14: if H = ∅then 15: return FAILURE 16: z ←arg miny∈H{costT (y) + cost to go(y)} 17: if costT (z) + cost to go(z) ≥cmax then 18: return FAILURE 19: return PATH III. ANYTIME FMT* (AFMT*) An algorithm is said to be anytime if it yields meaningful results even after a short time and it improves the quality of the solution as more computation time is available. We outline a straightforward enhancement to FMT* to make it anytime. As noted in previous work (see, e.g., [15]) one can turn a batch algorithm into an anytime one by the following general approach: choose an initial small number of samples n = n0 and apply the algorithm. As long as time permits, double n and repeat the process. The total running time is less than twice that of the running time for the largest n. Note that as FMT* is AO, aFMT* is also AO. We can further speed up this method by reusing both existing samples and connections from previous iterations. Due to lack of space, we omit these details and refer the interested reader to the extended version of our paper [16]. IV. ALGORITHMIC FRAMEWORK We are now ready to present our approach to exploiting lower bounds on cost in order to speed up sampling-based motion-planning algorithms. Given a random infinite sequence of collision-free samples S = s1, s2 . . . denote by Vi(S) the set of the first 2i elements of S. Let Gi(S) = G(Vi(S), r(|Vi(S)|)) and let Hi(S) ⊆Gi(S) be the subgraph containing collision-free edges only (here r(n) is the radius defined in Eq. 1). For brevity, we omit S when referring to Vi(S), Gi(S) and Hi(S). Moreover, when we compare our algorithm to the aFMT* algorithm, we do so for runs on the same random infinite sequence S. Clearly, for any two nodes x, y ∈Vi, costGi(x, y) ≤costHi(x, y). Thus for any node x ∈Vi, cost-to-goGi(x) ≤cost-to-goHi(x). Namely, the cost-to-go computed using the disk graph Gi is a lower bound on the cost-to-go that may be obtained using Hi. We call this the lower bound property. For an illustration, see Fig. 1. Xgoal xinit (a) Xgoal xinit (b) Fig. 1: This figure demonstrates that the part of the tree expanded when searching in cost-to-come space (shaded blue region, Fig. (a)) is larger than the one expanded when searching in cost-to-come+cost- to-go space (shaded green region, Fig. (b)). Obstacles in the C-space are depicted in red, start location and goal region are depicted by a purple circle and a turquoise region, respectively. Edges of the disk graph Gi that are contained and not contained in Hi are depicted in black and dashed red, respectively. The figure is best viewed in color. We present Motion Planning using Lower Bounds, or MPLB (outlined in Alg. 3). Similar to aFMT*, the algorithm runs in iterations and at the i’th iteration, uses Vi as its set of samples. Unlike aFMT*, each iteration consists of a preprocessing phase (line 4) of computing a lower bound on the cost-to-go values and a searching phase (line 5) where a modified version of FMT* is used. Let ci(ALG) denote the cost of the solution obtained by an algorithm ALG using Vi as the set of samples (set c0(ALG) ←∞). We now show that only a subset of the nodes sampled in each iteration need to be considered. We then proceed to describe the two phases of MPLB. A. Promising nodes We use the lower bound property to consider only a subset of Vi that will be used in the i’th iteration. Intuitively, we only wish to consider nodes that may produce a solution that is better than the solution obtained in previous iterations. This leads us to the definition of promising nodes: Definition 1: A node x∈Vi is promising (at iteration i) if cost-to-comeHi(x)+cost-to-goHi(x)