arXiv:1703.00812v1 [cs.RO] 1 Mar 2017 Improper Filter Reduction Fatemeh Zahra Saberifar ∗ Ali Mohades † Mohammadreza Razzazi ‡ Jason M. O’Kane § Abstract Combinatorial filters have been the subject of increasing interest from the robotics community in recent years. This paper considers automatic reduction of combinatorial filters to a given size, even if that reduction necessitates changes to the filter’s behavior. We introduce an algorithmic problem called improper filter reduction , in which the input is a combinatorial filter F along with an integer k representing the target size. The output is another combinatorial filter F ′ with at most k states, such that the difference in behavior between F and F ′ is minimal. We present two metrics for measuring the distance between pairs of filters, describe dynamic programming algorithms for computing these distances, and show that improper filter reduction is NP-hard under these metrics. We then describe two heuristic algorithms for improper filter reduction, one greedy sequential approach, and one randomized global approach based on prior work on weighted improper graph coloring. We have implemented these algorithms and analyze the results of three sets of experiments. Keywords: combinatorial filters, estimation, automata Mathematics subject classifications (2010): 68T40 93C85 68T20 68W01 68R01 ∗ Department of Mathematics and Computer Science, Amirkabir University of Technol- ogy, Tehran, Iran. fz.saberifar@aut.ac.ir † Department of Mathematics and Computer Science, Amirkabir University of Technol- ogy, Tehran, Iran. The corresponding author is A. Mohades. mohades@aut.ac.ir ‡ Department of Computer Engineering and IT, Amirkabir University of Technology, Tehran, Iran. razzazi@aut.ac.ir § Department of Computer Science and Engineering, University of South Carolina, Columbia, South Carolina, USA. jokane@cse.sc.edu 1 1 Introduction This paper builds upon the ongoing line of research on combinatorial filtering for robot tasks [26, 12, 1]. The intuition of that work is to carefully design filters that robots or other autonomous agents can use to retain only the information that is strictly necessary for completing an assigned task. This class of filters is an extremely general tool, applicable for any task that can be characterized by discrete transitions triggered by finite sets of observations. Recent research has considered the problem of automatic reduction of combinatorial filters: Given a combinatorial filter that correctly solves a problem of interest, can we algorithmically find the smallest equivalent fil- ter? Prior work, to which one of the present authors contributed, showed that this problem is NP-hard and described an efficient heuristic algorithm for finding small—but not necessarily the smallest—equivalent filters [7]. However, that prior work left a number of important questions unanswered, including these: Is there a useful notion of the behavior of a reduced filter being “close enough” to the original filter? If so, then given an input filter, can we find the most similar filter of a given fixed size? We refer to this problem as the improper filter reduction problem, by analogy to the existing work in improper graph coloring [5]. This paper addresses that problem, making several new contributions. 1. We introduce a family of metrics for measuring the difference between filters and describe dynamic programming algorithms for computing these metrics. 2. We argue that, for any metric, the improper filter reduction problem is NP-hard. 3. We present two heuristic algorithms for improper filter reduction. These are heuristic algorithms, in the sense that there is no guarantee that their output will fully minimize the distance from the original filter. 4. We present implementations of these algorithms, along with a series of experiments evaluating their performance. In broad terms, these answers are relevant because they provide some insight into one of the fundamental tasks in robot design, namely understanding how a robot should process and retain information collected from its sensors. 2 { 1 } { 2 , 3 } . . . { n − 1 , n } { 2 } . . . { n } { 1 , n } { 1 , 2 } { 1 , . . . , n } 1 , n 1 , n 2 , . . . , n − 1 1 , n − 1 n 2 , n 1 2 , . . . , n − 1 n 1 { 1 , . . . , n } { 1 , n − 1 } { 1 , 2 } { 2 , 3 } . . . { n − 1 , n } { 2 } . . . { n } { 1 } 1 , n 2 , . . . , n − 1 2 , . . . , n − 1 1 , n 2 , . . . , n − 1 1 , n Figure 1: [top left] An agent moves through an annulus divided into n regions by n beam sensors. [top right] The smallest filter, found by a prior algorithm [7], that can accurately track whether the agent is provably within region 1 or not. [bottom] A smaller filter produced, for any specific value of n , by the algorithm introduced in this paper. This reduced filter eliminates the extra states for region sets { 1 , n } and { 1 , 2 } , which are used only at the start. Figure 1 shows a simple example, in which an agent moves through a ring- shaped environment along a continuous but unpredictable path. A collection of n beam sensors throughout that environment can detect the agent passing by, but cannot determine whether that crossing was in the clockwise or counterclockwise direction. One pair of beams delimits a special “goal” region, called region 1. A natural question for this system is to ask what kinds of filters can determine when the agent is within region 1. That is, we are interested in designing filters that produce outputs, informally called colors, that indicate whether the agent is known to be in region 1 or not. In fact, it is possible to form a series of filters for this problem, each smaller than the last. 3 • A straightforward approach would be to from a filter, represented as a graph of 2 n − 1 nodes, each representing a nonempty set of “pos- sible states,” with directed edges indicating how the possible states change with each beam crossing. In that filter, we would choose the state corresponding to { 1 , . . . , n } as the initial state. The filter’s output would be defined as “yes” for the state corresponding to the set { 1 } —indicating that only state 1 is consistent with the history of observations—and “no” for each other state. • That na ̈ ıve filter can be made more compact by noticing that, if n > 3, only 2 n + 1 of those nodes can ever be reached: One initial state in which all regions are possible states; one state for each beam, in which only the two regions on opposite sides of that beam are possible states; and one state for each region, in which the agent is known to be in that region. The remaining 2 n − 2 n − 2 nodes can be discarded without changing the filter’s outputs, because no observation sequence can reach them. • That filter can be reduced even further, again without changing its outputs, by a carefully selected sequence of vertex contractions. Prior work shows that selecting those vertex contractions optimally is, in general, NP-hard [7], but also presents an efficient algorithm that per- forms this reduction well in practice. The resulting filter, which has 5 states regardless of the number of beams, is shown in the top portion of Figure 1. • This paper is concerned with additional reductions beyond this small- est equivalent filter, which requires us to tolerate some “mistakes” in which the filter produces incorrect outputs for some observation se- quences. Thus, the reduction is “improper.” The bottom portion of Figure 1 shows how the algorithm introduced in this paper can reduce the filter to three states, corresponding to “in region 1”(shown on the right), “not in region 1” (shown in the middle), and “maybe” (shown on the left). The filter may produce incorrect outputs for some obser- vation sequences—One such observation sequence is 1 , n, n, 1 , 1 , n, n, . . . — but behaves the same as the original after the first beam not adja- cent to region 1 is crossed. This paper examines the algorithmic problems latent in the final step of this series of reductions. The underlying objective is to understand the tradeoffs between a filter’s size and its ability to produce correct outputs. 4 The remainder of this paper obeys the following structure. Section 2 reviews related work, and Section 3 defines the improper filter reduction problem. Section 4 describes an algorithm for computing the distance be- tween two given filters, which is used as a subroutine in our two primary algorithms for the improper filter reduction problem, which are described in Sections 5 and 6. Section 7 describes our implementations and experimental evaluation of these algorithms. Concluding remarks, including a preview of future work, appear in Section 8. 2 Related Work 2.1 Combinatorial filters Combinatorial filters, as a general class, build upon the generalized infor- mation space formalism popularized by LaValle [25, 26]. The central idea is to perform filtering tasks in very small derived information spaces, thereby minimizing the computational burden of executing the filter, and illumi- nating the structure underlying the problem itself. Recent work describes combinatorial filters for navigation [23, 1], target tracking [12], and story validation [10, 11] problems. Tovar, Cohen, Bobadilla, Czarnowski, and LaValle [2, 3] introduced optimal combinatorial filters for solving some in- ference tasks in polygonal environments with beam sensors and obstacles. Kristek and Shell [27] showed how to extend existing methods for sensor- less manipulation [15, 13] to deformable objects. Song and O’Kane [29] investigated extensions to limited classes of infinite information spaces The deterministic filters we define in this paper that represent as I- state graph, are an special case of nondeterministic graphs explored by Erd- mann [16, 17]. He has studied topological planning with uncertainty. 2.2 Filter reduction The common thread through all of the prior work mentioned so far is to rely on human analysis generate efficient filters for specific problems. The first results on finding optimal filters automatically were presented by O’Kane and Shell [7]. They proved that the filter minimization problem is NP-hard and presented a heuristic algorithm to solve it. The same authors used sim- ilar techniques to solve concise planning [8] and discreet communication [9] problems. Our work extends those filter reduction results to consider the case in which the reduced filter need not be strictly equivalent to the original. 5 Our problem also has some similarity to the problem of measuring the similarity of two deterministic finite automata, as considered by Schwefel, Wegener, and Weinert [4], who solved it using evolutionary algorithms. Chen and Chow also used a scheme for comparing automata, specifically in the context of web services [14]. Our work is differs because of the unique challenges inherent in the differences between filters and automata—because the behavior of a filter may be undefined for certain state-action pairs, the problem of measuring similarities between filters is more challenging. 2.3 Probabilistic methods There are also some connections between the combinatorial filters consid- ered here and the Bayesian probabilistic filters commonly used in mobile robotics [28], including the Kalman filter [22] as a special case. In both cases, the filter’s operation can be described as a discrete-time transition system, in which the state of the system corresponds to the robot’s “belief” about the current state of the world, and transitions between such states are triggered by observations. The primary difference is that for combina- torial filters, we generally assume that both the state space and the obser- vation space are finite, which enables a number of interesting algorithmic questions—including, for example, the filter reduction problem addressed in this paper—to be reasonably posed. There exists some research in automatic probabilistic motion planning using partially-observable Markov decision process (POMDP) models. Roy, Gordon, and Thrun applied dimensionality-reduction techniques to reduce the computation needed for effective planning under such models [19]. Balles- teros, Wegener, and Weinert [6] improved the efficiency of online POMDPs in planning task by reduction of similar belief points based on a similarity measurement. Crook, Keizer, Wang, Tang, and Lemon [20] also presented an automatic POMDP-based approach for belief space compression. 3 Definitions and Formulation This section provides basic definitions for combinatorial filters and the im- proper filter reduction problem. 3.1 Filters and their languages A robot receives a discrete sequence of observations, drawn from a finite observation space. Each observation represents a single sensor reading or a 6 passively-observed action taken by the robot. In response to each of these observations, the robot produces an output, informally called a color, and modeled without loss of generality as a natural number. We model this type of system as a transition graph. Each vertex of the transition graph represents the knowledge, called an information state , retained by the robot at some particular time. Each directed edge is labeled with an observa- tion, showing how the information state changes in response to incoming observations. Definition 1 formalizes this idea. Definition 1. A filter F = ( V, E, l , Y , c, q 0 ) is 6-tuple, representing a di- rected graph with labels on both its vertices and edges, in which • the set V is a finite set of vertices called information states or simply states , • the multiset E is a finite collection of ordered pairs of states, called transitions , in which every state has at least one out-edge, • the function l : E → Y assigns a label l ( e ) to each edge e ∈ E , such that no two edges share both a source vertex and a label, • the domain Y of l is called the observation space , representing the sensor observations that act as the inputs to the filter, • the function c : V → N assigns an integer c ( q ) , informally called the color of q , to each state q ∈ V , representing the output of the filter at that state, and • the vertex q 0 ∈ V is called the initial state . Definition 1 makes two important assumptions about the edges in a filter. First, it requires that no two edges originating from the same vertex can share the same label. Thus, given a “current” state and a new observation, there is at most one resulting state reached by following the edge labeled with that observation. Second, the definition does not require that there must be an outgoing edge for each observation from each vertex. This corresponds to situations in which, based on the underlying structure of the problem, the robot is certain that a given observation cannot occur from a given state. During its execution, the filter starts at the initial state q 0 and immedi- ately outputs c ( q 0 ). After each observation y , the filter transitions from its current state q to a new state q ′ , following the edge q y −→ q ′ , if that edge exists. The filter then generates a new output, namely c ( q ′ ), and awaits the 7 next observation. For a given observation string s = y 1 y 2 · · · y m , there are two cases: 1. If all of the corresponding edges exist, then filter’s output is a sequence of m + 1 colors. We let F ( s, q ) denote the output sequence generated when the filter F processes s starting from state q . We also use the shorthand F ( s ) to denote F ( s, q 0 ). 2. If the filter ever reaches a current state for which no out-edge is labeled with the next observation, we say that the filter has failed for this input, and the filter output is undefined. The basic goal of this paper is to understand, given a filter that acts as a “specification” of the desired output, how to find small filters that produce the substantially similar outputs. However, this comparison only makes sense for observation sequences for which the output of the original filter is well-defined. The next definition formalizes this idea. Definition 2. The language L ( F ) of a filter F is the set of all observation sequences s for which F ( s ) is well-defined. 3.2 Defining distance between filters Given two filters F 1 and F 2 with the same observation space Y , we define the distance between F 1 and F 2 by considering their operation on identical ob- servation strings, and quantifying the difference between the corresponding output strings. Specifically, we choose a function m : ⋃ i ∈ N ( Y i × Y i ) → Z , representing a metric that measures the distance between two equal-length observation strings. We consider, both in the algorithms of Section 4 and the experi- ments in Section 7, two specific options for m : 1. The Hamming distance [21], denoted h, which counts the number of positions at which the two strings differ. 2. The edit distance [24], denoted e, which is the smallest number of single-character insert, delete, and substitute operations, weighted by given operation costs c ins , c del , and c sub respectively, needed to trans- form one string into another. 8 We then use this distance function over observation strings, either m = h or m = e , to define the distance D m ( F 1 , F 2 ) between a pair of filters F 1 and F 2 : D m ( F 1 , F 2 ) = sup s ∈ L ( F 1 ) ∩ L ( F 2 ) ( m ( F 1 ( s ) , F 2 ( s )) length( s ) + 1 ) . (1) The intuition is to consider the worst case distance between output strings, over all observation strings in L ( F 1 ) ∩ L ( F 2 )—that is, over all observation strings that can be processed by both filters—normalized by the length of the output strings. For the special case in which L ( F 1 ) ∩ L ( F 2 ) = ∅ , we define D m ( F 1 , F 2 ) = 0. The use of worst-case reasoning allows us to compare filters without the modelling burden of assigning probabilities to observation sequences, and the normalization is necessary to ensure that the distance between filters is finite. Because the denominator represents the length of both F 1 ( s ) and F 2 ( s )—that is, one more output than there are observations—this can be viewed as an “average cost-per-stage” model as described by LaValle [25]. 3.3 Improper filter reduction We can now state the central algorithmic problem addressed in this paper. Problem: Improper-FM Input: A filter F and an integer k . Output: A filter F ′ with at most k states, such that L ( F ) ⊆ L ( F ′ ) and D m ( F, F ′ ) is minimal. Note, however, that this problem can be proven NP-hard, regardless of the choice of m . Theorem 3. Improper-FM is NP-hard. Proof. Reduction from the basic (error-free) filter minimization problem, FM [7]. Given an instance F of FM , one can use binary search on the range { 0 , . . . , n } to find the smallest value of k for which there exists an F ′ with L ( F ) ⊆ L ( F ′ ) and D m ( F, F ′ ) = 0. This F ′ is, by definition, the correct output for FM . Therefore, a polynomial time algorithm for Improper-FM would imply a polynomial time algorithm for FM . Since FM is NP-hard, and we have a polynomial-time reduction from FM to Improper-FM , conclude that Improper-FM NP-hard as well. 9 Consequently, we restrict our attention in this paper to heuristic algo- rithms that attempt, but cannot guarantee, to minimize the distance be- tween the reduced filter and the original. 4 Computing distance between filters Before turning our attention to Improper-FM , we first consider the related problem of computing the distance D m ( F 1 , F 2 ) between a given pair of filters F 1 and F 2 . Attempting to evaluate Equation 1 directly would be futile, because it includes a supremum over L ( F 1 ) ∩ L ( F 2 ), which in general may be an infinite set. Instead, we introduce a dynamic programming algorithm that whose outputs converge to D m ( F 1 , F 2 ). The details of the algorithm differ depending on whether the underlying string distance function m is Hamming distance function h (addressed in Section 4.1) or edit distance function e (addressed in Section 4.2). 4.1 Hamming distance based metric Our algorithm for computing D h ( F 1 , F 2 ) is based on computing successive values of a simpler function d h , which only considers observation strings up to a certain given length. Our algorithm considers longer observation strings at each iteration. To facilitate a dynamic programming solution, we also must consider different starting states for each filter. Definition 4. Let d h ( q 1 , q 2 , k ) denote the maximum, over all strings s of length k in L ( F 1 ) ∩ L ( F 2 ) , of h ( F 1 ( s, q 1 ) , F 2 ( s, q 2 )) , or 0 if there are no such strings. The function d h is useful because values for D h can be derived from values of d h , as shown in the next lemma. Lemma 5. For any two filters F 1 and F 2 , with initial states q 0 1 and q 0 2 respectively, and with L ( F 1 ) ⊆ L ( F 2 ) , we have D h ( F 1 , F 2 ) = lim k →∞ ( max i ∈ 1 ,...,k d h ( q 0 1 , q 0 2 , i ) i + 1 ) . Proof. For any ǫ > 0, we must show that, for sufficiently large k , we have ∣ ∣ ∣ ∣ D h ( F 1 , F 2 ) − max i ∈ 1 ,...,k d h ( q 0 1 , q 0 2 , i ) i + 1 ∣ ∣ ∣ ∣ < ǫ. 10 First, we note that, for any k , we have D h ( F 1 , F 2 ) ≥ max i ∈ 1 ,...,k d h ( q 0 1 , q 0 2 , i ) i + 1 , because both the left and right expressions are defined to maximize the same quantity defined over some set of observation strings, but the left side considers a superset of the observation strings considered in the maximum operation in the right side. For the other direction, we let s denote an observation string for which D h ( F 1 , F 2 ) − h ( F 1 ( s ) , F 2 ( s )) length( s ) + 1 ≤ ǫ. (2) Such a string must exist by the definition of supremum. Then, choosing k = length( s ), we have max i ∈ 1 ,...,k d h ( q 0 1 , q 0 2 , i ) i + 1 ≥ h ( F 1 ( s ) , F 2 ( s )) k + 1 ≥ D h ( F 1 , F 2 ) − ǫ. (3) In Equation 3, the first step holds because the specific string s is among the strings considered in the maximum over all strings of length at most k , and the second steps proceeds directly from Equation 2. Therefore, the difference between D h ( F 1 , F 2 ) and max i ∈ 1 ,...,k d h ( q 0 1 , q 0 2 , i ) / ( i + 1) is at most ǫ , completing the proof. When k = 0, the filter receives no observations and produces a single output, so d h is trivial to compute in this case, depending only on whether the single outputs produced by each filter match each other: d h ( q 1 , q 2 , 0) = { 0 if c 1 ( q 1 ) = c 2 ( q 2 ) 1 otherwise . (4) Next we address the general case in which k > 0. From any state pair ( q 1 , q 2 ), we must consider the set of observations for which both q 1 and q 2 have same outgoing edges, denoted Y ( q 1 , q 2 ). Thus, for any observation in Y ( q 1 , q 2 ), we know that there exists an edge q 1 y −→ q ′ 1 in F 1 and an edge q 2 y −→ q ′ 2 in F 2 , labeled with the same observation y . d ( y ) h ( q 1 , q 2 , k ) = d h ( q 1 , q 2 , 0) + d h ( q ′ 1 , q ′ 2 , k − 1) . (5) Then we can express d h ( q 1 , q 2 , k ) recursively in terms of d ( y ) h values with shorter observation string lengths: d h ( q 1 , q 2 , k ) = max y ∈ Y ( q 1 ,q 2 ) d ( y ) h ( q 1 , q 2 , k ) , (6) 11 Algorithm 1: Dynamic programming to compute D h ( F 1 , F 2 ). k ← 0 repeat done ← True for ( q 1 , q 2 ) ∈ V 1 × V 2 do Compute d h ( q 1 , q 2 , k ) using Eq. 4 or 6. if ∣ ∣ ∣ d h ( q 1 ,q 2 ,k ) k +1 − d h ( q 1 ,q 2 ,k − 1) k ∣ ∣ ∣ > ǫ then done ← False end end k ← k + 1 until done return max i =0 ,...,k − 1 d h ( q 0 1 , q 0 2 , i ) i + 1 When Y ( q 1 , q 2 ) is empty, we use d h ( q 1 , q 2 , k ) = 0. (The intuition of this special case is that, in this case, q 1 and q 2 are a good choice to merge, because when forming a reduced filter in which q 1 is merged with q 2 , there will be no ambiguity in selecting the correct destination for any transition from the combined state.) Algorithm 1 shows the dynamic programming algorithm that uses this recurrence to compute D h ( F 1 , F 2 ). The intuition is to compute d h for in- creasing values of k , until those d h values converge, and then to find appro- priate normalized maximum. 4.2 Edit distance based metric We now turn our attention to algorithms that, given two filters F 1 and F 2 , compute D e . The approach is similar to the approach for D h in Section 4.1. However, the algorithm for D e is more complex because it must account for the different-length strings that can arise from insert and delete operations on the output sequences. The algorithm works by computing values for a function called d e , which we define below. Definition 6. Let d e ( q 1 , k 1 , q 2 , k 2 ) denote the largest edit distance , denoted e ( F 1 ( s 1 , q 1 ) , F 2 ( s 2 , q 2 )) , between F 1 ( s 1 , q 1 ) and F 2 ( s 2 , q 2 ) , over all strings s 1 of length k 1 and all strings s 2 of length k 2 , for which s 1 , s 2 ∈ L ( F 1 ) ∩ 12 L ( F 2 ) , and s 1 and s 2 are identical to each other for the first min( k 1 , k 2 ) observations. We can show that d e is useful for computing D e with a lemma analogous to Lemma 5. Lemma 7. For any two filters F 1 and F 2 , with initial states q 0 1 and q 0 2 respectively, and with L ( F 1 ) ⊆ L ( F 2 ) , we have D e ( F 1 , F 2 ) = lim k →∞ ( max i ∈ 1 ,...,k d e ( q 0 1 , i, q 0 2 , i ) i + 1 ) . Proof. The same argument from the proof of Lemma 5 applies. The fact that d e considers observation sequences of different lengths is irrelevant here, because the limit in Equation 7 uses the same value, namely k 1 = k 2 = i , for the two string lengths. We can now construct a recurrence for d e . There are three base cases. 1. When k 1 = 0 and k 2 = 0, both filters produce a single output. Editing one such string into the other can be one either by a substitution, or by one deletion and one insertion: d e ( q 1 , 0 , q 2 , 0) = { 0 if c 1 ( q 1 ) = c 2 ( q 2 ) min( c sub , c del + c ins ) otherwise . (7) 2. When k 1 = 0 and k 2 > 0, we have a single character c ( q 1 ) of output from F 1 and a longer output string F 2 ( s 2 , q 2 ), with length k 2 + 1, from F 2 . The edit distance between these strings depends on whether c ( q 1 ) appears in F ( s 2 , q 2 ). If c ( q 1 ) does not appear in F ( s 2 , q 2 ), then we need either (a) 1 deletion and k 2 + 1 insertions or (b) 1 substitution and k 2 insertions, whichever has smaller total cost. If c ( q 1 ) does appear in F ( s 2 , q 2 ), then c ( q 1 ) can be ‘reused’ in F ( s 2 , q 2 ), reducing the edits to simply k 2 insertions. Because Definition 6 calls for the largest edit distance, the relevant question is whether there exists any sequence of observations which, when given as input to F 2 starting at q 2 , avoids all states with color c ( q 1 ) for at least k 2 transitions. Let L ( q 2 , c ( q 1 )) denote the length of 13 the longest path in F 2 starting from q 2 that does not visit any states of color c ( q 1 ). 1 We can express d e for this case based on this value: d e ( q 1 , 0 , q 2 , k 2 ) = k 2 c ins + { min( c del + c ins , c sub ) if L ( q 2 , c ( q 1 )) > k 2 0 otherwise (8) 3. When k 1 > 0 and k 2 = 0, the situation is analogous, but with the roles of F 1 and F 2 swapped: d e ( q 1 , k 1 , q 2 , 0) = k 1 c del + { min( c ins + c del , c sub ) if L ( q 1 , c ( q 2 )) > k 1 0 otherwise (9) In the general case, we can express values for d e using a recurrence similar to the standard recurrence for edit distance [24], but accounting for the changes in state that accompany each observation. For given values of q 1 , k 1 , q 2 , and k 2 , we must consider the worst case over all observations for which both q 1 and q 2 have same outgoing edges. As in Section 4.1, we write Y ( q 1 , q 2 ) to denote this set of observations, and for each y ∈ Y ( q 1 , q 2 ) we write q 1 y −→ q ′ 1 and q 2 y −→ q ′ 2 for the two corresponding edges. To compute d e ( q 1 , k 1 , q 2 , k 2 ), we must consider the worst case over all observations y ∈ Y ( q 1 , q 2 ): d e ( q 1 , k 1 , q 2 , k 2 ) = max y ∈ Y ( q 1 ,q 2 ) d ( y ) e ( q 1 , k 1 , q 2 , k 2 ) , (10) in which we have introduced a shorthand notation d ( y ) e for the value of d e predicated receiving a specific observation y next. There are two cases for d ( y ) e . 1. If c 1 ( q 1 ) = c 2 ( q 2 ), no edits are needed, so we have d ( y ) e ( q 1 , k 1 , q 2 , k 2 ) = d ( y ) e ( q ′ 1 , k 1 − 1 , q ′ 2 , k 2 − 1) . 2. If c 1 ( q 1 ) 6 = c 2 ( q 2 ), we use the smallest cost from among insert, delete, and substitute operations: d ( y ) e ( q 1 , k 1 , q 2 , k 2 ) = min      d e ( q 1 , k 1 , q ′ 2 , k 2 − 1) + c ins , d e ( q ′ 1 , k 1 − 1 , q 2 , k 2 ) + c del , d e ( q ′ 1 , k 1 − 1 , q ′ 2 , k 2 − 1) + c sub      . 1 Note that this need not be a simple path, so the standard hardness result for longest paths in direct graphs [18] does not apply; in fact we can compute L ( q 2 , c ( q 1 )) efficiently using a variant of Dijkstra’s algorithm. 14 The intuition is that, in the case of a delete , F 1 will process one ob- servation, transitioning from q 1 to q ′ 1 and reducing the length of the remaining observation string by 1, whereas F 2 does not process any observations, remaining in state q 2 , with the same number of obser- vations remaining. For an insert, F 2 will process one observation, transitioning from q 2 to q ′ 2 and reducing the length of the remaining observation string by 1, whereas F 1 does not process any observations, remaining in state q 1 , with the same number of observations remain- ing. For a substitution, both F 1 and F 2 transition to new states, and both k 1 and k 2 are decreased. This matches the usual recurrence for edit distance, with the extra complication that we must consider the states reached by the two filters in addition to the string lengths. This recurrence leads directly to an algorithm for computing D e ( F 1 , F 2 ). Pseudocode appears as Algorithm 2. The algorithm applies the recurrence, starting from k 1 = k 2 = 0, and increasing those indices until the average er- ror per observation converges, as in Lemma 7. The most important subtlety is in the ordering: Before computing d e values with input string lengths ( k 1 , k 2 ), we must ensure that the corresponding computations have been completed for ( k 1 − 1 , k 2 ), ( k 1 , k 2 − 1), and ( k 1 − 1 , k 2 − 1). Note that Algorithm 2 is significantly slower than Algorithm 1 because of the need for an additional nested loop to accommodate the differing string lengths between F 1 and F 2 . This difference, which we also observe in the experiments described in Section 7, confirms the basic intuition that Ham- ming distance is a computationally simpler metric for string distance than edit distance. Note, however, that edit distance is a more ‘forgiving’ metric, in the sense that if c ins = c del = c sub = 1, then D e ( F 1 , F 2 ) ≤ D h ( F 1 , F 2 ) for any two filters F 1 and F 2 . 5 Local greedy sequential reduction In this section, we present a heuristic algorithm for improper filter reduction that efficiently reduces the input filter to the required size, while attempting to minimize the error. The algorithm performs a series of greedy “merge” operations, each of which reduces the filter size by one. We first describe the details of this merge operation (Section 5.1), and then propose an approach for selecting pairs of states to merge (Section 5.2). 15 Algorithm 2: Dynamic programming to compute D e ( F 1 , F 2 ). i ← 0 repeat done ← True for j ← 0 , . . . , i do for ( q 1 , q 2 ) ∈ V 1 × V 2 do Compute d e ( q 1 , i, q 2 , j ) using Eq. 7, 8, 9, or 10. Compute d e ( q 1 , j, q 2 , i ) using Eq. 7, 8, 9, or 10. end end for ( q 1 , q 2 ) ∈ V 1 × V 2 do if ∣ ∣ ∣ d e ( q 1 ,i,q 2 ,i ) i +1 − d e ( q 1 ,i − 1 ,q 2 ,i − 1) i ∣ ∣ ∣ > ǫ then done ← False end end k ← k + 1 until done return max i =0 ,..., k − 1 d e ( q 0 1 , i, q 0 2 , i ) i + 1 5.1 Merging states Suppose we have a filter F with n states, and we want form a new, nearly identical, filter F ′ with n − 1 states. One way to accomplish this is to select two states q 1 and q 2 from F —deferring to Section 5.2 the question of how to select q 1 and q 2 —and merge them, in the following way. • Replace q 1 and q 2 with a combined state, denoted q 1 , 2 . If either of q 1 or q 2 is the filter’s initial state, then q 1 , 2 becomes the new initial state. We assign the color of q 1 to the combined state q 1 , 2 . • Replace each in-edge q i y −→ q 1 of q 1 , with a new edge q i y −→ q 1 , 2 to the combined state. Repeat for q 2 , replacing each q i y −→ q 2 with q i y −→ q 1 , 2 . • Replace each out-edge q 1 y −→ q j of q 1 with a new edge q 1 , 2 y −→ q j . • For each out-edge q 2 y −→ q j , determine whether q 1 has an out-edge with the same label y . If so, then discard the edge q 2 y −→ q j . If not, 16 q 1 q 2 y 2 y 2 y 0 y 1 y 3 y 4 q 1 , 2 y 2 y 0 , y 1 y 3 y 4 q 1 , 2 y 2 y 0 , y 1 y 3 , y 4 y 4 Figure 2: An illustration of the need for post-processing after merging two states q 1 and q 2 , to ensure that L ( F ) ⊆ L ( F ′ ). [left] A filter F , before q 1 and q 2 are merged. [center] Another filter F ′ , formed by combining q 1 with q 2 , but without post-processing. In this example, F can process the observation string y 1 y 2 y 4 , but F ′ fails on that input. [right] The post-processing step adds an edge to resolve the problem. then replace that edge with q 1 , 2 y −→ q j . The intuition is to replace the two states q 1 and q 2 with just one state, with as little disruption to the filter as possible. The only complication—and the reason that this merge operation must be more complex than a standard vertex contraction—is that the combined node can have at most one out- edge for each observation label. When q 1 and q 2 both have out-edges with the same label, we arbitrarily resolve that conflict by giving priority to q 1 over q 2 . In the context of the Improper-FM problem, we also must ensure that the filter F ′ resulting from a merge in F has L ( F ) ⊆ L ( F ′ ). This may not be immediately true, as illustrated in Figure 5.1. To resolve this problem, we perform a forward search over state pairs ( q a , q b ), in which q a is from the original filter F and q b is from the merged filter F ′ , starting from ( q 2 , q 1 , 2 ) when we suppose that q 2 is merged into the q 1 to create q 1 , 2 . Each time this search finds a reachable state pair ( q a , q b ) and an observation y , for which F has an edge q a y −→ q ′ a but F ′ has no edge from q b labeled y , we insert an edge q b y −→ q ′ a into F ′ . This ensures that any observation sequence that can be processed by F can also be processed by F ′ . The runtime of this post-processing step is quadratic in the number of states in F . We write merge( F, q 1 , q 2 ) to denote the filter resulting from applying this complete merge operation to q 1 and q 2 in F . 17 5.2 Selecting pairs of states to merge We can use this merge operation to form a heuristic algorithm for improper filter reduction in a local greedy way. The basic idea is to consider all ordered pairs of distinct states in the current filter as candidates to merge, and to compare the filter resulting from each such merge to the original input filter. The merge that results in the smallest distance from the original filter is kept, and the algorithm repeats this process until filter is reduced to the desired size. Beyond this basic idea, we add two additional constraints to improve the quality of the final solution. First, we reject any merge operation that leaves some states unreachable, unless all available merges leave at least one unreachable state. Second, we also reject any merge operation that elimi- nates at least one output color from the resulting filter, unless all available merges eliminate a color. Algorithm 3 shows the complete approach. See Section 7 for an evaluation of this algorithm’s effectiveness. 6 Randomized global reduction The central limitation of Algorithm 3 is that it selects merges to perform in a sequential way, and therefore cannot account for the impact that each merge has on later steps in the algorithm. In this section, we present an alternative to Algorithm 3 that avoids this problem by selecting all of the merges to perform at once, in a global way. Algorithm 4 outlines the approach. The intuition is to cast the problem of choosing which states to merge with each other as an improper graph coloring problem, which we solve using an existing algorithm (Section 6.1). We then apply a randomized ‘voting’ process to construct the reduced filter from the colored graph (Section 6.2). 6.1 Selecting merges via improper graph coloring We can view the problem of reducing a given filter F down to k states as a question of partitioning the states of F into k groups. Our algorithm accomplishes this by solving a coloring problem on a complete undirected graph C ( F ), defined as follows: 1. For each state q in F , we create one vertex v ( q ) in C ( F ). 2. For each pair of distinct vertices v ( q 1 ) , v ( q 2 ) in C ( F ), we create a weighted edge. The intuition is that the weights should be estimates of how different the two corresponding states are. If the behavior of 18 Algorithm 3: A greedy sequential method to reduce F to k states. f ← False F orig ← F while | V ( F ) | > k do D ⋆ ← ∞ for ( q 1 , q 2 ) ∈ V ( F ) × V ( F ) do if q 1 = q 2 then continue end F ′ ← merge( F, q 1 , q 2 ) if f = False then if F ′ has unreachable states then continue end if F ′ has unreachable colors then continue end end D ← D m ( F orig , F ′ ) // Use Alg. 1 or Alg. 2. if D < D ⋆ then D ⋆ ← D F ⋆ ← F ′ end end if D ⋆ = ∞ then f ← True end else f ← False F ← F ⋆ end end return F 19 Algorithm 4: A randomized global method to reduce F to k states using r iterations. C ← C ( F ) c ← improper coloring of C with k colors[5]. D ⋆ ← ∞ for R ← 0 , . . . , r do F ′ ← empty filter for each color i in c do q ← state in F randomly selected from those colored i in C ( G ) Create a state q i in F ′ with same output color as q . end for each color i in c do for each observation y ∈ Y in F do q ← state in F randomly selected from those both colored i in C ( G ) and with an edge q y −→ w in F . Create an edge in F ′ from q i to q c ( w ) labeled y . end end D ← D m ( F, F ′ ) // Use Alg. 1 or Alg. 2. if D < D ⋆ then D ⋆ ← D F ⋆ ← F ′ end end return F ⋆ 20 A B C E D a b , c , d e a b , e b , c , d a , e a , e a , d e A B C D E 0 . 66 0 . 66 0 . 5 0 . 33 0 . 66 0 . 5 0 . 5 0 . 5 0 . 5 0 . 5 Figure 3: [left] An example five-state filter. (This filter was originally gen- erated by the algorithm of O’Kane and Shell [7]. [right] The corresponding complete graph, along with an improper coloring of that graph with two colors. F is very similar when started from q 1 as when started from q 2 , then we should assign a relatively small weight to the edge between v ( q 1 ) and v ( q 2 ). Conversely, if the behavior of F differs significantly when started from q 1 compared to starting from q 2 , then we should assign a relatively large weight to that edge. We can capture these kinds of differences by computing the distance of F with itself—That is, by computing either D h ( F, F ) or D e ( F, F )— and examining the intermediate values—either d h ( q 1 , q 2 , k ) or d e ( q 1 , k, q 2 , k ), in which k is the number of iterations of the outermost loops in Al- gorithm 1 or Algorithm 2—computed along the way. Specifically, we assign w ( q 1 , q 2 ) as w ( q 1 , q 2 ) = lim k →∞ ( max i ∈ 1 ,...,k d h ( q 1 , q 2 , i ) i + 1 ) for Hamming distance, and w ( q 1 , q 2 ) = lim k →∞ ( max i ∈ 1 ,...,k d e ( q 1 , i, q 2 , i ) i + 1 ) for edit distance. Figure 3 shows an example of this construction. 21 The idea is then to assign a color c ( v ) to each vertex v of C ( F ), using at most k colors, while minimizing the worst case over all vertices, of the total weight of edges to same-colored neighbors. That is, we want assign k colors to the vertices of C ( F ) in a way that minimizes this objective function: O ( F, c ) = max v ∈ V ( C ( F )) ∑ { u ∈ V ( C ( F )) −{ v } | c ( v )= c ( u ) } w ( u, v ) , (11) in which V ( C ( F )) denotes the vertex set of C ( F ). This optimization prob- lem, which is known as the Threshold Improper Coloring Problem , has been addressed by Araujo, Bermond, Giroire, Havet, Mazauric and Modrzejew- ski [5]. Though the problem is NP-hard even to approximate—Note that an efficient algorithm for this problem could be used to build an efficient algo- rithm for the standard proper graph coloring problem—that prior research presents a randomized heuristic algorithm that performs well in most cases. We use that algorithm to color C ( F ). 6.2 Randomized voting for improper filter reduction Next, we form the reduced filter F ′ by ‘merging’ the states in F correspond- ing to each group of same-colored nodes in C ( F ) into a single state in F ′ . The process is somewhat analogous to pairwise merging process described in Section 5.1, but must account for some important differences. Most im- portantly, because we want to merge the states within each of the k color groups simultaneously, when selecting the edges in the reduced filter F ′ , it only needs to consider which state in F ′ —that is, which color group—should be the target of that edge, rather than making a finer-grained selections of some state in some partially-reduced version of F . Note, however, that the states in each color group may not agree on which F ′ state should be reached under each observation. When, for a given observation y , such disagreements occur, we resolve them in a randomized way. The algorithm selects, using a uniform random distribution, one of the F states in this color group that has an edge labeled with y , and adds a transition in F ′ the state corresponding to the color group reached by that state. This forms a kind of ‘weighted voting,’ in which it is more likely to select transitions that correspond to larger number of states in the group. Similarly, we select the output color of each state in F ′ by randomly selecting one of its constituent states and using its color as the output for the combined state. Continuing the example from Figure 3, observe that to create reduced filter with two states, one of those states will correspond to the original filter’s A, B, and C states, and other one will consist of D and 22 E. Since D and E have different colors in the original filter, the algorithm will assign the new DE state to each of the two possible output colors with equal probability. Because the final filters produced by this process are not deterministic, we repeat the reduction several times and return the best filter resulting from those iterations. 7 Implementation and Experimental Results We have implemented Algorithms 1–4 in Python. The experiments de- scribed below were executed on a GNU/Linux computer with a 3GHz pro- cessor. Throughout, we used ǫ = 0 . 035 in Algorithms 1 and 2, c ins = c del = c sub = 1 in Algorithm 2, and r = 400 in Algorithm 4. 7.1 Distance and run time for varying filters First, we consider a family of filter reduction problems originally described by Tovar, Cohen, and LaValle [2]. In these problems, a pair of robots move through an annulus-shaped environment, occasionally crossing a beam sen- sor that detects a crossing of that beam has occurred, but cannot detect which robot has crossed, nor the direction of that crossing. The filter should process these observations and output 0 if the robots are together, or 1 if the robots are separated by at least one beam. (This is a different and more challenging problem than the single-robot variant described in Section 1.) We varied the number of beam sensors from 3 to 9, and tested two equiv- alent filters for each number of beams: one ‘unreduced’ na ̈ ıve filter, formed by directly computing the sets of possible states after each observation, and a smaller ‘reduced’ filter produced by applying the algorithm of O’Kane and Shell [7] to the unreduced versions. These reduced filters have the same be- havior as their unreduced counterparts, but are the smallest filters for which that equivalence holds. For each of these 7 · 2 = 14 filters, we executed both Algorithm 3 and Algorithm 4, using both D h and D e as the underlying metric for each al- gorithm. The target filter size was set to k = 2, the maximal meaningful reduction, for each of these trials. The results appear in Figure 4 for Ham- ming distance and in Figure 5 for edit distance. These results show that the final solution quality is somewhat better for Algorithm 4 in many cases. The global approach of Algorithm 4 is also faster the greedy sequential reduction of Algorithm 3. The difference, which is especially pronounced as the filter size grows large , is explained by the 23 fact that Algorithm 4 uses its subroutine for distance between filters—that is, Algorithm 1 or 2—only once, rather than many times for each reduction. Note that the computation time is substantially shorter for Hamming distance than for edit distance, resulting from the extra nested loop in Al- gorithm 2 that is not needed in Algorithm 1. 24 Unreduced, Alg. 3 Unreduced, Alg. 4 Reduced, Alg. 3 Reduced, Alg. 4 0 100 200 300 400 500 600 700 800 900 1000 3 4 5 6 7 8 9 Time (seconds) Number of Beams Runtime for Hamming distance, varying input filter size Unreduced, Alg. 3 Unreduced, Alg. 4 Reduced, Alg. 3 Reduced, Alg. 4 0.0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1.0 3 4 5 6 7 8 9 D h ( F, F ′ ) Number of Beams Solution quality for Hamming distance, varying input filter size Figure 4: Results for reduction of annulus filters under Hamming distance using Algorithm 3 and Algorithm 4. [top] Run time. [bottom] Final distance D h ( F, F ′ ) between original and reduced filters. 25 Unreduced, Alg. 3 Unreduced, Alg. 4 Reduced, Alg. 3 Reduced, Alg. 4 0 5000 10000 15000 20000 25000 30000 35000 40000 45000 50000 55000 60000 3 4 5 6 7 8 9 Time (seconds) Number of Beams Runtime for edit distance, varying input filter size Unreduced, Alg. 3 Unreduced, Alg. 4 Reduced, Alg. 3 Reduced, Alg. 4 0.0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1.0 3 4 5 6 7 8 9 D e ( F, F ′ ) Number of Beams Solution quality for edit distance, varying input filter size Figure 5: Results for reduction of annulus filters under edit distance using Algorithm 3 and Algorithm 4. [top] Run time. [bottom] Final distance D e ( F, F ′ ) between original and reduced filters. 26 7.2 Varying the target size for a single filter Next, we evaluated the impact of the target size k on the algorithms’ per- formance. We used the two-agent eight-beam filter described above in Sec- tion 7.1, in both its unreduced and reduced forms. We varied k from 2 to 10 and executed each of the algorithms for all k . Figures 6 and 7 show the results for Hamming distance and edit distance, respectively. The results show that, across all cases in this range, the run time is almost entirely unaffected by the target size. The solution quality shows a slight improving trend as the target size increases, as one might expect. Finally, to confirm that these results generalize to other kinds of filters, we repeated the analysis for the ‘L-shaped corridor’ filtering problem intro- duced by LaValle [25], which also appears in O’Kane and Shell [7]. In this problem, a sensorless robot moves along a long corridor with a right angle midway through it. The robot moves in steps that unpredictably vary be- tween 1 or 2 steps. The filter’s goal is to output 0 when the robot reaches the end of the corridor, or 1 otherwise. This problem is interesting because the size of the unreduced filters grows exponentially with the corridor’s length, whereas the reduced filter size grows only linearly. In this case, we use only the unreduced filters, because the reduced filter is too small to be of interest. Figures 8 and 9 show the results, which follow the same general trends as the previous tests. Of particular note that the resulting filters are the same for Hamming distance and edit distance, reflecting the fact that insert or delete operations are not useful in this problem, since the relevant informa- tion at each step is the furthest possible state from the goal. As a result, the error rates are identical between the two metrics. This is a clear instance in which Hamming distance, by virtue of the faster run time of Algorithm 1 is a better choice than edit distance. 27 Unreduced, Alg. 3 Unreduced, Alg. 4 Reduced, Alg. 3 Reduced, Alg. 4 0 50 100 150 200 250 300 350 400 450 2 3 4 5 6 7 8 9 10 Time (seconds) Reduced size ( k ) Runtime for Hamming distance, varying target filter size Unreduced, Alg. 3 Unreduced, Alg. 4 Reduced, Alg. 3 Reduced, Alg. 4 0.0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1.0 2 3 4 5 6 7 8 9 10 D h ( F, F ′ ) Reduced size ( k ) Solution quality for Hamming distance, varying target filter size Figure 6: Results for reduction of the eight-beam, two-robot annulus filter for varying target filter sizes under Hamming distance. [top] Run time. [bottom] Final distance D h ( F, F ′ ) between original and reduced filters. 28 Unreduced, Alg. 3 Unreduced, Alg. 4 Reduced, Alg. 3 Reduced, Alg. 4 0 3000 6000 9000 12000 15000 18000 21000 2 3 4 5 6 7 8 9 10 Time (seconds) Reduced size ( k ) Runtime for edit distance, varying target filter size Unreduced, Alg. 3 Unreduced, Alg. 4 Reduced, Alg. 3 Reduced, Alg. 4 0.0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1.0 2 3 4 5 6 7 8 9 10 D e ( F, F ′ ) Reduced size ( k ) Solution quality for edit distance, varying target filter size Figure 7: Results for reduction of the eight-beam, two-robot annulus filter for varying target filter sizes under edit distance. [top] Run time. [bottom] Final distance D e ( F, F ′ ) between original and reduced filters. 29 Alg. 3 Alg. 4 0 2 4 6 8 10 12 14 16 18 2 3 4 5 6 7 8 9 10 Time (seconds) Reduced size ( k ) Runtime for Hamming distance, L-shaped corridor Alg. 3 Alg. 4 0.0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1.0 2 3 4 5 6 7 8 9 10 D h ( F, F ′ ) Reduced size ( k ) Solution quality for Hamming distance, L-shaped corridor Figure 8: Results for reduction of L-shaped corridor filter for varying target filter sizes under Hamming distance. [top] Run time. [bottom] Final distance D h ( F, F ′ ) between original and reduced filters. 30 Alg. 3 Alg. 4 0 40 80 120 160 200 240 280 2 3 4 5 6 7 8 9 10 Time (seconds) Reduced size ( k ) Runtime for edit distance, L-shaped corridor Alg. 3 Alg. 4 0.0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1.0 2 3 4 5 6 7 8 9 10 D e ( F, F ′ ) Reduced size ( k ) Solution quality for edit distance, L-shaped corridor Figure 9: Results for reduction of L-shaped corridor filter for varying target filter sizes under edit distance. [top] Run time. [bottom] Final distance D e ( F, F ′ ) between original and reduced filters. 31 8 Conclusion In this paper, we introduced two metrics to measure the similarity between two filters and presented dynamic programming algorithms for computing these measurements . Then we presented two algorithms to reduce a filter to a given size automatically. There exist some future directions to extend this work. Most directly, a good extension of this work would be presenting more efficient algorithms. In particular, we anticipate that Algorithm 3 can be accelerated by reducing the number of filter distance queries it makes, possibly by borrowing the self- similarity idea from Algorithm 4. In addition, finding additional criteria for improving the merge operation, or even replacing that operation with some other form of reduction is another possible improvement. Though we proved that Improper-FM is NP-hard, it is worthy of study to determine whether it can be approximated efficiently. It may also be the case that certain interesting special classes of filters are efficiently solvable. Another direction is to investigate whether Improper-FM is fixed pa- rameter tractable. That is, does there exist a algorithm whose run time is polynomial in the input size, but exponential in some other parameter that measures the complexity of a given instance. Acknowledgment The authors are grateful to Dylan Shell for helpful feedback on an earlier version of this work. This material is based upon work supported by the Na- tional Science Foundation under Grant Nos. IIS-0953503 and IIS-1526862. References [1] B. Tovar, Minimalist Models and Methods for Visibility- based Tasks, Ph.D. Thesis, University of Illinois at Ur- bana Champaign, (2009) [2] B. Tovar and F. Cohen and S. M. LaValle, Sensor beams, obstacles, and possible paths, Algorithmic Foundations of Robotics VIII, Springer-Verlag, Berlin, (2009) [3] B. Tovar and F. Cohen and L. Bobadilla and J. Czarnowski and S. M. LaValle, Combinatorial Filters: 32 Sensor Beams, Obstacles, and Possible Paths, TOSN, 10, 47, (2014) [4] H.-P. Schwefel and I. Wegener and K. Weinert, Ad- vances in Computational Intelligence: Theory and Prac- tice, Springer, (2003) [5] J. Araujo and J-C. Bermond and F. Giroire and F. Havet and D. Mazauric and R. Modrzejewski, Weighted Im- proper Colouring, International Workshop on Combina- torial Algorithms, 7056, 1–18, (2011) [6] J. Ballesteros and L. Merino and M. A. Trujillo and A. Viguria and A. Ollero, Improving the efficiency of online POMDPs by using belief similarity measures, IEEE Inter- national Conference on Robotics and Automation, 1792- 1798, (2013) [7] J. M. O’Kane and D. A. Shell, Automatic Reduction of Combinatorial Filters, Proc. IEEE International Confer- ence on Robotics and Automation, (2013) [8] J. M. O’Kane and D. A. Shell, Finding concise plans: Hardness and algorithms, Proc. IEEE/RSJ International Conference on Intelligent Robots and Systems, (2013) [9] J. M. O’Kane and D. A. Shell, Automatic design of dis- creet discrete filters, Proc. IEEE International Conference on Robotics and Automation, (2015) [10] J. Yu and S. M. LaValle, Story Validation and Approx- imate Path Inference with a Sparse Network of Het- erogeneous Sensors, IEEE International Conference on Robotics and Automation, (2011) [11] J. Yu and S. M. LaValle, Cyber Detectives: Determining When Robots or People Misbehave, Algorithmic Foun- dations of Robotics IX, Springer Tracts in Advanced Robotics (STAR), 68, Springer Berlin/Heidelberg, 391- 407, (2011) [12] J. Yu and S. M. LaValle, Shadow Information Spaces: Combinatorial Filters for Tracking Targets, IEEE Trans- actions on Robotics, 28, (2012) 33 [13] K. Y. Goldberg, Orienting Polygonal Parts Without Sen- sors, Algorithmica, 10, 201–225, (1993) [14] L. Chen and R. Chow, A web service similarity refine- ment framework using automata comparison, Proc. Inter- national Conference on Information Systems, Technology and Management, (2010) [15] M. Erdmann and M. T. Mason, An Exploration of Sen- sorless Manipulation, IEEE Transactions on Robotics and Automation, 4, 369–379, (1988) [16] M. Erdmann, On the Topology of Discrete Strategies, In- ternational Journal of Robotics Research, 29, 855–896, (2010) [17] M. Erdmann, On the Topology of Discrete Planning with Uncertainty, In Advances in Applied and Computational Topology, Proc. Symposia in Applied Mathematics, 70, American Mathematical Society, (2012) [18] M. R. Garey and D. S. Johnson, Computers and In- tractability: A Guide to the Theory of NP-Completeness, W.H. Freeman and Company, New York, (1979) [19] N. Roy and G. Gordon and S. Thrun, Finding Ap- proximate POMDP solutions Through Belief Compres- sion, Journal of Artificial Intelligence Research, 23, 1–40, (2005) [20] P. A. Crook and S. Keizer and Z. Wang and W. Tang and O. Lemon, Real user evaluation of a POMDP spo- ken dialogue system using automatic belief compression, Computer Speech & Language, 28, 873–887, (2014) [21] R. W. Hamming, Bell System Technical Journal, Er- ror Detecting and Error Correcting Codes, 26, 147–160, (1950) [22] R. Kalman, A new approach to linear filtering and pre- diction problems, Transactions of the ASME, Journal of Basic Engineering, 35–45, (1960) 34 [23] R. Lopez-Padilla and R. Murrieta-Cid and S. M. LaValle, Optimal gap navigation for a disc robot, Proc. Workshop on the Algorithmic Foundations of Robotics, (2012) [24] R. A. Wagner and M. J. Fischer, The String-to-String Correction Problem, Journal of the Association for Com- puting Machinery, 21, 168–173, (1974) [25] S. M. LaValle, Planning Algorithms, Cambridge University Press, Cambridge, U.K., Available at http://planning.cs.uiuc.edu/, (2006) [26] S. M. LaValle, Sensing and Filtering: A Fresh Perspective Based on Preimages and Information Spaces, Foundations and Trends in Robotics, 1, 253–372, (2012) [27] S. Kristek and D. Shell, Orienting Deformable Polygonal Parts without Sensors, Proc. International Conference on Intelligent Robots and Systems, (2012) [28] S. Thrun and W. Burgard and D. Fox, Probabilistic Robotics, Cambridge University Press, MIT Press, Cam- bridge, MA, (2005) [29] Y. Song and J. M. O’Kane, Comparison of Constrained Geometric Approximation Strategies for Planar Informa- tion States, Proc. International Conference on Robotics and Automation, (2012) 35