UA12003 STEWART ET AL. 551 The Revisiting Problem in Mobile Robot Map Building: A Hierarchical Bayesian Approach Benjamin Stewart! Jonathan Ko t Dept. of Computer Science & Engineering t University of Washington Seattle, WA Abstract We present an application of hierarchical Bayesian estimation to robot map building. The revisiting problem occurs when a robot has to decide whether it is seeing a previously-built portion of a map, or is exploring new territory. This is a difficult decision problem, requiring the probability of being outside of the current known map. To estimate this probability, we model the structure of a "typical" environment as a hidden Markov model that generates sequences of views observed by a robot navigating through the envi­ ronment. A Dirichlet prior over structural mod­ els is learned from previously explored environ­ ments. Whenever a robot explores a new en­ vironment, the posterior over the model is es­ timated by Dirichlet hyperparameters. Our ap­ proach is implemented and tested in the context of multi-robot map merging, a particularly dif­ ficult instance of the revisiting problem. Ex­ periments with robot data show that the tech­ nique yields strong improvements over alterna­ tive methods. 1 Introduction Building maps of unknown environments is one of the fun­ damental problems in mobile robotics. As a robot explores an unknown environment, it incrementally builds a map consisting of the locations of objects or landmarks. Typ­ ically, as it explores larger areas, its uncertainty relative to older portions of the map increases; for example, in closing a large loop. Thus, a key problem is determining whether the current position of the robot is in an unexplored area or in the already-constructed map (the revisiting problem). The revisiting problem for single robots is illustrated in Fig. !(a). Shown there is a map built by a robot during exploration. The robot started in the lower right hallway Dieter Fox! Kurt Konolige+ Artificial Intelligence Center+ SRI International Menlo Park, CA and moved clockwise around the large loop. At the end, it moves down the right hallway, but due to the accumu­ lated uncertainty in its own position, 1t can not determme whether it is in the same hallway as in the beginning or whether it is in a parallel hallway. Multiple robots exploring the same environment from un­ known start locations face a particularly difficult instance of the revisiting problem. For coordinated exploration, the robots have to merge their maps so as to build a shared world model. Map merging requires the determination of the robots' relative location. Consider the situation shown in Fig. !(b). Here, two robots have explored parts of the large environment shown below. In order to merge the partial maps, they have to determine whether they visited the same locations in the environment and if so, they have to determine the offset between their maps. The diffi­ culty of this problem lies in the first step, i.e. in decid­ ing whether there is an overlap between the two maps or not. To avoid this decision problem, most existing ap­ proaches assume knowledge about the robots' relative start locations [ 5, 15, 14]. At the minimum, these techniques require that one robot is known to start in the map already built by the other robot. If we consider the revisiting problem in a Bayesian context, then to make an informed decision, we require probabili­ ties for two different hypotheses, one for the robot moving through area that has already been mapped, and one for the robot moving through unexplored area. While it is well­ understood how to compute the likelihood of sensor mea­ surements in areas already mapped by a robot, this problem additionally requires to compute the likelihood of sensor measurements in areas the robot has not yet explored. Vir­ tually all existing approaches to map building implicitly de­ termine the likelihood for "out of map" measurements un­ der the assumption that objects are distributed uniformly, i.e. they assign fixed, identical likelihoods to all observa­ tions in unexplored areas [13, 4, 12, 8]. However, such ap­ proaches ignore valuable information since most environ­ ments are structured rather than randomly patched together. The key contribution of this paper is a method for estimat- 552 STEWART ET AL. UA12003 Roi position;', ·. · · A ! · . _.ell (II) Start position : : ' k,.; ,-iŽ-:1 • .Jli '' err) ri :1i, .:;·.,:1-l i·J t,N_; f· 1 [·.it· _c];l [ ;r ·_;,"!'':(b) Figure 1: (a) Loop closing: A robot explores an environment and has to decide whether it returned to the hallway it started in (I) or whether it is in a parallel hallway (II). (b) Multi-robot map merging: Two robots built the partial maps (I) and (II) and have to decide whether they explored an overlapping part of the envi­ ronment, i.e. whether they can merge their maps or not. ing the probability of the out-of-map hypothesis 1. In a nutshell, we construct a structural model of a typical en­ vironment; when the robot is outside the partial map, we use the model to predict what a typical view would look like, given the robot's history of observations. The current observation is then compared against the generated view to compute a likelihood. More specifically, we introduce a hierarchical Bayesian ap­ proach that captures the structure of an environment by a hidden Markov process that represents transitions between views of the environment. An offline learning process takes a set of maps and generates a Dirichlet prior over map structures. The prior is the "typical" generative map used by the robot at the start of exploration. An adaptation pro­ cess refines the model distribution online, as the robot en­ counters views of its environment. To prove the validity of the approach, we have constructed an efficient implementation, using a particle filter that de­ rives the likelihoods of the out-of-map hypothesis under the structural model. Views are discrete features extracted from laser range-finder scans. Experiments using a multi­ robot exploration scenario show that our technique clearly outperforms alternative approaches to map merging. This paper is organized as follows. In the next section, we will describe the Bayesian approach to learning and esti­ mating the structure of environments. Section 3 presents the generative model for partial map merging and imple­ mentational details are given in Section 4. Experiments are described in Section 5, followed by a discussion. 2 Hierarchical Model for Map Structures Our model of map structures is based on the idea that in­ door environments consist of collections of local patches. 1 In the context of hypothesis testing, for example, this proba­ bility can be used to evaluate the null hypothesis for the question, "Am I in this previously mapped area?". Figure 2: Hierarchical Bayesian model: The hyperparameter a represents the prior distribution over maps. The structure of each map is captured by a Dirichlet process Qt describing how map patches are connected. The Dirichlet processes generate data d1, sequences of views observed by a robot during exploration. These patches, especially the way they are connected, gen­ erate sequences of views observed by a robot as it moves through an environment. For example, many indoor envi­ ronments consist of straight hallways, hallway crossings, and rooms. These local pieces are not patched together by pure chance, but rather according to the global structure of the environment. For instance, while there is no surprise if a robot observes a straight hallway piece directly after another straight hallway piece, it is rather uncommon to observe two hallway crossings next to each other. The hierarchical Bayesian model for map structures is il­ lustrated in Fig. 2. Shown there are maps of typical indoor environments. Each map generates sequences of views dt distributed according to the transition parameters Ql of the map structure. All maps share a common hyperparameter a that serves as a prior distribution from which the different map structures are drawn. The key idea of our hierarchical approach is to learn this hyperparameter a based on a col­ lection of previously encountered maps. Whenever a robot explores a new environment, it can then use a as a prior for the structure of this new environment. The estimate of the structure is updated as the robots observes data in the new environment. More specifically, we assume that a robot can observe a finite number v of distinctive views. The structure of an environment is captured by parameters Qili• 1::::; i,j::::; v, which describe the probability of observing view i given that the robot previously saw view j. Let qli denote the multinomial distribution over views following view j. The complete structure of an individual environment l is thus represented by a collection of v multinomial distributions qli. The model prior a is a v x v matrix, where each a j = (a1;, a2;, ... , av;) serves as a conjugate Dirichlet prior for the multinomial qli. 2.1 Inference Let us first describe how to update the map structure param­ eters based on observations made during exploration of an UAI2003 STEWART ET AL. 553 environment. To do so, we assume that the priors o: j over model parameters are known. As a robot moves through the environment, it makes a sequence of observations, de­ noted by zu. We make the simplifying assumption that it is possible to extract from such an observation sequence frequency counts filj, which describe how often the robot observed view i after observing view j 2• Correspondingly, flj = (fllj• hij• ... , fvjj) denotes the vector of frequency counts following view j. Given the Dirichlet prior O:j and the counts flj up to time t, the posterior distribution over qlj is Dirichlet with parameters O:j + f1j [7]: (I) The posterior predictive probability that view i follows view j can be determined by integrating over the posterior of the transition probabilities qiJ: p(v,=i I Vt-l=j,o:j,flj) = J P(%1o:j,fjj) qilj d% J Dirichlet(% I 0: j + r,j) qijj dqjj (2) where (2) follows from the properties of the Dirichlet dis­ tribution [I 0]. Thus, the prior and the frequency counts are sufficient statistics for the posterior over the parameters of our structural model. The individual o:i; 's are often re­ ferred to as prior samples, since they serve as initial counts added to the observed frequencies filj. As can be seen, whenever a robot makes an observation in the new environment, the posterior over the structural model is updated by simply incrementing the frequency count fijj of the most recently observed view transition. 2.2 Learning Priors Over Map Structures It remains to be shown how to learn the prior for transitions between views. To do so, we use data d collected in typ­ ical indoor environments previously explored by a robot. While a full Bayesian treatment would require to learn a distribution over hyperparameters o: = (o:1, 0:2, ... , o:v), we restrict our model to the MAP estimate o:*: o:' = argmaxp(o: I d) = p(d I o:) p(o:) "=' p(d I o:) (3) "' p(d) Here the rightmost term follows from a uniform prior over the hyperparameter o: and the fact that p( d) has no impact on the MAP estimate. The data d = (d1, ... , dk) consists of frequency counts observed in the k previously explored maps. Assuming independence between the different maps 2Note that the robot actually does not observe discrete views, but rather continuous, noisy versions thereof. In our approach, we determine the frequency counts J;11 using the views that are most likely to have generated the observations. See [I, 16] for approaches dealing with partially observable views. and between the individual Dirichlet priors, we can maxi­ mize (3) over the individual priors O:j. A rather straight­ forward derivation similar to [I 0] shows that the likelihood functionp(d I O:j) is given by where r is the gamma distribution, Jf,j denotes how often view i follows view j in the data observed in map !, and n and Cij are the sums over all Jf1. and O:i;, respectively. The MAP o:* can be found by maximizing the log of(4) using a conjugate gradients method (see also [10, II]). To summarize, the structure of an environment is captured by a collection of multinomial distributions qli describ­ ing the sequence of views observed by a robot as it nav­ igates through the environment. A Dirichlet prior o: over these structural parameters is learned from data collected in previously explored environments. As the robot moves through a new environment, it estimates the posterior over the structure of this environment. Sufficient statistics for the posterior over multinomials are given by the Dirichlet prior and the frequency counts of view transitions observed in the new environment. In the next section we show how this predictive model can be used in the context of multi­ robot map merging. 3 Generative Model for Map Merging As described in Section I, the multi-robot map merging problem is a particularly difficult instance of the revisiting problem. Imagine two robots exploring an environment from different, unknown start locations. As soon as they can communicate via wireless connection, the robots try to determine whether they can merge their maps by estimating the relative offset between the maps (the robots can not see each other). To do so, one robot transmits the sensor data it collected so far and the other robot estimates the location of this robot relative to its own, partial map. Once the rel­ ative offset between the maps is determined, map merging can be performed by a mapping algorithm such as [14]. Existing approaches to map merging assume knowledge about the robots' relative start locations [5, 15, 14]. At the minimum, these techniques require that one robot is known to start in the map built by the other robot. In this case, map merging can be solved by localizing one robot in the other robot's map using a localization approach capable of global localization [6]. To the best of our knowledge, map merg­ ing has not been addressed for completely unknown start locations including a chance that the partial maps do not overlap at all. Since the map merging problem is closely related to robot localization, we start with a brief discus­ sion of Bayes filters for localization. 554 STEWART ET AL. UA/2003 3.1 Bayes Filters for Robot Localization Consider the recursive Bayes filter, which underlies virtu­ ally all probabilistic robot localization techniques [6]: p(x, I zu,uu-d cx:p(zt I Vx,) · J p(x, I Xt-1,Ut-1) p(Xt-1 I zu-1,uu-2) d.xt-1· (5) Here Xt denotes the position of the robot at timet, typically given in continuous two-dimensional Cartesian coordinates and orientation. Z1:t is the history of all sensor measure­ ments obtained up to time t, and Ul:t-1 is the control in­ formation. In robot localization the term p(xt l xt-1• Ut-1) is a probabilistic model of robot motion. vx, denotes the expected view, or observation, given a map of the envi­ ronment and the robot's location x, therein. p(zt I Vx,) describes the likelihood of making observation Zt given that the robot is expected to observe view vx,. In a nut­ shell, the Bayes filter recursively updates a posterior over the robot's location whenever the robot moves or new sen­ sor information is available. Sensor observations Zt are in­ corporated by multiplying the probability of each location with the likelihood p( Zt I vx,) of making the observation at this location. Observations are typically obtained from a robot's cameras, ultrasound sensors, or laser range-finders. Posteriors over robot locations can be represented by (mix­ tures of) Gaussians, discrete grids, or samples drawn from the posterior (see [6] for a discussion). In our experiments we use data collected by a laser range-finder and a sample­ based posterior representation. 3.2 Partial Map Localization The Bayes filter described above assumes that a complete map of an environment is known. In the context of esti­ mating a robot's location relative to a partial map, loca­ tions Xt can be both inside and outside the map. This raises the question of how to determine the expected view vx, for positions outside the partial map, i.e. in unexplored areas. Existing approaches to map merging assume that views are uniformly distributed throughout the environment. Such an approach corresponds to using a fixed likelihood for all observations Zt made at locations Xt outside the partial map. Obviously, this technique ignores valuable informa­ tion about the structure of an environment and results in brittle estimates for map merging. We will now show how to use the structural model de­ scribed in Section 2 to estimate the likelihood of observa­ tions outside a partial map. The generative model for our technique is shown in Fig. 3. Here, x, denotes the posi­ tion of the other robot in the partial map at time t (xt is not restricted to positions within the partial map). Just as in regular robot localization, the robot's position x, solely depends on its previous position and the control u,_1. The position determines the expected view Vt. which itself gen­ erates a noisy observation Zt. If Xt is inside the partial map, Figure 3: Generative model for partial map localization. The hyperparameter q, estimates the structure of the environment and emits transition probabilities p(v, I Vt-1). Depending on whether the robot is inside or outside the partial map, views are generated by the structural model or the partial map. then v, can be extracted deterministically from the map. If, however, Xt is outside the explored area, then Vt is not di­ rectly observable and has to be extracted from the structural model of the environment. This model is estimated by the structural parameter q,, as described in the previous sec­ tion. The key idea of our hierarchical model is that the node q, outputs transition probabilities p(v, = i I Vt-1 = j) for views according to (2). These transitions can be used to predict the expected view at time t. 3 According to the model shown in Fig. 3, the posterior over the robot's loca­ tion Xt is given by p(xt I zu, Uu-1) ex: L J J Jp(ztlvt) p(vdq,,x,,v,_!)p(qdqt-1,Vt-1) · tJt,Vt-1 p(x, I Xt-1, Ut-1) ji(x,_!) ji( Vt-1) fi( qt-1) dq,dqt-1 d.xt-1(6) where p(-) is short for p(·l Zl:t-1• ul:t-2). This equation can be simplified significantly if we split the update into two different cases, one for locations inside and one for locations outside the partial map. We will now discuss the two cases. Locations inside the partial map: If x, is in the partial map, then the expected view Vt is uniquely determined by Xt and the partial map, i.e. p(v, I q., x., Vt-1) becomes a Dirac delta function at v, = Vx,. Accordingly, the sum­ mation over Vt and Vt-1 and the integrations over q, and q,_1 collapse and, not surprisingly, it can be shown that (6) becomes identical to the Bayes filter update rule for robot localization in complete maps given in (5). Locations outside the partial map: In this case it is not possible to extract the expected view from the partial map. Rather, v, has to be predicted using the previous view Vt-1 and the structural model encoded in q,. We make the 30bviously, the transitions between Vt-1 and v, also depend on how far the robot moved. In our current implementation we update the view whenever the robot moved two meters, which makes the transition probabilities sufficiently stable. UAI2003 STEWART ET AL. 555 assumption that for locations outside the partial map, Vt is independent of the actual location Xt (it only depends on the previous view and the structure). Thus, the term J J p(v, I Qt, Xt, Vt-1)p(qt I Qt-1, Vt-1)jj(q,_I)dq,dqt-1 in ( 6) can be solved analytically for our Dirichlet model described in Section 2.1. As shown in (I), the posterior over the structural parameter q, can be computed by incre­ menting the transition frequency count of the most recently observed view transition. The views used for the transi­ tion counts are those that are most likely to have generated the raw observations z,_1 and z,. Once q, is updated, the predictive probability for v, is computed by normalization of the obtained counts, as given in (2). To emphasize the simplicity of these update steps, we replace the double inte­ gration term by p(v, I Vt-1, a, j,), where a and ft are the Dirichlet prior and the frequency counts used for the pos­ terior over the map structure at time t. These modifications yield the following, more simple update rule for locations outside the partial map. p(x, izu,uu-I) ex: L jp(z,lv,)p(v,lv,_,a,f,)· Vt,Vt-1 p(x, lxt-1, Ut-I) jj(Xt-I) p(v,_!) dx,_, (7) To summarize, the key idea of our approach to map merg­ ing is to sequentially estimate a robot's location both inside and outside the partial map built by the other robot. Lo­ cations inside the map are updated using ( 5) and locations outside the map are updated based on (7). To estimate the likelihood of observations outside the map, the technique estimates a structural parameter q, along with the robot's location. At each iteration, this parameter is updated us­ ing the frequency counts based on the most likely views extracted from the observations. 4 Implementation 4.1 Particle filter for partial map localization The generative model for map merging is implemented us­ ing a particle filter [6, 3]. A detailed description of this implementation can be found in [9]. Particle filters repre­ sent posteriors over a robot's continuous position by sets S, = { (x)il, wii)) I i = 1, ... , N} of N weighted sam­ ples distributed according to the posterior. Here each xii) is a sample (or state), and the wi i) are non-negative nu­ merical factors called importance weights, which sum up to one. Sets at time t are generated from previous sets St-1 by a sampling procedure often referred to as SISR, se­ quential importance sampling with re-sampling [3]. SISR implements the recursive Bayes filter update rule (5) in a three stage process: First, draw states x)O1 from the pre­ vious sample set with probability given by the importance weights w)O1, then draw for each such state a new state from the predictive distribution p(xt I x)M1, Ut-1), and fi- nally weight these new states/samples proportional to the observation likelihood p(z, I v,,) . The last step, impor­ tance sampling, adjusts for the fact that samples are not drawn from the actual posterior distribution but from the predictive distribution. The generative model for map merging described in the previous section requires to estimate the posterior over robot locations both inside and outside the partial map. We assume that the size of the area outside the partial map can be set based on an estimate of the total size of the environ­ ment. Clearly, a representation of all locations outside the map would require too many samples for online estimation. Our solution to this problem is based on the idea that, along with its history, a sample can be seen as the end point of a robot trajectory. This allows us represent only those sam­ ples (trajectories) for which the robot was inside the partial map at some point in time. To do so, our approach initially generates samples uniformly distributed inside the partial map. At later iterations, samples enter and exit the map, de­ pending on their location and the robot's motion (see [9]). At each iteration, the samples inside the map are weighted by p(zt I v,,), i.e. likelihood of the observation given the robot's position in the partial map. All samples outside the partial map are weighted by p(zt I outside), the likelihood of the observation computed from the structural model: Vt,Vt-1 This term for the importance weight follows directly from (7). The structural parameter q, and the distribution over views Vt is updated as described in Section 2. After each iteration, the samples represent a robot's location rel­ ative to the partial map built by another robot. Each sample along with its history represents a unique match between the partial maps built by the two robots. Once a match with sufficiently high probability is found, map merging can be performed by a mapping algorithm such as [14]. The par­ tial map localization algorithm is highly efficient and can be computed in real time on a state-of-the-art laptop. 4.2 View extraction To test our approach using data collected by real robots we have to extract discrete views from sensor data. Since this is not the current focus of our work, we implemented a rather simple technique that extracts structural information from laser range-scans. To do so, the approach sequen­ tially evaluates the individual beams of a laser scan and checks for differences between neighboring beams. De­ pending on their relationship, consecutive beams are clus­ tered into groups denoted w, g, m, and c. Group w (for wall or fiat obstacle) is assigned to groups of beams for which all neighboring beams measure similar distances, g for large gaps between two beams, m for max range readings, and c for comers (based on lines extracted from the scan). Thus 556 STEWART ET AL. UAI2003 ...... > ' ,' '. lL ; ' ï" Ill '" ' • ! • •• • ill Oburved Vhw " (a) ' &nent'View 1}< " (b) 0 &nant'vhw \1, " (d) Figure 5: Learned models: (a) Observation model. (b}--{d) View transitions q. They-axes represent views v,_, and the x-axes give the following view Vt. Shown are only the 15 most frequent views, higher probabilities are darker. (b) Prior a extracted from all maps, (c) posterior for map 4, and (d) posterior for map k in Fig. 2. . a!l•- _ w__;::::_=;:' ] . ' j . r"- L .. _______ ·---------------¸-=---·-····-····-.1 m m w 00 00 Figure 4: Two laser scans corresponding to the most frequently observed views. The robot is on the left side moving to the right. (a) wmw, is typically observed when a robot moves down a hall­ way. (b) wmwgw, indicates that the robot approaches an opening (gap) on its left. each laser scan is represented as a string of these four let­ ters. Fig. 4 shows two example laser scans along with the corresponding feature strings (counterclockwise). The key advantage of this model is that it is extremely ro­ bust in capturing the main structural elements of an envi­ ronment such as hallways, junctions, rooms, and comers. Furthermore, the detected features are robust with respect to rescaling (e.g. different widths of hallways). A disadvan­ tage of these views is that they do not provide accurate loca­ tion information. We overcome this problem by weighting samples inside the partial map using the raw laser scans, which provide highly accurate location information [6]. This technique has no impact on our solution to the revisit­ ing problem, since these samples are still weighted against samples outside using the views as described in the previ­ ous section. The parameters of the model were hand-tuned so as to get satisfying results. After merging symmetric views, in the 35,000 laser scans collected in the environments shown in Fig. 3, only v = 37 different scan "strings" occurred. 5 Experiments The experiments were carried out using data collected in the five environments shown in Fig. 3. 5.1 Learning structural models To Jearn structural models, we used 35,000 pairs of con­ secutive views (strings) collected by mobile robots when mapping the different environments. The parameters of the learned modes are shown in Fig. 5. Each graph plots the probability matrix for the 15 most frequent views. These 15 views cover approximately 80% of all observed scans . View I and 2 are the strings wmw and wmwgw, illustrated in Fig. 4(a) and (b), respectively. Fig. 5(a) shows the obser­ vation model p(z;lvi) extracted from the data. This model was learned using the same hierarchical approach as the one described in Section 2.2 for map structures. In this context the hyperparameters smooth the extracted counts of p( z; I vi). The high probabilities on the diagonal indicate that our view extraction is very robust. The prior transition model a extracted from all maps is shown in Fig. 5(b ). Not surprisingly, most views have a high probability to transi­ tion to the hallway view 1, since the training environments contain many long corridors. When comparing the poste­ riors shown in Fig. 4( c) and (d), it becomes clear that the approach was able to extract the fact that environment 4 has far less hallways than environment k in Fig. 2. 5.2 Partial map localization We systematically evaluated our approach to map merg­ ing under global uncertainty using the following scenario. Imagine two robots are placed at random locations in an unknown environment. Both robots start to explore the en­ vironment and at some point they can communicate. At that point, one robot localizes the other robot in its own map so as to determine whether there is an overlap be­ tween the two maps. We generated 15 partial maps based on data collected in the three environments labeled 2, 4, and k in Fig. 2. Some of these maps are shown in Fig. 6. In our scenario, one robot used these partial maps to localize the other robot based on data collected in the same environ­ ment. For each environment, we generated a prior struc­ tural model a based on the other environments only. The data of the other robot consisted of 25 data sequences for each environment, resulting in a total of 5 · 25 = 125 map­ trajectory pairs for each environment. The results given be­ low are averaged over the average performances in the three different environments. For each pair we proceeded as fol­ lows. Let A denote the robot with the partial map and B the other robot. At each iteration of the particle filter, robot A determines the probability of the most likely hypothesis for B's position in its map. A considers a hypothesis to be valid if its probability exceeds a certain threshold e. The solid line in Fig. 7 shows the resulting precision-recall UAI2003 STEWART ET AL. 557 t:;.1 n n PJ ,.· 11 r1 Cl  F.,  Ul IB IJ II  Figure 6: Partial maps used for evaluation of map merging. The maps were taken from three different environments. In each experiment, one robot built such a partial map and receives data from another robot collected in the same environment. The robot has to determine when and if so where the other robot is in its partial map. 0.8 " 0 0.6 ., ·‹ 0. 0.4 0.2 0 0.1 Fix˜d p(zlŒutj · · - - ­ H1erarcfuca  0.4 Recall 0.6 Figure 7: Precision vs. recall: Each point represents an average over 375 pairs of partial maps and trajectories. Each curve shows the trade-off for different thresholds e (0.05-0.99). The dashed lines indicate results obtained with different fixed values for p(z I outside) and the solid line represents the results of our approach. trade-off for different thresholds using our approach (each point on the line represents a different threshold). For each threshold (},precision measures the fraction of the correct matches among those hypotheses that are considered valid, i.e. above the threshold. Correctness is tested by compar­ ing the position of the hypothesis to a ground truth estimate computed offline. To determine recall, we first checked at what times robot B was in robot A's partial map. Recall, then, measures the fraction of this time for which robot A generated a correct hypothesis, i.e. at the correct posi­ tion and with probability above the threshold e. To show the advantage of our approach, we compared it to an al­ ternative method that uses a fixed likelihood p(z1loutside) for locations outside the partial map (compare to (8) for our approach). The trade-offs resulting from different fixed likelihoods are plotted as dashed lines in Fig. 7 (data points are omitted for clarity). The graph clearly shows the supe­ rior performance of our approach. It achieves 26% higher precision than the best likelihood value for the alternative method. Note that high precision values are more impor­ tant than high recalls since low precision results in wrong map merges while low recall only delays the map merging decision. Note also that one cannot expect very high re­ call values since robot B has to be in the partial map for a certain duration before a valid hypothesis can be generated. Fig. 8 shows the same evaluation for different ways of up­ dating and learning map structures. The dashed line de­ noted by "Frequency" represents the results obtained with­ out considering the transition model for views. This ap­ proach uses frequency counts obtained from the training maps to compute the likelihood p( z1 I outside) of a view. This likelihood is computed by dividing the number of 0.8 " 0.6 -¶ ·‹ 0. 0.4 0.2 0 u U.2 Hierarchical  Extrapolate - Prior only Frequency - - - 0.4 Recall 0.6 Figure 8: Precision vs. recall for different adaptive techniques. The black, dashed line is obtained when using pure frequency counts for the individual features. times this view was observed in the training data by the total number of observations. The bad performance of this method confirms our belief that it is crucial to consider the connective structure of environments as modeled by our Dirichlet process. The dotted line represents the results ob­ tained without updating the structural parameter q during map merging, i.e. q1 is set to the prior a. It can be seen that adjusting the estimation process during map merging increases the robustness of the approach. Finally, the short dashed, best curve shows a variant of our hierarchical ap­ proach that weights the observed frequencies proportional to the ratio between the size of the partial map and the size of the entire environment. In essence, this approach extrap­ olates the observations made in the partial map assuming that the unexplored areas have the same structure. In these experiments we only tested the quality of the es­ timation process underlying the decision problem in multi­ robot map merging. Our current project aims to field l 00 robots in an indoor exploration and reconnaissance task. To achieve maximum robustness against false positive map merges, our multi-robot control system additionally veri­ fies the hypotheses generated by the partial localization ap­ proach described here. Robots verify a match hypothesis by meeting at a location that follows from the hypothesis. The integration of this approach into a decision-theoretic robot exploration strategy is described in [9]. 6 Conclusions and Future Work In this paper, we introduced a novel approach to addressing the revisiting problem in mobile robot map building. Multi­ robot map merging, a particularly difficult instance of this problem, requires the localization of one robot relative to a partial map built by another robot. The key problem in 558 STEWART ET AL. UAI2003 map merging without knowledge about the robot's relative locations is to get accurate estimates for the likelihoods of observations outside the partial map. To solve this problem, we introduce a structural model of an environment that can be used to predict the observations made by the robot. The structural model is a hidden Markov model that generates sequences of views observed by a robot when navigating through the environment. The parameters of the model are updated during exploration via Dirichlet hyperparameters. A Dirichlet prior is learned from previously encountered environments. The structural model is integrated into a particle filter that uses samples to represent a robot's location and that up­ dates the structural parameters as more data becomes avail­ able. Extensive experiments show that our approach sig­ nificantly outperforms an alternative technique that uses a fixed likelihood for observations outside the partial map. We were not able to find a likelihood that yielded results comparable to our method. The approach presented here can be readily applied to the loop closing problem in single robot mapping (see Fig. l(b)). Here, a robot has to decide whether it came back to a previously explored location, or whether it moves through a similar, unexplored area. Especially map­ ping approaches based on Rao-Blackwellised particle fil­ ters [2, 12, 4] can easily incorporate our structural modeL Just like in multi-robot map merging, the model can then be used to assign appropriate probabilities to location hy­ potheses (particles) in unexplored areas. Despite these encouraging results, this is only the first step towards using structural models of environments. For ex­ ample, our current approach uses maximum likelihood esti­ mates to update the parameters of the modeL More sophis­ ticated EM-based techniques such as [1, 16] might yield further improvements. Other areas for improvement are better algorithms for extracting views from sensor data. Another application of our method is to improve robot ex­ ploration strategies by predicting partial maps into unex­ plored areas. Thereby, for example, a robot can actively try to close loops so as to improve map quality. We consider hierarchical Bayesian techniques such as the one used in this paper to be an extremely promising tool for achieving more robust estimation and reasoning processes in robotics. Most existing approaches to state estimation in robotics are fixed in that they do not adapt to the envi­ ronment. For example, if a map building approach is based on the assumption that the environment is rectilinear it will fail in environments that violate this assumption. On the other hand, not making use of the fact that most environ­ ments are rectilinear obviously discards valuable informa­ tion. Using a hyperparameter modeling the type of environ­ ment, a mapping approach would work reliably in different types of environments while still being able to make use of the structure underlying a specific environment. Acknowledgments T his work has partly been supported by the NSF under grant number IIS-0093406 and by DARPA's SDR and MICA programs (contract numbers NBCHC020073 and AFRL F30602-0l¨- 0219). The authors would also like to thank Aaron Hertzmann, Geoff Gordon, and Sebastian Thrun for stimulating discussions. References [I] M.J. Beal, Z. Ghahramani, and C.E. Rasmussen. The infi­ nite hidden Markov model. In Advances in Neural Informa­ tion Processing Systems (NIPS), 2002. [2] A. Doucet, J.F.G. de Freitas, K. Murphy, and S. Russell. Rao-Blackwellised particle filtering for dynamic bayesian networks. In Proc. of the Conference on Uncertainty in Ar­ tificial Intelligence (UAI), 2000. [3] A. Doucet, N. de Freitas, and N. Gordon, editors. Sequen­ tial Monte Carlo in Practice. Springer-Verlag, New York, 2001. [4] A. Eliazar and R. Parr. DP-SLAM: Fast, robust simultane­ ous localization and mapping without predetermined land­ marks. In Proc. of the International Joint Conference on Artificial Intelligence (IJCAI), 2003. [5] J.W. Fenwick, P.M. Newman, and J.J. Leonard. Cooperative concurrent mapping and localization. In Proc. of the IEEE International Conference on Robotics & Automation, 2002. [6] D. Fox. Adapting the sample size in particle filters through KLD-sampling. International Journal of Robotics Research (IJRR), 22, 2003. [7] A. Gelman, J. B. Carlin, H. S. Stern, and D. B. Rubin. Bayesian Data Analysis. Chapman and Hall!CRC, 1997. [8] J.S. Gutmann and K. Konolige. Incremental mapping of large cyclic environments. In Proc. of the IEEE In­ ternational Symposium on Computational Intelligence in Robotics and Automation (CIRA), 2000. [9] J. Ko, B. Stewart, D. Fox, K. Konolige, and B. Limketkai. A practical, decision-theoretic approach to multi-robot map­ ping and exploration. Technical Re­ port UW-CSE-03-05-02, University of Washington, 2003. http://www.cs.washington.edu/robotics/exploration. [10] D.J.C. MacKay and L.C. Bauman Peto. A hierarchical Dirichlet language modeL Natural language engineering, 1(3), 1995. [II] T. Minka. Estimating a Dirichlet distribution. Technical report, MIT, 2000. [12] M. Montemerlo and S. Thrun. Simultaneous localization and mapping with unknown data association using Fast­ SLAM. In Proc. of the IEEE International Conference on Robotics & Automation, 2003. [13] J. Neira and J.D. Tard6s. Data association in stochastic mapping using the joint compatibility test. IEEE Transac­ tions on Robotics and Automation, 17(6), 2001. [14] S. Thrun. A probabilistic online mapping algorithm for teams of mobile robots. International Journal of Robotics Research, 20(5), 2001. [15] S.B. Williams, G. Dissanayake, and H. Durrant-Whyte. Towards multi-vehicle simultaneous localisation and map­ ping. In Proc. of the IEEE International Conference on Robotics & Automation, 2002. [16] E.P. Xing, M. Jordan, R.M. Karp, and S. RusselL A hierar­ chical Bayesian Markovian model for motifs in biopolymer sequences. In Advances in Neural Information Processing Systems (NIPS), 2003.