This paper proposes a vision-based Semantic Unscented FastSLAM (UFastSLAM) algorithm for mobile service robot combining the semantic relationship and the Unscented FastSLAM. The landmark positions and the semantic relationships among landmarks are detected by a binocular vision. Then the semantic observation model can be created by transforming the semantic relationships into the semantic metric map. Semantic Unscented FastSLAM can be used to update the locations of the landmarks and robot pose even when the encoder inherits large cumulative errors that may not be corrected by the loop closure detection of the vision system. Experiments have been carried out to demonstrate that the Semantic Unscented FastSLAM algorithm can achieve much better performance in indoor autonomous surveillance than Unscented FastSLAM.
1. Introduction
Visual simultaneous localization and mapping (SLAM) uses the cameras as the only exteroceptive sensors to recover a representation of the environment and achieve localization of the robot complemented with information from the proprioceptive sensors with the aim of increasing accuracy and robustness. To the mobile robotic, vision has proved to be an effective and inexpensive sensing device for localization and mapping. Sim et al. solved the SLAM problem with a stereo pair of cameras [1, 2]. Schleicher et al. used a top-down Bayesian method to perform a vision-based mapping process where identification and localization of the natural landmarks from the images were provided by a wide-angle stereo camera [3]. In this paper, a new semantic vision SLAM framework is proposed to improve the performance without increasing the complexity of the algorithms dramatically.
Literature on visual SLAM have focused on feature-based SLAM where a feature could be described by the points with its 2D position (SIFT [4], SURF [5]) or 3D position [6, 7], and also edge segments [8, 9]. But feature extractions from the natural visual scenes were heavily dependent on the environment where the sparse features might be found. These features could be occasionally too few to fully constrain the pose of the robot. Hence the appearance-based SLAM was proposed to represent the recorded images of the environment with prominent features as a whole [10]. Morita et al. reported another novel appearance-based localization approach for outdoor navigation with feature or object learning, recognition, and classification using SVM [11]. However, the usage of rich sensorial information in these appearance-based SLAM solutions has resulted in very time-consuming computation especially for larger-scale environments. To allow real-time operation in more moderately sized environments, one method was proposed to observe the interframe motion of every other corner feature in a visual odometry style [12, 13]. Also, some researchers proposed the method of discovering and incorporating higher level map structure in the forms of lines [14] and planes [15, 16].
Different kinds of maps have been applied in SLAM. Metric maps capture the geometric properties of the environment whereas topological maps describe the connectivity between different locations [17]. Topological maps can represent the environment as a list of the significant places which has simplified the problem of large-scale mapping [18]. However, one limitation of the topological representation was the lack of metric information. So the strategy of mixing the metric and topological information in a single consistent model was proposed [19]. Fernández et al. also developed a hybrid metric-topological algorithm to build a metric map while maintaining a topological graph and to detect loop closures [20]. Thrun and Buecken combined the grid based and topological based methods to map indoor robot environments [21]. Such hybrid algorithms took advantage of the local metric grids for enhanced local planning while avoiding the computation of a complete global grid map. However, these maps are very limited in describing the environment other than distinguishing between occupied and empty areas. In order to explore richer information of the environment, semantic mapping has become a research topic recently. Wolf and Sukhatme proposed a semantic classification method based on HMMs and SVMs to tackle the problem of terrain mapping and activity-based mapping [22]. Ranganathan and Dellaert described a technique to model and recognize the places using objects as the basic semantic concept [23]. Yi et al. proposed a semantic representation and Bayesian model for robot localization using spatial contexts among objects [24, 25]. This paper will take advantage of semantic relationship of features in the visual SLAM framework.
Early work on SLAM was done by Smith et al., where the Extended Kalman Filter (EKF) was applied [26]. Later Doucet et al. introduced the Rao-Blackwellized particle filter (RBPF) as an efficient solution to the SLAM problem which is also called FastSLAM [27]. The Unscented FastSLAM algorithm was then proposed to overcome the drawbacks of FastSLAM where the scaled unscented transformation (SUT) was applied to replace the linearization in the FastSLAM framework [28]. The SLAM solution in this paper will be based on Unscented FastSLAM.
Hence the main contribution of this paper includes a novel Semantic Unscented FastSLAM algorithm to improve accuracy of localization and mapping while maintaining the sparse map for real-time implementation. The semantic relationship and topological metric map are combined to form a new kind of map for SLAM. Few experiments have been carried out for validation of the proposed technique.
The rest of the paper is organized as follows: Section 2 describes the semantic topological metric map and observation model. Framework of the Semantic Unscented FastSLAM is presented in Section 3. The experimental results and discussion are presented in Section 4. The concluding remarks are presented in Section 5.
Semantic topological metric map is defined as the combination of the topological metric map and the semantic relationships between the landmarks where the assumption is that such semantic relationships can be represented by some mathematical equations. The spatial semantic relationship between the available landmarks is always invariant with respect to the robot location. Denote the semantic topological metric map as M and the semantic metric relationship as τ. Figure 1 shows the process of creating the semantic topological map. The procedures are summarized as follows. (i) When a robot starts to move and the first landmark is observed, the semantic topological map M only includes the position vector of the m1, (1). (ii) As the robot moves forward, more landmarks are observed. If there are no semantic relationships between any pair of landmarks, the semantic topological metric map M will be the same as the regular topological metric map. If the number of the observed landmarks is P, the semantic topological map M includes the position vectors of m1,…mp, (2). (iii) When the robot observes landmark P+1, the semantic relationship between landmark P+1 and landmark P, τmp+1,mp, is also found. If all the semantic relationships with the observed landmark P+1 are defined as the set Tsemantic,mp+1, the semantic topological metric map M is then updated with the addition of the semantic metric relationship as in (3)-(4):(1)M=m1,(2)M=m1,…,mp,(3)M=m1,…,mp,mp+1,Tsemantic,mp+1,(4)Tsemantic,mp+1=τmp+1,mp.(iv) When the robot observes landmark P+2, if the semantic relationship between landmarks P+2 and P+1 is found, then the total semantic topological metric map would be(5)M=m1,…,mp,mp+1,mp+2,Tsemantic,mp+2,(6)Tsemantic,mp+2=τmp+2,mp+1.Since more-than-one semantic relationships between different landmarks, (4) and (6), have been observed, the extended new semantic topological relationship will be created, Figure 2, where landmarks P+2, P+1 and P are associated together. The semantic topological metric map at the time being will become(7)Tsemantic,mp+2=τmp+2,mp+1,τmp+2,mp,where τmp+2,mp is the extended semantic relationship between landmark P+2 and landmark P. When τmp+2,mp+1 has the same semantic relationship as τmp+1,mp coincidently, we can associate them together as(8)τmp+2,mp+1=τmp+1,mp⟹τmp+2,mp=τmp+2,mp+1=τmp+1,mp.
The process of creating the semantic map.
Extension of the semantic topological relationship.
2.2. Semantic Observation Model
A semantic observation model is the observation model of the vision sensor with implicit of the semantic relationships. Hence the semantic observation model consists of not only the metric distance σ and the bearing ω of each landmark, but also the semantic metric relationships between different landmarks. The dimension of the semantic observation model could be N+1 where N is the total number of the landmarks observed so far. In this case, the semantic observation model can be represented as(9)zsemantic=σωη1η2⋮ηN-1=x-mN,x2+y-mN,y2tan-1y-mN,yx-mN,xHτmN,m1,mN,x,mN,y,m1,x,m1,yHτmN,m2,mN,x,mN,y,m2,x,m2,y⋮HτmN,mN-1,mN,x,mN,y,mN-1,x,mN-1,y=φX,m,Tsementic,mN,where H(∗) is the mathematical expression of the semantic metric relationship, (x,y) is the coordinates of the current robot pose, and (mN,x,mN,y) is the coordinates of the landmark N observed at the current time period. τmN,m1,τmN,m2,…,τmN,mN-1 are the series of the possible semantic metric relationships associated with mN. The position vector of the robot is defined as X=[x,y,θ]T. m is defined as the position set of all the landmarks observed at the current time as follows:(10)m=m1,…,mNT.
3. Semantic Unscented FastSLAM
Semantic unscented FastSLAM partitions the SLAM posterior into a localization problem and independent landmark position estimation problem conditioned on the robot pose estimate and the semantic metric relationships between the landmarks as follows:(11)pXt,mt,rt∣zt,semantic,ut=pXt∣zt,semantic,ut·∏j=1Npmj,t,Tsemantic,mj∣Xt,zt,semantic,where Xt is the robot pose at time t and r denotes the full semantic metric map at the current time period as follows:(12)rt=Tsemantic,m1,Tsemantic,m2,…,Tsemantic,mNT.Suppose the control vector of the robot is u=[v,w]T where v and w represent the linear and angular velocities of the robot. According to the kinematics of the wheeled mobile robot [29], the motion model is represented as follows:(13)Xt=Xt-1+-vtwtsinθt-1+vtwtsinθt-1+wtΔtvtwtcosθt-1-vtwtcosθt-1+wtΔtwtΔt=fut,Xt-1.
3.1. Robot Pose Estimation
Since particle filter is incorporated into the FastSLAM frame, the following derivation will be associated with only one particle as an example. Then the robot pose at time t for a kth particle can be estimated as (14)pXtk∣zt,semantic,ut~pXtk∣Xt-1k,utpXt-1k∣zt-1,semantic,ut-1,where pXt-1k∣zt-1,semantic,ut-1 is represented by a Gaussian with the mean Xt-1k and covariance Pt-1[k]. The p(Xtk∣Xt-1k,ut) can be predicted in the following according to the motion model of the robot. In order to integrate the robot pose and the map update, the state vector is augmented with a control input and the observation vector as(15)Xt-1ak=Xt-1k,0,0T,Pt-1ak=Pt-1k000Qt000Rt,where Xt-1a[k] is the augmented state vector, Qt is the motion noise covariance and Rt is the observation noise covariance, and Pt-1a[k] is the augmented covariance matrix.
In order to apply the unscented transformation, a symmetric set of 2n+1 sigma points (n is the dimension of the augmented state vector) need to be extracted first as follows [30]: (16)χt-1a0k=Xt-1ak,χt-1aik=Xt-1ak+n+λPt-1ki,fori=1,…,n,χt-1aik=Xt-1ak-n+λPt-1ki-n,fori=n+1,…,2n,where the subscript i means the ith column of a matrix. The λ is computed by λ=α2n+κ-n and α is a small number to avoid the sampling nonlocal effects for high nonlinearities. κ is a scaling parameter determining how far the sigma points are separated from the mean value. Each sigma point χt-1aik contains the robot pose, control noise, and semantic observation noise components as(17)χt-1aik=χt-1ikχtuikχtzik.So the prediction of the robot pose can be derived by passing the above sigma points through the motion model, f in (13). The transformed sigma points of the robot pose, χ-tik, are calculated as (18)χ-tik=futk+χtuik,χt-1ik,where the current control vector is the sum of the ut[k] and the control noise component χtuik of each sigma point. Then the prediction of the robot pose can be calculated as(19)Xt∣t-1k=∑i=02nωbiχ-tik,Pt∣t-1k=∑i=12nωciχ-tik-Xt∣t-1kχ-tik-Xt∣t-1kT.The weights are calculated by the following equations:(20)ωb0=λn+λ,ωc0=λn+λ+1-α2+β,ωci=ωbi=12n+λ,fori=1,…,2n,where the weight ωci is used to compute the mean of the predicted robot pose, and the weight ωbi is used to recover the covariance of the Gaussian. The parameter β is used to incorporate the knowledge of the higher order moments of the posterior distribution.
Suppose the jth landmark and its semantic relationships are observed; the transformed sigma points of the semantic observation vector can be derived as(21)N-tik=φχ-tik,mt-1k,Tsemantic,mjk+χtzik,where the semantic metric relationships, Tsemantic,mj[k], are included in the semantic observation model φ(·) in (9) for robot pose update. So this new update will result in the improvement of robot localization. Then the prediction of the semantic observation vector can be calculated as(22)n^tk=∑i=02nωbiN-tik.The Kalman gain can then be obtained by the following equations as usual: (23)Stk=∑i=02nωciN-tik-n^tkN-tik-n^tkT,ηtx,nk=∑i=02nωciχ-tik-Xt∣t-1kN-tik-n^tkT,Ktk=ηtx,nkStk-1,where Stk is the innovation covariance and ηtx,n[k] is the cross-covariance.
Therefore, the mean and covariance of the robot pose are estimated at the time period t by(24)Xtk=Xt∣t-1k+Ktkzt-n^tk,(25)Ptk=Pt∣t-1k-KtkStkKtkT.
3.2. Landmark Position Estimate with Semantic Constraints
For the observed landmark j, the probability of the landmark position estimate can be represented as(26)pmj,tk,Tsemantpic,mj∣Xtk,zt,semantick~pzt,semantick∣Xtk,mj,tk,Tsemantpic,mj·pmj,t-1k,Tsemantpic,mj∣Xt-1k,zt-1,semantick,where the probability pmj,t-1[k],Tsemantpic,mj∣Xt-1[k],zt-1,semantic[k] is represented by a Gaussian with the mean mj,t-1k and covariance Σj,t-1[k]. pzt,semantic[k]∣Xt[k],mj,t[k],Tsemantpic,mj will be derived in the following. Likewise, the sigma points of the observed landmark position, mj, are initialized as (27)χm,t-10k=mj,t-1k,χm,t-1ik=mj,t-1k+n+λΣj,t-1ki,fori=1,…,n,χm,t-1ik=mj,t-1k-n+λΣj,t-1ki-n,fori=n+1,…,2n.The transformed sigma points of the landmark position estimation with semantic relationships can be derived as (28)Z-t,semanticik=φXtk,χm,t-1ik,Tsemantic,mjk,i=0,…,2n,where φ(·) is the observation model in (9), Xt[k] is the current estimation of the robot pose in (24). Hence the predicted semantic observation vector, z^t,semantic[k], is(29)z^t,semantick=∑i=02nωbiZ-t,semanticik.Then the Kalman gain K-t,semantick is calculated as follows: (30)S-t,semantick=∑i=02nωciZ-t,semanticik-z^t,semantick·Z-t,semanticik-z^t,semantickT+Rt,Σ-t,semantick=∑i=02nωciχmik-mj,t-1k·Z-t,semanticik-z^t,semantickT,K-t,semantick=Σ-t,semantickS-t,semantick-1.Note that the weights ωb[i] and ωci are the same as (20). Finally, the mean mj,tk and the covariance Σj,t[k] of the jth landmark position are updated by(31)mj,tk=mj,t-1k+K-t,semantickzt,semantic-z^t,semantick,Σj,tk=Σj,t-1k-K-t,semantickS-t,semantickK-t,semantickT.Note that zt,semantic includes the true observation of the relative position of the landmark and the robot and the associated semantic relationships with this landmark. These observation values are obtained from the image process of the vision sensor data. If more landmarks are observed at one time, the derivation would be similar except that more semantic relationships would be included in the observation model.
As mentioned at the beginning, all the above derivation is with respect to the particle k. Then the traditional resampling procedure will be taken, and the robot pose and the landmark positions will be estimated finally.
4. Experiments and Discussions4.1. Experiment Procedures
The platform used in the experiments was a Pioneer 3-DX robot equipped with a binocular camera system. The camera was the only exteroceptive sensor to recover the representation of the environment. The sampling period was 0.5 seconds. The proposed technique has been evaluated by three different types of the experiments. In Experiment 1, the robot moved along a simple rectangular trajectory (8 m × 14 m) in a neat lab environment. The environment in Experiment 2 was a regular office area that was more general to most indoor service robots to verify the superior performance of the Semantic Unscented FastSLAM. Experiment 3 was conducted in a messy environment where the robot had to move along a zig-zag path to go through aisles.
In the experiments, three kinds of the semantic metric relationships were found. One semantic relationship was that the new observed landmark and another landmark existing in the previous map were both along the x-axis (x-line). The second semantic relationship was that two landmarks were both along y-axis (y-line). Such two kinds of semantic relationships are denoted by x-line,y-line. The third one was that three landmarks were collinear such as the walls of neighboring cubes in an office. Suppose mp, ms, and mq are three landmarks with the above semantic relationships; the semantic observation model can be represented, respectively, as(32)Hτmp,ms,mp,x,mp,y,ms,x,ms,yτmp,ms=x-line=mp,y-ms,y,Hτmp,ms,mp,x,mp,y,ms,x,ms,yτmp,ms=y-line=mp,x-ms,x,Hτmp,ms,mq,mp,x,mp,y,ms,x,ms,y,mq,x,mq,yτmp,ms,mq=collinear=mp,y-ms,ymp,x-mq,x-mp,y-mq,ymp,x-ms,x.
4.2. Experiment Results and DiscussionsExperiment 1.
Figure 3 shows one image taken by the vision sensor on the robot with three landmarks P, P+1, and P+2 where the landmarks P+2 and P+1 were located along the x-axis. This semantic relationship will be applied for localization and mapping. Figure 4 shows the comparison of the system performance using the Unscented FastSLAM (Figure 4(a)) and the proposed method (Figure 4(b)). As shown in Figure 4(a), the error of robot pose became larger especially after the robot was turning. This error could not be corrected by the loop closure detection because all the landmarks observed after turning have not been observed before. In Figure 4(b), the localization error has been eliminated greatly after the semantic topological metric map was applied. Figure 5 is the partially enlarged view of Figure 4 where A1 and B1 are the estimations by odometer only, A2 and B2 are the estimation from Unscented FastSLAM, and A3 and B3 are the estimation from the proposed Semantic Unscented FastSLAM. When Landmark #6 was observed by the robot at the first time, it was also found that Landmark #6 has the semantic topological relationship “y-line” with the landmarks #4, #2, and #3 in the existing map. Hence this semantic relationship has resulted in much better robot pose estimation, B3 in Figure 5(b), which has pulled the dead reckoning estimate B1 back from the deviation comparing with B2 without taking advantage of semantic relationships.
The observation result of vision system.
Comparison of SLAM results, Experiment 1.
Experiment result using Unscented FastSLAM
Experiment result using Semantic Unscented FastSLAM
Partially enlarged view of SLAM results in Experiment 1.
Experiment result using Unscented FastSLAM
Experiment result using Semantic Unscented FastSLAM
Experiment 2.
Figure 6 shows the experimental environment in Experiment 2 where the reference trajectory started from the circle and ended at the same point after a complex surveillance along the arrow directions. The start point was defined as the origin of the inertial frame (0,0). It is worth noting that this office was composed of a few cubes that were higher than the robot. Hence when the robot moved along the reference trajectory, most landmarks could not be observed more than once before the robot was close to the end point.
The experimental results using the Unscented FastSLAM and the proposed Semantic Unscented FastSLAM are shown in Figure 7. In this experiment, when the robot moved close to the end point, Landmark #1 should be observed after a long period for the loop closure detection. The estimation of the end point, C1, by odometer only was far away from the real end point. As shown in Figure 7(a), the end point estimated by the Unscented FastSLAM before the loop closure detection of Landmark #1, C2, was better but still had a huge error. This error was too large to be corrected by the loop closure detection (see D2 for the estimation after loop closure detection). Figure 7(b) shows that the end point estimated by the proposed Semantic Unscented FastSLAM before the loop closure detection, C3, was much smaller because of the semantic updates in the algorithm. Therefore, after the loop closure detection, the error was reduced close to the reference point (see D3 for the estimation by Semantic Unscented FastSLAM in Figure 7(b)).
Floor map of the office environment, Experiment 2.
Comparison of SLAM results, Experiment 2.
Experiment result using Unscented FastSLAM
Experiment result using Semantic Unscented FastSLAM
Experiment 3.
The experiment environment is shown in Figure 8 where small triangles represent a few locations along the reference path, and the solid line represents the wall of cubes. Notice that the reference path is not straightforward during each aisle because the aisle had irregular width and the robot also needs to avoid chairs and boxes on both sides of the aisle. Figure 9 shows an example of two pictures captured by the camera where three green landmarks were detected as collinear relationship. Figure 10 illustrates the performance of the surveillance robot in Experiment 3. As shown in Figure 10(b), the locations of robot and landmarks are much closer to the reference path using the proposed Semantic Unscented FastSLAM than without considering semantic relationships (Figure 10(a)).
Map of the office environment with reference path, Experiment 3.
Two pictures with three collinear landmarks, Experiment 3.
Comparison of SLAM results, Experiment 3.
Experiment result using Unscented FastSLAM
Experiment result using Semantic Unscented FastSLAM
5. Conclusions
This paper has proposed a vision-based Semantic Unscented FastSLAM for mobile robot. The semantic relationship is combined with the traditional topological metric map to improve the accuracy of localization and mapping. Experiments were conducted to verify that the Semantic Unscented FastSLAM is more robust and applicable to more general indoor autonomous surveillance.
Conflict of Interests
The authors declare that there is no conflict of interests regarding the publication of this paper.
Acknowledgments
This research is currently supported by the National Science and Technology Ministry of China under Grant no. 2013BAK01B02 and was supported by the State Key Laboratory of Robotics and System (HIT) under SKLRS-2011-ZD-04. The partial support of the National Natural Science Foundation of China (no. 61273335) and National High-Tech Research and Development Program of China (863 Program, no. 2015AA042303) is also appreciated.
SimR.ElinasP.GriffinM.ShyrA.LittleJ. J.Design and analysis of a framework for real-time vision-based SLAM using Rao-Blackwellised particle filtersProceedings of the 3rd Canadian Conference on Computer and Robot Vision (CRV '06)June 2006IEEE2110.1109/crv.2006.252-s2.0-33845363600SimR.LittleJ. J.Autonomous vision-based exploration and mapping using hybrid maps and Rao-Blackwellised particle filtersProceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS '06)October 20062082208910.1109/iros.2006.2824852-s2.0-34250624811SchleicherD.BergasaL. M.BareaR.LópezE.OcañaM.Real-time simultaneous localization and mapping using a wide-angle stereo camera and adaptive patchesProceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS '06)October 20062090209510.1109/iros.2006.2824862-s2.0-34250683385FrintropS.JensfeltP.Attentional landmarks and active gaze control for visual SLAM20082451054106510.1109/TRO.2008.20049772-s2.0-56049103514MagnaboscoM.BreckonT. P.Cross-spectral visual simultaneous localization and mapping (SLAM) with sensor handover201361219520810.1016/j.robot.2012.09.0232-s2.0-84874116209WilliamsB.SmithP.ReidI.Automatic relocalisation for a single-camera simultaneous localisation and mapping systemProceedings of the IEEE International Conference on Robotics and Automation (ICRA '07)April 20072784279010.1109/robot.2007.3638932-s2.0-35648983652ChekhlovD.PupilliM.MayolW.CalwayA.Robust real-time visual SLAM using scale prediction and exemplar based feature descriptionProceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR '07)June 2007Minneapolis, Minn, USA1710.1109/cvpr.2007.3830262-s2.0-34948872069EadeE.DrummondT.Edge landmarks in monocular SLAM200927558859610.1016/j.imavis.2008.04.0122-s2.0-62949217778ZhouC.WeiY.TanT.Mobile robot self-localization based on global visual appearance featuresProceedings of the IEEE International Conference on Robotics and AutomationSeptember 2003127112762-s2.0-0344013450BaccaB.SalviJ.CufíX.Appearance-based SLAM for mobile robotsProceedings of the 12th International Conference of the Catalan Association for Artificial Intelligence (CCIA '09)October 2009Cardona, Spain5564MoritaH.HildM.MiuraJ.ShiraiY.Panoramic view-based navigation in outdoor environments based on support vector learningProceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS '06)October 20062302230710.1109/iros.2006.2826362-s2.0-34250671043WilliamsB.ReidI.On combining visual SLAM and visual odometryProceedings of the IEEE International Conference on Robotics and Automation (ICRA '10)May 20103494350010.1109/robot.2010.55092482-s2.0-77955777516Martínez-CarranzaJ.CalwayA.Efficient visual odometry using a structure-driven temporal mapProceedings of the IEEE International Conference on Robotics and Automation (ICRA '12)May 2012Saint Paul, Minn, USAIEEE5210521510.1109/icra.2012.62251002-s2.0-84864469296GeeA. P.ChekhlovD.CalwayA.Mayol-CuevasW.Discovering higher level structure in visual SLAM200824598099010.1109/TRO.2008.20046412-s2.0-56049123363Fernandez-MoralE.Mayol-CuevasW.ArevaloV.Gonzalez-JimenezJ.Fast place recognition with plane-based mapsProceedings of the IEEE International Conference on Robotics and Automation (ICRA '13)May 20132719272410.1109/icra.2013.66309512-s2.0-84887286309KwonJ.LeeK. M.Monocular SLAM with locally planar landmarks via geometric rao-blackwellized particle filtering on lie groupsProceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR '10)June 20101522152910.1109/cvpr.2010.55397892-s2.0-77955995670Fuentes-PachecoJ.Ruiz-AscencioJ.Rendón-ManchaJ. M.Visual simultaneous localization and mapping: a survey2015431558110.1007/s10462-012-9365-82-s2.0-84868605430AngeliA.DoncieuxS.MeyerJ.-A.FilliatD.Visual topological SLAM and global localizationProceedings of the IEEE International Conference on Robotics and Automation (ICRA '09)May 2009Kobe, Japan4300430510.1109/robot.2009.51525012-s2.0-70350352770KonoligeK.Marder-EppsteinE.MarthiB.Navigation in hybrid metric-topological mapsProceedings of the IEEE International Conference on Robotics and Automation (ICRA '11)May 20113041304710.1109/icra.2011.59800742-s2.0-84871681087FernándezL.PayáL.ReinosoO.GilA.ValienteD.Visual hybrid SLAM: an appearance-based approach to loop closure2014252Springer693701Advances in Intelligent Systems and Computing10.1007/978-3-319-03413-3_51ThrunS.BueckenA.Integrating grid-based and topological maps for mobile robot navigationProceedings of the 13th National Conference on Artificial Intelligence (AAAI '96)August 19969449502-s2.0-0030370231WolfD. F.SukhatmeG. S.Semantic mapping using mobile robots200824224525810.1109/tro.2008.9170012-s2.0-42549116685RanganathanA.DellaertF.Semantic modeling of places using objectsProceedings of the Robotics: Science and Systems ConferenceJune 2007Atlanta, Ga, USA2730YiC.SuhI. H.LimG. H.ChoiB.-U.Active-semantic localization with a single consumer-grade cameraProceedings of the IEEE International Conference on Systems, Man and Cybernetics (SMC '09)October 20092161216610.1109/icsmc.2009.53462582-s2.0-74849129863YiC.SuhI. H.LimG. H.ChoiB.-U.Bayesian robot localization using spatial object contextsProceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS '09)October 2009St. Louis, Mo, USAIEEE3467347310.1109/iros.2009.53544622-s2.0-76249115613SmithR.SelfM.CheesemanP.Estimating uncertain spatial relationships in robotics1990Berlin, GermanySpringer167193DoucetA.De FreitasN.MurphyK.RussellS.Rao-blackwellised particle filtering for dynamic Bayesian networksProceedings of the 16th Conference on Uncertainty in Artificial Intelligence (UAI '00)2000176183KimC.SakthivelR.ChungW. K.Unscented FastSLAM: a robust and efficient solution to the SLAM problem200824480882010.1109/tro.2008.9249462-s2.0-50649085675CampionG.BastinG.Dandrea-NovelB.Structural properties and classification of kinematic and dynamic models of wheeled mobile robots1996121476210.1109/70.4817502-s2.0-0030081857JulierS.UhlmannJ.Durrant-WhyteH. F.A new method for the nonlinear transformation of means and covariances in filters and estimators200045347748210.1109/9.847726MR17628592-s2.0-0033723743