Global registration of multiview robot data is a challenging task. Appearance-based global localization approaches often fail under drastic view-point changes, as representations have limited view-point invariance. This letter is based… Click to show full abstract
Global registration of multiview robot data is a challenging task. Appearance-based global localization approaches often fail under drastic view-point changes, as representations have limited view-point invariance. This letter is based on the idea that human-made environments contain rich semantics that can be used to disambiguate global localization. Here, we present X-View, a multiview semantic global localization system. X-View leverages semantic graph descriptor matching for global localization, enabling localization under drastically different view-points. While the approach is general in terms of the semantic input data, we present and evaluate an implementation on visual data. We demonstrate the system in experiments on the publicly available SYNTHIA dataset, on a realistic urban dataset recorded with a simulator, and on real-world StreetView data. Our findings show that X-View is able to globally localize aerial-to-ground, and ground-to-ground robot data of drastically different view-points. Our approach achieves an accuracy of up to $\text{85}\,\%$ on global localizations in the multiview case, while the benchmarked baseline appearance-based methods reach up to $\text{75}\,\%$.
Share on Social Media:
  
        
        
        
Sign Up to like & get recommendations! 0
Related content
More Information
            
News
            
Social Media
            
Video
            
Recommended
               
Click one of the above tabs to view related content.