LAUSR.org creates dashboard-style pages of related content for over 1.5 million academic articles. Sign Up to like articles & get recommendations!

Cross-Modal Knowledge Distillation for Depth Privileged Monocular Visual Odometry

Photo by lukejonesdesign from unsplash

Most self-supervised monocular visual odometry (VO) suffer from the scale ambiguity problem. A promising way to address this problem is to introduce additional information for training. In this work, we… Click to show full abstract

Most self-supervised monocular visual odometry (VO) suffer from the scale ambiguity problem. A promising way to address this problem is to introduce additional information for training. In this work, we propose a new depth privileged framework to learn a monocular VO. It assumes that sparse depth is provided during training time but not available at the test stage. To make full use of the privileged depth information, we propose a cross-modal knowledge distillation method, which utilizes a well-trained visual-lidar odometry (VLO) as a teacher to guide the training of the VO network. Knowledge distillation is conducted at both output and hint levels. Besides, a distillation condition check is also designed to leave out the noise that may be contained in the teacher’s predictions. Experiments on the KITTI odometry benchmark show that the proposed method produces accurate pose estimation results with a recovered actual scale. It also outperforms most stereo privileged monocular VOs.

Keywords: monocular visual; distillation; depth; knowledge distillation; visual odometry

Journal Title: IEEE Robotics and Automation Letters
Year Published: 2022

Link to full text (if available)


Share on Social Media:                               Sign Up to like & get
recommendations!

Related content

More Information              News              Social Media              Video              Recommended



                Click one of the above tabs to view related content.