Enhanced Indoor Visual Navigation using Sensor Fusion and Semantic Information

Public Deposited
Resource Type
Creator
Abstract
  • Accurate and robust indoor navigation systems are crucial in fields like robotics and autonomous vehicles. In the absence of an absolute positioning system like GPS, there is no single sensor that can provide an accurate and robust indoor navigation solution. The presented thesis tackles the indoor navigation challenge using two approaches; multi-sensor fusion and semantic information. In the first approach, visual odometry is enhanced by the fusion of inertial sensors and wireless ranging measurements. The fusion filter is based on Extended Kalman Filter (EKF). Stereo vision can provide 3D positioning by triangulating visual features. However, depth estimation errors and expensive computation are key challenges. The developed multi-sensor system has dual-mode where stereo vision is applied first to estimate inertial sensor biases. Once converged, the estimated biases help the system to switch to a monocular mode which reduces the system complexity and enables the tracking of faster movements with higher frame rates. As both visual and inertial tracking are drifting solutions, wireless ranging/positioning is integrated into the system to provide absolute global positioning and ensure overall accuracy. In the second approach, an improved Visual Simultaneous Localization and Mapping (VSLAM) solution using semantic segmentation and layout estimation is developed. The system utilizes advanced semantic segmentation and indoor layout estimation to optimize map representation and increase positioning accuracy. A testbed has been developed to collect indoor multi-sensor data and to perform experiments and analysis. Out of this thesis work, three conference papers were published, one journal paper was published, in addition to one journal paper and one conference to be submitted.

Subject
Language
Publisher
Thesis Degree Level
Thesis Degree Name
Thesis Degree Discipline
Identifier
Rights Notes
  • Copyright © 2022 the author(s). Theses may be used for non-commercial research, educational, or related academic purposes only. Such uses include personal study, research, scholarship, and teaching. Theses may only be shared by linking to Carleton University Institutional Repository and no part may be used without proper attribution to the author. No part may be used for commercial purposes directly or indirectly via a for-profit platform; no adaptation or derivative works are permitted without consent from the copyright owner.

Date Created
  • 2022

Relations

In Collection:

Items