A Robust Approach to Vision-Based Terrain-Aided Localization

Dan Navon, Ehud Rivlin, and Hector Rotstein

Peer Reviewed

Abstract: Terrain-aided navigation, which combines radar altitude with a digital terrain map (DTM), was developed before the era of the Global Positioning System to prevent error growth resulting from inertial navigation. Recently, cameras and substantial computational power have become ubiquitous in flying platforms, prompting interest in studying whether the radar altimeter can be replaced by a visual sensor. This paper presents a novel approach to vision-based terrain-aided localization by revisiting the correspondence and DTM (C-DTM) problem. We demonstrate that we can simplify the C-DTM problem by dividing it into a structure-from-motion (SFM) problem and then anchoring the solution to the terrain. The SFM problem can be solved using existing techniques such as feature detection, matching, and triangulation wrapped with a bundle adjustment algorithm. Anchoring is achieved by matching the point cloud to the terrain using ray-tracing and a variation of the iterative closest point method. One of the advantages of this two-step approach is that an innovative outlier filtering scheme can be included between the two stages to enhance overall robustness. The resulting algorithm consistently demonstrated high precision and statistical independence in the presence of initial errors across various simulations. The impact of different filtering methods was also studied, showing an improvement of 50% compared with the unfiltered case. The new algorithm has the potential to improve localization in real-world scenarios, making it a suitable candidate for pairing with an inertial navigation system and a Kalman f ilter to construct a comprehensive navigation system.
Video Abstract: