Airborne SLAM Using High-Speed Vision : The construction of a fully self-contained micro air vehicle localized and controlledusing computer vision based Simultaneous Localization And Mapping.

University essay from KTH/Skolan för datavetenskap och kommunikation (CSC)

Author: Daniel Henell; [2013]

Keywords: ;

Abstract: A helicopter platform was built where the all the controls, localization and other calculations are performed onboard the helicopter making it fully self-contained. The localization is made only by using a monocular camera (with an option to use a stereo pair for easier initialization) and processing the video feed with computer vision algorithms. The helicopter’s pose is estimated by a computer vision algorithm which is an extended version of PTAM, a Simultaneous Localization and Mapping (SLAM) algorithm published 2007 by G. Klein and D. Murray. The program was changed to be able to track different kinds of self-similar ground textures and to be integrated with the helicopter hardware. The algorithm was also modified to be able to auto-initialize and to keep the map size constant by pruning out far away key frames to not be confined only to small areas. The impact on tracking using high-speed vision at 60 Hz was investigated and compared to tracking at 30 Hz, respectively 10 Hz. The impact was not as big as hypothesized. Tracking stability increases a lot when going from 10 Hz to 30 Hz video. However increasing the frame rate from 30 Hz to 60 Hz has a very small effect. In 60 Hz the difference between frames becomes smaller, but does not seem to affect the tracking stability very much. The reason for this is most likely that 30 Hz is adequate for the velocities in which the helicopter flies and the limiting factor is the algorithm in itself that it cannot track in every possible setting, and this will not be fixed by increasing the frame rate further but will require changes in the algorithm. The computer vision localization works well as long as there are good salient features to track. The tracking accuracy in such cases is measured to have a RMS error of 2.4 cm compared to motion capture data that can be assumed to be ground truth.

  AT THIS PAGE YOU CAN DOWNLOAD THE WHOLE ESSAY. (follow the link to the next page)