A SLAM Model Based on Omni-Directional Vision and Odometer

Abstract:

Article Preview

A new SLAM measurement model based on omni-directional vision and odometer is proposed in this paper. A virtual stereo vision composed of an omni-directional vision sensor and an odometer. Scale Invariant Feature Transform is used to extract stable and available vision features from the omni-images. The 3-D locations of these features are initialized by the pixel coordinates and the odometer data by stereo projection, and the locations will be corrected during the SLAM process when they are observed again. It is demonstrated that the new model can make a good accuracy with FastSLAM algorithm, and the accuracy is greatly improved corresponding to the classical vision sensor.

Info:

Periodical:

Edited by:

Qiancheng Zhao

Pages:

128-133

DOI:

10.4028/www.scientific.net/AMM.103.128

Citation:

Y. Q. Wang and Y. H. Li, "A SLAM Model Based on Omni-Directional Vision and Odometer", Applied Mechanics and Materials, Vol. 103, pp. 128-133, 2012

Online since:

September 2011

Export:

Price:

$35.00

In order to see related information, you need to Login.

In order to see related information, you need to Login.