Title :
Vision-based localization algorithm based on landmark matching, triangulation, reconstruction, and comparison
Author :
Yuen, David C K ; MacDonald, Bruce A.
Author_Institution :
Dept. of Electr. & Comput. Eng., Univ. of Auckland, New Zealand
fDate :
4/1/2005 12:00:00 AM
Abstract :
Many generic position-estimation algorithms are vulnerable to ambiguity introduced by nonunique landmarks. Also, the available high-dimensional image data is not fully used when these techniques are extended to vision-based localization. This paper presents the landmark matching, triangulation, reconstruction, and comparison (LTRQ global localization algorithm, which is reasonably immune to ambiguous landmark matches. It extracts natural landmarks for the (rough) matching stage before generating the list of possible position estimates through triangulation. Reconstruction and comparison then rank the possible estimates. The LTRC algorithm has been implemented using an interpreted language, onto a robot equipped with a panoramic vision system. Empirical data shows remarkable improvement in accuracy when compared with the established random sample consensus method. LTRC is also robust against inaccurate map data.
Keywords :
image sensors; mobile robots; position control; robot vision; generic position-estimation algorithms; high-dimensional image data; landmark matching; panoramic vision system; vision-based localization algorithm; Data mining; Image reconstruction; Image sensors; Insects; Machine vision; Mobile robots; Navigation; Robot localization; Robot sensing systems; Robot vision systems; Landmark matching, triangulation, reconstruction, and comparison (LTRC); natural landmark; panoramic image; random sample consensus (RANSAC); triangulation; vision-based localization;
Journal_Title :
Robotics, IEEE Transactions on
DOI :
10.1109/TRO.2004.835452