Get 20M+ Full-Text Papers For Less Than $1.50/day. Start a 14-Day Trial for You or Your Team.

Learn More →

An improved binocular visual odometry algorithm based on the Random Sample Consensus in visual navigation systems

An improved binocular visual odometry algorithm based on the Random Sample Consensus in visual... PurposeThe purpose of this paper is to propose a binocular visual odometry algorithm based on the Random Sample Consensus (RANSAC) in visual navigation systems.Design/methodology/approachThe authors propose a novel binocular visual odometry algorithm based on features from accelerated segment test (FAST) extractor and an improved matching method based on the RANSAC. Firstly, features are detected by utilizing the FAST extractor. Secondly, the detected features are roughly matched by utilizing the distance ration of the nearest neighbor and the second nearest neighbor. Finally, wrong matched feature pairs are removed by using the RANSAC method to reduce the interference of error matchings.FindingsThe performance of this new algorithm has been examined by an actual experiment data. The results shown that not only the robustness of feature detection and matching can be enhanced but also the positioning error can be significantly reduced by utilizing this novel binocular visual odometry algorithm. The feasibility and effectiveness of the proposed matching method and the improved binocular visual odometry algorithm were also verified in this paper.Practical implicationsThis paper presents an improved binocular visual odometry algorithm which has been tested by real data. This algorithm can be used for outdoor vehicle navigation.Originality/valueA binocular visual odometer algorithm based on FAST extractor and RANSAC methods is proposed to improve the positioning accuracy and robustness. Experiment results have verified the effectiveness of the present visual odometer algorithm. http://www.deepdyve.com/assets/images/DeepDyve-Logo-lg.png Industrial Robot: An International Journal Emerald Publishing

An improved binocular visual odometry algorithm based on the Random Sample Consensus in visual navigation systems

Loading next page...
 
/lp/emerald-publishing/an-improved-binocular-visual-odometry-algorithm-based-on-the-random-lJkpT0ytmS
Publisher
Emerald Publishing
Copyright
Copyright © Emerald Group Publishing Limited
ISSN
0143-991X
DOI
10.1108/IR-11-2016-0280
Publisher site
See Article on Publisher Site

Abstract

PurposeThe purpose of this paper is to propose a binocular visual odometry algorithm based on the Random Sample Consensus (RANSAC) in visual navigation systems.Design/methodology/approachThe authors propose a novel binocular visual odometry algorithm based on features from accelerated segment test (FAST) extractor and an improved matching method based on the RANSAC. Firstly, features are detected by utilizing the FAST extractor. Secondly, the detected features are roughly matched by utilizing the distance ration of the nearest neighbor and the second nearest neighbor. Finally, wrong matched feature pairs are removed by using the RANSAC method to reduce the interference of error matchings.FindingsThe performance of this new algorithm has been examined by an actual experiment data. The results shown that not only the robustness of feature detection and matching can be enhanced but also the positioning error can be significantly reduced by utilizing this novel binocular visual odometry algorithm. The feasibility and effectiveness of the proposed matching method and the improved binocular visual odometry algorithm were also verified in this paper.Practical implicationsThis paper presents an improved binocular visual odometry algorithm which has been tested by real data. This algorithm can be used for outdoor vehicle navigation.Originality/valueA binocular visual odometer algorithm based on FAST extractor and RANSAC methods is proposed to improve the positioning accuracy and robustness. Experiment results have verified the effectiveness of the present visual odometer algorithm.

Journal

Industrial Robot: An International JournalEmerald Publishing

Published: Jun 19, 2017

References