Affiliation:
1. Electromechanical Engineering Technology , Abu Dhabi Polytechnic , AbuDhabi , UAE
2. School of Engineering and Information Technology , The University of New South Wales , Canberra , ACT , Australia
Abstract
Abstract
The novel framework for estimating dense scene flow using depth camera data is demonstrated in the article. Using these estimated flow vectors to identify obstacles improves the path planning module of the autonomous vehicle's (AV) intelligence. The primary difficulty in the development of AVs has been thought to be path planning in cluttered environments. These vehicles must possess the intelligence to recognize their surroundings and successfully navigate around obstacles. The AV needs a thorough understanding of the surroundings to detect and avoid obstacles in a cluttered environment. Therefore, when determining the course, it is preferable to be aware of the kinematic behavior (position and the direction) of the obstacles. As a result, by comparing the depth images between different time frames, the position and direction of the obstacles are calculated using a 3D vision sensor. The current study focuses on the extraction of the flow vectors in 3D coordinates from the differential scene flow method. Generally, the evaluation of scene flow algorithms is crucial in determining their accuracy and effectiveness in different applications. The gradient of the vector field snake model, which extracts changes in pixel values in three directions, is combined with the scene flow technique to identify both static and dynamic obstacles. Our goal is to create a single-vision sensor-based real-time obstacle avoidance method based on scene flow estimation. In addition, the common evaluation metrics such as endpoint error (EPE), average angular error (AAE), and standard deviation angular error (STDAE) are used to measure the accuracy of different algorithms in terms of computational errors with the benchmark Middlebury datasets. The proposed technique is validated with different experiments using a Pixel-Mixed-Device (PMD) camera and a Kinect sensor as 3D sensors. Finally, the numerical and experimental results are displayed and reported.
Subject
Electrical and Electronic Engineering,Control and Systems Engineering
Reference30 articles.
1. A survey on rgb-d datasets. Computer Vision and Image Understanding, 222:103489, 2022.
2. Mats Ahlskog. 3D Vision. Master's thesis, Department of Computer Science and Electronics, Mälardalen University, 2007.
3. Alefs, Bram, David Schreiber, and Markus Clabian. Hypothesis based vehicle detection for increased simplicity in multi sensor. In IEEE Intelligent Vehicles Symposium, pages 261–266, 2005.
4. J.L. Barron, D.J. Fleet, S.S. Beauchemin, and T.A. Burkitt. Performance of optical flow techniques. In IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Proceedings CVPR ’92, pages 236–242, June 1992.
5. N. Bauer, P. Pathirana, and P. Hodgson. Robust Optical Flow with Combined Lucas-Kanade/Horn-Schunck and Automatic Neighborhood Selection. In Information and Automation, 2006. ICIA 2006. International Conference on, pages 378–383, December 2006.