Free Space Detection Using Camera-LiDAR Fusion in a Bird's Eye View Plane

Sensors (Basel). 2021 Nov 17;21(22):7623. doi: 10.3390/s21227623.

Abstract

Although numerous road segmentation studies have utilized vision data, obtaining robust classification is still challenging due to vision sensor noise and target object deformation. Long-distance images are still problematic because of blur and low resolution, and these features make distinguishing roads from objects difficult. This study utilizes light detection and ranging (LiDAR), which generates information that camera images lack, such as distance, height, and intensity, as a reliable supplement to address this problem. In contrast to conventional approaches, additional domain transformation to a bird's eye view space is executed to obtain long-range data with resolutions comparable to those of short-range data. This study proposes a convolutional neural network architecture that processes data transformed to a bird's eye view plane. The network's pathways are split into two parts to resolve calibration errors in the transformed image and point cloud. The network, which has modules that operate sequentially at various scaled dilated convolution rates, is designed to quickly and accurately handle a wide range of data. Comprehensive empirical studies using the Karlsruhe Institute of Technology and Toyota Technological Institute's (KITTI's) road detection benchmarks demonstrate that this study's approach takes advantage of camera and LiDAR information, achieving robust road detection with short runtimes. Our result ranks 22nd in the KITTI's leaderboard and shows real-time performance.

Keywords: autonomous vehicle; bird’s eye view transformation; convolutional neural network; heterogeneous sensor fusion; road detection; semantic segmentation.

MeSH terms

  • Neural Networks, Computer*