Presentation + Paper
12 April 2021 RGB pixel-block point-cloud fusion for object detection
Timothy Foster, Ajaya Dalal, John E. Ball
Author Affiliations +
Abstract
This paper explores a multimodal deep learning network based on SqueezeSeg. We extend the standard SqueezeSeg architecture to enable camera and lidar fusion. The sensor processing method is termed pixel-block point-cloud fusion. Using co-registered camera and lidar sensors, the input section of the proposed network creates a feature vector by extracting information from a block of RGB pixels from each point in the point-cloud that is also in the camera’s field of view. Essentially, each lidar point is paired with neighboring RGB data so the feature extractor has more meaningful information from the image. This fusion method adds rich information on object color and texture from the camera data to enhance the overall performance. The image pixel blocks will not only add color information to the lidar data, but it will also add information about texture. The proposed pixel-block point-cloud fusion method yields better results than single-pixel fusion.
Conference Presentation
© (2021) COPYRIGHT Society of Photo-Optical Instrumentation Engineers (SPIE). Downloading of the abstract is permitted for personal use only.
Timothy Foster, Ajaya Dalal, and John E. Ball "RGB pixel-block point-cloud fusion for object detection", Proc. SPIE 11748, Autonomous Systems: Sensors, Processing, and Security for Vehicles and Infrastructure 2021, 1174804 (12 April 2021); https://doi.org/10.1117/12.2587510
Advertisement
Advertisement
RIGHTS & PERMISSIONS
Get copyright permission  Get copyright permission on Copyright Marketplace
KEYWORDS
LIDAR

RGB color model

Sensors

Clouds

Image fusion

Data fusion

Image processing

Back to Top