Author
Listed:
- Yongsheng Wang
(School of Information Engineering, Wuhan University of Technology, Wuhan 430070, China
Current address: Nanhu Campus College Students Innovation Park, Wuhan University of Technology, Wuhan 430070, China.)
- Xiaobo Han
(School of Automation, Wuhan University of Technology, Wuhan 430070, China
Current address: Nanhu Campus College Students Innovation Park, Wuhan University of Technology, Wuhan 430070, China.)
- Xiaoxu Wei
(School of Automotive Engineering, Wuhan University of Technology, Wuhan 430070, China
Current address: Nanhu Campus College Students Innovation Park, Wuhan University of Technology, Wuhan 430070, China.)
- Jie Luo
(School of Automation, Wuhan University of Technology, Wuhan 430070, China)
Abstract
The fusion of camera and LiDAR perception has become a research focal point in the autonomous driving field. Existing image–point cloud fusion algorithms are overly complex, and processing large amounts of 3D LiDAR point cloud data requires high computational power, which poses challenges for practical applications. To overcome the above problems, herein, we propose an Instance Segmentation Frustum (ISF)–PointPillars method. Within the framework of our method, input data are derived from both a camera and LiDAR. RGB images are processed using an enhanced 2D object detection network based on YOLOv8, thereby yielding rectangular bounding boxes and edge contours of the objects present within the scenes. Subsequently, the rectangular boxes are extended into 3D space as frustums, and the 3D points located outside them are removed. Afterward, the 2D edge contours are also extended to frustums to filter the remaining points from the preceding stage. Finally, the retained points are sent to our improved 3D object detection network based on PointPillars, and this network infers crucial information, such as object category, scale, and spatial position. In pursuit of a lightweight model, we incorporate attention modules into the 2D detector, thereby refining the focus on essential features, minimizing redundant computations, and enhancing model accuracy and efficiency. Moreover, the point filtering algorithm substantially diminishes the volume of point cloud data while concurrently reducing their dimensionality, thereby ultimately achieving lightweight 3D data. Through comparative experiments on the KITTI dataset, our method outperforms traditional approaches, achieving an average precision (AP) of 88.94% and bird’s-eye view (BEV) accuracy of 90.89% in car detection.
Suggested Citation
Yongsheng Wang & Xiaobo Han & Xiaoxu Wei & Jie Luo, 2024.
"Instance Segmentation Frustum–PointPillars: A Lightweight Fusion Algorithm for Camera–LiDAR Perception in Autonomous Driving,"
Mathematics, MDPI, vol. 12(1), pages 1-22, January.
Handle:
RePEc:gam:jmathe:v:12:y:2024:i:1:p:153-:d:1312470
Download full text from publisher
Corrections
All material on this site has been provided by the respective publishers and authors. You can help correct errors and omissions. When requesting a correction, please mention this item's handle: RePEc:gam:jmathe:v:12:y:2024:i:1:p:153-:d:1312470. See general information about how to correct material in RePEc.
If you have authored this item and are not yet registered with RePEc, we encourage you to do it here. This allows to link your profile to this item. It also allows you to accept potential citations to this item that we are uncertain about.
We have no bibliographic references for this item. You can help adding them by using this form .
If you know of missing items citing this one, you can help us creating those links by adding the relevant references in the same way as above, for each refering item. If you are a registered author of this item, you may also want to check the "citations" tab in your RePEc Author Service profile, as there may be some citations waiting for confirmation.
For technical questions regarding this item, or to correct its authors, title, abstract, bibliographic or download information, contact: MDPI Indexing Manager (email available below). General contact details of provider: https://www.mdpi.com .
Please note that corrections may take a couple of weeks to filter through
the various RePEc services.