Go to:
Logótipo
Comuta visibilidade da coluna esquerda
Você está em: Start > Publications > View > Multimodal PointPillars for Efficient Object Detection in Autonomous Vehicles
Publication

Publications

Multimodal PointPillars for Efficient Object Detection in Autonomous Vehicles

Title
Multimodal PointPillars for Efficient Object Detection in Autonomous Vehicles
Type
Article in International Scientific Journal
Year
2024
Authors
Oliveira M.
(Author)
Other
The person does not belong to the institution. The person does not belong to the institution. The person does not belong to the institution. Without AUTHENTICUS Without ORCID
Cerqueira R.
(Author)
Other
The person does not belong to the institution. The person does not belong to the institution. The person does not belong to the institution. Without AUTHENTICUS Without ORCID
Pinto J.R.
(Author)
Other
The person does not belong to the institution. The person does not belong to the institution. The person does not belong to the institution. View Authenticus page Without ORCID
Fonseca J.
(Author)
Other
The person does not belong to the institution. The person does not belong to the institution. The person does not belong to the institution. Without AUTHENTICUS Without ORCID
Journal
Pages: 1-11
Publisher: IEEE
Indexing
Other information
Authenticus ID: P-010-JQT
Abstract (EN): Autonomous Vehicles aim to understand their surrounding environment by detecting relevant objects in the scene, which can be performed using a combination of sensors. The accurate prediction of pedestrians is a particularly challenging task, since the existing algorithms have more difficulty detecting small objects. This work studies and addresses this often overlooked problem by proposing Multimodal PointPillars (M-PP), a fast and effective novel fusion architecture for 3D object detection. Inspired by both MVX-Net and PointPillars, image features from a 2D CNN-based feature map are fused with the 3D point cloud in an early fusion architecture. By changing the heavy 3D convolutions of MVX-Net to a set of convolutional layers in 2D space, along with combining LiDAR and image information at an early stage, M-PP considerably improves inference time over the baseline, running at 28.49 Hz. It achieves inference speeds suitable for real-world applications while keeping the high performance of multimodal approaches. Extensive experiments show that our proposed architecture outperforms both MVX-Net and PointPillars for the pedestrian class in the KITTI 3D object detection dataset, with 62.78% in <inline-formula><tex-math notation="LaTeX">$AP_{BEV}$</tex-math></inline-formula> (moderate difficulty), while also outperforming MVX-Net in the nuScenes dataset. Moreover, experiments were conducted to measure the detection performance based on object distance. The performance of M-PP surpassed other methods in pedestrian detection at any distance, particularly for faraway objects (more than 30 meters). Qualitative analysis shows that M-PP visibly outperformed MVX-Net for pedestrians and cyclists, while simultaneously making accurate predictions of cars.
Language: English
Type (Professor's evaluation): Scientific
Documents
We could not find any documents associated to the publication.
Related Publications

Of the same journal

Weather and Meteorological Optical Range Classification for Autonomous Driving (2024)
Article in International Scientific Journal
Pereira, C; Cruz, RPM; Fernandes, JND; Pinto, JR; Jaime S Cardoso
Map-Aided Dead-Reckoning Using Only Measurements of Speed (2016)
Article in International Scientific Journal
Johan Wahlstroem; João G. P. Rodrigues; Isaac Skog; Ana Aguiar; Peter Haendel
Recommend this page Top
Copyright 1996-2025 © Faculdade de Direito da Universidade do Porto  I Terms and Conditions  I Acessibility  I Index A-Z
Page created on: 2025-08-10 at 19:01:54 | Privacy Policy | Personal Data Protection Policy | Whistleblowing