Accurate 3D Object Detection from Point Cloud Data using Bird’s Eye View Representations

Autores: Nerea Aranjuelo Ansa Marcos Nieto Doncel Luis Unzueta Irurtia Oihana Otaegui Madurga David Montero Guus Engels Ignacio Arganda

Fecha: 25.10.2021


Abstract

In this paper, we show that accurate 3D object detection is possible using deep neural networks and a Bird’s Eye View (BEV) representation of the LiDAR point clouds. Many recent approaches propose complex neural network architectures to process directly the point cloud data. The good results obtained by these methods have left behind the research of BEV-based approaches. However, BEV-based detectors can take advantage of the advances in the 2D object detection field and need to handle much less data, which is important in real-time automotive applications. We propose a two-stage object detection deep neural network, which takes BEV representations as input and validate it in the KITTI BEV benchmark, outperforming state-of-the-art methods. In addition, we show how additional information can be added to our model to improve the accuracy of the smallest and most challenging object classes. This information can come from the same point cloud or an additional sensor’s data, such as the camera.

BIB_text

@Article {
title = {Accurate 3D Object Detection from Point Cloud Data using Bird’s Eye View Representations},
pages = {246-253},
keywds = {
Point Cloud, Object Detection, Deep Neural Networks, LiDAR
}
abstract = {

In this paper, we show that accurate 3D object detection is possible using deep neural networks and a Bird’s Eye View (BEV) representation of the LiDAR point clouds. Many recent approaches propose complex neural network architectures to process directly the point cloud data. The good results obtained by these methods have left behind the research of BEV-based approaches. However, BEV-based detectors can take advantage of the advances in the 2D object detection field and need to handle much less data, which is important in real-time automotive applications. We propose a two-stage object detection deep neural network, which takes BEV representations as input and validate it in the KITTI BEV benchmark, outperforming state-of-the-art methods. In addition, we show how additional information can be added to our model to improve the accuracy of the smallest and most challenging object classes. This information can come from the same point cloud or an additional sensor’s data, such as the camera.


}
isbn = {978-989-758-534-0},
date = {2021-10-25},
}
Vicomtech

Parque Científico y Tecnológico de Gipuzkoa,
Paseo Mikeletegi 57,
20009 Donostia / San Sebastián (España)

+(34) 943 309 230

Zorrotzaurreko Erribera 2, Deusto,
48014 Bilbao (España)

close overlay

Las cookies de publicidad comportamental son necesarias para cargar el contenido

Aceptar cookies de publicidad comportamental