The University of Southampton
University of Southampton Institutional Repository

End-to-end multimodal sensor dataset collection framework for autonomous vehicles

End-to-end multimodal sensor dataset collection framework for autonomous vehicles
End-to-end multimodal sensor dataset collection framework for autonomous vehicles
Autonomous driving vehicles rely on sensors for the robust perception of their surroundings. Such vehicles are equipped with multiple perceptive sensors with a high level of redundancy to ensure safety and reliability in any driving condition. However, multi-sensor, such as camera, LiDAR, and radar systems raise requirements related to sensor calibration and synchronization, which are the fundamental blocks of any autonomous system. On the other hand, sensor fusion and integration have become important aspects of autonomous driving research and directly determine the efficiency and accuracy of advanced functions such as object detection and path planning. Classical model-based estimation and data-driven models are two mainstream approaches to achieving such integration. Most recent research is shifting to the latter, showing high robustness in real-world applications but requiring large quantities of data to be collected, synchronized, and properly categorized. However, there are two major research gaps in existing works: (i) they lack fusion (and synchronization) of multi-sensors, camera, LiDAR and radar; and (ii) generic scalable, and user-friendly end-to-end implementation. To generalize the implementation of the multi-sensor perceptive system, we introduce an end-to-end generic sensor dataset collection framework that includes both hardware deploying solutions and sensor fusion algorithms. The framework prototype integrates a diverse set of sensors, such as camera, LiDAR, and radar. Furthermore, we present a universal toolbox to calibrate and synchronize three types of sensors based on their characteristics. The framework also includes the fusion algorithms, which utilize the merits of three sensors, namely, camera, LiDAR, and radar, and fuse their sensory information in a manner that is helpful for object detection and tracking research. The generality of this framework makes it applicable in any robotic or autonomous applications and suitable for quick and large-scale practical deployment.
autonomous driving, dataset collection framework, multimodal sensors, sensor calibration and synchronization, sensor fusion
1424-8220
Gu, Junyi
1bf3f1cd-ba82-4a5a-b5e8-9b3817055b1a
Lind, Artjom
03b6e2d1-3bf4-41d6-917e-052d4cd7ba9e
Chhetri, Tek Raj
c3431de5-4860-43e5-b09f-3dbb752c8490
Bellone, Mauro
b63674da-41d3-426b-89b9-ab8e36e16904
Sell, Raivo
9384ed47-c8ae-4029-8a58-b8512fc17e27
Gu, Junyi
1bf3f1cd-ba82-4a5a-b5e8-9b3817055b1a
Lind, Artjom
03b6e2d1-3bf4-41d6-917e-052d4cd7ba9e
Chhetri, Tek Raj
c3431de5-4860-43e5-b09f-3dbb752c8490
Bellone, Mauro
b63674da-41d3-426b-89b9-ab8e36e16904
Sell, Raivo
9384ed47-c8ae-4029-8a58-b8512fc17e27

Gu, Junyi, Lind, Artjom, Chhetri, Tek Raj, Bellone, Mauro and Sell, Raivo (2023) End-to-end multimodal sensor dataset collection framework for autonomous vehicles. Sensors, 23 (15), [6783]. (doi:10.3390/s23156783).

Record type: Article

Abstract

Autonomous driving vehicles rely on sensors for the robust perception of their surroundings. Such vehicles are equipped with multiple perceptive sensors with a high level of redundancy to ensure safety and reliability in any driving condition. However, multi-sensor, such as camera, LiDAR, and radar systems raise requirements related to sensor calibration and synchronization, which are the fundamental blocks of any autonomous system. On the other hand, sensor fusion and integration have become important aspects of autonomous driving research and directly determine the efficiency and accuracy of advanced functions such as object detection and path planning. Classical model-based estimation and data-driven models are two mainstream approaches to achieving such integration. Most recent research is shifting to the latter, showing high robustness in real-world applications but requiring large quantities of data to be collected, synchronized, and properly categorized. However, there are two major research gaps in existing works: (i) they lack fusion (and synchronization) of multi-sensors, camera, LiDAR and radar; and (ii) generic scalable, and user-friendly end-to-end implementation. To generalize the implementation of the multi-sensor perceptive system, we introduce an end-to-end generic sensor dataset collection framework that includes both hardware deploying solutions and sensor fusion algorithms. The framework prototype integrates a diverse set of sensors, such as camera, LiDAR, and radar. Furthermore, we present a universal toolbox to calibrate and synchronize three types of sensors based on their characteristics. The framework also includes the fusion algorithms, which utilize the merits of three sensors, namely, camera, LiDAR, and radar, and fuse their sensory information in a manner that is helpful for object detection and tracking research. The generality of this framework makes it applicable in any robotic or autonomous applications and suitable for quick and large-scale practical deployment.

Text
sensors-23-06783 - Version of Record
Available under License Creative Commons Attribution.
Download (9MB)

More information

Accepted/In Press date: 20 July 2023
Published date: 29 July 2023
Additional Information: Funding Information: The financial support from the Estonian Ministry of Education and Research and the Horizon 2020 Research and Innovation Programme is gratefully acknowledged. Publisher Copyright: © 2023 by the authors.
Keywords: autonomous driving, dataset collection framework, multimodal sensors, sensor calibration and synchronization, sensor fusion

Identifiers

Local EPrints ID: 481451
URI: http://eprints.soton.ac.uk/id/eprint/481451
ISSN: 1424-8220
PURE UUID: 60d86651-e396-4a7b-8d7b-f0c8f3874b82
ORCID for Tek Raj Chhetri: ORCID iD orcid.org/0000-0002-3905-7878

Catalogue record

Date deposited: 29 Aug 2023 16:50
Last modified: 17 Mar 2024 04:21

Export record

Altmetrics

Contributors

Author: Junyi Gu
Author: Artjom Lind
Author: Tek Raj Chhetri ORCID iD
Author: Mauro Bellone
Author: Raivo Sell

Download statistics

Downloads from ePrints over the past year. Other digital versions may also be available to download e.g. from the publisher's website.

View more statistics

Atom RSS 1.0 RSS 2.0

Contact ePrints Soton: eprints@soton.ac.uk

ePrints Soton supports OAI 2.0 with a base URL of http://eprints.soton.ac.uk/cgi/oai2

This repository has been built using EPrints software, developed at the University of Southampton, but available to everyone to use.

We use cookies to ensure that we give you the best experience on our website. If you continue without changing your settings, we will assume that you are happy to receive cookies on the University of Southampton website.

×