Efficient Ingestion, Labeling, and Storage of LiDAR, Radar, and Camera Datasets: Optimizing Storage Formats for Autonomous Vehicle Development
DOI:
https://doi.org/10.22399/ijcesen.4839Keywords:
Autonomous Vehicle Perception, Multi-Modal Sensor Fusion, Lidar Point Cloud Processing, Dataset Management Infrastructure, Storage OptimizationAbstract
The development of autonomous vehicle technologies has placed new standards of managing the multimodal sensor data that includes LiDAR point clouds, radar measurements, and camera images. The contemporary autonomous platforms produce large volumes of non-uniform perception data that need advanced infrastructure to ingest, label, and store in the long term. Combining different modalities of sensing creates significant technical issues throughout the data lifetime, both during the initial capture and archival preservation. Infrastructure Edge preprocessing hardware will need to synchronize heterogeneous sensor streams with less than millisecond accuracy and apply real-time compression to achieve a lower bandwidth in transmission. Automated labeling workflows based on pre-trained perception models can help reduce the manual annotation load by a significant factor, and label quality is likely to be high enough to be used in training. Storage format optimization trades off conflicting demands such as compression performance, random access performance, and compatibility with the distributed processing model. The use of cloud-scale deployment allows managing the petabyte-scale volumes of data and optimizing the costs of infrastructure based on the tiered storage approaches that match the storage performance properties with the access patterns. Benchmark datasets have played a vital role in achieving advances in autonomous vehicle perception, and have allowed the performance to be systematically compared and the outstanding challenges to be reflected. Direct neural network processing of point cloud data has permitted significant progress in performance in detection and segmentation. Further progress will be realised through further growing multi-modal datasets with multi-modes of operation and persistent advances in algorithms that fuse sensor data and build scene understanding.
References
[1] Dimitris Zermas et al., "Fast Segmentation of 3D Point Clouds: A Paradigm on LiDAR Data for Autonomous Vehicle Applications," ResearchGate, 2017. [Online]. Available: https://www.researchgate.net/publication/318325507_Fast_Segmentation_of_3D_Point_Clouds_A_Paradigm_on_LiDAR_Data_for_Autonomous_Vehicle_Applications
[2] Jens Behley et al., "SemanticKITTI: A Dataset for Semantic Scene Understanding of LiDAR Sequences," arXiv:1904.01416, 2019. [Online]. Available: https://arxiv.org/abs/1904.01416
[3] Jakob Geyer et al., "A2D2: Audi Autonomous Driving Dataset," arXiv:2004.06320, 2020. [Online]. Available: https://arxiv.org/abs/2004.06320
[4] Pei Sun et al., "Scalability in Perception for Autonomous Driving: Waymo Open Dataset," CVF, 2020. [Online]. Available: https://openaccess.thecvf.com/content_CVPR_2020/papers/Sun_Scalability_in_Perception_for_Autonomous_Driving_Waymo_Open_Dataset_CVPR_2020_paper.pdf
[5] Holger Caesar et al., "nuScenes: A multimodal dataset for autonomous driving," arXiv:1903.11027, 2020. [Online]. Available: https://arxiv.org/abs/1903.11027
[6] Bin Yang et al., "PIXOR: Real-time 3D Object Detection from Point Clouds," arXiv:1902.06326, 2019. [Online]. Available: https://arxiv.org/abs/1902.06326
[7] Hang Su et al., "Multi-view Convolutional Neural Networks for 3D Shape Recognition," arXiv:1505.00880, 2015. [Online]. Available: https://arxiv.org/abs/1505.00880
[8] Charles R. Qi et al., "PointNet: Deep Learning on Point Sets for 3D Classification and Segmentation," arXiv:1612.00593, 2017. [Online]. Available: https://arxiv.org/abs/1612.00593
[9] Marius Cordts et al., "The Cityscapes Dataset for Semantic Urban Scene Understanding," arXiv:1604.01685, 2016. [Online]. Available: https://arxiv.org/abs/1604.01685
[10] Andreas Geiger et al., "Are we ready for autonomous driving? The KITTI vision benchmark suite". [Online]. Available: https://www.cs.toronto.edu/~urtasun/publications/geiger_et_al_cvpr12.pdf
Downloads
Published
How to Cite
Issue
Section
License
Copyright (c) 2025 International Journal of Computational and Experimental Science and Engineering

This work is licensed under a Creative Commons Attribution 4.0 International License.