File Download

There are no files associated with this item.

  Links for fulltext
     (May Require Subscription)
Supplementary

Conference Paper: 3D Data Augmentation for Driving Scenes on Camera

Title3D Data Augmentation for Driving Scenes on Camera
Authors
Keywords3D Perception
Autonomous Driving
Data Augmentation
NeRF
Issue Date2025
Citation
Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics), 2025, v. 15036 LNCS, p. 46-63 How to Cite?
AbstractDriving scenes are extremely diverse and complicated that it is impossible to collect all cases with human effort alone. While data augmentation is an effective technique to enrich the training data, existing methods for camera data in autonomous driving applications are confined to the 2D image plane, which may not optimally increase data diversity in 3D real-world scenarios. To this end, we propose a 3D data augmentation approach termed Drive-3DAug, aiming at augmenting the driving scenes on camera in the 3D space. We first utilize Neural Radiance Field (NeRF) to reconstruct the 3D models of background and foreground objects. Then, augmented driving scenes can be obtained by placing the 3D objects with adapted location and orientation at the pre-defined valid region of backgrounds. As such, the training database could be effectively scaled up. However, the 3D object modeling is constrained to the image quality and the limited viewpoints. To overcome these problems, we modify the original NeRF by introducing a geometric rectified loss and a symmetric-aware training strategy. We evaluate our method for the camera-only monocular 3D detection task on the Waymo and nuScences datasets. The proposed data augmentation approach contributes to a gain of and in terms of detection accuracy, on Waymo and nuScences respectively. Furthermore, the constructed 3D models serve as digital driving assets and could be recycled for different detectors or other 3D perception tasks.
Persistent Identifierhttp://hdl.handle.net/10722/352484
ISSN
2023 SCImago Journal Rankings: 0.606

 

DC FieldValueLanguage
dc.contributor.authorTong, Wenwen-
dc.contributor.authorXie, Jiangwei-
dc.contributor.authorLi, Tianyu-
dc.contributor.authorLi, Yang-
dc.contributor.authorDeng, Hanming-
dc.contributor.authorDai, Bo-
dc.contributor.authorLu, Lewei-
dc.contributor.authorZhao, Hao-
dc.contributor.authorYan, Junchi-
dc.contributor.authorLi, Hongyang-
dc.date.accessioned2024-12-16T03:59:22Z-
dc.date.available2024-12-16T03:59:22Z-
dc.date.issued2025-
dc.identifier.citationLecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics), 2025, v. 15036 LNCS, p. 46-63-
dc.identifier.issn0302-9743-
dc.identifier.urihttp://hdl.handle.net/10722/352484-
dc.description.abstractDriving scenes are extremely diverse and complicated that it is impossible to collect all cases with human effort alone. While data augmentation is an effective technique to enrich the training data, existing methods for camera data in autonomous driving applications are confined to the 2D image plane, which may not optimally increase data diversity in 3D real-world scenarios. To this end, we propose a 3D data augmentation approach termed Drive-3DAug, aiming at augmenting the driving scenes on camera in the 3D space. We first utilize Neural Radiance Field (NeRF) to reconstruct the 3D models of background and foreground objects. Then, augmented driving scenes can be obtained by placing the 3D objects with adapted location and orientation at the pre-defined valid region of backgrounds. As such, the training database could be effectively scaled up. However, the 3D object modeling is constrained to the image quality and the limited viewpoints. To overcome these problems, we modify the original NeRF by introducing a geometric rectified loss and a symmetric-aware training strategy. We evaluate our method for the camera-only monocular 3D detection task on the Waymo and nuScences datasets. The proposed data augmentation approach contributes to a gain of and in terms of detection accuracy, on Waymo and nuScences respectively. Furthermore, the constructed 3D models serve as digital driving assets and could be recycled for different detectors or other 3D perception tasks.-
dc.languageeng-
dc.relation.ispartofLecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)-
dc.subject3D Perception-
dc.subjectAutonomous Driving-
dc.subjectData Augmentation-
dc.subjectNeRF-
dc.title3D Data Augmentation for Driving Scenes on Camera-
dc.typeConference_Paper-
dc.description.naturelink_to_subscribed_fulltext-
dc.identifier.doi10.1007/978-981-97-8508-7_4-
dc.identifier.scopuseid_2-s2.0-85209359372-
dc.identifier.volume15036 LNCS-
dc.identifier.spage46-
dc.identifier.epage63-
dc.identifier.eissn1611-3349-

Export via OAI-PMH Interface in XML Formats


OR


Export to Other Non-XML Formats