File Download

There are no files associated with this item.

  Links for fulltext
     (May Require Subscription)
Supplementary

Article: Delving into the Devils of Bird's-Eye-View Perception: A Review, Evaluation and Recipe

TitleDelving into the Devils of Bird's-Eye-View Perception: A Review, Evaluation and Recipe
Authors
Keywords3D detection and segmentation
autonomous driving challenge
birds-eye-view (BEV) perception
Issue Date2024
Citation
IEEE Transactions on Pattern Analysis and Machine Intelligence, 2024, v. 46, n. 4, p. 2151-2170 How to Cite?
AbstractLearning powerful representations in bird's-eye-view (BEV) for perception tasks is trending and drawing extensive attention both from industry and academia. Conventional approaches for most autonomous driving algorithms perform detection, segmentation, tracking, etc., in a front or perspective view. As sensor configurations get more complex, integrating multi-source information from different sensors and representing features in a unified view come of vital importance. BEV perception inherits several advantages, as representing surrounding scenes in BEV is intuitive and fusion-friendly; and representing objects in BEV is most desirable for subsequent modules as in planning and/or control. The core problems for BEV perception lie in (a) how to reconstruct the lost 3D information via view transformation from perspective view to BEV; (b) how to acquire ground truth annotations in BEV grid; (c) how to formulate the pipeline to incorporate features from different sources and views; and (d) how to adapt and generalize algorithms as sensor configurations vary across different scenarios. In this survey, we review the most recent works on BEV perception and provide an in-depth analysis of different solutions. Moreover, several systematic designs of BEV approach from the industry are depicted as well. Furthermore, we introduce a full suite of practical guidebook to improve the performance of BEV perception tasks, including camera, LiDAR and fusion inputs. At last, we point out the future research directions in this area. We hope this report will shed some light on the community and encourage more research effort on BEV perception.
Persistent Identifierhttp://hdl.handle.net/10722/351485
ISSN
2023 Impact Factor: 20.8
2023 SCImago Journal Rankings: 6.158

 

DC FieldValueLanguage
dc.contributor.authorLi, Hongyang-
dc.contributor.authorSima, Chonghao-
dc.contributor.authorDai, Jifeng-
dc.contributor.authorWang, Wenhai-
dc.contributor.authorLu, Lewei-
dc.contributor.authorWang, Huijie-
dc.contributor.authorZeng, Jia-
dc.contributor.authorLi, Zhiqi-
dc.contributor.authorYang, Jiazhi-
dc.contributor.authorDeng, Hanming-
dc.contributor.authorTian, Hao-
dc.contributor.authorXie, Enze-
dc.contributor.authorXie, Jiangwei-
dc.contributor.authorChen, Li-
dc.contributor.authorLi, Tianyu-
dc.contributor.authorLi, Yang-
dc.contributor.authorGao, Yulu-
dc.contributor.authorJia, Xiaosong-
dc.contributor.authorLiu, Si-
dc.contributor.authorShi, Jianping-
dc.contributor.authorLin, Dahua-
dc.contributor.authorQiao, Yu-
dc.date.accessioned2024-11-20T03:56:38Z-
dc.date.available2024-11-20T03:56:38Z-
dc.date.issued2024-
dc.identifier.citationIEEE Transactions on Pattern Analysis and Machine Intelligence, 2024, v. 46, n. 4, p. 2151-2170-
dc.identifier.issn0162-8828-
dc.identifier.urihttp://hdl.handle.net/10722/351485-
dc.description.abstractLearning powerful representations in bird's-eye-view (BEV) for perception tasks is trending and drawing extensive attention both from industry and academia. Conventional approaches for most autonomous driving algorithms perform detection, segmentation, tracking, etc., in a front or perspective view. As sensor configurations get more complex, integrating multi-source information from different sensors and representing features in a unified view come of vital importance. BEV perception inherits several advantages, as representing surrounding scenes in BEV is intuitive and fusion-friendly; and representing objects in BEV is most desirable for subsequent modules as in planning and/or control. The core problems for BEV perception lie in (a) how to reconstruct the lost 3D information via view transformation from perspective view to BEV; (b) how to acquire ground truth annotations in BEV grid; (c) how to formulate the pipeline to incorporate features from different sources and views; and (d) how to adapt and generalize algorithms as sensor configurations vary across different scenarios. In this survey, we review the most recent works on BEV perception and provide an in-depth analysis of different solutions. Moreover, several systematic designs of BEV approach from the industry are depicted as well. Furthermore, we introduce a full suite of practical guidebook to improve the performance of BEV perception tasks, including camera, LiDAR and fusion inputs. At last, we point out the future research directions in this area. We hope this report will shed some light on the community and encourage more research effort on BEV perception.-
dc.languageeng-
dc.relation.ispartofIEEE Transactions on Pattern Analysis and Machine Intelligence-
dc.subject3D detection and segmentation-
dc.subjectautonomous driving challenge-
dc.subjectbirds-eye-view (BEV) perception-
dc.titleDelving into the Devils of Bird's-Eye-View Perception: A Review, Evaluation and Recipe-
dc.typeArticle-
dc.description.naturelink_to_subscribed_fulltext-
dc.identifier.doi10.1109/TPAMI.2023.3333838-
dc.identifier.pmid37976193-
dc.identifier.scopuseid_2-s2.0-85178071901-
dc.identifier.volume46-
dc.identifier.issue4-
dc.identifier.spage2151-
dc.identifier.epage2170-
dc.identifier.eissn1939-3539-

Export via OAI-PMH Interface in XML Formats


OR


Export to Other Non-XML Formats