File Download

There are no files associated with this item.

  Links for fulltext
     (May Require Subscription)
Supplementary

Conference Paper: Embodied Understanding of Driving Scenarios

TitleEmbodied Understanding of Driving Scenarios
Authors
Issue Date2025
Citation
Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics), 2025, v. 15120 LNCS, p. 129-148 How to Cite?
AbstractEmbodied scene understanding serves as the cornerstone for autonomous agents to perceive, interpret, and respond to open driving scenarios. Such understanding is typically founded upon Vision-Language Models (VLMs). Nevertheless, existing VLMs are restricted to the 2D domain, devoid of spatial awareness and long-horizon extrapolation proficiencies. We revisit the key aspects of autonomous driving and formulate appropriate rubrics. Hereby, we introduce the Embodied Language Model (ELM), a comprehensive framework tailored for agents’ understanding of driving scenes with large spatial and temporal spans. ELM incorporates space-aware pre-training to endow the agent with robust spatial localization capabilities. Besides, the model employs time-aware token selection to accurately inquire about temporal cues. We instantiate ELM on the reformulated multi-faced benchmark, and it surpasses previous state-of-the-art approaches in all aspects. All code, data, and models are accessible at https://github.com/OpenDriveLab/ELM.
Persistent Identifierhttp://hdl.handle.net/10722/351503
ISSN
2023 SCImago Journal Rankings: 0.606

 

DC FieldValueLanguage
dc.contributor.authorZhou, Yunsong-
dc.contributor.authorHuang, Linyan-
dc.contributor.authorBu, Qingwen-
dc.contributor.authorZeng, Jia-
dc.contributor.authorLi, Tianyu-
dc.contributor.authorQiu, Hang-
dc.contributor.authorZhu, Hongzi-
dc.contributor.authorGuo, Minyi-
dc.contributor.authorQiao, Yu-
dc.contributor.authorLi, Hongyang-
dc.date.accessioned2024-11-20T03:56:47Z-
dc.date.available2024-11-20T03:56:47Z-
dc.date.issued2025-
dc.identifier.citationLecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics), 2025, v. 15120 LNCS, p. 129-148-
dc.identifier.issn0302-9743-
dc.identifier.urihttp://hdl.handle.net/10722/351503-
dc.description.abstractEmbodied scene understanding serves as the cornerstone for autonomous agents to perceive, interpret, and respond to open driving scenarios. Such understanding is typically founded upon Vision-Language Models (VLMs). Nevertheless, existing VLMs are restricted to the 2D domain, devoid of spatial awareness and long-horizon extrapolation proficiencies. We revisit the key aspects of autonomous driving and formulate appropriate rubrics. Hereby, we introduce the Embodied Language Model (ELM), a comprehensive framework tailored for agents’ understanding of driving scenes with large spatial and temporal spans. ELM incorporates space-aware pre-training to endow the agent with robust spatial localization capabilities. Besides, the model employs time-aware token selection to accurately inquire about temporal cues. We instantiate ELM on the reformulated multi-faced benchmark, and it surpasses previous state-of-the-art approaches in all aspects. All code, data, and models are accessible at https://github.com/OpenDriveLab/ELM.-
dc.languageeng-
dc.relation.ispartofLecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)-
dc.titleEmbodied Understanding of Driving Scenarios-
dc.typeConference_Paper-
dc.description.naturelink_to_subscribed_fulltext-
dc.identifier.doi10.1007/978-3-031-73033-7_8-
dc.identifier.scopuseid_2-s2.0-85208541186-
dc.identifier.volume15120 LNCS-
dc.identifier.spage129-
dc.identifier.epage148-
dc.identifier.eissn1611-3349-

Export via OAI-PMH Interface in XML Formats


OR


Export to Other Non-XML Formats