File Download

There are no files associated with this item.

  Links for fulltext
     (May Require Subscription)
Supplementary

Conference Paper: SemGrasp : Semantic Grasp Generation via Language Aligned Discretization

TitleSemGrasp : Semantic Grasp Generation via Language Aligned Discretization
Authors
KeywordsDiscrete representation
MLLM
Semantic Grasp Generation
Issue Date2025
Citation
Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics), 2025, v. 15060 LNCS, p. 109-127 How to Cite?
AbstractGenerating natural human grasps necessitates consideration of not just object geometry but also semantic information. Solely depending on object shape for grasp generation confines the applications of prior methods in downstream tasks. This paper presents a novel semantic-based grasp generation method, termed SemGrasp, which generates a static human grasp pose by incorporating semantic information into the grasp representation. We introduce a discrete representation that aligns the grasp space with semantic space, enabling the generation of grasp postures in accordance with language instructions. A Multimodal Large Language Model (MLLM) is subsequently fine-tuned, integrating object, grasp, and language within a unified semantic space. To facilitate the training of SemGrasp, we compile a large-scale, grasp-text-aligned dataset named CapGrasp, featuring over 300k detailed captions and 50k diverse grasps. Experimental findings demonstrate that SemGrasp efficiently generates natural human grasps in alignment with linguistic intentions. Our code, models, and dataset are available publicly at: https://kailinli.github.io/SemGrasp.
Persistent Identifierhttp://hdl.handle.net/10722/352480
ISSN
2023 SCImago Journal Rankings: 0.606

 

DC FieldValueLanguage
dc.contributor.authorLi, Kailin-
dc.contributor.authorWang, Jingbo-
dc.contributor.authorYang, Lixin-
dc.contributor.authorLu, Cewu-
dc.contributor.authorDai, Bo-
dc.date.accessioned2024-12-16T03:59:20Z-
dc.date.available2024-12-16T03:59:20Z-
dc.date.issued2025-
dc.identifier.citationLecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics), 2025, v. 15060 LNCS, p. 109-127-
dc.identifier.issn0302-9743-
dc.identifier.urihttp://hdl.handle.net/10722/352480-
dc.description.abstractGenerating natural human grasps necessitates consideration of not just object geometry but also semantic information. Solely depending on object shape for grasp generation confines the applications of prior methods in downstream tasks. This paper presents a novel semantic-based grasp generation method, termed SemGrasp, which generates a static human grasp pose by incorporating semantic information into the grasp representation. We introduce a discrete representation that aligns the grasp space with semantic space, enabling the generation of grasp postures in accordance with language instructions. A Multimodal Large Language Model (MLLM) is subsequently fine-tuned, integrating object, grasp, and language within a unified semantic space. To facilitate the training of SemGrasp, we compile a large-scale, grasp-text-aligned dataset named CapGrasp, featuring over 300k detailed captions and 50k diverse grasps. Experimental findings demonstrate that SemGrasp efficiently generates natural human grasps in alignment with linguistic intentions. Our code, models, and dataset are available publicly at: https://kailinli.github.io/SemGrasp.-
dc.languageeng-
dc.relation.ispartofLecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)-
dc.subjectDiscrete representation-
dc.subjectMLLM-
dc.subjectSemantic Grasp Generation-
dc.titleSemGrasp : Semantic Grasp Generation via Language Aligned Discretization-
dc.typeConference_Paper-
dc.description.naturelink_to_subscribed_fulltext-
dc.identifier.doi10.1007/978-3-031-72627-9_7-
dc.identifier.scopuseid_2-s2.0-85208231251-
dc.identifier.volume15060 LNCS-
dc.identifier.spage109-
dc.identifier.epage127-
dc.identifier.eissn1611-3349-

Export via OAI-PMH Interface in XML Formats


OR


Export to Other Non-XML Formats