File Download

There are no files associated with this item.

  Links for fulltext
     (May Require Subscription)
Supplementary

Article: Object-centric Representation Learning for Video Scene Understanding

TitleObject-centric Representation Learning for Video Scene Understanding
Authors
KeywordsDepth estimation
Estimation
Feature extraction
Generators
IP networks
object-centric representation
Pipelines
scene understanding
Semantics
Task analysis
tracking
video panoptic segmentation
Issue Date15-May-2024
PublisherInstitute of Electrical and Electronics Engineers
Citation
IEEE Transactions on Pattern Analysis and Machine Intelligence, 2024, p. 1-13 How to Cite?
Abstract

Depth-aware Video Panoptic Segmentation (DVPS) is a challenging task that requires predicting the semantic class and 3D depth of each pixel in a video, while also segmenting and consistently tracking objects across frames. Predominant methodologies treat this as a multi-task learning problem, tackling each constituent task independently, thus restricting their capacity to leverage interrelationships amongst tasks and requiring parameter tuning for each task. To surmount these constraints, we present Slot-IVPS, a new approach employing an object-centric model to acquire unified object representations, thereby facilitating the model's ability to simultaneously capture semantic and depth information. Specifically, we introduce a novel representation, Integrated Panoptic Slots (IPS), to capture both semantic and depth information for all panoptic objects within a video, encompassing background semantics and foreground instances. Subsequently, we propose an integrated feature generator and enhancer to extract depth-aware features, alongside the Integrated Video Panoptic Retriever (IVPR), which iteratively retrieves spatial-temporal coherent object features and encodes them into IPS. The resulting IPS can be effortlessly decoded into an array of video outputs, including depth maps, classifications, masks, and object instance IDs. We undertake comprehensive analyses across four datasets, attaining state-of-the-art performance in both Depth-aware Video Panoptic Segmentation and Video Panoptic Segmentation tasks. Codes will be available at https://github.com/SAITPublic/.


Persistent Identifierhttp://hdl.handle.net/10722/350740
ISSN
2023 Impact Factor: 20.8
2023 SCImago Journal Rankings: 6.158

 

DC FieldValueLanguage
dc.contributor.authorZhou, Yi-
dc.contributor.authorZhang, Hui-
dc.contributor.authorPark, Seung In-
dc.contributor.authorYoo, Byung In-
dc.contributor.authorQi, Xiaojuan-
dc.date.accessioned2024-11-02T00:36:48Z-
dc.date.available2024-11-02T00:36:48Z-
dc.date.issued2024-05-15-
dc.identifier.citationIEEE Transactions on Pattern Analysis and Machine Intelligence, 2024, p. 1-13-
dc.identifier.issn0162-8828-
dc.identifier.urihttp://hdl.handle.net/10722/350740-
dc.description.abstract<p>Depth-aware Video Panoptic Segmentation (DVPS) is a challenging task that requires predicting the semantic class and 3D depth of each pixel in a video, while also segmenting and consistently tracking objects across frames. Predominant methodologies treat this as a multi-task learning problem, tackling each constituent task independently, thus restricting their capacity to leverage interrelationships amongst tasks and requiring parameter tuning for each task. To surmount these constraints, we present Slot-IVPS, a new approach employing an object-centric model to acquire unified object representations, thereby facilitating the model's ability to simultaneously capture semantic and depth information. Specifically, we introduce a novel representation, Integrated Panoptic Slots (IPS), to capture both semantic and depth information for all panoptic objects within a video, encompassing background semantics and foreground instances. Subsequently, we propose an integrated feature generator and enhancer to extract depth-aware features, alongside the Integrated Video Panoptic Retriever (IVPR), which iteratively retrieves spatial-temporal coherent object features and encodes them into IPS. The resulting IPS can be effortlessly decoded into an array of video outputs, including depth maps, classifications, masks, and object instance IDs. We undertake comprehensive analyses across four datasets, attaining state-of-the-art performance in both Depth-aware Video Panoptic Segmentation and Video Panoptic Segmentation tasks. Codes will be available at https://github.com/SAITPublic/.</p>-
dc.languageeng-
dc.publisherInstitute of Electrical and Electronics Engineers-
dc.relation.ispartofIEEE Transactions on Pattern Analysis and Machine Intelligence-
dc.rightsThis work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License.-
dc.subjectDepth estimation-
dc.subjectEstimation-
dc.subjectFeature extraction-
dc.subjectGenerators-
dc.subjectIP networks-
dc.subjectobject-centric representation-
dc.subjectPipelines-
dc.subjectscene understanding-
dc.subjectSemantics-
dc.subjectTask analysis-
dc.subjecttracking-
dc.subjectvideo panoptic segmentation-
dc.titleObject-centric Representation Learning for Video Scene Understanding -
dc.typeArticle-
dc.identifier.doi10.1109/TPAMI.2024.3401409-
dc.identifier.scopuseid_2-s2.0-85193287200-
dc.identifier.spage1-
dc.identifier.epage13-
dc.identifier.eissn1939-3539-
dc.identifier.issnl0162-8828-

Export via OAI-PMH Interface in XML Formats


OR


Export to Other Non-XML Formats