File Download

There are no files associated with this item.

  Links for fulltext
     (May Require Subscription)
Supplementary

Article: High-level Feature Guided Decoding for Semantic Segmentation

TitleHigh-level Feature Guided Decoding for Semantic Segmentation
Authors
KeywordsCircuits and systems
Cityscapes
Decoding
Feature extraction
Representation Learning
Semantic segmentation
Semantic Segmentation
Spatial resolution
Task analysis
Training
Issue Date2024
Citation
IEEE Transactions on Circuits and Systems for Video Technology, 2024 How to Cite?
AbstractExisting pyramid-based upsamplers (e.g. SemanticFPN), although efficient, usually produce less accurate results compared to dilation-based models when using the same backbone. This is partially caused by the contaminated high-level features since they are fused and fine-tuned with noisy low-level features on limited data. To address this issue, we propose to use powerful pre-trained high-level features as guidance (HFG) so that the upsampler can produce robust results. Specifically, only the high-level features from the backbone are used to train the class tokens, which are then reused by the upsampler for classification, guiding the upsampler features to more discriminative backbone features. One crucial design of the HFG is to protect the high-level features from being contaminated by using proper stop-gradient operations so that the backbone does not update according to the noisy gradient from the upsampler. To push the upper limit of HFG, we introduce a context augmentation encoder (CAE) that can efficiently and effectively operate on the low-resolution high-level feature, resulting in improved representation and thus better guidance. We named our complete solution as the High-Level Features Guided Decoder (HFGD). We evaluate the proposed HFGD on three benchmarks: Pascal Context, COCOStuff164k, and Cityscapes. HFGD achieves state-of-the-art results among methods that do not use extra training data, demonstrating its effectiveness and generalization ability.
Persistent Identifierhttp://hdl.handle.net/10722/345383
ISSN
2023 Impact Factor: 8.3
2023 SCImago Journal Rankings: 2.299

 

DC FieldValueLanguage
dc.contributor.authorHuang, Ye-
dc.contributor.authorKang, Di-
dc.contributor.authorGao, Shenghua-
dc.contributor.authorLi, Wen-
dc.contributor.authorDuan, Lixin-
dc.date.accessioned2024-08-15T09:27:00Z-
dc.date.available2024-08-15T09:27:00Z-
dc.date.issued2024-
dc.identifier.citationIEEE Transactions on Circuits and Systems for Video Technology, 2024-
dc.identifier.issn1051-8215-
dc.identifier.urihttp://hdl.handle.net/10722/345383-
dc.description.abstractExisting pyramid-based upsamplers (<italic>e.g</italic>. SemanticFPN), although efficient, usually produce less accurate results compared to dilation-based models when using the same backbone. This is partially caused by the <italic>contaminated</italic> high-level features since they are fused and fine-tuned with noisy low-level features on limited data. To address this issue, we propose to use powerful pre-trained <italic>h</italic>igh-level <italic>f</italic>eatures as <italic>g</italic>uidance (HFG) so that the upsampler can produce robust results. Specifically, <italic>only</italic> the high-level features from the backbone are used to train the class tokens, which are then reused by the upsampler for classification, guiding the upsampler features to more discriminative backbone features. One crucial design of the HFG is to protect the high-level features from being contaminated by using proper stop-gradient operations so that the backbone does not update according to the noisy gradient from the upsampler. To push the upper limit of HFG, we introduce a <italic>c</italic>ontext <italic>a</italic>ugmentation <italic>e</italic>ncoder (CAE) that can efficiently and effectively operate on the low-resolution high-level feature, resulting in improved representation and thus better guidance. We named our complete solution as the High-Level Features Guided Decoder (HFGD). We evaluate the proposed HFGD on three benchmarks: Pascal Context, COCOStuff164k, and Cityscapes. HFGD achieves state-of-the-art results among methods that do not use extra training data, demonstrating its effectiveness and generalization ability.-
dc.languageeng-
dc.relation.ispartofIEEE Transactions on Circuits and Systems for Video Technology-
dc.subjectCircuits and systems-
dc.subjectCityscapes-
dc.subjectDecoding-
dc.subjectFeature extraction-
dc.subjectRepresentation Learning-
dc.subjectSemantic segmentation-
dc.subjectSemantic Segmentation-
dc.subjectSpatial resolution-
dc.subjectTask analysis-
dc.subjectTraining-
dc.titleHigh-level Feature Guided Decoding for Semantic Segmentation-
dc.typeArticle-
dc.description.naturelink_to_subscribed_fulltext-
dc.identifier.doi10.1109/TCSVT.2024.3393632-
dc.identifier.scopuseid_2-s2.0-85191715595-
dc.identifier.eissn1558-2205-

Export via OAI-PMH Interface in XML Formats


OR


Export to Other Non-XML Formats