File Download

There are no files associated with this item.

  Links for fulltext
     (May Require Subscription)
Supplementary

Article: Learning with privileged multimodal knowledge for unimodal segmentation

TitleLearning with privileged multimodal knowledge for unimodal segmentation
Authors
KeywordsContrastive learning
Knowledge distillation
Multimodal segmentation
Privileged knowledge
Issue Date2022
Citation
IEEE Transactions on Medical Imaging, 2022, v. 41, n. 3, p. 621-632 How to Cite?
AbstractMultimodal learning usually requires a complete set of modalities during inference to maintain performance. Although training data can be well-prepared with high-quality multiple modalities, in many cases of clinical practice, only one modality can be acquired and important clinical evaluations have to be made based on the limited single modality information. In this work, we propose a privileged knowledge learning framework with the 'Teacher-Student' architecture, in which the complete multimodal knowledge that is only available in the training data (called privileged information) is transferred from a multimodal teacher network to a unimodal student network, via both a pixel-level and an image-level distillation scheme. Specifically, for the pixel-level distillation, we introduce a regularized knowledge distillation loss which encourages the student to mimic the teacher's softened outputs in a pixel-wise manner and incorporates a regularization factor to reduce the effect of incorrect predictions from the teacher. For the image-level distillation, we propose a contrastive knowledge distillation loss which encodes image-level structured information to enrich the knowledge encoding in combination with the pixel-level distillation. We extensively evaluate our method on two different multi-class segmentation tasks, i.e., cardiac substructure segmentation and brain tumor segmentation. Experimental results on both tasks demonstrate that our privileged knowledge learning is effective in improving unimodal segmentation and outperforms previous methods.
Persistent Identifierhttp://hdl.handle.net/10722/349616
ISSN
2023 Impact Factor: 8.9
2023 SCImago Journal Rankings: 3.703

 

DC FieldValueLanguage
dc.contributor.authorChen, Cheng-
dc.contributor.authorDou, Qi-
dc.contributor.authorJin, Yueming-
dc.contributor.authorLiu, Quande-
dc.contributor.authorHeng, Pheng Ann-
dc.date.accessioned2024-10-17T06:59:43Z-
dc.date.available2024-10-17T06:59:43Z-
dc.date.issued2022-
dc.identifier.citationIEEE Transactions on Medical Imaging, 2022, v. 41, n. 3, p. 621-632-
dc.identifier.issn0278-0062-
dc.identifier.urihttp://hdl.handle.net/10722/349616-
dc.description.abstractMultimodal learning usually requires a complete set of modalities during inference to maintain performance. Although training data can be well-prepared with high-quality multiple modalities, in many cases of clinical practice, only one modality can be acquired and important clinical evaluations have to be made based on the limited single modality information. In this work, we propose a privileged knowledge learning framework with the 'Teacher-Student' architecture, in which the complete multimodal knowledge that is only available in the training data (called privileged information) is transferred from a multimodal teacher network to a unimodal student network, via both a pixel-level and an image-level distillation scheme. Specifically, for the pixel-level distillation, we introduce a regularized knowledge distillation loss which encourages the student to mimic the teacher's softened outputs in a pixel-wise manner and incorporates a regularization factor to reduce the effect of incorrect predictions from the teacher. For the image-level distillation, we propose a contrastive knowledge distillation loss which encodes image-level structured information to enrich the knowledge encoding in combination with the pixel-level distillation. We extensively evaluate our method on two different multi-class segmentation tasks, i.e., cardiac substructure segmentation and brain tumor segmentation. Experimental results on both tasks demonstrate that our privileged knowledge learning is effective in improving unimodal segmentation and outperforms previous methods.-
dc.languageeng-
dc.relation.ispartofIEEE Transactions on Medical Imaging-
dc.subjectContrastive learning-
dc.subjectKnowledge distillation-
dc.subjectMultimodal segmentation-
dc.subjectPrivileged knowledge-
dc.titleLearning with privileged multimodal knowledge for unimodal segmentation-
dc.typeArticle-
dc.description.naturelink_to_subscribed_fulltext-
dc.identifier.doi10.1109/TMI.2021.3119385-
dc.identifier.pmid34633927-
dc.identifier.scopuseid_2-s2.0-85117259830-
dc.identifier.volume41-
dc.identifier.issue3-
dc.identifier.spage621-
dc.identifier.epage632-
dc.identifier.eissn1558-254X-

Export via OAI-PMH Interface in XML Formats


OR


Export to Other Non-XML Formats