File Download
There are no files associated with this item.
Links for fulltext
(May Require Subscription)
- Publisher Website: 10.1109/TMI.2021.3119385
- Scopus: eid_2-s2.0-85117259830
- PMID: 34633927
- Find via
Supplementary
- Citations:
- Appears in Collections:
Article: Learning with privileged multimodal knowledge for unimodal segmentation
Title | Learning with privileged multimodal knowledge for unimodal segmentation |
---|---|
Authors | |
Keywords | Contrastive learning Knowledge distillation Multimodal segmentation Privileged knowledge |
Issue Date | 2022 |
Citation | IEEE Transactions on Medical Imaging, 2022, v. 41, n. 3, p. 621-632 How to Cite? |
Abstract | Multimodal learning usually requires a complete set of modalities during inference to maintain performance. Although training data can be well-prepared with high-quality multiple modalities, in many cases of clinical practice, only one modality can be acquired and important clinical evaluations have to be made based on the limited single modality information. In this work, we propose a privileged knowledge learning framework with the 'Teacher-Student' architecture, in which the complete multimodal knowledge that is only available in the training data (called privileged information) is transferred from a multimodal teacher network to a unimodal student network, via both a pixel-level and an image-level distillation scheme. Specifically, for the pixel-level distillation, we introduce a regularized knowledge distillation loss which encourages the student to mimic the teacher's softened outputs in a pixel-wise manner and incorporates a regularization factor to reduce the effect of incorrect predictions from the teacher. For the image-level distillation, we propose a contrastive knowledge distillation loss which encodes image-level structured information to enrich the knowledge encoding in combination with the pixel-level distillation. We extensively evaluate our method on two different multi-class segmentation tasks, i.e., cardiac substructure segmentation and brain tumor segmentation. Experimental results on both tasks demonstrate that our privileged knowledge learning is effective in improving unimodal segmentation and outperforms previous methods. |
Persistent Identifier | http://hdl.handle.net/10722/349616 |
ISSN | 2023 Impact Factor: 8.9 2023 SCImago Journal Rankings: 3.703 |
DC Field | Value | Language |
---|---|---|
dc.contributor.author | Chen, Cheng | - |
dc.contributor.author | Dou, Qi | - |
dc.contributor.author | Jin, Yueming | - |
dc.contributor.author | Liu, Quande | - |
dc.contributor.author | Heng, Pheng Ann | - |
dc.date.accessioned | 2024-10-17T06:59:43Z | - |
dc.date.available | 2024-10-17T06:59:43Z | - |
dc.date.issued | 2022 | - |
dc.identifier.citation | IEEE Transactions on Medical Imaging, 2022, v. 41, n. 3, p. 621-632 | - |
dc.identifier.issn | 0278-0062 | - |
dc.identifier.uri | http://hdl.handle.net/10722/349616 | - |
dc.description.abstract | Multimodal learning usually requires a complete set of modalities during inference to maintain performance. Although training data can be well-prepared with high-quality multiple modalities, in many cases of clinical practice, only one modality can be acquired and important clinical evaluations have to be made based on the limited single modality information. In this work, we propose a privileged knowledge learning framework with the 'Teacher-Student' architecture, in which the complete multimodal knowledge that is only available in the training data (called privileged information) is transferred from a multimodal teacher network to a unimodal student network, via both a pixel-level and an image-level distillation scheme. Specifically, for the pixel-level distillation, we introduce a regularized knowledge distillation loss which encourages the student to mimic the teacher's softened outputs in a pixel-wise manner and incorporates a regularization factor to reduce the effect of incorrect predictions from the teacher. For the image-level distillation, we propose a contrastive knowledge distillation loss which encodes image-level structured information to enrich the knowledge encoding in combination with the pixel-level distillation. We extensively evaluate our method on two different multi-class segmentation tasks, i.e., cardiac substructure segmentation and brain tumor segmentation. Experimental results on both tasks demonstrate that our privileged knowledge learning is effective in improving unimodal segmentation and outperforms previous methods. | - |
dc.language | eng | - |
dc.relation.ispartof | IEEE Transactions on Medical Imaging | - |
dc.subject | Contrastive learning | - |
dc.subject | Knowledge distillation | - |
dc.subject | Multimodal segmentation | - |
dc.subject | Privileged knowledge | - |
dc.title | Learning with privileged multimodal knowledge for unimodal segmentation | - |
dc.type | Article | - |
dc.description.nature | link_to_subscribed_fulltext | - |
dc.identifier.doi | 10.1109/TMI.2021.3119385 | - |
dc.identifier.pmid | 34633927 | - |
dc.identifier.scopus | eid_2-s2.0-85117259830 | - |
dc.identifier.volume | 41 | - |
dc.identifier.issue | 3 | - |
dc.identifier.spage | 621 | - |
dc.identifier.epage | 632 | - |
dc.identifier.eissn | 1558-254X | - |