File Download
There are no files associated with this item.
Links for fulltext
(May Require Subscription)
- Publisher Website: 10.1109/TMI.2023.3287256
- Scopus: eid_2-s2.0-85162901224
- PMID: 37335798
- WOS: WOS:001099088700016
- Find via
Supplementary
- Citations:
- Appears in Collections:
Article: Shared-Specific Feature Learning With Bottleneck Fusion Transformer for Multi-Modal Whole Slide Image Analysis
Title | Shared-Specific Feature Learning With Bottleneck Fusion Transformer for Multi-Modal Whole Slide Image Analysis |
---|---|
Authors | |
Keywords | knowledge transfer multi-modal multi-instance learning transformer Whole slide image |
Issue Date | 1-Nov-2023 |
Publisher | Institute of Electrical and Electronics Engineers |
Citation | IEEE Transactions on Medical Imaging, 2023, v. 42, n. 11, p. 3374-3383 How to Cite? |
Abstract | The fusion of multi-modal medical data is essential to assist medical experts to make treatment decisions for precision medicine. For example, combining the whole slide histopathological images (WSIs) and tabular clinical data can more accurately predict the lymph node metastasis (LNM) of papillary thyroid carcinoma before surgery to avoid unnecessary lymph node resection. However, the huge-sized WSI provides much more high-dimensional information than low-dimensional tabular clinical data, making the information alignment challenging in the multi-modal WSI analysis tasks. This paper presents a novel transformer-guided multi-modal multi-instance learning framework to predict lymph node metastasis from both WSIs and tabular clinical data. We first propose an effective multi-instance grouping scheme, named siamese attention-based feature grouping (SAG), to group high-dimensional WSIs into representative low-dimensional feature embeddings for fusion. We then design a novel bottleneck shared-specific feature transfer module (BSFT) to explore the shared and specific features between different modalities, where a few learnable bottleneck tokens are utilized for knowledge transfer between modalities. Moreover, a modal adaptation and orthogonal projection scheme were incorporated to further encourage BSFT to learn shared and specific features from multi-modal data. Finally, the shared and specific features are dynamically aggregated via an attention mechanism for slide-level prediction. Experimental results on our collected lymph node metastasis dataset demonstrate the efficiency of our proposed components and our framework achieves the best performance with AUC (area under the curve) of 97.34%, outperforming the state-of-the-art methods by over 1.27%. |
Persistent Identifier | http://hdl.handle.net/10722/340954 |
ISSN | 2023 Impact Factor: 8.9 2023 SCImago Journal Rankings: 3.703 |
ISI Accession Number ID |
DC Field | Value | Language |
---|---|---|
dc.contributor.author | Wang, ZH | - |
dc.contributor.author | Yu, LQ | - |
dc.contributor.author | Ding, X | - |
dc.contributor.author | Liao, XH | - |
dc.contributor.author | Wang, LS | - |
dc.date.accessioned | 2024-03-11T10:48:33Z | - |
dc.date.available | 2024-03-11T10:48:33Z | - |
dc.date.issued | 2023-11-01 | - |
dc.identifier.citation | IEEE Transactions on Medical Imaging, 2023, v. 42, n. 11, p. 3374-3383 | - |
dc.identifier.issn | 0278-0062 | - |
dc.identifier.uri | http://hdl.handle.net/10722/340954 | - |
dc.description.abstract | <p>The fusion of multi-modal medical data is essential to assist medical experts to make treatment decisions for precision medicine. For example, combining the whole slide histopathological images (WSIs) and tabular clinical data can more accurately predict the lymph node metastasis (LNM) of papillary thyroid carcinoma before surgery to avoid unnecessary lymph node resection. However, the huge-sized WSI provides much more high-dimensional information than low-dimensional tabular clinical data, making the information alignment challenging in the multi-modal WSI analysis tasks. This paper presents a novel transformer-guided multi-modal multi-instance learning framework to predict lymph node metastasis from both WSIs and tabular clinical data. We first propose an effective multi-instance grouping scheme, named siamese attention-based feature grouping (SAG), to group high-dimensional WSIs into representative low-dimensional feature embeddings for fusion. We then design a novel bottleneck shared-specific feature transfer module (BSFT) to explore the shared and specific features between different modalities, where a few learnable bottleneck tokens are utilized for knowledge transfer between modalities. Moreover, a modal adaptation and orthogonal projection scheme were incorporated to further encourage BSFT to learn shared and specific features from multi-modal data. Finally, the shared and specific features are dynamically aggregated via an attention mechanism for slide-level prediction. Experimental results on our collected lymph node metastasis dataset demonstrate the efficiency of our proposed components and our framework achieves the best performance with AUC (area under the curve) of 97.34%, outperforming the state-of-the-art methods by over 1.27%.</p> | - |
dc.language | eng | - |
dc.publisher | Institute of Electrical and Electronics Engineers | - |
dc.relation.ispartof | IEEE Transactions on Medical Imaging | - |
dc.rights | This work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License. | - |
dc.subject | knowledge transfer | - |
dc.subject | multi-modal multi-instance learning | - |
dc.subject | transformer | - |
dc.subject | Whole slide image | - |
dc.title | Shared-Specific Feature Learning With Bottleneck Fusion Transformer for Multi-Modal Whole Slide Image Analysis | - |
dc.type | Article | - |
dc.identifier.doi | 10.1109/TMI.2023.3287256 | - |
dc.identifier.pmid | 37335798 | - |
dc.identifier.scopus | eid_2-s2.0-85162901224 | - |
dc.identifier.volume | 42 | - |
dc.identifier.issue | 11 | - |
dc.identifier.spage | 3374 | - |
dc.identifier.epage | 3383 | - |
dc.identifier.eissn | 1558-254X | - |
dc.identifier.isi | WOS:001099088700016 | - |
dc.publisher.place | PISCATAWAY | - |
dc.identifier.issnl | 0278-0062 | - |