File Download

There are no files associated with this item.

  Links for fulltext
     (May Require Subscription)
Supplementary

Article: ADFound: A Foundation Model for Diagnosis and Prognosis of Alzheimer's Disease

TitleADFound: A Foundation Model for Diagnosis and Prognosis of Alzheimer's Disease
Authors
KeywordsAlzheimer's Disease
Contrastive Learning
Dementia
Foundation Model
Multi-modal Learning
Issue Date3-Jun-2025
PublisherInstitute of Electrical and Electronics Engineers
Citation
IEEE Journal of Biomedical and Health Informatics, 2025 How to Cite?
Abstract

Alzheimer's disease (AD) is an incurable neurodegenerative disorder characterized by progressive cognitive and functional decline. Consequently, early diagnosis and accurate prediction of disease progression are of paramount importance and inherently complex, necessitating the integration of multi-modal data. However, most existing methods are task-specific models that lack generalization ability, addressing only one task at a time and failing to simultaneously assess disease diagnosis and progression. In this paper, we introduce ADFound, the first foundation model for AD that serves as a basis for various downstream tasks, such as diagnosis and prognosis, with high generalization capability. ADFound leverages a substantial amount of unlabeled 3D multi-modal neuroimaging, including paired and unpaired data, to achieve its objectives. Specifically, ADFound is developed upon the Multi-modal Vim encoder by Vision Mamba block to capture long-range dependencies inherent in 3D multi-modal medical images. To efficiently pre-train ADFound on unlabeled paired and upaired multi-modal neuroimaging data, we proposed a novel self-supervised learning framework that integrates multi-modal masked autoencoder (MAE) and contrastive learning. The multi-modal MAE aims to learn local relations among modalities by reconstructing images with unmasked image patches. Additionally, we introduce a Dual Contrastive Learning for Multi-modal Data to enhance the discriminative capabilities of multi-modal representations from intra-modal and inter-modal perspectives. Our experiments demonstrate that ADFound outperforms stateof-the-art methods across a wide range of downstream tasks relevant to the diagnosis and prognosis of AD. Furthermore, the results indicate that our foundation model can be extended to more modalities, such as non-image data, showing its versatility.


Persistent Identifierhttp://hdl.handle.net/10722/360529
ISSN
2023 Impact Factor: 6.7
2023 SCImago Journal Rankings: 1.964

 

DC FieldValueLanguage
dc.contributor.authorYang, Guangqian-
dc.contributor.authorDu, Kangrui-
dc.contributor.authorYang, Zhihan-
dc.contributor.authorDu, Ye-
dc.contributor.authorCheung, Eva Yi Wah-
dc.contributor.authorZheng, Yongping-
dc.contributor.authorYang, Mo-
dc.contributor.authorKourtzi, Zoe-
dc.contributor.authorSchonlieb, Carola Bibiane-
dc.contributor.authorWang, Shujun-
dc.date.accessioned2025-09-12T00:36:34Z-
dc.date.available2025-09-12T00:36:34Z-
dc.date.issued2025-06-03-
dc.identifier.citationIEEE Journal of Biomedical and Health Informatics, 2025-
dc.identifier.issn2168-2194-
dc.identifier.urihttp://hdl.handle.net/10722/360529-
dc.description.abstract<p>Alzheimer's disease (AD) is an incurable neurodegenerative disorder characterized by progressive cognitive and functional decline. Consequently, early diagnosis and accurate prediction of disease progression are of paramount importance and inherently complex, necessitating the integration of multi-modal data. However, most existing methods are task-specific models that lack generalization ability, addressing only one task at a time and failing to simultaneously assess disease diagnosis and progression. In this paper, we introduce ADFound, the first foundation model for AD that serves as a basis for various downstream tasks, such as diagnosis and prognosis, with high generalization capability. ADFound leverages a substantial amount of unlabeled 3D multi-modal neuroimaging, including paired and unpaired data, to achieve its objectives. Specifically, ADFound is developed upon the Multi-modal Vim encoder by Vision Mamba block to capture long-range dependencies inherent in 3D multi-modal medical images. To efficiently pre-train ADFound on unlabeled paired and upaired multi-modal neuroimaging data, we proposed a novel self-supervised learning framework that integrates multi-modal masked autoencoder (MAE) and contrastive learning. The multi-modal MAE aims to learn local relations among modalities by reconstructing images with unmasked image patches. Additionally, we introduce a Dual Contrastive Learning for Multi-modal Data to enhance the discriminative capabilities of multi-modal representations from intra-modal and inter-modal perspectives. Our experiments demonstrate that ADFound outperforms stateof-the-art methods across a wide range of downstream tasks relevant to the diagnosis and prognosis of AD. Furthermore, the results indicate that our foundation model can be extended to more modalities, such as non-image data, showing its versatility.</p>-
dc.languageeng-
dc.publisherInstitute of Electrical and Electronics Engineers-
dc.relation.ispartofIEEE Journal of Biomedical and Health Informatics-
dc.rightsThis work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License.-
dc.subjectAlzheimer's Disease-
dc.subjectContrastive Learning-
dc.subjectDementia-
dc.subjectFoundation Model-
dc.subjectMulti-modal Learning-
dc.titleADFound: A Foundation Model for Diagnosis and Prognosis of Alzheimer's Disease-
dc.typeArticle-
dc.identifier.doi10.1109/JBHI.2025.3576436-
dc.identifier.scopuseid_2-s2.0-105007424227-
dc.identifier.eissn2168-2208-
dc.identifier.issnl2168-2194-

Export via OAI-PMH Interface in XML Formats


OR


Export to Other Non-XML Formats