Your browser doesn't support javascript.
loading
Multimodal integration of neuroimaging and genetic data for the diagnosis of mood disorders based on computer vision models.
Lee, Seungeun; Cho, Yongwon; Ji, Yuyoung; Jeon, Minhyek; Kim, Aram; Ham, Byung-Joo; Joo, Yoonjung Yoonie.
Afiliación
  • Lee S; Department of Mathematics, Korea University, Anamro 145, Seoungbuk-gu, Seoul, 02841, Republic of Korea.
  • Cho Y; Department of Computer Science and Engineering, Soonchunhyang University, South Korea, Republic of Korea.
  • Ji Y; Division of Life Science, Korea University, Anamro 145, Seoungbuk-gu, Seoul, 02841, Republic of Korea.
  • Jeon M; Division of Biotechnology, Korea University, Anamro 145, Seoungbuk-gu, Seoul, 02841, Republic of Korea; Computational Biology Department, Carnegie Mellon University, Pittsburgh, PA, 15213, United States.
  • Kim A; Department of Biomedical Sciences, Korea University College of Medicine, Seoul, 02841, Republic of Korea.
  • Ham BJ; Department of Psychiatry, Korea University Anam Hospital, 73, Goryeodae-ro, Seoungbuk-gu, Seoul, 02841, Republic of Korea. Electronic address: byungjoo.ham@gmail.com.
  • Joo YY; Department of Digital Health, Samsung Advanced Institute for Health Sciences & Technology (SAIHST), Sungkyunkwan University, Samsung Medical Center, 115 Irwon-Ro, Gangnam-Gu, Seoul, 06355, Republic of Korea. Electronic address: helloyjjoo@gmail.com.
J Psychiatr Res ; 172: 144-155, 2024 Apr.
Article en En | MEDLINE | ID: mdl-38382238
ABSTRACT
Mood disorders, particularly major depressive disorder (MDD) and bipolar disorder (BD), are often underdiagnosed, leading to substantial morbidity. Harnessing the potential of emerging methodologies, we propose a novel multimodal fusion approach that integrates patient-oriented brain structural magnetic resonance imaging (sMRI) scans with DNA whole-exome sequencing (WES) data. Multimodal data fusion aims to improve the detection of mood disorders by employing established deep-learning architectures for computer vision and machine-learning strategies. We analyzed brain imaging genetic data of 321 East Asian individuals, including 147 patients with MDD, 78 patients with BD, and 96 healthy controls. We developed and evaluated six fusion models by leveraging common computer vision models in image classification Vision Transformer (ViT), Inception-V3, and ResNet50, in conjunction with advanced machine-learning techniques (XGBoost and LightGBM) known for high-dimensional data analysis. Model validation was performed using a 10-fold cross-validation. Our ViT ⊕ XGBoost fusion model with MRI scans, genomic Single Nucleotide polymorphism (SNP) data, and unweighted polygenic risk score (PRS) outperformed baseline models, achieving an incremental area under the curve (AUC) of 0.2162 (32.03% increase) and 0.0675 (+8.19%) and incremental accuracy of 0.1455 (+25.14%) and 0.0849 (+13.28%) compared to SNP-only and image-only baseline models, respectively. Our findings highlight the opportunity to refine mood disorder diagnostics by demonstrating the transformative potential of integrating diverse, yet complementary, data modalities and methodologies.
Asunto(s)
Palabras clave

Texto completo: 1 Colección: 01-internacional Base de datos: MEDLINE Asunto principal: Trastorno Bipolar / Trastorno Depresivo Mayor Límite: Humans Idioma: En Revista: J Psychiatr Res Año: 2024 Tipo del documento: Article

Texto completo: 1 Colección: 01-internacional Base de datos: MEDLINE Asunto principal: Trastorno Bipolar / Trastorno Depresivo Mayor Límite: Humans Idioma: En Revista: J Psychiatr Res Año: 2024 Tipo del documento: Article