Integrating images and genomics for multi-modal cancer survival analysis via mixture of experts

Wei Zhang, Wenxin Xu, Tong Chen, Collin Sakal, Xinyue Li*

*Corresponding author for this work

Research output: Journal Publications and ReviewsRGC 21 - Publication in refereed journalpeer-review

14 Downloads (CityUHK Scholars)

Abstract

Survival prediction seeks to provide patient prognosis by measuring the time span from diagnosis or the first treatment to the occurrence of a specific event of interest. It represents a challenging ordinal regression task that often involves modeling the intricate interactions among multiple data modalities, such as genomic profiles and Whole Slide Images (WSIs). Despite recent advancements, two critical obstacles persist: (i) learning effective representations for each modality, and (ii) capturing the intricate interactions and heterogeneity among different features. To address these challenges, we propose SurMoE (Survival analysis with Mixture of Experts), a novel framework that designs a Mixture of Experts (MoE) architecture for multi-modal survival prediction. Specifically, we introduce a patch clustering layer to identify morphological prototypes from the vast collection of WSI patches and incorporate gene set enrichment analysis to capture biological associations among pathways and gene sets, yielding more robust modality representations. To model the intrinsic relationships within pathological WSIs and genomic profiles, we integrate multiple experts that dynamically adapt to diverse input patterns through a routing mechanism. Additionally, we employ cross-modal attention to seamlessly integrate multi-modal data and introduce a self-attention pooling module to refine modality-specific insights, thereby enhancing the accuracy of survival prediction. We conduct extensive experiments on five public TCGA datasets, supplemented by ablation studies, statistical analyses, and visualizations. The results demonstrate that our proposed SurMoE outperforms existing State-Of-The-Art (SOTA) methods, with an average increase of 2.29% in C-index across all datasets compared to the SOTA, underscoring the potential of MoE layer for refined multi-modal survival outcome prediction. Our code is publicly available at https://github.com/coffeeNtv/SurMoE. © 2025 The Authors. 
Original languageEnglish
Article number103521
JournalInformation Fusion
Volume126, Part A
Online published21 Jul 2025
DOIs
Publication statusPublished - Feb 2026

Funding

This work was supported by the Hong Kong Innovation and Technology Commission (InnoHK Project CIMDA), the City University of Hong Kong Institute of Digital Medicine, and City University of Hong Kong Internal Grant 7005967.

UN SDGs

This output contributes to the following UN Sustainable Development Goals (SDGs)

  1. SDG 3 - Good Health and Well-being
    SDG 3 Good Health and Well-being

Research Keywords

  • Computational pathology
  • Survival analysis
  • Multi-modal fusion
  • Mixture of experts

Publisher's Copyright Statement

  • This full text is made available under CC-BY-NC-ND 4.0. https://creativecommons.org/licenses/by-nc-nd/4.0/

Fingerprint

Dive into the research topics of 'Integrating images and genomics for multi-modal cancer survival analysis via mixture of experts'. Together they form a unique fingerprint.

Cite this