Balancing Performance and Explainability in Deep Learning for Lung Disease Detection Using Unimodal and Multimodal Approaches
/Ghadir Moemen Helmy Ali
- 2025
- 214 p. ill. 21 cm.
Abstract: The adoption of deep learning in healthcare applications has highlighted a critical trade-off between model performance and explainability, particularly in clinical decision-making contexts. This thesis explores methods to address this trade-off through two key studies focusing on lung disease detection: pneumonia classification using chest X-rays and metastatic classification for non-small-cell lung cancer using CT imaging and genomics. In our work with pneumonia, attention mechanisms were integrated into DL models, serving as a balance between post-hoc and inherently explainable approaches. The use of attention improved classification performance while enhancing model focus on clinically relevant regions, as demonstrated through Grad-CAM visualizations. Attention proved to be an effective explainability choice when chest X-ray was the sole data source. For the non-small-cell lung cancer study, we adopted a multimodal approach combining CT imaging and RNA-sequencing data. Ablation studies revealed a novel clinical insight: surrounding tissue regions were more predictive of metastasis than tumor regions, aligning with Grad-CAM visualizations that emphasized tissue edges. This finding correlates with medical knowledge where tissue involvement signals metastasis progression. Genomic features alone exhibited limited predictive power but improved model performance when fused with imaging data. In contrast, handcrafted features, such as shape and size metrics, showed poor predictive capability. Given access to multimodal data, vii this study opens the door for exploring more clinically meaningful explainability techniques to further align AI model outputs with clinical reasoning. These results underscore the importance of context-specific explainability approaches in healthcare AI. While attention mechanisms are effective for unimodal data like chest X-ray, complex tasks with rich multimodal inputs, such as non-small-cell lung cancer classification, can benefit from advanced explainability methods. This thesis contributes to developing clinically explainable AI systems that strike a balance between performance and interpretability, ensuring their utility in real-world medical applications. Keywords Deep Learning, Convolutional Neural Networks, Attention Mechanisms, Explainable Artificial Intelligence, Lung Diseases, Medical Imagining, Bioinformatics, Radiogenomics, Clinical Applicability