![]()
Multimodal Deep Learning for Lung Cancer Analysis Using Data Visualisation
Dharmaiah Devarapalli1, M. Kavitha2, J Adithya Narasimha3, M Soma Shekar4, K Jaswanth5
1Dr. Dharmaiah Devarapalli, Department of Computer Science and Engineering, Koneru Lakshmaiah Education Foundation, Vaddeswaram, (A.P.), India.
2Dr. M. Kavitha, Department of Computer Science and Engineering, Koneru Lakshmaiah Education Foundation, Vaddeswaram, (A.P.), India.
3J. Adithya Narasimha, Department of Computer Science and Engineering, Koneru Lakshmaiah Education Foundation, Vaddeswaram, (A.P.), India.
4M. Soma Shekar, Department of Computer Science and Engineering, Koneru Lakshmaiah Education Foundation, Vaddeswaram, (A.P.), India.
5K. Jaswanth, Department of Computer Science and Information Technology, Koneru Lakshmaiah Education Foundation, Vaddeswaram, (A.P.), India.
Manuscript received on 28 September 2025 | Revised Manuscript received on 05 October 2025 | Manuscript Accepted on 15 October 2025 | Manuscript published on 30 October 2025 | PP: 6-11 | Volume-13 Issue-11, October 2025 | Retrieval Number: 100.1/ijese.L262113121125 | DOI: 10.35940/ijese.L2621.13111025
Open Access | Editorial and Publishing Policies | Cite | Zenodo | OJS | Indexing and Abstracting
© The Authors. Blue Eyes Intelligence Engineering and Sciences Publication (BEIESP). This is an open access article under the CC-BY-NC-ND license (http://creativecommons.org/licenses/by-nc-nd/4.0/)
Abstract: Lung cancer is still one of the leading causes of cancer deaths worldwide. Early and accurate risk prediction can help doctors make better decisions and improve patient outcomes. In this work, we develop a deep-learning framework that combines clinical records, imaging features, and survey data to predict lung cancer and its prognosis. For imaging, we use pretrained convolutional neural networks to extract features from CT and X-ray images. For clinical history, we use recurrent models, and for structured data, we apply gradient-ensemble models. We combine these features into a fully connected layer and fine-tune the model end-to-end. We test our model on several open datasets, including Kaggle lung CT sets, IQ-OTHNCCD, and a diagnostic survey dataset. We report accuracy, precision, recall, F1, and ROC-AUC. To ensure a fair evaluation, we use stratified cross-validation, tune hyperparameters, and run ablation studies to see how each data type contributes. Our combined model consistently outperforms both image-only and tabular-only models. It improves ROC-AUC and F1 scores and reduces false negatives, which is especially important for diagnosis. We also provide interactive visualisations in Looker Studio to show which features matter most, how risk is distributed, and confusion matrices, helping clinicians understand the model. We use statistical tests to confirm our improvements and discuss ways to make predictions more understandable for clinical use. Our results demonstrate that combining different data types enhances accuracy and provides valuable insights that support informed clinical decisions. We also discuss limitations, such as differences between datasets and label noise, and suggest further experiments and external validation to move closer to clinical use.
Keywords: Machine Learning, Artificial Intelligence, Convolutional Neural Network, Recurrent Neural Network, Receiver Operating Characteristic, Area Under the Curve, Computed Tomography, Comma-Separated Values.
Scope of the Article: Artificial Intelligence and Methods
