Login (DCU Staff Only)
Login (DCU Staff Only)

DORAS | DCU Research Repository

Explore open access research and scholarly works from DCU

Advanced Search

Developing multimodal cervical cancer risk assessment and prediction model based on LMIC hospital patient card sheets and histopathological images

Chane Jemane, Kelebet, Bedaso Kuyu, Muktar and Sahle Tegenaw, Geletaw (2025) Developing multimodal cervical cancer risk assessment and prediction model based on LMIC hospital patient card sheets and histopathological images. BMC Medical Informatics and Decision Making, 25 (322). ISSN 1472-6947

Abstract
Cervical cancer remains a significant global health burden, particularly in low- and middle-income countries (LMICs) where access to early diagnostic tools is limited. In Ethiopia, cervical cancer diagnosis often relies on manual interpretation of biopsies, which can be time-consuming and subjective. This study aims to develop a multimodal machine learning model that integrates histopathological images and associated patient clinical records to improve cervical cancer risk prediction and biopsy detection. The dataset comprises 404 biopsy images and corresponding clinical records from 499 patients, collected at Jimma Medical Center. The preprocessing of histopathological images and clinical records involved image enhancement, data augmentation, imputation of missing values, and class balancing techniques. Subsequently, (I) a pre-trained convolutional neural network deep learning (VGG16) model was applied on the histopathological dataset, (II) a Random Forest classifier was trained on the patient clinical records, and (III) a late fusion strategy was employed to integrate the outputs of both classifiers for multimodal analysis. Recursive Feature Elimination was used to identify key predictive factors from the patient data, and the model’s performance was thoroughly validated using accuracy, AUC-ROC curves, and confusion matrices, ensuring reliability across all classes. As a result, convolutional neural networks and Random Forest classifiers achieved accuracies of 91% and 96%, respectively. The integrated multimodal model achieved 92% accuracy, demonstrating enhanced robustness and clinical relevance by combining complementary data sources. These findings suggest that multimodal approaches hold promise for improving cervical cancer diagnostics in resource-limited settings. Future work will focus on validating the model with diverse datasets and integrating it into clinical workflows to support healthcare providers in LMICs.
Metadata
Item Type:Article (Published)
Refereed:Yes
Uncontrolled Keywords:Cervical cancer, Risk analysis, Deep learning, Predictive model, Machine learning, Multimodal model, LMICs
Subjects:Computer Science > Machine learning
Computer Science > Machine translating
DCU Faculties and Centres:Research Institutes and Centres > INSIGHT Centre for Data Analytics
Publisher:BioMed Central Ltd.
Official URL:https://bmcmedinformdecismak.biomedcentral.com/art...
Copyright Information:Authors
ID Code:31547
Deposited On:17 Sep 2025 09:48 by Gordon Kennedy . Last Modified 17 Sep 2025 09:48
Documents

Full text available as:

[thumbnail of s12911-025-03174-6.pdf]
Preview
PDF - Requires a PDF viewer such as GSview, Xpdf or Adobe Acrobat Reader
Creative Commons: Attribution-Noncommercial-No Derivative Works 4.0
2MB
Metrics

Altmetric Badge

Dimensions Badge

Downloads

Downloads

Downloads per month over past year

Archive Staff Only: edit this record