Submitted:
23 July 2025
Posted:
25 July 2025
You are already at the latest version
Abstract

Keywords:
1. Introduction
2. Materials and Methods
2.1. Public CT Datasets for Kidney-Tumor Research
| Dataset / Collection | Year | Modality | Size (CT studies / images) | # Patients |
|---|---|---|---|---|
| C4KC-KiTS19 | 2019 | CT | 210 | 210 |
| KiTS21 (MICCAI 2021) | 2021 | CT | ≈ 300 | ≈ 300 |
| KiTS23 (MICCAI 2023) | 2023 | CT | 599 | 599 |
| CPTAC-CCRCC (TCIA[1])* | 2024 | CT | 85 studies† | 262 |
| TCGA-KIRC | 2020 | CT | 439 | 267 |
| TCGA-KIRP | 2020 | CT | 47 | 33 |
| TCGA-KICH | 2020 | CT | 15 | 15 |
| CT-Kidney (Normal–Cyst–Tumor–Stone) | 2022 | CT | 12 446 images | n/a |
2.2. Data Preprocessing
2.3. Model Architecture and Transfer Learning
2.4. Transfer Learning
2.5. Proposed M16+ Model for Tumor Classification
- Paired Conv–BN–ReLU blocks inserted before each max-pool layer (“dual-Conv”) lowered validation loss by 18 % and cut the generalization gap from 4.3 pp to 1.1 pp (see Supplementary Table S2), thereby mitigating over-fitting on heterogeneous renal CT data.
- Pre-trained weights accelerate convergence and reduce computational requirements.
- Binary Cross-Entropy loss, an 80 / 10 / 10 train–validation–test split, and early stopping, M16⁺ achieved 98.0 % accuracy, surpassing the fine-tuned VGG16 baseline (92.0 %)

2.6. Extended Model Comparisons and Cross Validation
3. Results
3.1. Performance Evaluation

3.2. Comparative Analysis
3.3. Training and Validation
3.4. Model Interpretability Using Grad-CAM++
4. Discussion
4.1. Clinical Implications of Key Findings
4.2. Comparative Analysis with the Latest Technology
4.3. Limitations and Mitigation Solutions
5. Conclusions
6. Patents
| 1 | * Dataset received a major update in May 2024; the CT radiology subset now contains 85 studies within a 262-subject cohort. |
| 2 | † Study count reported by TCIA; full collection comprises 727 series / 99 098 DICOM images. |
| 3 | GitHub repository: https://github.com/DaliaAlzubi/Kidney_Tumor_Detection_And_Classification
|
| 4 | Google Drive mirror: https://drive.google.com/file/d/1zp6b2o99_SdTEgxQwArlZWrMUfzLJj0/view
|
Supplementary Materials
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Acknowledgments
Conflicts of Interest
Abbreviations
| KAUH | King Abdullah University Hospital |
| CT | Computed tomography |
| MRI | Magnetic resonance imaging |
References
- Padala, S.A.; Barsouk, A.; Thandra, K.C.; Saginala, K.; Mohammed, A.; Vakiti, A.; et al. Epidemiology of renal cell carcinoma. World J Oncol. 2020, 11, 79–87. [Google Scholar] [CrossRef] [PubMed]
- Noriaki, S.; Uchida, E.; Ozawa, Y. Artificial intelligence in kidney pathology. In: Dey N, Ashour AS, eds. Artificial Intelligence in Medicine. Singapore: Springer; 2022. p. 539-549. [CrossRef]
- Liu, J.; Cao, L.; Akin, O.; Tian, Y. 3DFPN-HS²: 3D feature pyramid network based high-sensitivity and specificity pulmonary nodule detection. Lect Notes Comput Sci. 2019, 11769, 513–521. [Google Scholar] [CrossRef]
- Nadimi-Shahraki, M.H.; Taghian, S.; Mirjalili, S.; Abualigah, L. Binary Aquila optimizer for selecting effective features from medical data: a COVID-19 case study. Mathematics. 2022, 10, 1929. [Google Scholar] [CrossRef]
- Abualigah, L.; Diabat, A. Chaotic binary reptile search algorithm and its feature-selection applications. J Ambient Intell Human Comput. 2023, 14, 13931–13947. [Google Scholar] [CrossRef]
- Meenakshi, S.; Suganthi, M.; Sureshkumar, P. Segmentation and boundary detection of fetal kidney images in second and third trimesters using kernel-based fuzzy clustering. J Med Syst. 2019, 43, 243. [Google Scholar] [CrossRef] [PubMed]
- L; Zhao, Y. ; Wang, R.; Chang, M.; Purkayastha, S.; Chang, K.; et al. Deep learning to distinguish benign from malignant renal lesions on routine MR imaging. Clin Cancer Res. 2020, 26, 1944–1952. [CrossRef]
- Habibi-Aghdam, H.; Jahani-Heravi, E.; Shirazi-Parvazian, A.G. *Guide to Convolutional Neural Networks: A Practical Application to Traffic-Sign Detection and Classification.* Cham: Springer; 2018.
- Liu, H.; Cao, H.; Chen, L.; Fang, L.; Liu, Y.; Zhan, J.; et al. Quantitative evaluation of contrast-enhanced ultrasound in differentiating small renal cell carcinoma subtypes and angiomyolipoma. Quant Imaging Med Surg. 2022, 12, 106–118. [Google Scholar] [CrossRef] [PubMed]
- Alzu’bi, D.; Abdullah, M.; Hmeidi, I.; Alazab, R.; Gharaibeh, M.; El-Heis, M.; et al. Kidney tumor detection and classification based on deep learning approaches: a new dataset in CT scans. J Healthc Eng. 2022, 2022, 3861161. [Google Scholar] [CrossRef] [PubMed]
- Praveen, S.P.; Sidharth, S.R.; Priya, T.K.; Kavuri, Y.S.; Sindhura, S.M.; Donepudi, S. ResNet and ResNeXt-powered kidney tumor detection: a robust approach on a subset of the KAUH dataset. In: Proc ICACRS; Pudukkottai, India; 2023. p. 749-757. [CrossRef]
- Zhou, L.; Zhang, Z.; Chen, Y.C.; Zhao, Z.Y.; Yin, X.D.; Jiang, H.B. A deep-learning radiomics model for differentiating benign and malignant renal tumors. Transl Oncol. 2019, 12, 292–300. [Google Scholar] [CrossRef] [PubMed]
- Mahmud, S.; Abbas, T.O.; Mushtak, A.; Prithula, J.; Chowdhury, M.E.H. Kidney cancer diagnosis and surgery selection by machine learning from CT scans combined with clinical metadata. Cancers (Basel). 2023, 15, 3189. [Google Scholar] [CrossRef] [PubMed]
- Heller, N.; Isensee, F.; Trofimova, D.; Tejpaul, R.; Zhao, Z.; Chen, H.; et al. The KiTS21 challenge: automatic segmentation of kidneys, renal tumors, and renal cysts in corticomedullary-phase CT. 2023; arXiv:2307.01984. [Google Scholar]
- Ghalib, M.R.; Bhatnagar, S.; Jayapoorani, S.; Pande, U. Artificial neural network-based detection of renal tumors using CT-scan image processing. Int J Eng Technol. 2014, 6, 28–35. [Google Scholar]
- Zabihollahy, F.; Schieda, N.; Krishna, S.; Ukwatta, E. Automated classification of solid renal masses on contrast-enhanced CT images using a CNN with decision fusion. Eur Radiol. 2020, 30, 5183–5190. [Google Scholar] [CrossRef] [PubMed]
- Kukačka, J.; Golkov, V.; Cremers, D. Regularization for deep learning: a taxonomy. 2017; arXiv:1710.10686. [Google Scholar]
- Shorten, C.; Khoshgoftaar, T.M. A survey on image data augmentation for deep learning. J Big Data. 2019, 6, 60. [Google Scholar] [CrossRef]
- Weiss, K.; Khoshgoftaar, T.M.; Wang, D.D. A survey of transfer learning. J Big Data. 2016, 3, 9. [Google Scholar] [CrossRef]
- Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. 2015; arXiv:1409.1556. [Google Scholar]
- Selvaraju, R.R.; Das, A.; Vedantam, R.; Cogswell, M.; Parikh, D.; Batra, D. Grad-CAM: visual explanations from deep networks via gradient-based localization. Int J Comput Vis. 2020, 128, 336–359. [Google Scholar] [CrossRef]
- Sokolova, M.; Lapalme, G. A systematic analysis of performance measures for classification tasks. Inf Process Manag. 2009, 45, 427–437. [Google Scholar] [CrossRef]



| Model | Tumor Type | Precision | Recall | F1-Score | Accuracy% |
|---|---|---|---|---|---|
| Dalia(Alzu’Bi et al., 2022) | Benign | 0.99 | 0.89 | 0.94 | 92.0 |
| Malign | 0.80 | 0.98 | 0.88 | ||
| (Mahmud et al. al2023) | Benign | 84.18 | 85.66 | 84.92 | 90.6 |
| Malign | 90.83 | 90.61 | 90.50 | ||
| Praveen (2023) | Benign | - | - | - | 94.9 |
| Malign | - | - | - | ||
| M16+ (Proposed_Vgg16) | Benign | 0.99 | 0.98 | 0.98 | 98.0 |
| Malign | 0.99 | 0.98 | 0.98 | ||
| EfficientNet-B4 | Benign | 0.97 | 0.98 | 0.98 | 97.0 |
| Malign | 0.97 | 0.97 | 0.97 | ||
| Densnet 201 | Benign | 0.98 | 0.99 | 0.98 | 98.0 |
| Malign | 0.98 | 0.97 | 0.97 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).