Submitted:
21 October 2024
Posted:
23 October 2024
You are already at the latest version
Abstract
Keywords:
1. Introduction
2. Materials and Methods
2.1. Data Preprocessing
- Image Resizing: Each image is resized to a standard dimension, certifying uniformity across the dataset. This is important for consistency when feeding the images into the neural network models.
- Grayscale Conversion: The resized images are converted to grayscale to moderate computational complexity. Color information is typically redundant for the task at hand, as the focus is on texture and shape within the images.
- Gaussian Blurring: To reduce high-frequency noise, a Gaussian blur is applied to the grayscale images. This smoothing technique aids in highlighting the more significant structures within the brain by softening edges and details.
- Otsu's Thresholding: We implement Otsu's thresholding to separate the brain tissue from the background. This method automatically computes a threshold value for image binarization, which is employed to detect contours.
- Contour Detection and Selection: By applying contour detection, we identify the boundaries of all the objects in the binary image. We assume the largest contour to be the brain's boundary, a reasonable assumption in a typical MRI scan.
- Extreme Points and Cropping: Once the largest contour is identified, we verify its extreme points. These points represent the furthest pixels in the horizontal and vertical directions within the contour, which we employ to define a cropping boundary.
- Image Cropping: The image is cropped using the extreme points as vertices, with an optional padding added to ensure no part of the brain is excluded.
- Displaying the Process: We visualize the preprocessing steps, displaying the original image, the contour of the largest detected region, the extreme points, and finally, the cropped brain region.
2.2. Optimization Method
- Optimization of Data Augmentation Parameters: MVO was utilized to uncover the best data augmentation techniques and their respective parameter values. This step was crucial to enhance the dataset variability without deviating from the realistic transformations applicable to MRI images. By doing so, we aimed to maximize the performance of the classification models while avoiding overfitting.
- Layer-wise Trainability in CNNs: We also applied the MVO strategy to find the optimal configuration of trainable layers within VGG16 and ResNet50 architectures. The method helped in identifying which layers should be frozen (weights not updated during training) and which should be trainable (weights updated during training) to improve the models' accuracy and efficiency. This is particularly important as deep neural networks can be computationally expensive to train, and freezing certain layers can significantly diminish the number of parameters that essential to be updated, thus speeding up the training process and potentially improving generalization.
- Universe Size (Population Size): This parameter defines the number of potential solutions (universes) that the algorithm will consider simultaneously. A larger universe size allows for greater exploration of the solution space. A universe size of 30 achieves a compromise between computing efficiency and diversity within the search space. More extensive populations enhance diversity but may impede the optimization process, whereas smaller populations might converge rapidly yet risk overlooking ideal solutions due to inadequate research. A universe size of 30 facilitates a sufficiently diversified population to investigate many solutions while minimizing computational expenses.
- Wormhole Existence Probability (WEP): This probability controls the creation of wormholes, which are mechanisms for sharing information between universes. A value of 0.6 signifies a considerable probability of the formation of these wormholes. This likelihood facilitates an effective combination of exploration (investigating new regions of the solution space) and exploitation (enhancing current solutions). A high WEP (approaching 1) may result in premature convergence due to excessive emphasis on exploitation, whereas a low WEP (approaching 0) restricts exploitation, hence obstructing convergence to an ideal solution. A WEP of 0.6 guarantees that the MVO algorithm can efficiently navigate the solution space while simultaneously enhancing viable solutions.
- Travelling Distance Rate (TDR): This rate determines how much a solution can be altered when a wormhole event occurs. A value of 0.4 signifies that alterations to the solutions are moderate. If the TDR is overly elevated, solutions may be excessively disturbed, resulting in irregular exploration that could overlook excellent solutions. If TDR is excessively low, the algorithm may adopt an overly conservative approach and become trapped in local optima. A TDR of 0.4 offers a balanced methodology, enabling the algorithm to implement significant modifications to solutions while remaining close to attractive regions.
- Maximum/Minimum WEP: These parameters set the lower and upper bounds for the wormhole existence probability, introducing dynamic variability in the algorithm. The interval [0.2, 1.0] guarantees that during the initial phases of optimization, the method prioritizes exploration (with a reduced WEP at 0.2), so avoiding premature convergence. As the algorithm advances, WEP approaches 1.0, highlighting exploitation in the latter phases to optimize the results. The incremental rise in WEP enables the early investigation of several options, while refining the search in the concluding iterations for enhanced precision.
- Maximum/Minimum TDR: Similar to WEP, these parameters set the boundaries for the travelling distance rate, controlling the extent of solution alterations through the optimization process. Commencing with a minimum TDR of 0.4 guarantees that initial exploration remains closely aligned with the existing optimal solutions. As optimization progresses, the TDR may rise to 1.0, facilitating more substantial alterations in subsequent phases. This adaptive modification allows the method to maintain flexibility throughout the initial phases of optimization, while intensifying solution refinement as it nears convergence.
2.3. Data Augmentation
- Mitigates Overfitting: Augmentation spreads the variety of the training samples, which helps in preventing the model from memorizing specific images and overfitting.
- Improves Generalization: By simulating various scenarios, data augmentation allows the model to generalize better to new, unseen data.
- Compensates for Imbalanced Datasets: In cases where some classes are underrepresented, augmentation can help to balance the samples without the need to collect more data.
- Enhances Model Robustness: Augmented data aids the model to learn more robust features that are invariant to certain transformations, which is significant for real-world applications.
2.4. Model Building with VGG16 and ResNet50
2.5. Model Training
2.6. Dataset
-
No Tumor:
- This class encompasses brain MRI scans attained from individuals who do not demonstrate any detectable brain tumors. These scans serve as a crucial baseline for comparative analysis against scans that do depict tumors.
- Individuals within this class are typically those seeking routine brain examinations or experiencing neurological symptoms unrelated to tumor presence.
-
Tumor:
- The "Tumor" class comprises brain MRI scans from individuals who have been clinically diagnosed with brain tumors. Brain tumors marked in various locations, sizes, and types, making their accurate finding and grouping a formidable challenge in the field of medical imaging.
- The tumors within this class may include metastatic tumors, gliomas, meningiomas, and other forms of abnormal growths within the brain.
2.7. Model Evaluation
-
Precision (Equation 1):
- Description: Precision assesses the accuracy of the positive predictions made by the structure. In simpler terms, it answers the question: Of all the instances where the model predicted 'positive', how many were actually positive?
-
Components:
- True Positives (TP): These are the instances where the model precisely predicts the positive class.
- False Positives (FP): These are the instances where the model wrongly predicts the positive class.
- Usage: Precision is particularly essential in situations where the cost of a false positive is high. For example, in medical diagnostics, falsely recognizing a healthy patient as having a tumor can lead to unnecessary stress and treatment.
-
Accuracy (Equation 2):
- Description: Accuracy measures the proportion of true outcomes (both true negatives and positives) among the total number of cases examined. It essentially quantifies how often the model is correct.
-
Components:
- True Negatives (TN): These are the instances where the model properly predicts the negative class.
- False Negatives (FN): These are the instances where the model mistakenly predicts the negative class.
- Usage: Accuracy is a valuable measure when the classes in the dataset are balanced. However, its usefulness reduces when dealing with imbalanced datasets, as it can be misleadingly high in cases where the model mostly predicts the majority class correctly.
-
Recall (Sensitivity) (Equation 3):
- Description: Recall measures the proportion of actual positives that were accurately recognized by the model. It answers the question: Of all the instances that were actually positive, how many did the model identify?
- Usage: Recall is critical in contexts where missing a positive instance is significantly worse than falsely detecting a negative instance as positive. For example, in medical diagnostics, a false negative (failing to detect a tumor) can be more dangerous than a false positive.
3. Results and Discussion
4. Conclusions and Outlook
Author Contributions
Conflicts of interest/Competing interests
References
- M. M. Badža and M. Barjaktarović, “Segmentation of Brain Tumors from MRI Images Using Convolutional Autoencoder,” Applied Sciences 2021, Vol. 11, Page 4317, vol. 11, no. 9, p. 4317, May 2021. [CrossRef]
- S.-G. Choi, C.-B. Sohn, and P. D. Candidate, “Detection of HGG and LGG Brain Tumors using U-Net,” Medico Legal Update, vol. 19, no. 1, pp. 560–565, Feb. 2019. [CrossRef]
- R. Ranjbarzadeh, A. Keles, M. Crane, S. Anari, and M. Bendechache, “Secure and Decentralized Collaboration in Oncology: A Blockchain Approach to Tumor Segmentation,” 2024 IEEE 48th Annual Computers, Software, and Applications Conference (COMPSAC), pp. 1681–1686, Jul. 2024. [CrossRef]
- R. Ranjbarzadeh, A. Caputo, E. B. Tirkolaee, S. Jafarzadeh Ghoushchi, and M. Bendechache, “Brain tumor segmentation of MRI images: A comprehensive review on the application of artificial intelligence tools,” Comput Biol Med, vol. 152, p. 106405, Jan. 2023. [CrossRef]
- N. Tataei Sarshar et al., “Glioma Brain Tumor Segmentation in Four MRI Modalities Using a Convolutional Neural Network and Based on a Transfer Learning Method,” pp. 386–402, 2023. [CrossRef]
- R. Ranjbarzadeh, P. Zarbakhsh, A. Caputo, E. B. Tirkolaee, and M. Bendechache, “Brain Tumor Segmentation based on an Optimized Convolutional Neural Network and an Improved Chimp Optimization Algorithm,” Oct. 2022. [CrossRef]
- L. Zhang, J. Tan, D. Han, and H. Zhu, “From machine learning to deep learning: progress in machine intelligence for rational drug discovery,” Nov. 01, 2017, Elsevier Ltd. [CrossRef]
- R. Raj et al., “Machine learning-based dynamic mortality prediction after traumatic brain injury,” Sci Rep, vol. 9, no. 1, pp. 1–13, Dec. 2019. [CrossRef]
- A. B. Kasgari, S. Safavi, M. Nouri, J. Hou, N. T. Sarshar, and R. Ranjbarzadeh, “Point-of-Interest Preference Model Using an Attention Mechanism in a Convolutional Neural Network,” Bioengineering 2023, Vol. 10, Page 495, vol. 10, no. 4, p. 495, Apr. 2023. [CrossRef]
- A. R. Ranjbarzadeh et al., “Breast tumor localization and segmentation using machine learning techniques: Overview of datasets, findings, and methods,” Comput Biol Med, vol. 152, p. 106443, Jan. 2023. [CrossRef]
- A. Karkehabadi, M. Bakhshi, and S. B. Razavian, “Optimizing Underwater IoT Routing with Multi-Criteria Decision Making and Uncertainty Weights,” May 2024, Accessed: Jun. 06, 2024. [Online]. Available: https://arxiv.org/abs/2405.11513v1.
- A. Karkehabadi, H. Homayoun, and A. Sasan, “FFCL: Forward-Forward Net with Cortical Loops, Training and Inference on Edge Without Backpropagation,” May 2024. [CrossRef]
- S. mohammad E. Saryazdi, A. Etemad, A. Shafaat, and A. M. Bahman, “Data-driven performance analysis of a residential building applying artificial neural network (ANN) and multi-objective genetic algorithm (GA),” Build Environ, vol. 225, p. 109633, Nov. 2022. [CrossRef]
- S. Anari, N. Tataei Sarshar, N. Mahjoori, S. Dorosti, and A. Rezaie, “Review of Deep Learning Approaches for Thyroid Cancer Diagnosis,” Math Probl Eng, vol. 2022, pp. 1–8, Aug. 2022. [CrossRef]
- B. Pandey, D. Kumar Pandey, B. Pratap Mishra, and W. Rhmann, “A comprehensive survey of deep learning in the field of medical imaging and medical natural language processing: Challenges and research directions,” Journal of King Saud University - Computer and Information Sciences, Jan. 2021. [CrossRef]
- A. Conneau, D. Kiela, H. Schwenk, L. Barrault, and A. Bordes, “Supervised Learning of Universal Sentence Representations from Natural Language Inference Data,” EMNLP 2017 - Conference on Empirical Methods in Natural Language Processing, Proceedings, pp. 670–680, May 2017, Accessed: Jun. 15, 2020. [Online]. Available: http://arxiv.org/abs/1705.02364.
- A. Bagherian Kasgari, R. Ranjbarzadeh, A. Caputo, S. Baseri Saadi, and M. Bendechache, “Brain Tumor Segmentation Based on Zernike Moments, Enhanced Ant Lion Optimization, and Convolutional Neural Network in MRI Images,” pp. 345–366, 2023. [CrossRef]
- S. M. Mousavi, A. Asgharzadeh-Bonab, and R. Ranjbarzadeh, “Time-Frequency Analysis of EEG Signals and GLCM Features for Depth of Anesthesia Monitoring,” Comput Intell Neurosci, vol. 2021, pp. 1–14, Aug. 2021. [CrossRef]
- N. T. Sarshar and M. Mirzaei, “Premature Ventricular Contraction Recognition Based on a Deep Learning Approach,” J Healthc Eng, vol. 2022, 2022. [CrossRef]
- M. Abdossalehi and N. Tataei Sarshar, “Automated Cardiovascular Arrhythmia Classification Based on Through Nonlinear Features and Tunable-Q Wavelet Transform (TQWT) Based Decomposition,” 2021. [CrossRef]
- C. Gambella, B. Ghaddar, and J. Naoum-Sawaya, “Optimization problems for machine learning: A survey,” Eur J Oper Res, vol. 290, no. 3, pp. 807–828, May 2021. [CrossRef]
- J. Ver Berne, S. B. Saadi, C. Politis, and R. Jacobs, “A deep learning approach for radiological detection and classification of radicular cysts and periapical granulomas,” J Dent, vol. 135, p. 104581, Aug. 2023. [CrossRef]
- Z. Liu et al., “Deep learning based brain tumor segmentation: a survey,” Complex and Intelligent Systems, pp. 1–26, Jul. 2022. [CrossRef]
- Z. Liu et al., “Deep learning based brain tumor segmentation: a survey,” Complex & Intelligent Systems 2022, pp. 1–26, Jul. 2022. [CrossRef]
- P. K. Chahal, S. Pandey, and S. Goel, “A survey on brain tumor detection techniques for MR images,” Multimedia Tools and Applications 2020 79:29, vol. 79, no. 29, pp. 21771–21814, May 2020. [CrossRef]
- S. Vadhnani and N. Singh, “Brain tumor segmentation and classification in MRI using SVM and its variants: a survey,” Multimed Tools Appl, pp. 1–26, Apr. 2022. [CrossRef]
- R. Raza, U. Ijaz Bajwa, Y. Mehmood, M. Waqas Anwar, and M. Hassan Jamal, “dResU-Net: 3D deep residual U-Net based brain tumor segmentation from multimodal MRI,” Biomed Signal Process Control, vol. 79, p. 103861, Jan. 2023. [CrossRef]
- Z. Zhu, X. He, G. Qi, Y. Li, B. Cong, and Y. Liu, “Brain tumor segmentation based on the fusion of deep semantics and edge information in multimodal MRI,” Information Fusion, vol. 91, pp. 376–387, Mar. 2023. [CrossRef]
- M. Parhizkar and M. Amirfakhrian, “Car detection and damage segmentation in the real scene using a deep learning approach,” International Journal of Intelligent Robotics and Applications 2022, pp. 1–15, Apr. 2022. [CrossRef]
- S. Safavi and M. Jalali, “RecPOID: POI Recommendation with Friendship Aware and Deep CNN,” Future Internet 2021, Vol. 13, Page 79, vol. 13, no. 3, p. 79, Mar. 2021. [CrossRef]
- S. Safavi and M. Jalali, “DeePOF: A hybrid approach of deep convolutional neural network and friendship to Point-of-Interest (POI) recommendation system in location-based social networks,” Concurr Comput, vol. 34, no. 15, p. e6981, Jul. 2022. [CrossRef]
- R. Ranjbarzadeh, M. Crane, and M. Bendechache, “The Impact of Backbone Selection in Yolov8 Models on Brain Tumor Localization”. [CrossRef]
- R. Vankdothu and M. A. Hameed, “Brain tumor MRI images identification and classification based on the recurrent convolutional neural network,” Measurement: Sensors, vol. 24, p. 100412, Dec. 2022. [CrossRef]
- M. H. Siddiqi et al., “A Precise Medical Imaging Approach for Brain MRI Image Classification,” Comput Intell Neurosci, vol. 2022, 2022. [CrossRef]
- Z. Ullah, M. U. Farooq, S. H. Lee, and D. An, “A hybrid image enhancement based brain MRI images classification technique,” Med Hypotheses, vol. 143, p. 109922, Oct. 2020. [CrossRef]
- H. H. N. Alrashedy, A. F. Almansour, D. M. Ibrahim, and M. A. A. Hammoudeh, “BrainGAN: Brain MRI Image Generation and Classification Framework Using GAN Architectures and CNN Models,” Sensors 2022, Vol. 22, Page 4297, vol. 22, no. 11, p. 4297, Jun. 2022. [CrossRef]
- M. A. Khan et al., “Multimodal brain tumor detection and classification using deep saliency map and improved dragonfly optimization algorithm,” Int J Imaging Syst Technol, vol. 33, no. 2, pp. 572–587, Mar. 2023. [CrossRef]
- H. M. Balaha and A. E. S. Hassan, “A variate brain tumor segmentation, optimization, and recognition framework,” Artif Intell Rev, vol. 56, no. 7, pp. 7403–7456, Jul. 2023. [CrossRef]
- S. Deepa, J. Janet, S. Sumathi, and J. P. Ananth, “Hybrid Optimization Algorithm Enabled Deep Learning Approach Brain Tumor Segmentation and Classification Using MRI,” J Digit Imaging, vol. 36, no. 3, pp. 847–868, Jun. 2023. [CrossRef]
- A. E. Ezugwu et al., “Metaheuristics: a comprehensive overview and classification along with bibliometric analysis,” Artificial Intelligence Review 2021 54:6, vol. 54, no. 6, pp. 4237–4316, Mar. 2021. [CrossRef]
- R. P. França, A. C. B. Monteiro, V. V. Estrela, and N. Razmjooy, “Using Metaheuristics in Discrete-Event Simulation,” in Lecture Notes in Electrical Engineering, vol. 696, Springer Science and Business Media Deutschland GmbH, 2021, pp. 275–292. [CrossRef]
- N. Razmjooy, M. Ashourian, and Z. Foroozandeh, Eds., Metaheuristics and Optimization in Computer and Electrical Engineering, vol. 696. in Lecture Notes in Electrical Engineering, vol. 696. Cham: Springer International Publishing, 2021. [CrossRef]
- A. Hu and N. Razmjooy, “Brain tumor diagnosis based on metaheuristics and deep learning,” Int J Imaging Syst Technol, vol. 31, no. 2, pp. 657–669, Jun. 2021. [CrossRef]
- S. Mirjalili, S. M. Mirjalili, and A. Hatamlou, “Multi-Verse Optimizer: a nature-inspired algorithm for global optimization,” Neural Computing and Applications 2015 27:2, vol. 27, no. 2, pp. 495–513, Mar. 2015. [CrossRef]
- P. V. H. Son and N. T. Nguyen Dang, “Solving large-scale discrete time–cost trade-off problem using hybrid multi-verse optimizer model,” Scientific Reports 2023 13:1, vol. 13, no. 1, pp. 1–24, Feb. 2023. [CrossRef]
- Y. Han, W. Chen, A. A. Heidari, H. Chen, and X. Zhang, “A solution to the stagnation of multi-verse optimization: An efficient method for breast cancer pathologic images segmentation,” Biomed Signal Process Control, vol. 86, p. 105208, Sep. 2023. [CrossRef]
- A. Haseeb, U. Waleed, M. M. Ashraf, F. Siddiq, M. Rafiq, and M. Shafique, “Hybrid Weighted Least Square Multi-Verse Optimizer (WLS–MVO) Framework for Real-Time Estimation of Harmonics in Non-Linear Loads,” Energies 2023, Vol. 16, Page 609, vol. 16, no. 2, p. 609, Jan. 2023. [CrossRef]
- W. Xu and X. Yu, “A multi-objective multi-verse optimizer algorithm to solve environmental and economic dispatch,” Appl Soft Comput, vol. 146, p. 110650, Oct. 2023. [CrossRef]
- C. Shorten and T. M. Khoshgoftaar, “A survey on Image Data Augmentation for Deep Learning,” J Big Data, vol. 6, no. 1, pp. 1–48, Dec. 2019. [CrossRef]
- J. Nalepa, M. Marcinkiewicz, and M. Kawulok, “Data Augmentation for Brain-Tumor Segmentation: A Review,” Front Comput Neurosci, vol. 13, p. 83, Dec. 2019. [CrossRef]
- F. A. Zeiser et al., “Segmentation of Masses on Mammograms Using Data Augmentation and Deep Learning,” J Digit Imaging, vol. 33, no. 4, pp. 858–868, Aug. 2020. [CrossRef]
- F. López de la Rosa, J. L. Gómez-Sirvent, R. Sánchez-Reolid, R. Morales, and A. Fernández-Caballero, “Geometric transformation-based data augmentation on defect classification of segmented images of semiconductor materials using a ResNet50 convolutional neural network,” Expert Syst Appl, vol. 206, p. 117731, Nov. 2022. [CrossRef]
- N. Deepa and S. P. Chokkalingam, “Optimization of VGG16 utilizing the Arithmetic Optimization Algorithm for early detection of Alzheimer’s disease,” Biomed Signal Process Control, vol. 74, p. 103455, Apr. 2022. [CrossRef]
- F. Zhu, J. Li, B. Zhu, H. Li, and G. Liu, “UAV remote sensing image stitching via improved VGG16 Siamese feature extraction network,” Expert Syst Appl, vol. 229, p. 120525, Nov. 2023. [CrossRef]
- W. Bakasa and S. Viriri, “VGG16 Feature Extractor with Extreme Gradient Boost Classifier for Pancreas Cancer Prediction,” Journal of Imaging 2023, Vol. 9, Page 138, vol. 9, no. 7, p. 138, Jul. 2023. [CrossRef]
- S. Sarker, S. N. B. Tushar, and H. Chen, “High accuracy keyway angle identification using VGG16-based learning method,” J Manuf Process, vol. 98, pp. 223–233, Jul. 2023. [CrossRef]
- L. Mpova, T. Calvin Shongwe, and A. Hasan, “The Classification and Detection of Cyanosis Images on Lightly and Darkly Pigmented Individual Human Skins Applying Simple CNN and Fine-Tuned VGG16 Models in TensorFlow’s Keras API,” pp. 1–6, Sep. 2023. [CrossRef]
- Y. Zhang et al., “Deep Learning-based Automatic Diagnosis of Breast Cancer on MRI Using Mask R-CNN for Detection Followed by ResNet50 for Classification,” Acad Radiol, vol. 30, pp. S161–S171, Sep. 2023. [CrossRef]
- A. K. Sharma et al., “Brain tumor classification using the modified ResNet50 model based on transfer learning,” Biomed Signal Process Control, vol. 86, p. 105299, Sep. 2023. [CrossRef]
- J.-R. Lee, K.-W. Ng, and Y.-J. Yoong, “Face and Facial Expressions Recognition System for Blind People Using ResNet50 Architecture and CNN,” Journal of Informatics and Web Engineering, vol. 2, no. 2, pp. 284–298, Sep. 2023. [CrossRef]
- M. B. Hossain, S. M. H. S. Iqbal, M. M. Islam, M. N. Akhtar, and I. H. Sarker, “Transfer learning with fine-tuned deep CNN ResNet50 model for classifying COVID-19 from chest X-ray images,” Inform Med Unlocked, vol. 30, p. 100916, Jan. 2022. [CrossRef]
- “Brain Tumor Classification (MRI).” Accessed: May 04, 2024. [Online]. Available: https://www.kaggle.com/datasets/sartajbhuvaji/brain-tumor-classification-mri/data.
- S. Krishnapriya and Y. Karuna, “Pre-trained deep learning models for brain MRI image classification,” Front Hum Neurosci, vol. 17, p. 1150120, Apr. 2023. [CrossRef]
- D. V. Gore, A. K. Sinha, and V. Deshpande, “Automatic CAD System for Brain Diseases Classification Using CNN-LSTM Model,” pp. 623–634, 2023. [CrossRef]
- S. Anari, G. G. de Oliveira, R. Ranjbarzadeh, A. M. Alves, G. C. Vaz, and M. Bendechache, “EfficientUNetViT: Efficient Breast Tumor Segmentation utilizing U-Net Architecture and Pretrained Vision Transformer,” Aug. 2024. [CrossRef]
- S. Baseri Saadi, N. Tataei Sarshar, S. Sadeghi, R. Ranjbarzadeh, M. Kooshki Forooshani, and M. Bendechache, “Investigation of Effectiveness of Shuffled Frog-Leaping Optimizer in Training a Convolution Neural Network,” J Healthc Eng, vol. 2022, pp. 1–11, Mar. 2022. [CrossRef]
- S. Mirjalili and A. Lewis, “The Whale Optimization Algorithm,” Advances in Engineering Software, vol. 95, pp. 51–67, May 2016. [CrossRef]
- A. M. Fathollahi-Fard, M. Hajiaghaei-Keshteli, and R. Tavakkoli-Moghaddam, “Red deer algorithm (RDA): a new nature-inspired meta-heuristic,” Soft comput, vol. 24, no. 19, pp. 14637–14665, Oct. 2020. [CrossRef]













| Hyperparameters | Description | Best value |
|---|---|---|
| Universe Size | Number of solutions in the population | 30 |
| Wormhole Existence Probability (WEP) | Probability of wormholes' appearance | 0.6 |
| Travelling Distance Rate (TDR) | How far a wormhole can alter a solution | 0.4 |
| Max/Min WEP | Bounds the occurrence of wormholes | Max: 1.0, Min: 0.2 |
| Max/Min TDR | Limits the modification extent by the wormholes | Max: 1.0, Min: 0.4 |
| Index | Augmentation Method | Description | Used in Study | Suggested values by MVO optimizer |
|---|---|---|---|---|
| 1 | Rotation range | Random rotation within a specified range of degrees. | Yes | 10 |
| 2 | Width Shifting range | Random horizontal shift. | Yes | 0.1 |
| 3 | Height Shifting range | Random vertical shift. | Yes | 0.1 |
| 4 | Shear range | Random shearing. | Yes | 0.1 |
| 5 | Brightness range | Random brightness adjustments. | Yes | [0.5 1.5] |
| 6 | Horizontal Flip | Random horizontal flipping. | Yes | True |
| 7 | Vertical Flip | Random vertical flipping. | Yes | True |
| 8 | Zoom range | Random zooming into the images. | Yes | 0.1 |
| 9 | Channel Shift* | Random channel shifting. | No | - |
| 10 | Fill Mode* | Method for filling points outside the boundaries of the input. | No | - |
| Data | Model | Accuracy (%) | Precision (%) | Recall(%) | Run results |
|---|---|---|---|---|---|
| Validation |
VGG16 |
89 | 91 | 96 | ![]() |
| Test | 88 | 92 | 94 | ||
| Validation |
ResNet50 |
92 | 95 | 96 | ![]() |
| Test | 92 | 96 | 94 | ||
| Validation |
VGG16 + ResNet50 |
92 | 93 | 98 | ![]() |
| Test | 89 | 92 | 96 | ||
| Validation | VGG16 + ResNet50 + MVO | 93 | 97 | 95 | ![]() |
| Test | 94 | 97 | 96 |
| Data | Model | Accuracy (%) | Precision (%) | Recall (%) |
|
|---|---|---|---|---|---|
| Validation | Our model + SFLA [66] | 91 | 93 | 96 | ![]() |
| Test | 91 | 94 | 96 | ||
| Validation | Our model + WOA [67] | 85 | 97 | 85 | ![]() |
| Test | 88 | 98 | 87 | ||
| Validation | Our model + RDA [68] | 89 | 97 | 89 | ![]() |
| Test | 91 | 89 | 91 | ||
| Validation | Our model + MVO [44] | 93 | 97 | 95 | ![]() |
| Test | 94 | 97 | 96 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).








