Submitted:
01 September 2023
Posted:
04 September 2023
You are already at the latest version
Abstract
Keywords:
1. Introduction
- An in-depth comparison and evaluation of three methods for building CNN ensembles, both standalone and in combination, verified across five different data sets;
- The introduction of several new methods for perturbing network weights;
- Free access to all resources, including the MATLAB source code, used in our experiments.
2. Related Work
3. Methods
3.1. Activation Functions
3.2. Data Augmentation
- APP1: This augmentation generates three new images based on a given image. It randomly reflects the image vertically and horizontally, resulting in two new images. The third transformation involves linearly scaling the original image along both axes with two factors randomly selected from a uniform distribution ranging from 1 to 2.
- APP2: Building upon APP1, this augmentation generates six new images. It includes the transformations of APP1 and adds three additional manipulations. First, image rotation is applied with a random angle extracted from the range of -10 to 10 degrees. Second, translation is performed by shifting the image along both axes with values randomly sampled from the interval of 0 to 5 pixels. Last, shear transformation is applied, with vertical and horizontal angles randomly selected from the range of 0 to 30 degrees.
- APP3: This augmentation replicates APP2 but excludes the shear and the scale transformations, resulting in four new images.
- APP4: This augmentation approach generates three new images by applying a transform based on Principal Component Analysis (PCA). The PCA coefficients extracted from a given image are subjected to three perturbations that generate three new images. For the first image, each element of the feature vector has a 50% probability of being randomly set to zero. For the second, noise is added to each component based on the standard deviation of the projected image. For the third, five images from the same class as the original image are selected, and their PCA vectors are computed. With a 5% probability, components from the original PCA vector are swapped with corresponding components from the other five PCA vectors. The three perturbed PCA vectors are then transformed back using the inverse PCA transform to produce the augmented images.
- APP5: Similar to APP4, this augmentation generates three new images using the perturbation method described above. However, instead of using PCA, the Discrete Cosine Transform (DCT) is applied. It should be noted that the DC coefficient is never changed during this transformation.
- APP6: This augmentation is designed specifically for color images. It creates three new images by color shifting and by altering contrast and sharpness. Contrast alteration is achieved by linearly scaling the original image’s contrast between the lowest value (a) and the highest value (b) allowed for the augmented image. Any pixel in the original image outside this range is mapped to 0 if it is lower than a or 255 if it is greater than b. Sharpness is modified by blurring the original image with a Gaussian filter (variance = 1) and subtracting the blurred image from the original. Color shifting is performed by applying integer shifts to the three RGB filters, and each shift is added to one of the three channels in the original image.
3.3. Parameter Ensembling via Perturbation (PEP)
- Dout: similar to drop-out: 2% of the weights zeroed out;
- DCTa: each set of weights is projected onto a Discrete Cosine Transform (DCT) space, with (3.33%) randomly chosen DCT coefficients set to zero (the DC component is never zeroed out), after which the inverse DCT is applied.
- DCTb: each set of weights is projected onto a Discrete Cosine Transform (DCT) space where a small amount of random noise is injected (the DC component is never perturbed), after which the inverse DCT is applied.
- PEPa: method similar to the original version, but where a small amount of random noise is injected.
- PEPb: the same idea as PEPa, but noise is injected in a different manner.
| Listing 1. Dout: similar to dropout filter |
| Perturbation = rand(size(Weights)); % Weights are the weights of the given net % Perturbation is a tensor of the same size as the % set of weights of the net, randomly initialized to [0,1] Weights = Weights.*Perturbation; % some weights are zeroed out |
| Listing 2. DCTa: DCT based perturbation approach |
| for each layer for each channel IMG = Weights(layer,channel); % weights of a given channel-layer are stored dctProj = dct2(IMG); % DCT projection dctProj_reset = dctProj; % reset some random dct coefficients dctProj_reset("random indexes") = 0; % DC component is never zeroed out dctProj_reset(1,1) = dctProj(1,1); Weights(layer,channel) = idct2(dctProj_reset); % retroprojection end end |
| Listing 3. DCTb: DCT based perturbation approach |
| for each layer for each channel IMG = Weights(layer,channel); % weights of a given channel-layer are stored dctProj = dct2(IMG); % DCT projection % standard deviation of the values of the weights noise = std(dctProj) / 4; % random noise dctProjNew = dctProj + (rand-0.5) .* noise; % rand is random between 0 and 1 dctProjNew(1,1) = dctProj(1,1); % DC component is never zeroed out Weights(layer, channel) = idct2(dctProjNew); % retroprojection end end |
| Listing 4. PEPa: Method 3 is similar to DCT1 |
| sigma = 0.002; Weights = Weights + rand(size(Weights)) .* sigma; % Weights are the weights of the given net |
| Listing 5. PEPb: Method 3 is similar to DCT1 |
| sigma = 0.2; Weights = Weights .* (1 + rand(size(Weights)) .* sigma); % Weights are the weights of the given net |
4. Experimental Results
4.1. Datasets
- HE (2D HeLa data set [60]): This data set has 862 fluorescence microscopy images of HeLa cells stained with different fluorescent dyes specific to various organelles. The data set is well-balanced and divided into ten classes representing different organelles, including DNA (Nuclei), ER (Endoplasmic reticulum), Giantin (cis/medial Golgi), GPP130 (cis Golgi), Lamp2 (Lysosomes), Nucleolin (Nucleoli), Actin, TfR (Endosomes), Mitochondria, and Tubulin. A 5-fold cross-validation is applied.
- MA (C. elegans Muscle Age data set [61]): This data set focuses on classifying the age of C. elegans nematodes. It has 257 images of C. elegans muscles collected at four different ages, representing distinct classes based on age. A 5-fold cross-validation is applied.
- BG (Breast Grading Carcinoma [62]): This data set, obtained from Zenodo (record: 834910#.Wp1bQ-jOWUl), has 300 annotated histological images of breast tissues from patients diagnosed with invasive ductal carcinoma. The data set is categorized into three classes representing different grades (1-3) of carcinoma. A 5-fold cross-validation is applied.
- LAR (Laryngeal data-set [63]): Obtained from Zenodo (record: 1003200#.WdeQcnBx0nQ), has 1320 images of laryngeal tissues. It includes both healthy and early-stage cancerous tissues, representing a total of four tissue classes. This data set is split into three folds by the original authors.
- POR (portrait dataset) data set [64] focuses specifically on portrait images of humans. It is designed to evaluate segmentation performance in the context of portrait photography, considering factors such as facial features, skin tones, and background elements. This dataset includes 1447 images for training and 289 images for validation. POR can be accessed at https://github.com/HYOJINPARK/ExtPortraitSeg.
4.2. Results
- Data augmentation is useful both for RE and SE.
- Both RE and SE increase performance only slightly from x = 14 to x = 30.
- SE outperforms RE.
- StocDA_PEP(18), which was created as follows: for each DA method, we train three networks for a total of eighteen. Each network is then coupled with one of the five PEP variants (randomly chosen).
- StocDA(18), which was created as follows: For each DA method, we train three SE networks for a total of eighteen.
- PEST, [65], It is a dataset of 563 pest images, 10 classes, commonly found on plants. We use the split training-test sets suggested by the original authors.
- InfLAR, [66], it is a dataset of 720 images, four classes, extracted from laryngoscopic videos. We use the split training-test sets (three different folds) suggested by the original authors.
- TRIZ, [67], it is a dataset of 574 gastric lesion type images, four classes; as suggested by the original authors we apply a 10-fold cross-validation.
- In data sets with gray-level images, StocDA_PEP(18) and StocDA(18) are less useful (probably because some data augmentation methods are suitable for color images), whereas in LAR, POR, PEST, InfLAR, and TRIZ they fare better;
- There is a noticeable difference in performance between each ensemble compared to the original single ResNet50.
- considering the well-known Wilcoxon-signed rank test, considering both the performance indicators (i.e. accuracy and EUC): RE(18)-DA outperforms RE(1)-DA with a p-value of 0.01; SE(18)-DA outperforms RE(18)-DA with a p-value of 0.05; StocDA(18) obtains similar performance with respect to SE(18)-DA (due to the gray level images datasets); StocDA_PEP(18) outperforms StocDA(18) with a p-value of 0.05.
5. Conclusion
Author Contributions
Funding
Data Availability Statement
Acknowledgments
Conflicts of Interest
References
- Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. Advances in neural information processing systems 2012, 25. [Google Scholar] [CrossRef]
- Haggenmüller, S.; Maron, R.C.; Hekler, A.; Utikal, J.S.; Barata, C.; Barnhill, R.L.; Beltraminelli, H.; Berking, C.; Betz-Stablein, B.; Blum, A.; others. Skin cancer classification via convolutional neural networks: systematic review of studies involving human experts. European Journal of Cancer 2021, 156, 202–216. [Google Scholar] [CrossRef] [PubMed]
- Esteva, A.; Kuprel, B.; Novoa, R.A.; Ko, J.; Swetter, S.M.; Blau, H.M.; Thrun, S. Dermatologist-level classification of skin cancer with deep neural networks. nature 2017, 542, 115–118. [Google Scholar] [CrossRef]
- Horie, Y.; Yoshio, T.; Aoyama, K.; Yoshimizu, S.; Horiuchi, Y.; Ishiyama, A.; Hirasawa, T.; Tsuchida, T.; Ozawa, T.; Ishihara, S.; others. Diagnostic outcomes of esophageal cancer by artificial intelligence using convolutional neural networks. Gastrointestinal endoscopy 2019, 89, 25–32. [Google Scholar] [CrossRef]
- Qummar, S.; Khan, F.G.; Shah, S.; Khan, A.; Shamshirband, S.; Rehman, Z.U.; Khan, I.A.; Jadoon, W. A deep learning ensemble approach for diabetic retinopathy detection. Ieee Access 2019, 7, 150530–150539. [Google Scholar] [CrossRef]
- Pan, D.; Zeng, A.; Jia, L.; Huang, Y.; Frizzell, T.; Song, X. Early detection of Alzheimer’s disease using magnetic resonance imaging: a novel approach combining convolutional neural networks and ensemble learning. Frontiers in neuroscience 2020, 14, 259. [Google Scholar] [CrossRef]
- Nanni, L.; Loreggia, A.; Lumini, A.; Dorizza, A. A Standardized Approach for Skin Detection: Analysis of the Literature and Case Studies. Journal of Imaging 2023, 9, 35. [Google Scholar] [CrossRef] [PubMed]
- Thanapol, P.; Lavangnananda, K.; Bouvry, P.; Pinel, F.; Leprévost, F. Reducing overfitting and improving generalization in training convolutional neural network (CNN) under limited sample sizes in image recognition. 2020-5th International Conference on Information Technology (InCIT). IEEE, 2020, pp. 300–305.
- Campagner, A.; Ciucci, D.; Svensson, C.M.; Figge, M.T.; Cabitza, F. Ground truthing from multi-rater labeling with three-way decision and possibility theory. Information Sciences 2021, 545, 771–790. [Google Scholar] [CrossRef]
- Panch, T.; Mattie, H.; Celi, L.A. The “inconvenient truth” about AI in healthcare. NPJ digital medicine 2019, 2, 77. [Google Scholar] [CrossRef]
- Bravin, R.; Nanni, L.; Loreggia, A.; Brahnam, S.; Paci, M. Varied Image Data Augmentation Methods for Building Ensemble. IEEE Access 2023, 11, 8810–8823. [Google Scholar] [CrossRef]
- Nanni, L.; Cuza, D.; Lumini, A.; Loreggia, A.; Brahman, S. Polyp Segmentation with Deep Ensembles and Data Augmentation. In Artificial Intelligence and Machine Learning for Healthcare: Vol. 1: Image and Data Analytics; Springer, 2022; pp. 133–153.
- Nanni, L.; Fantozzi, C.; Loreggia, A.; Lumini, A. Ensembles of Convolutional Neural Networks and Transformers for Polyp Segmentation. Sensors 2023, 23, 4688. [Google Scholar] [CrossRef]
- Nanni, L.; Lumini, A.; Loreggia, A.; Brahnam, S.; Cuza, D. Deep ensembles and data augmentation for semantic segmentation. In Diagnostic Biomedical Signal and Image Processing Applications with Deep Learning Methods; Elsevier, 2023; pp. 215–234.
- Papanastasopoulos, Z.; Samala, R.K.; Chan, H.P.; Hadjiiski, L.; Paramagul, C.; Helvie, M.A.; Neal, C.H. Explainable AI for medical imaging: deep-learning CNN ensemble for classification of estrogen receptor status from breast MRI. Medical imaging 2020: Computer-aided diagnosis. SPIE, 2020, Vol. 11314, pp. 228–235.
- He, X.; Zhou, Y.; Wang, B.; Cui, S.; Shao, L. Dme-net: Diabetic macular edema grading by auxiliary task learning. International Conference on Medical Image Computing and Computer-Assisted Intervention. Springer, 2019, pp. 788–796.
- Coupé, P.; Mansencal, B.; Clément, M.; Giraud, R.; de Senneville, B.D.; Ta, V.T.; Lepetit, V.; Manjon, J.V. AssemblyNet: A large ensemble of CNNs for 3D whole brain MRI segmentation. NeuroImage 2020, 219, 117026. [Google Scholar] [CrossRef]
- Savelli, B.; Bria, A.; Molinara, M.; Marrocco, C.; Tortorella, F. A multi-context CNN ensemble for small lesion detection. Artificial Intelligence in Medicine 2020, 103, 101749. [Google Scholar] [CrossRef]
- Cornelio, C.; Donini, M.; Loreggia, A.; Pini, M.S.; Rossi, F. Voting with random classifiers (VORACE): theoretical and experimental analysis. Auton. Agent 2021, 35, 2. [Google Scholar] [CrossRef]
- Yao, X.; Liu, Y. Making use of population information in evolutionary artificial neural networks. IEEE Transactions on Systems, Man, and Cybernetics, Part B (Cybernetics) 1998, 28, 417–425. [Google Scholar]
- Opitz, D.; Shavlik, J. Generating accurate and diverse members of a neural-network ensemble. Advances in neural information processing systems 1995, 8. [Google Scholar]
- Liu, Y.; Yao, X.; Higuchi, T. Evolutionary ensembles with negative correlation learning. IEEE Transactions on Evolutionary Computation 2000, 4, 380–387. [Google Scholar]
- Rosen, B.E. Ensemble learning using decorrelated neural networks. Connection science 1996, 8, 373–384. [Google Scholar] [CrossRef]
- Liu, Y.; Yao, X. Ensemble learning via negative correlation. Neural networks 1999, 12, 1399–1404. [Google Scholar] [CrossRef] [PubMed]
- He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2016, pp. 770–778. [CrossRef]
- Kassani, S.H.; Kassani, P.H.;Wesolowski, M.J.; Schneider, K.A.; Deters, R. Classification of histopathological biopsy images using ensemble of deep learning networks. arXiv preprint arXiv:1909.11870 2019.
- Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:1409.1556 2014. arXiv:1409.1556 2014.
- Sandler, M.; Howard, A.; Zhu, M.; Zhmoginov, A.; Chen, L.C. Mobilenetv2: Inverted residuals and linear bottlenecks. Proceedings of the IEEE conference on computer vision and pattern recognition, 2018, pp. 4510–4520.
- Huang, G.; Liu, Z.; Van Der Maaten, L.; Weinberger, K.Q. Densely connected convolutional networks. Proceedings of the IEEE conference on computer vision and pattern recognition, 2017, pp. 4700–4708.
- He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. Proceedings of the IEEE conference on computer vision and pattern recognition, Las Vegas, NV, USA, 26 June–, pp. 770–778. 1 July.
- Szegedy, C.; Vanhoucke, V.; Ioffe, S.; Shlens, J.; Wojna, Z. Rethinking the inception architecture for computer vision. Proceedings of the IEEE conference on computer vision and pattern recognition, 2016, pp. 2818–2826.
- Chollet, F. Xception: Deep learning with depthwise separable convolutions. Proceedings of the IEEE conference on computer vision and pattern recognition, 2017, pp. 1251–1258.
- Liu, K.; Zhang, M.; Pan, Z. Facial expression recognition with CNN ensemble. 2016 international conference on cyberworlds (CW). IEEE, 2016, pp. 163–166.
- Goodfellow, I.J.; Erhan, D.; Carrier, P.L.; Courville, A.; Mirza, M.; Hamner, B.; Cukierski, W.; Tang, Y.; Thaler, D.; Lee, D.H.; others. Challenges in representation learning: A report on three machine learning contests. Neural Information Processing: 20th International Conference, ICONIP 2013, Daegu, Korea, -7, 2013. Proceedings, Part III 20. Springer, 2013, pp. 117–124. 3 November.
- Kumar, A.; Kim, J.; Lyndon, D.; Fulham, M.; Feng, D. An ensemble of fine-tuned convolutional neural networks for medical image classification. IEEE journal of biomedical and health informatics 2016, 21, 31–40. [Google Scholar] [CrossRef]
- Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going deeper with convolutions. Proceedings of the IEEE conference on computer vision and pattern recognition, 2015, pp. 1–9.
- Gilbert, A.; Piras, L.; Wang, J.; Yan, F.; Ramisa, A.; Dellandrea, E.; Gaizauskas, R.J.; Villegas, M.; Mikolajczyk, K.; others. Overview of the ImageCLEF 2016 Scalable Concept Image Annotation Task. CLEF (Working Notes), 2016, pp. 254–278.
- Pandey, P.; Deepthi, A.; Mandal, B.; Puhan, N.B. FoodNet: Recognizing foods using ensemble of deep networks. IEEE Signal Processing Letters 2017, 24, 1758–1762. [Google Scholar] [CrossRef]
- Breiman, L. Bagging predictors. Machine learning 1996, 24, 123–140. [Google Scholar] [CrossRef]
- Wolpert, D.H.; Macready, W.G. An efficient method to estimate bagging’s generalization error. Machine Learning 1999, 35, 41–55. [Google Scholar] [CrossRef]
- Bauer, E.; Kohavi, R. An empirical comparison of voting classification algorithms: Bagging, boosting, and variants. Machine learning 1999, 36, 105–139. [Google Scholar] [CrossRef]
- Kim, P.K.; Lim, K.T. Vehicle type classification using bagging and convolutional neural network on multi view surveillance image. Proceedings of the IEEE conference on computer vision and pattern recognition workshops, 2017, pp. 41–46.
- Dong, X.; Qian, L.; Huang, L. A CNN based bagging learning approach to short-term load forecasting in smart grid. 2017 IEEE SmartWorld, Ubiquitous Intelligence & Computing, Advanced & Trusted Computed, Scalable Computing & Communications, Cloud & Big Data Computing, Internet of People and Smart City Innovation (SmartWorld/SCALCOM/UIC/ATC/CBDCom/IOP/SCI). IEEE, 2017, pp. 1–6.
- Guo, J.; Gould, S. Deep CNN ensemble with data augmentation for object detection. arXiv preprint arXiv:1506.07224 2015. arXiv:1506.07224 2015.
- Gan, Y.; Chen, J.; Xu, L. Facial expression recognition boosted by soft label with a diverse ensemble. Pattern Recognition Letters 2019, 125, 105–112. [Google Scholar] [CrossRef]
- Antipov, G.; Berrani, S.A.; Dugelay, J.L. Minimalistic CNN-based ensemble model for gender prediction from face images. Pattern recognition letters 2016, 70, 59–65. [Google Scholar] [CrossRef]
- Nanni, L.; Brahnam, S.; Paci, M.; Ghidoni, S. Comparison of different convolutional neural network activation functions and methods for building ensembles for small to midsize medical data sets. Sensors 2022, 22, 6129. [Google Scholar] [CrossRef] [PubMed]
- Nanni, L.; Lumini, A.; Ghidoni, S.; Maguolo, G. Stochastic selection of activation layers for convolutional neural networks. Sensors 2020, 20, 6. [Google Scholar] [CrossRef]
- Berno, F.; Nanni, L.; Maguolo, G.; Brahnam, S. Ensembles of convolutional neural networks with different activation functions for small to medium size biomedical datasets. Machine Learning in Medicine; CRC Press Taylor & Francis Group: Boca Raton, FL, USA 2021. [Google Scholar]
- Ju, C.; Bibaut, A.; van der Laan, M. The relative performance of ensemble methods with deep convolutional neural networks for image classification. Journal of Applied Statistics 2018, 45, 2800–2818. [Google Scholar] [CrossRef]
- Harangi, B. Skin lesion classification with ensembles of deep convolutional neural networks. Journal of biomedical informatics 2018, 86, 25–32. [Google Scholar] [CrossRef]
- Lyksborg, M.; Puonti, O.; Agn, M.; Larsen, R. An ensemble of 2D convolutional neural networks for tumor segmentation. Image Analysis: 19th Scandinavian Conference, SCIA 2015, Copenhagen, Denmark, -17, 2015. Proceedings 19. Springer, 2015, pp. 201–211. 15 June.
- Minetto, R.; Segundo, M.P.; Sarkar, S. Hydra: An ensemble of convolutional neural networks for geospatial land classification. IEEE Transactions on Geoscience and Remote Sensing 2019, 57, 6530–6541. [Google Scholar] [CrossRef]
- Dong, X.; Yu, Z.; Cao, W.; Shi, Y.; Ma, Q. A survey on ensemble learning. Frontiers of Computer Science 2020, 14, 241–258. [Google Scholar] [CrossRef]
- Brown, G.; Wyatt, J.; Harris, R.; Yao, X. Diversity creation methods: a survey and categorisation. Information fusion 2005, 6, 5–20. [Google Scholar] [CrossRef]
- Sagi, O.; Rokach, L. Ensemble learning: A survey. Wiley Interdisciplinary Reviews: Data Mining and Knowledge Discovery 2018, 8, e1249. [Google Scholar] [CrossRef]
- Lumini, A.; Nanni, L.; Maguolo, G. Deep learning for plankton and coral classification. Applied Computing and Informatics 2020. [Google Scholar] [CrossRef]
- Duch, W.; Jankowski, N. Survey of neural transfer functions. Neural computing surveys 1999, 2, 163–212. [Google Scholar]
- Mehrtash, A.; Abolmaesumi, P.; Golland, P.; Kapur, T.; Wassermann, D.; Wells, W. Pep: Parameter ensembling by perturbation. Advances in neural information processing systems 2020, 33, 8895–8906. [Google Scholar]
- Boland, M.V.; Murphy, R.F. A neural network classifier capable of recognizing the patterns of all major subcellular structures in fluorescence microscope images of HeLa cells. Bioinformatics 2001, 17, 1213–1223. [Google Scholar] [CrossRef] [PubMed]
- Shamir, L.; Orlov, N.; Mark Eckley, D.; Macura, T.J.; Goldberg, I.G. IICBU 2008: a proposed benchmark suite for biological image analysis. Medical & biological engineering & computing 2008, 46, 943–947. [Google Scholar]
- Dimitropoulos, K.; Barmpoutis, P.; Zioga, C.; Kamas, A.; Patsiaoura, K.; Grammalidis, N. Grading of invasive breast carcinoma through Grassmannian VLAD encoding. PloS one 2017, 12, e0185110. [Google Scholar] [CrossRef] [PubMed]
- Moccia, S.; De Momi, E.; Guarnaschelli, M.; Savazzi, M.; Laborai, A.; Guastini, L.; Peretti, G.; Mattos, L.S. Confident texture-based laryngeal tissue classification for early stage diagnosis support. Journal of Medical Imaging 2017, 4, 034502–034502. [Google Scholar] [CrossRef] [PubMed]
- Kim, Y.W.; Byun, Y.C.; Krishna, A.V.N. Portrait Segmentation Using Ensemble of Heterogeneous Deep-Learning Models. Entropy 2021, 23. [Google Scholar] [CrossRef] [PubMed]
- Deng, L.; Wang, Y.; Han, Z.; Yu, R. Research on insect pest image detection and recognition based on bio-inspired methods. Biosystems Engineering 2018, 169, 139–148. [Google Scholar] [CrossRef]
- Patrini, I.; Ruperti, M.; Moccia, S.; Mattos, L.S.; Frontoni, E.; De Momi, E. Transfer learning for informative-frame selection in laryngoscopic videos through learned features. Medical and Biological Engineering and Computing 2020, 58, 1225–1238. [Google Scholar] [CrossRef] [PubMed]
- Zhao, R.; Zhang, R.; Tang, T.; Feng, X.; Li, J.; Liu, Y.; Zhu, R.; Wang, G.; Li, K.; Zhou, W.; Yang, Y.; Wang, Y.; Ba, Y.; Zhang, J.; Liu, Y.; Zhou, F. TriZ-a rotation-tolerant image feature and its application in endoscope-based disease diagnosis. Computers in Biology and Medicine 2018, 99, 182–190. [Google Scholar] [CrossRef] [PubMed]
| HE | MA | BG | LAR | POR | Average | |
|---|---|---|---|---|---|---|
| RE(1)-noDA | 94.65 | 92.50 | 91.67 | 90.98 | 85.74 | 91.11 |
| RE(14)-noDA | 96.05 | 95.00 | 90.33 | 94.02 | 87.15 | 92.51 |
| RE(30)-noDA | 95.81 | 94.58 | 90.67 | 94.02 | 87.15 | 92.44 |
| RE(1)-DA | 95.93 | 95.83 | 92.67 | 94.77 | 86.29 | 93.10 |
| RE(14)-DA | 96.63 | 97.50 | 94.33 | 95.76 | 88.24 | 94.49 |
| RE(30)-DA | 96.33 | 98.33 | 94.00 | 95.83 | 88.56 | 94.61 |
| SE(14)-noDA | 95.47 | 95.42 | 92.67 | 94.62 | 88.02 | 93.24 |
| SE(30)-noDA | 95.58 | 96.25 | 92.67 | 95.00 | 88.77 | 93.65 |
| SE(14)-DA | 96.63 | 98.33 | 94.67 | 95.98 | 88.67 | 94.86 |
| SE(30)-DA | 96.33 | 98.33 | 95.00 | 96.21 | 89.00 | 94.97 |
| DataAUG | HE | MA | BG | LAR | POR | Average |
|---|---|---|---|---|---|---|
| DA1 | 95.12 | 95.00 | 93.00 | 92.95 | 87.05 | 92.62 |
| DA2 | 96.63 | 95.83 | 94.00 | 95.08 | 85.97 | 93.50 |
| DA3 | 95.93 | 95.83 | 92.67 | 94.77 | 86.29 | 93.10 |
| DA4 | 95.23 | 93.33 | 92.33 | 94.62 | 84.90 | 92.08 |
| DA5 | 95.35 | 91.25 | 91.33 | 95.45 | 86.41 | 91.95 |
| DA6 | 92.44 | 91.25 | 92.33 | 94.39 | 87.37 | 91.55 |
| ALL | 96.74 | 97.50 | 94.00 | 96.06 | 89.00 | 94.66 |
| RE(6)-DA | 96.40 | 97.08 | 93.67 | 95.98 | 88.45 | 94.31 |
| HE | MA | BG | LAR | POR | Average | |
|---|---|---|---|---|---|---|
| DropOut | 94.53 | 95.00 | 88.33 | 92.65 | 84.57 | 91.10 |
| DCTa | 94.88 | 93.33 | 92.00 | 94.09 | 85.11 | 91.88 |
| DCTb | 93.95 | 94.17 | 90.00 | 92.35 | 84.79 | 91.05 |
| PEPa | 95.58 | 93.33 | 89.67 | 92.20 | 85.22 | 91.20 |
| PEPb | 94.77 | 92.08 | 89.33 | 92.58 | 85.11 | 90.77 |
| PEPa(5) | 95.93 | 96.25 | 90.33 | 94.02 | 86.94 | 92.69 |
| ALL | 96.05 | 97.08 | 90.67 | 94.24 | 86.95 | 93.00 |
| RE(5)-noDA | 95.47 | 94.58 | 91.33 | 93.48 | 86.82 | 92.33 |
| HE | MA | BG | LAR | POR | PEST | InfLAR | TRIZ | Average | |
|---|---|---|---|---|---|---|---|---|---|
| RE(1)-DA | 95.93 | 95.83 | 92.67 | 94.77 | 86.29 | 93.70 | 95.56 | 98.78 | 94.19 |
| RE(18)-DA | 96.33 | 98.33 | 94.33 | 95.61 | 88.13 | 93.87 | 96.30 | 98.78 | 95.21 |
| SE(18)-DA | 96.51 | 98.33 | 95.00 | 96.06 | 88.56 | 94.36 | 96.67 | 98.95 | 95.55 |
| StocDA(18) | 96.10 | 96.67 | 94.33 | 96.81 | 89.96 | 94.48 | 96.53 | 98.95 | 95.47 |
| StocDA_PEP(18) | 96.40 | 97.50 | 94.00 | 96.82 | 91.68 | 94.14 | 97.08 | 99.13 | 95.84 |
| HE | MA | BG | LAR | POR | PEST | InfLAR | TRIZ | Average | |
|---|---|---|---|---|---|---|---|---|---|
| RE(1)-DA | 0.40 | 0.79 | 2.74 | 0.41 | 2.69 | 0.75 | 0.54 | 0.10 | 1.05 |
| RE(18)-DA | 0.22 | 0.16 | 2.32 | 0.18 | 2.05 | 0.71 | 0.49 | 0.13 | 0.78 |
| SE(18)-DA | 0.14 | 0.06 | 2.72 | 0.14 | 1.88 | 0.57 | 0.49 | 0.05 | 0.75 |
| StocDA(18) | 0.15 | 0.10 | 2.96 | 0.09 | 1.36 | 0.53 | 0.41 | 0.04 | 0.70 |
| StocDA_PEP(18) | 0.10 | 0.07 | 1.67 | 0.07 | 1.31 | 0.52 | 0.40 | 0.03 | 0.52 |
| GPU | GPU Year | Single ResNet50 | Ensemble 15 ResNet50 |
|---|---|---|---|
| GTX 1080 | 2016 | 0.36 sec | 5.58 sec |
| Titan Xp | 2017 | 0.31 sec | 4.12 sec |
| Titan RTX | 2018 | 0.22 sec | 2.71 sec |
| Titan V100 | 2018 | 0.20 sec | 2.42 sec |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2023 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).
