Submitted:
23 August 2024
Posted:
27 August 2024
You are already at the latest version
Abstract
Keywords:
MSC: 68U10
1. Introduction
2. Literature Review, Research Gaps, and Contributions
- Most available studies relied on individual model prediction to perform the crack semantic segmentation. Nevertheless, it is well-known that the individual model might suffer from low variance and low generalization ability in case of data alteration.
- To overcome the overfitting of crack image data, many studies focus on various hybridizations or modifications of existing models as well as transfer learning which still do not incorporate the knowledge of multiple learning to perform the concrete semantic segmentation task.
- Crack semantic segmentation underlies several problems, particularly when dealing with complex and highly contaminated image backgrounds, blurring, shadows, etc. Therefore, it is necessary to improve the existing identification method and include novel techniques.
- The ensemble learning is a very effective method to improve the performance of individual learners by combining their knowledge using some well-established methods, such as weighted averaging, stacking, bagging, and boosting.
- For pixel-level semantic segmentation especially in case of crack images, the abovementioned ensemble learning methods are less popular among researchers. This is mainly due to problems related to computational cost and difficulties in optimizing ensemble learning parameters.
- The traditional weighted average ensemble learning for pixel-level semantic segmentation might suffer from pixel blurring of crack boundaries resulting high bias of predicted crack map than the ground truth.
- It is well-known that Pixel-level semantic segmentation is of high spatial correlation features which do not highly suit the independent sampling of supervised learning. Moreover, as most pixels belong to background and to crack area, class imbalance is inevitable in pixel level crack detection. These two reasons make the use of traditional ensemble learning such as boosting and stacking difficult.
- Hence, it is of great significance to improve the existing ensemble learning methods for pixel-level semantic segmentation, especially when considering crack images that naturally include various background contaminations.
- By leveraging the ensemble deep learning philosophy, a novel corporative semantic segmentation of concrete cracks method called Co-CrackSegment is proposed.
- Five models, namely the U-net, SegNet, DeepCrack19, and DeepLabV3 with ResNet50, and ResNet101 backbones are trained to serve as core models for the Co-CrackSegment.
- To build the corporative model, a new iterative approach based best evaluation metrics, namely the dice score, IoU, pixel accuracy, precision, and recall metrics is developed.
- Finally, a detailed numerical and visual comparisons between the Co-CrackSegment and the core models as well as the weighted average ensemble learning model is presented.
3. Materials and Methods
3.1. Crack Semantic Segmentation Framework.
3.2. Datasets


3.3. The Core Models
3.2.1. The U-net.
3.2.2. The SegNet
3.2.3. DeepCrack19
3.2.4. The DeepLabV3 with Backbones
3.4. Training Procedure
3.5. Evaluation Metrics
3.6. The Proposed Method
- Load N trained semantic segmentation models in the model_list.
- Choose one Co-CrackSegment framework, namely Co-CrackSegment/dice, Co-CrackSegment/IoU, Co-CrackSegment/Pixel_Acc, Co-CrackSegment/Precision, or Co-CrackSegment/Recall.
- Set best_evaluation_metric_score to -1, and best_model_metrics to an empty matrix.
-
For each test image do the following:
- (a)
- For each current_model in the model_list (N times)
- (b)
- Set the trainer.model to the current_model.
- (c)
- Evaluate current_model with test image and compute the segmentation prediction output.
- (d)
- Compute the overall evaluation metric scores including the current_evaluation_metric_score of the test image (current_model_metrics).
- (e)
- If (current_evaluation_metric_score >best_ evaluation_metric_score)
- i.
- best_ evaluation_metric_score = current_ evaluation_metric_score
- ii.
- best_model_metrics= current_model_metrics
- iii.
- Add trainer.model to the evaluation results matrix.
- Show the results

- 1-
- Load N trained semantic segmentation models in the model_list.
- 2-
- Set current_model to model1, and model_outputs to an empty matrix.
- 3-
-
For each test image do the following:For each current_model in the model_list
- Compute the prediction of the current model.
-
Multiply predictions by the weight of the modelweighted_predictions = predictions * weights[j]
-
Add the weighted predictions to the listmodel_outputs.append(weighted_predictions).
- 4-
- Perform weighted average sum:
- 5-
- Compute metrics for the ensemble output
- 6-
- Show the results
4. Results and Discussion
4.1. Performances of the Core Models
4.2. Performances of Co-CrackSegment
4.3. Visual Comparison and Discussion
| Test sample 1 | Ground Truth | DeepLabV3Plus resnet50 | DeepCrack19 | SegNet | Unet | DeepLabV3Plus resnet101 |
![]() | ||||||
| Co-CrackSegment/ Dice | Co-CrackSegment/ IoU | Co-CrackSegment/ Pixel_Acc | Co-CrackSegment/ Precision | Co-CrackSegment/ Recall | Weighted average ensemble | |
![]() |
||||||
| Test sample 2 | Ground Truth | DeepLabV3Plus resnet50 | DeepCrack19 | SegNet | Unet | DeepLabV3Plus resnet101 |
![]() | ||||||
| Co-CrackSegment/ Dice | Co-CrackSegment/ IoU | Co-CrackSegment/ Pixel_Acc | Co-CrackSegment/ Precision | Co-CrackSegment/ Recall | Weighted average ensemble | |
![]() |
||||||
| Test sample 3 | Ground Truth | DeepLabV3Plus resnet50 | DeepCrack19 | SegNet | Unet | DeepLabV3Plus resnet101 |
![]() | ||||||
| Co-CrackSegment/ Dice | Co-CrackSegment/ IoU | Co-CrackSegment/ Pixel_Acc | Co-CrackSegment/ Precision | Co-CrackSegment/ Recall | Weighted average ensemble | |
![]() |
||||||
| Test sample 4 | Ground Truth | DeepLabV3Plus resnet50 | DeepCrack19 | SegNet | Unet | DeepLabV3Plus resnet101 |
![]() | ||||||
| Co-CrackSegment/ Dice | Co-CrackSegment/ IoU | Co-CrackSegment/ Pixel_Acc | Co-CrackSegment/ Precision | Co-CrackSegment/ Recall | Weighted average ensemble | |
![]() |
||||||
| Test sample 5 | Ground Truth | DeepLabV3Plus resnet50 | DeepCrack19 | SegNet | Unet | DeepLabV3Plus resnet101 |
![]() | ||||||
| Co-CrackSegment/ Dice | Co-CrackSegment/ IoU | Co-CrackSegment/ Pixel_Acc | Co-CrackSegment/ Precision | Co-CrackSegment/ Recall | Weighted average ensemble | |
![]() |
||||||
| Test sample 6 | Ground Truth | DeepLabV3Plus resnet50 | DeepCrack19 | SegNet | Unet | DeepLabV3Plus resnet101 |
![]() | ||||||
| Co-CrackSegment/ Dice | Co-CrackSegment/ IoU | Co-CrackSegment/ Pixel_Acc | Co-CrackSegment/ Precision | Co-CrackSegment/ Recall | Weighted average ensemble | |
![]() |
||||||
| Test sample 7 | Ground Truth | DeepLabV3Plus resnet50 | DeepCrack19 | SegNet | Unet | DeepLabV3Plus resnet101 |
![]() | ||||||
| Co-CrackSegment/ Dice | Co-CrackSegment/ IoU | Co-CrackSegment/ Pixel_Acc | Co-CrackSegment/ Precision | Co-CrackSegment/ Recall | Weighted average ensemble | |
![]() |
||||||
| Test sample 1 | Ground Truth | DeepLabV3Plus resnet50 | DeepCrack19 | SegNet | Unet | DeepLabV3Plus resnet101 |
![]() | ||||||
| Co-CrackSegment/ Dice | Co-CrackSegment/ IoU | Co-CrackSegment/ Pixel_Acc | Co-CrackSegment/ Precision | Co-CrackSegment/ Recall | Weighted average ensemble | |
![]() |
||||||
| Test sample 2 | Ground Truth | DeepLabV3Plus resnet50 | DeepCrack19 | SegNet | Unet | DeepLabV3Plus resnet101 |
![]() | ||||||
| Co-CrackSegment/ Dice | Co-CrackSegment/ IoU | Co-CrackSegment/ Pixel_Acc | Co-CrackSegment/ Precision | Co-CrackSegment/ Recall | Weighted average ensemble | |
![]() |
||||||
| Test sample 3 | Ground Truth | DeepLabV3Plus resnet50 | DeepCrack19 | SegNet | Unet | DeepLabV3Plus resnet101 |
![]() | ||||||
| Co-CrackSegment/ Dice | Co-CrackSegment/ IoU | Co-CrackSegment/ Pixel_Acc | Co-CrackSegment/ Precision | Co-CrackSegment/ Recall | Weighted average ensemble | |
![]() |
||||||
| Test sample 4 | Ground Truth | DeepLabV3Plus resnet50 | DeepCrack19 | SegNet | Unet | DeepLabV3Plus resnet101 |
![]() | ||||||
| Co-CrackSegment/ Dice | Co-CrackSegment/ IoU | Co-CrackSegment/ Pixel_Acc | Co-CrackSegment/ Precision | Co-CrackSegment/ Recall | Weighted average ensemble | |
![]() |
||||||
| Test sample 5 | Ground Truth | DeepLabV3Plus resnet50 | DeepCrack19 | SegNet | Unet | DeepLabV3Plus resnet101 |
![]() | ||||||
| Ensemble (Dice) | Ensemble (IoU) | Ensemble (Pixel_Acc) | Ensemble (Precision) | Ensemble (Recall) | Ensemble (weighted average) | |
![]() |
||||||
| Test sample 6 | Ground Truth | DeepLabV3Plus resnet50 | DeepCrack19 | SegNet | Unet | DeepLabV3Plus resnet101 |
![]() | ||||||
| Co-CrackSegment/ Dice | Co-CrackSegment/ IoU | Co-CrackSegment/ Pixel_Acc | Co-CrackSegment/ Precision | Co-CrackSegment/ Recall | Weighted average ensemble | |
![]() |
||||||
| Test sample 7 | Ground Truth | DeepLabV3Plus resnet50 | DeepCrack19 | SegNet | Unet | DeepLabV3Plus resnet101 |
![]() | ||||||
| Co-CrackSegment/ Dice | Co-CrackSegment/ IoU | Co-CrackSegment/ Pixel_Acc | Co-CrackSegment/ Precision | Co-CrackSegment/ Recall | Weighted average ensemble | |
![]() |
||||||
5. Conclusion
- Under the theme of the core models, it has been reported that the U-net has achieved a prominent performance by means of loss, pixel accuracy, IoU, dice, and mAP when trained using dataset1. It has also been observed that the DLV3/ResNet50 and DLV3/ResNet101 have also achieved high performances when compromising the overall evaluation metrics. Moreover, the iteration per second score of the DeepCrack19 has given it competitive advantages as a computationally efficient model. Moreover, when trained using dataset2 and similar to dataset1, the U-net has also achieved an outstanding by means of loss, accuracy, IoU, recall, dice, and mAP. Also, the DeepCrack19 has shown better computational performance than the other models when considering the number of iterations per seconds.
- When studying the proposed corporative semantic segmentation Co-CrackSegment approach, the Co-CrackSegment/dice and Co-CrackSegment/IoU have shown the best trade off evaluation scores comparing to other Co-CrackSegment frameworks. Furthermore, when comparing to weighted average method, most Co-CrackSegment frameworks outperformed the weighted average ensemble as well as the core models by means of all evaluation metrics. This is because the traditional weighted average ensemble learning for pixel-level semantic segmentation suffer from pixel blurring of crack boundaries due to average predictions resulting in high bias of predicted crack map than the ground truth. Furthermore, when comparing the results of core models with the Co-CrackSegment frameworks, it has been observed that the corporative learning approach has boosted the performance of the individual models by means of all evaluation metrics. This proofs the efficiency of Co-CrackSegment approach for pixel level semantic segmentation of surface cracks.
- When studying feeding the developed models with test samples that contain many challenges, such as crack-like scaling, foreign objects, this cracks, bulges, voids, spalling, etc. It has been reported that all the developed Co-CrackSegment approach for pixel level semantic segmentation of surface cracks have given very enhanced crack maps even in challenging cases. Also, the Co-CrackSegment/Dice and Co-CrackSegment/IoU frameworks have achieved the best performance comparing to other Co-CrackSegment frameworks and the weighted average method as well as the core models. This confirms the results presented in the previous discussion.
- Finally, several future improvements can be done to improve the proposed method. Firstly, the Co-CrackSegment approach can accept the insertion of any semantic segmentation model. This mainly due to its flexibility to add core models to its main framework. Moreover, the Co-CrackSegment method can be boosted by improving the utilized performance metrics to make a better trade-off between the original performance metrics that have already been used in its framework. Furthermore, the proposed Co-CrackSegment method can be further improved for multi-level semantic segmentation of structural surface defects. Lastly, the Co-CrackSegment can be easily adapted to be used in other semantic segmentation applications.
Supplementary Materials
Author Contributions
Funding
Data Availability Statement
Acknowledgments
Conflicts of Interest
References
- N. F. Alkayem, M. Cao, Y. Zhang, M. Bayat and Z. Su, "Structural damage detection using finite element model updating with evolutionary algorithms: a survey," Neural Computing and Applications, vol. 30, no. 2, p. 389–411, 2018. [CrossRef]
- S. D. Nguyen, T. S. Tran, V. P. Tran, H. J. Lee, M. J. Piran and V. P. Le, "Deep Learning-Based Crack Detection: A Survey," International Journal of Pavement Research and Technology, vol. 16, p. 943–967, 2023. [CrossRef]
- P. M. Bhatt, R. K. Malhan, P. Rajendran, B. C. Shah, S. Thakar, Y. J. Yoon and S. K. Gupta, "Image-Based Surface Defect Detection Using Deep Learning: A Review," Journal of Computing and Information Science in Engineering, vol. 21, no. 4, p. 040801, 2021. [CrossRef]
- A. T. G. Tapeh and M. Z. Naser, "Artificial Intelligence, Machine Learning, and Deep Learning in Structural Engineering: A Scientometrics Review of Trends and Best Practices," Archives of Computational Methods in Engineering, vol. 30, pp. 115-159, 2023. [CrossRef]
- H.-T. Thai, "Machine learning for structural engineering: A state-of-the-art review," Structures, vol. 38, p. 448–491, 2022. [CrossRef]
- M. Cao, N. F. Alkaye m, L. Pan and D. Novák, "Advanced methods in neural networks-based sensitivity analysis with their applications in civil engineering," in Artificial neural networks: models and applications, Rijeka, Croatia, IntechOpen, 2016.
- D. H. Nguyen and M. A. Wahab, "Damage detection in slab structures based on two-dimensional curvature mode shape method and Faster R-CNN," Advances in Engineering Software, vol. 176, p. 103371, 2023. [CrossRef]
- L. Yu, S. He, X. Liu, S. Jiang and S. Xiang, "Intelligent Crack Detection and Quantification in the Concrete Bridge: A Deep Learning-Assisted Image Processing Approach," Advances in Civil Engineering, vol. 2022, p. 1813821, 2022. [CrossRef]
- P. Kaewniam, M. Cao, et al. "Recent advances in damage detection of wind turbine blades: A state-of-the-art review," Renewable and Sustainable Energy Reviews, vol. 167, p. 112723, 2022. [CrossRef]
- S.-J. Wang, J.-K. Zhang and X.-Q. Lu, "Research on Real-Time Detection Algorithm for Pavement Cracks Based on SparseInst-CDSM," Mathematics, vol. 11, no. 5, p. 3277, 2023.
- G. Yu and X. Zhou, "An Improved YOLOv5 Crack Detection Method Combined with a Bottleneck Transformer," Mathematics 2023, 11(10), 2377, vol. 11, no. 10, p. 2377, 2023. [CrossRef]
- T. S. Tran, S. D. Nguyen, H. J. Lee and V. P. Tran, "Advanced crack detection and segmentation on bridge decks using deep learning," Construction and Building Materials, vol. 400, p. 132839, 2023. [CrossRef]
- J. Zhang, Y.-Y. Cai, D. Yang, Y. Yuan, W.-Y. He and Y.-J. Wang, "MobileNetV3-BLS: A broad learning approach for automatic concrete surface crack detection," Construction and Building Materials, vol. 392, p. 131941, 2023. [CrossRef]
- N. F. Al kayem, L. Shen, A. Mayya, P. G. Asteris, R. Fu, G. D. Luzio, A. Strauss and M. Cao, "Prediction of concrete and FRC properties at high temperature using machine and deep learning: A review of recent advances and future perspectives," Journal of Building Engineering, vol. 83, p. 108369, 2024. [CrossRef]
- R. Fu, M. Cao, D. Novák, et al. "Extended efficient convolutional neural network for concrete crack detection with illustrated merits," Automation in Construction, vol. 156, p. 105098, 2023. [CrossRef]
- C. Xiong, T. Zayed and E. M. Abdelkader, "A novel YOLOv8-GAM-Wise-IoU model for automated detection of bridge surface cracks," Construction and Building Materials, vol. 414, p. 135025, 2024.
- N. F. Al kayem, M. Cao and M. Ragulskis, "Damage Diagnosis in 3D Structures Using a Novel Hybrid Multiobjective Optimization and FE Model Updating Framework," Complexity, vol. 2018, p. 3541676 , 2018. [CrossRef]
- M. Cao, P. Qiao and Q. Ren, "Improved hybrid wavelet neural network methodology for time-varying behavior prediction of engineering structures," Neural Computing and Applications, vol. 18, pp. 821-832, 2009. [CrossRef]
- N. F. A lkayem and M. Cao, "Damage identification in three-dimensional structures using single-objective evolutionary algorithms and finite element model updating: evaluation and comparison," Engineering Optimization, vol. 50, no. 10, pp. 1695-1714, 2018. [CrossRef]
- P. Arafin, A. M. Billah and A. Issa, "Deep learning-based concrete defects classification and detection using semantic segmentation," Structural Health Monitoring, vol. 23, no. 1, p. 383–409, 2024. [CrossRef]
- J. Hang, Y. Wu, Y. Li, T. Lai, J. Zhang and Y. Li, "A deep learning semantic segmentation network with attention mechanism for concrete crack detection," Structural Health Monitoring, vol. 22, no. 5, pp. 3006-3026, 2023. [CrossRef]
- D. Tabernik, M. Šuc and D. Skočaj, "Automated detection and segmentation of cracks in concrete surfaces using joined segmentation and classification deep neural network," Construction and Building Materials, vol. 408, p. 133582, 2023. [CrossRef]
- J. Shang, J. Xu, A. A. Zhang, Y. Liu, K. C. Wang, D. Ren, H. Zhang, Z. Dong and A. He, "Automatic Pixel-level pavement sealed crack detection using Multi-fusion U-Net network," Measurement, vol. 208, p. 112475, 2023. [CrossRef]
- B. Chen, H. Zhang, G. Wang, J. Huo, Y. Li and L. Li, "Automatic concrete infrastructure crack semantic segmentation using deep learning," Automation in Construction, vol. 152, p. 104950, 2023.
- L. M. Dang, H. Wang, Y. Li, L. Q. Nguyen, T. N. Nguyen, H.-K. Song and H. Moon, "Lightweight pixel-level semantic segmentation and analysis for sewer defects using deep learning," Construction and Building Materials, vol. 371, p. 130792, 2023. [CrossRef]
- D. Joshi, T. P. Singh and G. Sharma, "Automatic surface crack detection using segmentation-based deep-learning approach," Engineering Fracture Mechanics, vol. 268, p. 108467, 2022. [CrossRef]
- M. Mishra, V. Jain, S. K. Singh and D. Maity, "Two stage method based on the you only look once framework and image segmentation for crack detection in concrete structures," Architecture, Structures and Construction, vol. 3, p. 429–446, 2023. [CrossRef]
- P. Shi, S. Shao, X. Fan, Z. Zhou and Y. Xin, "MCL-CrackNet: A Concrete Crack Segmentation Network Using Multilevel Contrastive Learning," IEEE TRANSACTIONS ON INSTRUMENTATION AND MEASUREMENT, vol. 72, p. 5030415, 2023.
- P. Savino and F. Tondolo, "Civil infrastructure defect assessment using pixel wise segmentation based on deep learning," Journal of Civil Structural Health Monitoring, vol. 13, p. 35–48, 2023. [CrossRef]
- P. N. Hadinata, D. Simanta, L. Eddy and K. Nagai, "Multiclass Segmentation of Concrete Surface Damages Using U-Net and DeepLabV3+," Applied Sciences, vol. 13, p. 2398, 2023. [CrossRef]
- Z. Al-Huda, B. Peng, R. N. A. Algburi, M. A. Al-antari, R. AL-Jarazi and D. Zhai, "A hybrid deep learning pavement crack semantic segmentation," Engineering Applications of Artificial Intelligence, vol. 122 , p. 106142, 2023. [CrossRef]
- R. Ali, J. H. Chuah, M. S. A. Talip, N. Mokhtar and M. A. Shoaib, "Automatic pixel-level crack segmentation in images using fully convolutional neural network based on residual blocks and pixel local weights," Engineering Applications of Artificial Intelligence, vol. 104, p. 104391, 2021. [CrossRef]
- D. Kang, S. S. Benipal, D. L. Gopal and Y.-J. Cha, "Hybrid pixel-level concrete crack segmentation and quantification across complex backgrounds using deep learning," Automation in Construction, vol. 118, p. 103291, 2020. [CrossRef]
- C. Sha, C. Yue and W. Wang, "Ensemble 1D DenseNet Damage Identification Method Based on Vibration Acceleration," Structural Durability & Health Monitoring, vol. 17, no. 5, pp. 369-381, 2023. [CrossRef]
- V. Kailkhura, S. Aravindh, S. S. Jha and N. Jayanth, "Ensemble learning-based approach for crack detection using CNN," in Proceedings of the Fourth International Conference on Trends in Electronics and Informatics (ICOEI 2020), 2020.
- Z. Fan, C. Li, Y. Chen, P. Mascio, X. Chen, G. Zhu and G. Loprencipe, "Ensemble of Deep Convolutional Neural Networks for Automatic Pavement Crack Detection and Measurement," Coatings, vol. 10, p. 152, 2020. [CrossRef]
- Y. Hong and S. B. Yoo, "OASIS-Net: Morphological Attention Ensemble Learning for Surface Defect Detection," Mathematics, vol. 10, p. 4114, 2022. [CrossRef]
- M. S. Barkhordari, D. J. Armaghani and P. G. Asteris, "Structural Damage Identification Using Ensemble Deep Convolutional Neural Network Models," Computer Modeling in Engineering & Sciences, vol. 134, no. 2, pp. 835-855, 2023. [CrossRef]
- A. A. Maarouf and F. Hachouf, "Transfer Learning-based Ensemble Deep Learning for Road Cracks Detection," in International Conference on Advanced Aspects of Software Engineering (ICAASE), Constantine, Algeria, 2022 .
- W. Bousselham, G. Thibault, L. Pagano and A. Machireddy, "Efficient Self-Ensemble for Semantic Segmentation," arXiv , p. arXiv:cs.CV/2111.13280, 2022.
- I. Nigam, C. Huang and D. Ramanan, "Ensemble Knowledge Transfer for Semantic Segmentation," in IEEE Winter Conference on Applications of Computer Vision, Lake Tahoe, NV, USA, 2018.
- L. Zhang, S. Slade, C. P. Lim, H. Asadi, S. Nahavandi, H. Huang and H. Ruan, "Semantic segmentation using Firefly Algorithm-based evolving ensemble deep neural networks," Knowledge-Based Systems, vol. 277, p. 110828, 2023. [CrossRef]
- C. Lee, S. Yoo, S. Kim and J. Lee, "Progressive Weighted Self-Training Ensemble for Multi-Type Skin Lesion Semantic Segmentation," IEEE Access, vol. 10, pp. 132376-132383, 2022. [CrossRef]
- T. Lee, J.-H. Kim, S.-J. Lee, S.-K. Ryu and B.-C. Joo, "Improvement of Concrete Crack Segmentation Performance Using Stacking Ensemble Learning," Applied Sciences, vol. 13, p. 2367, 2023. [CrossRef]
- S. Li and X. Zhao, "A Performance Improvement Strategy for Concrete Damage Detection Using Stacking Ensemble Learning of Multiple Semantic Segmentation Networks," Sensors, vol. 22, p. 3341, 2022. [CrossRef]
- G. E. Amieghemen and M. M. Sherif, "Deep convolutional neural network ensemble for pavement crack detection using high elevation UAV images," Structure and Infrastructure Engineering, p. 2023. [CrossRef]
- G. Cyganov, A. Rychenkov, A. Sinitca and D. Kaplun, "Using the fuzzy integrals for the ensemble based segmentation of asphalt cracks," Industrial Artificial Intelligence, vol. 1, p. 5, 2023. [CrossRef]
- Y. Chen, Y. Mo, A. Readie, G. Ligozio, I. Mandal, F. Jabbar, T. Coroller and B. W. Papież, "VertXNet: an ensemble method for vertebral body segmentation and identification from cervical and lumbar spinal X rays," Scientific Reports, vol. 14, p. 3341, 2024. [CrossRef]
- R. Bao, K. Palaniappan, Y. Zhao and G. Seetharaman, "GLSNet++: Global and Local-Stream Feature Fusion for LiDAR Point Cloud Semantic Segmentation Using GNN Demixing Block," IEEE SENSORS JOURNAL, vol. 24, no. 7, pp. 11610-11624, 2024. [CrossRef]
- D. Dais, I. E. Bal, E. Smyrou and V. Sarhosis, "Automatic crack classification and segmentation on masonry surfaces using convolutional neural networks and transfer learning," Automation in Construction, vol. 125, p. 103606, 2021. [CrossRef]
- R. Vij and S. Arora, "A hybrid evolutionary weighted ensemble of deep transfer learning models for retinal vessel segmentation and diabetic retinopathy detection," Computers and Electrical Engineering , vol. 115, p. 109107, 2024. [CrossRef]
- Z. Fan, C. Li, Y. Chen, P. D. Mascio, X. Chen, G. Zhu and G. Loprencipe, "Ensemble of Deep Convolutional Neural Networks for Automatic Pavement Crack Detection and Measurement," Coatings, vol. 10, p. 152, 2020. [CrossRef]
- K. S. Devan, H. A. Kestler, C. Read and P. Walther, "Weighted average ensemble based semantic segmentation in biological electron microscopy images," Histochemistry and Cell Biology, vol. 158, p. 447–462, 2022. [CrossRef]
- F. Panella, A. Lipani and J. Boehm, "Semantic segmentation of cracks: Data challenges and architecture," Automation in Construction, vol. 135 , p. 104110, 2022. [CrossRef]
- Y. Liu, J. Yao, X. Lu, R. Xie and L. Li, "DeepCrack: A deep hierarchical feature learning architecture for crack segmentation," Neurocomputing, vol. 338, pp. 139-153, 2019. [CrossRef]
- S. Kulkarni, S. Singh, D. Balakrishnan, S. Sharma, S. Devunuri and S. C. R. Korlapati, "CrackSeg9k: A Collection and Benchmark for Crack Segmentation Datasets and Frameworks," in Karlinsky, L., Michaeli, T., Nishino, K. (eds) Computer Vision – ECCV 2022 Workshops. ECCV 2022. Lecture Notes in Computer Science, vol 13807, Springer, Cham, 2022.
- M. Pak and S. Kim, "Crack Detection Using Fully Convolutional Network in Wall-Climbing Robot," in Park, J.J., Fong, S.J., Pan, Y., Sung, Y. (eds) Advances in Computer Science and Ubiquitous Computing. Lecture Notes in Electrical Engineering, vol 715., Springer, Singapore. , 2021.
- O. Ronneberger, P. Fischer and T. Brox, "U-Net: Convolutional Networks for Biomedical Image Segmentation," in In: Navab, N., Hornegger, J., Wells, W., Frangi, A. (eds) Medical Image Computing and Computer-Assisted Intervention – MICCAI 2015. MICCAI 2015. Lecture Notes in Computer Science, vol 9351. , Springer, Cham, 2015.
- K. Simonyan and A. Zisserman, "Very Deep Convolutional Networks for Large-Scale Image Recognition," arXiv, p. arXiv:1409.1556, 2014.
- L.-C. Chen, G. Papandreou, F. Schroff and H. Adam, "Rethinking Atrous Convolution for Semantic Image Segmentation," arXiv, p. arXiv:1706.05587, 2017.











| Epochs | 40 |
| Loss function | Dice loss |
| batch_size | 8 |
| Initial Learning rate | 1e-3 |
| weight_decay | 5e-5 |
| Classification layer activation function | Sigmoid |
| Input image dimensions | 448*448*3 |
| Data augmentation operations | Normalization Random rotation Horizontal flip Vertical flip Random color jittering |
| Optimizer | Adam |
| Epochs | 40 |
| Loss function | Dice loss |
| batch_size | 8 |
| Initial Learning rate | 1e-3 |
| weight_decay | 5e-5 |
| Classification layer activation function | Sigmoid |
| Input image dimensions | 448*448*3 |
| Data augmentation operations | Normalization Random rotation Horizontal flip Vertical flip Random color jittering |
| Optimizer | Adam |
| Model | Loss | Pixel ACC% | IoU% | Precision% | Recall% | Dice% | mAP% | It/Sec |
| DLV3/ ResNet50 | 0.201 | 98.72 | 68.41 | 73.85 | 90.92 | 80 | 84.87 | 1.09 |
| Unet | 0.178 | 98.89 | 71.15 | 81.1 | 85.58 | 82.2 | 84.9 | 1.01 |
| DeepCrack19 | 1.06 | 98.88 | 70.1 | 81.31 | 83.79 | 81.6 | 84.79 | 1.31 |
| SegNet | 0.191 | 98.78 | 69.4 | 79.23 | 85.48 | 81 | 84.522 | 1.17 |
| DLV3/ ResNet101 | 0.185 | 98.9 | 70.2 | 80.38 | 84.6 | 81.7 | 84.69 | 1.11 |
| Model | Loss | Pixel ACC% | IoU% | Precision% | Recall% | Dice% | mAP% | It/Sec |
| DLV3/ResNet50 | 0.347 | 98.39 | 49.79 | 64.12 | 69.43 | 65.6 | 68.07 | 2.44 |
| Unet | 0.33 | 98.5 | 51.2 | 65.9 | 70.46 | 67.04 | 69.36 | 2.37 |
| DeepCrack19 | 1.8 | 98.4 | 51.2 | 66.24 | 69.21 | 67.1 | 68.1 | 3.06 |
| SegNet | 0.339 | 98.4 | 50.22 | 65.1 | 69.93 | 66.1 | 68.35 | 2.66 |
| DLV3/ResNet101 | 0.346 | 98.4 | 49.82 | 64.16 | 69.38 | 65.7 | 68.0 | 1.56 |
| Model | Pixel_ACC | IoU | Precision | Recall | Dice | mAP |
| Co-CrackSegment/dice | 99.03 | 72.98 | 82.22 | 86 | 83.62 | 85.8 |
| Co-CrackSegment/IoU | 99.038 | 72.98 | 82.24 | 86 | 83.62 | 85.8 |
| Co-CrackSegment/Pixel_Acc | 99.042 | 72.88 | 82.61 | 85.3 | 83.52 | 85.57 |
| Co-CrackSegment/Precision | 98.96 | 71.67 | 83.29 | 83.22 | 82.74 | 85.1 |
| Co-CrackSegment/Recall | 98.96 | 71.85 | 79.84 | 87.31 | 82.75 | 85.12 |
| Weighted average | 98.91 | 70.56 | 80.61 | 85.38 | 81.91 | 83.24 |
| Model | Pixel ACC | IoU | Precision | Recall | Dice | mAP |
| Co-CrackSegment/dice | 98.52 | 53.28 | 67.37 | 71.925 | 68.9 | 70.31 |
| Co-CrackSegment/IoU | 98.52 | 53.28 | 67.41 | 71.88 | 68.9 | 70.31 |
| Co-CrackSegment/Pixel_Acc | 98.536 | 52.97 | 68.21 | 70.43 | 68.6 | 70.5 |
| Co-CrackSegment/Precision | 98.527 | 52.31 | 68.46 | 69.08 | 68.03 | 69.45 |
| Co-CrackSegment/Recall | 97.96 | 52.14 | 64.72 | 73.33 | 67.8 | 69.4 |
| Weighted average | 98.48 | 51.84 | 67.59 | 68.89 | 67.61 | 68.62 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).



























