Preprint Article Version 1 Preserved in Portico This version is not peer-reviewed

Making more with Less: Improving Software Testing Outcomes Using a Cross-Project and Cross-Language ML Classifier Based on Cost-Sensitive Training

Version 1 : Received: 14 May 2024 / Approved: 15 May 2024 / Online: 16 May 2024 (08:17:00 CEST)

How to cite: Nascimento, A. M.; Shimanuki, G. K. G.; Dias, L. A. V. Making more with Less: Improving Software Testing Outcomes Using a Cross-Project and Cross-Language ML Classifier Based on Cost-Sensitive Training. Preprints 2024, 2024051028. https://doi.org/10.20944/preprints202405.1028.v1 Nascimento, A. M.; Shimanuki, G. K. G.; Dias, L. A. V. Making more with Less: Improving Software Testing Outcomes Using a Cross-Project and Cross-Language ML Classifier Based on Cost-Sensitive Training. Preprints 2024, 2024051028. https://doi.org/10.20944/preprints202405.1028.v1

Abstract

As digitalization expands across all sectors, the economic toll of software defects on the U.S. economy reaches up to $2.41 trillion annually. High-profile incidents like the Boeing 787-Max 8 crash have shown the devastating potential of these defects, highlighting the critical importance of software testing within quality assurance frameworks. However, due to its complexity and resource intensity, the exhaustive nature of comprehensive testing often surpasses budget constraints. This research utilizes a machine learning (ML) model to enhance software testing decisions by pinpointing areas most susceptible to defects and optimizing scarce resource allocation. Previous studies have shown promising results using cost-sensitive training to refine ML models, improving predictive accuracy by reducing false negatives through addressing class imbalances in defect prediction datasets. This approach facilitates more targeted and effective testing efforts. Nevertheless, the generalizability of these models across different projects (cross-project) and programming languages (cross-language) remained untested. This study validates the model's applicability across diverse development environments by integrating various datasets from distinct projects into a unified, using a more interpretable ML approach. The results demonstrate that ML can support software testing decisions, enabling teams to identify up to seven times more defective modules with the same testing effort as a benchmark.

Keywords

Machine Learning; Imbalance; Software Defect Prediction; NASA MDP; Random Forest; Software Quality; Generalization; Cost-Sensitive; Cross-language; Cross-Project

Subject

Engineering, Safety, Risk, Reliability and Quality

Comments (0)

We encourage comments and feedback from a broad range of readers. See criteria for comments and our Diversity statement.

Leave a public comment
Send a private comment to the author(s)
* All users must log in before leaving a comment
Views 0
Downloads 0
Comments 0
Metrics 0


×
Alerts
Notify me about updates to this article or when a peer-reviewed version is published.
We use cookies on our website to ensure you get the best experience.
Read more about our cookies here.