Preprint
Article

This version is not peer-reviewed.

Build Outcome Prediction for Continuous Integration: Preventing Data Leakage in Machine Learning Models Using Pre-Execution SDLC Metrics

Submitted:

27 October 2025

Posted:

31 October 2025

You are already at the latest version

Abstract
The execution of millions of builds through CI/CD pipelines every day makes it difficult to determine build results before starting the process. Software delivery becomes delayed because build failures consume resources and create disruptions in workflow operations. The research develops a machine learning system which uses SDLC metrics from before builds to predict success rates across 1000+ open-source projects. The research solution tackles the problem of temporal data leakage which occurs when model performance becomes artificially high because of outcome-dependent features. The research uses TravisTorrent dataset containing 2.64 million builds to develop a method which eliminates 15 build outcome-dependent features while keeping 31 pre-execution features for analysis. The features used in this study include project development stages and code complexity measures and test organization and team work dynamics and time-based patterns. The research investigates three machine learning models including Logistic Regression and Random Forest and Gradient Boosting. Random Forest produces the highest performance results with 85.24% accuracy and 91.38% ROC-AUC and 89.80% F1-score when analyzing 100,000 builds. The analysis of feature importance shows that project maturity (9.49%) and repository age (9.46%) provide better predictions than recent code modifications do. The model achieves identical performance results when testing different programming languages including Java and Ruby and Python and JavaScript because the accuracy difference remains within 3.4 percentage points. The model shows 93.6% success in predicting successful builds and 66.2% success in identifying failed builds which allows organizations to plan their resources in advance. The research provides an effective data leakage prevention approach through its analysis of 1000+ projects which shows that project history patterns outperform code-specific metrics for build outcome prediction. The research provides essential knowledge for build optimization through its demonstration of how project history patterns outperform code-specific metrics for build outcome prediction.
Keywords: 
;  ;  ;  ;  ;  ;  ;  

1. Introduction

Modern software development relies on continuous integration and continuous deployment (CI/CD) as essential practices which help organizations deliver features quickly while shortening their market entry time. The fast pace of development through CI/CD pipelines leads to major problems because they run millions of builds every day across worldwide software systems while using up large amounts of processing power and developer resources. The occurrence of build failures during 20-40% of typical open source project [1,2] builds creates development workflow interruptions which extend feature release timelines and waste cloud resources while decreasing team performance. A failed build in extensive projects requires developers to spend 30-60 minutes for investigation while simultaneously using $5-10 worth of cloud infrastructure resources[3].
The ability to forecast build results before starting construction work would create revolutionary improvements in CI/CD pipeline operational speed. Organizations would use build confidence levels to select important builds while dedicating more testing resources to dangerous modifications and they could run failing builds when system traffic is minimal and developers could receive immediate feedback about their code quality. The achievement of precise build predictions depends on solving the essential data quality problem which involves temporal data leakage.
The TravisTorrent [4] dataset unites Travis CI build logs with GitHub repository metadata to create an extensive research base for build prediction studies. The dataset contains 2.64 million builds from more than 1000 open-source projects which use different programming languages including Java Ruby Python and JavaScript to show actual CI/CD operational patterns. The analysis of TravisTorrent and similar datasets fails to address a critical problem which affects data quality through temporal data leakage. The model training process becomes perfect because `tr_status` and `tr_log_tests_failed` and test execution results serve as direct build outcome indicators but become inaccessible during future build predictions. The reported accuracy rates above 95-99% in certain studies [5] remain unattainable for real-world deployment environments.
Recent machine learning advances have demonstrated strong capabilities for software engineering tasks. Random Forest and Gradient Boosting achieve state-of-the-art results for defect prediction [6,7], while deep learning architectures including LSTM networks [8,9] and Graph Neural Networks have shown promise for code analysis and defect prediction tasks. Just-in-time defect prediction approaches use commit-level features to identify risky code changes immediately after submission [10,11]. Industrial applications demonstrate practical viability of machine learning-based defect prediction across diverse organizational contexts [12]. Build prediction specifically has been explored using ensemble methods [2,13] and deep learning [5], with reported accuracies ranging from 75–84% on real data when leakage is avoided [1]. However, systematic analysis of which SDLC metrics provide genuine predictive power—available before build execution—and how to prevent temporal data leakage remains an open research challenge.

1.1. Research Questions

The research investigates machine learning effectiveness for CI/CD build prediction through four specific questions which evaluate both reliability and practical implementation.
RQ1: Can SDLC metrics effectively predict build outcomes using only pre-build features?
The research examines whether project characteristics measured before build execution including maturity level and code complexity and test organization and past build results can successfully forecast build completion status. The research needs to establish a method for detecting and eliminating temporal data leakage which causes features that depend on outcomes to produce false positive results.
RQ2: Which SDLC phases and metrics contribute most to build prediction accuracy?
The research evaluates feature importance to identify which project characteristics (maturity and history) or code metrics (complexity and churn) and test indicators (coverage and density) and team collaboration patterns and member contribution patterns generate the most accurate predictions.
RQ3: How well do prediction models generalize across different programming languages and project types?
The research evaluates model performance through an analysis of Java and Ruby and Python and JavaScript projects to determine their ability to generalize between programming languages while detecting unique modeling needs for each language.
RQ4: What is the impact of data leakage prevention on model performance and practical deployment?
The research establishes the exact difference in model accuracy between models trained with complete feature sets including outcome-dependent variables and models trained with only pre-built features to prove that proper feature selection remains essential for deployment success.

1.2. Research Contributions

The research presents four main contributions which enhance the current understanding of CI/CD build prediction.
  • Rigorous Data Leakage Prevention Methodology: The research team successfully eliminated 15 outcome-dependent features from TravisTorrent dataset which contained build results to show how temporal data leakage creates artificial accuracy increases from 85.24% to 95-99%. The research establishes a method to stop data leakage in temporal software engineering datasets which protects model performance by using only pre-build features available during prediction time.
  • Real-World Validation at Scale: The research validates build prediction through analysis of 100,000 builds from 1,000+ open-source projects which include Java, Ruby, Python and JavaScript programming languages. The Random Forest model achieved 85.24% accuracy through analysis of 100,000 builds from 1,000+ open-source projects. The study provides evidence of practical build prediction applications through its evaluation of multiple project types and programming languages at a large scale.
  • The research findings show that project maturity (9.49%) and repository age (9.46%) predict build outcomes better than immediate code changes which contradicts the conventional belief that code-level metrics control build results. The research shows that project history and development context generate more accurate signals than the changes made in individual commits.
  • Cross-Language Generalization Evidence: The research shows that the model achieves similar results when analyzing different programming languages (Java: 84.2%, Ruby: 82.1%, Python: 81.5%, JavaScript: 80.8%) with a maximum difference of 4.9 percentage points. The results show that models trained on multiple programming languages can predict build outcomes for various software systems without needing language-specific model updates.

1.3. Paper Organization

The following structure guides the rest of this paper. The paper reviews build prediction research alongside CI/CD analytics and software engineering machine learning and data leakage prevention methods in Section II. The paper describes its methodology through TravisTorrent dataset evaluation and data leakage protection system development and pre-build prediction feature creation and classification model construction. The section demonstrates experimental findings through model evaluation and confusion matrix evaluation and feature importance assessment and cross-language prediction assessment. The paper examines CI/CD practice benefits and data leakage risks and validity threats and outlines research avenues for future study in Section V. The paper ends with essential results and useful implementation advice in Section VI.

2. Related Work

Research into machine learning applications for software engineering and CI/CD build prediction and development lifecycle metrics has evolved through separate research areas. Machine learning technology has transformed software engineering through its implementation in testing systems and prediction models and analytics platforms [14,15]. Systematic literature reviews demonstrate that machine learning and deep learning have advanced software engineering operations yet defect prediction requires improved solutions to address its persistent challenges [16,17,18] and [19]. The following section examines three fundamental research areas which combine machine learning in software development with SDLC and DevOps metric analysis and temporal prediction and data leakage protection.

2.1. Machine Learning for Software Performance Prediction

The field of performance prediction using machine learning algorithms has shown substantial advancement through the last few years across multiple performance evaluation criteria. The following section reviews neural network architectures and ensemble methods and transformer-based approaches and hybrid techniques.

2.1.1. Neural Networks and Deep Learning Architectures

Deep learning approaches including LSTM networks [8,9] and recurrent architectures have achieved strong results for software defect prediction and code analysis tasks. Component-based prediction models use modular software structure for granular performance forecasting [20]. However, these methods require extensive training data, longer training times, and careful hyperparameter configuration [21], while offering limited interpretability compared to ensemble methods like Random Forest. For tabular SDLC features with moderate dataset sizes, ensemble approaches provide competitive accuracy with superior explainability and lower computational cost, which the approach employs in this work.

2.1.2. Ensemble Methods and Gradient Boosting

Research studies demonstrate that ensemble learning methods generate superior prediction results than individual models when solving different software engineering prediction tasks [22,23,24,25]. Ibraigheeth et al. created weighted ensemble learning which assigns higher weights to base models that achieve superior results in challenging situations [26]. The stacking ensemble method which employs multiple base learners generates dependable results for software effort prediction and defect classification applications [27,28].
XGBoost (Extreme Gradient Boosting) continues to be the top choice for software engineering applications. The combination of Recursive Feature Elimination with ensemble methods allows for systematic feature selection of high-dimensional software metrics according to [29,30]. Random Forest-based methods produce outstanding results when applied to software fault prediction and effort estimation tasks according to [6]. Research findings demonstrate LightGBM outperforms XGBoost and CatBoost for time-series forecasting when Grid Search optimizes hyperparameters [31,32]. Ensemble stacking methods that combine multiple base learners generate reliable predictions for software effort estimation and defect classification tasks according to [22,27,28]. The achievement of superior than 99% AUC performance in specific studies depends on thorough feature validation to stop temporal data leakage during prediction.
Transformer architectures have shown promise for time-series forecasting tasks [33,34], achieving improvements over LSTM through variable selection and self-attention mechanisms. Transformer-based quality-of-service prediction demonstrates applicability to software system performance tasks [35]. However, for tabular SDLC features in CI/CD build prediction, ensemble methods provide superior performance-to-complexity ratios. Transfer learning approaches have demonstrated value for cross-project defect prediction [36], relevant to the cross-language generalization evaluation in RQ3.

2.2. SDLC Metrics and DevOps Analytics

Research now focuses on systematic SDLC metric and DevOps practice analysis because it produces essential results for performance prediction.

2.2.1. Performance Code Metrics and Bug Prediction

The 2024 MSR conference analyzed 80 established Java projects through machine learning to detect performance bugs while creating performance code metrics which expand past traditional metrics including cyclomatic complexity [37]. The models Random Forest and XGBoost achieved average results of 0.84 for AUC and 0.21 for PR-AUC and 0.38 for MCC. The introduction of performance code metrics led to a 7.7% decrease in median AUC and PR-AUC decreased by 25.4% and MCC decreased by 20.2%. The research demonstrates that performance-specific metrics generate significant changes in prediction outcomes.
Research indicates Graph Neural Networks successfully predict software defects because they extract structural data from code representations which include Abstract Syntax Trees and Control Flow Graphs ([38]).

2.2.2. DevOps Metrics and Organizational Performance

Research studies have identified all necessary DevOps metrics which monitor software development lifecycle practices and industry reports demonstrate that DevOps implementation produces enhanced organizational performance [39].
The DORA (DevOps Research and Assessment) metrics framework continues to serve as the primary assessment tool for software delivery performance evaluation. The 2024 State of DevOps Report showed that elite performers deploy at 182 times the rate of low performers while achieving 8 times better change failure rates and deploying changes 127 times faster and recovering from failed deployments 2,293 times quicker [39]. The CD Foundation’s State of CI/CD Report which surveyed 125,000+ developers revealed that 83% of developers engage in DevOps work and the number of DevOps technologies used determines the likelihood of becoming a top performer [40].

2.2.3. Build Optimization and CI/CD Pipeline Analytics

Build Failure Characterization. The identification of CI/CD build failure patterns serves as the fundamental requirement for developing prediction systems. Vassallo et al. studied build failure data from open-source projects and financial institutions to find that standard projects experience 20-40% build failures because of compilation errors and test failures and dependency problems [1]. Rausch et al. analyzed Java-based open-source software build failures to discover recurring patterns which resulted in 15-25% of failures because of flaky tests and infrastructure timeouts and dependency resolution problems [2]. Research on continuous integration timeout builds demonstrates that environmental factors and restricted system resources lead to extended build completion times [41]. Zampetti et al. conducted empirical research to identify CI anti-patterns which revealed that poor test organization and wrong build configurations and missing dependency definitions result in elevated build failure rates [42]. Gallaba et al. analyzed Travis CI build data to discover that platform updates and external service dependencies and environmental changes generate significant unpredictability in build outcomes [43].
Build Prediction Approaches and Data Leakage Concerns. Machine learning for build prediction has become widespread but researchers apply different levels of methodological detail to their work. TravisTorrent [4] serves as the primary dataset because scientists have analyzed its 2.64 million builds from 1,000+ projects through more than 40 research papers. The assessment of data quality problems reveals that multiple previous studies generated wrong results because of these issues. The research by DL-CIBuild employed LSTM-based RNN with Genetic Algorithm hyperparameter optimization to analyze 91,330 Travis CI builds which resulted in accuracy performance exceeding 95% [5]. The exceptional results stem from the inclusion of build execution-dependent variables in their feature set which becomes accessible after build completion. Sun et al. created RavenBuild as a context-aware system which predicts build outcomes at Ubisoft through relevance-aware methods and dependency-aware methods that achieved a 19.8 percentage points F1-score improvement by integrating build context and dependency graph features [44]. Research studies have shown that models which use only pre-build features achieve accuracy between 75% and 84% according to previous studies [1,2]. The existing methodological variations between studies demand our research to create a standardized approach which prevents data leakage by restricting analysis to pre-build information.
Build Time Prediction and Test Optimization. The research community has developed two new areas which focus on build duration forecasting and test optimization. Macho et al. developed prediction models which forecast build co-changes to identify when code modifications need build script updates with 70-80% accuracy for build configuration needs prediction [45]. Hilton et al. analyzed open-source continuous integration projects to demonstrate that CI deployment reduces integration issues by 35-50% while requiring $5-10 per build for large projects to sustain their infrastructure [3]. Luo et al. performed research on test prioritization methods for large projects which demonstrated that demonstrated focused testing reduces feedback time by 40-60% without affecting fault detection precision [46]. The Google Chrome test analysis revealed that test batching outperformed test selection because it reduced feedback time to 99% without failing any tests while decreasing machine utilization to 88% [47]. The study of test flakiness effects on build stability has become a popular research topic during the last few years. Gruber et al. investigated how automatic test generation tools produce unstable tests during ICSE 2024 [48]. Silva et al. investigated how different computational resource levels impact test flakiness in IEEE TSE 2024 and found that resource constraints strongly influence test reliability [49]. The system which uses our leakage-free build prediction method operates under three separate research advancements that concentrate on test optimization and build time prediction and failure characterization.
GitHub Actions and Modern CI/CD Practices. The introduction of GitHub Actions has revolutionized CI/CD automation methods which researchers now study extensively. Bernardo et al. studied CI practices between ML and non-ML projects on GitHub Actions to discover workflow complexity and testing pattern differences which help build prediction features [50]. Bouzenia and Pradel studied 1.3 million GitHub Actions workflow runs to discover major optimization potential which leads to 30-40% CI cost reduction through workflow optimization [51]. Zhang et al. studied 6,590 Stack Overflow questions about GitHub Actions to create a developer challenge taxonomy which shows typical failure types [52,53,54]. Mastropaolo et al. developed GH-WCOM which uses Transformers to generate GitHub workflow configuration completions which produce 34% accurate results and simplify developer workflow creation [55]. Research studies between GitHub Actions and GitLab CI and Jenkins platforms show how different platforms affect workflow failures and resource usage patterns [56,57]. The use of Infrastructure as Code allows developers to create reproducible CI/CD environments through declarative configuration management [58]. The current research using TravisTorrent fits into the GitHub Actions-based CI/CD development path which scientists can use to test new CI platforms in the future. The implementation of continuous delivery methods together with deployment speed determines how well organizations deliver their software products [59,60,61].

2.2.4. Commit Metrics and Developer Productivity

Research studies about code-based metrics (lines of code and code quality) and commit-based metrics (commit frequency and activity) showed these metrics offer different views of developer productivity [62,63]. The assessment of code quality depends on three fundamental software complexity metrics which include McCabe’s cyclomatic complexity [64] and Halstead metrics [65]. The measurement of code churn through change frequency and extent reveals direct links between software system fault susceptibility [66]. Software quality metrics frameworks enable developers to perform organized measurement operations [67]. The quality of commit messages directly affects development speed according to ICSE 2023 research which used machine learning classifiers to evaluate message quality [68]. The predictive value of software quality metrics depends on complexity measurements and coupling and cohesion assessments and technical debt indicators which help estimate defect risk and maintenance requirements [69,70,71]. The quality of software code depends on review practices which also enable knowledge transfer between developers [72]. Bug localization methods use past defect records together with code modification patterns to determine which helps teams determine their inspection work priorities [73].
The SPACE framework established a five-dimensional model to measure developer productivity through its Satisfaction and Well-being and Performance and Activity and Communication and Collaboration and Efficiency and Flow dimensions [74]. The framework demonstrated that organizations need to balance multiple performance indicators because tracking activity metrics alone proves insufficient.

2.3. Data Leakage in Machine Learning for Software Engineering

The main data quality issue in predictive modeling stems from temporal data leakage because features that contain prediction outcome information boost training results but become inaccessible during actual prediction operations. Kaufman et al. established the fundamental classification system for data mining leakage patterns through their research which identified three main categories of leakage: target leakage and train-test contamination and temporal leakage [75]. The authors created official definitions and detection methods which serve as essential components for conducting rigorous machine learning studies in various fields..
In software engineering, temporal data leakage manifests particularly in CI/CD build prediction and defect forecasting tasks where outcome-dependent features are inadvertently included. Features such as test execution results, build duration, log analysis outputs, and failure counts are only computable after build completion, yet prior research frequently includes these variables when training predictive models [5]. This practice creates models with unrealistically high reported accuracies (95-99%) during training and cross-validation, but complete failure when deployed to production environments where such features are definitionally unavailable. The temporal nature of software development data exacerbates leakage risks: rolling averages computed using non-causal windows, cumulative metrics incorporating current observations, and features aggregated without respecting temporal ordering all introduce subtle future information into historical predictions [76,77].
Cross-validation strategies must respect temporal dependencies to prevent leakage in time-series software metrics. Standard k-fold cross-validation randomly shuffles data, enabling future builds to influence predictions for historical builds. Time-series cross-validation using forward chaining progressively expands training sets while strictly respecting chronological order, ensuring models train only on past data when predicting future outcomes [77]. Feature engineering for temporal data requires explicit validation that all computed features can be derived exclusively from information available before the prediction timestamp [76,78]. Missing value imputation must use only historical statistics, never future observations. Normalization parameters (mean, standard deviation) must derive solely from training data, preventing test set leakage into preprocessing [79].
The research solves data leakage problems by performing full feature evaluation for CI/CD build prediction with TravisTorrent dataset reduction to 15 outcome-independent variables which get validated for computational readiness before build execution. The research conducts an exact evaluation method to produce genuine deployment-ready performance results which differ from previous studies that used incorrect features to achieve artificial accuracy results.

2.4. Evaluation Metrics and Statistical Testing

Performance prediction research requires strict evaluation systems to operate effectively. Machine learning testing requires systematic validation procedures to confirm model reliability and achieve generalization according to [18]. The field of empirical software engineering uses established methods to achieve both rigor and reproducibility according to [80,81].
Matthews Correlation Coefficient emerged as essential for imbalanced data, providing balanced treatment of all confusion matrix quadrants [82,83]. Class imbalance presents fundamental challenges requiring specialized learning approaches [25]. MASE (Mean Absolute Scaled Error) addresses time-series evaluation challenges through scale-independence [84]. Statistical significance testing standards emphasize Wilcoxon signed-rank test for two-model comparison [85], Friedman test for K≥3 models with Nemenyi or Bonferroni-corrected post-hoc tests, and effect sizes (Cohen’s d, Cliff’s delta) alongside p-values [86,87]. For imbalanced classification tasks, precision and recall must be evaluated jointly to avoid misleading conclusions from accuracy alone [88]. ROC-AUC interpretation requires understanding that values near 0.5 indicate random performance while values exceeding 0.9 demonstrate strong discriminative capability [89].
The research determined forward chaining (time series split) as the best approach for temporal data analysis because it enables training set growth through time-dependent sequence preservation [90]. The NASA MDP defect prediction dataset serves as a standard time-series forecasting benchmark together with Electricity ETT Exchange Rate Traffic Weather and ILI datasets [91].

2.5. Research Gaps and Novel Contributions

The research identifies multiple critical knowledge gaps which this study aims to address. The research fails to provide a complete solution for comprehensive performance prediction through multi-phase SDLC metric integration even though machine learning dominates defect prediction and individual performance assessment. Research has developed new deep learning models but scientists need to investigate how to optimize basic models through complete metric integration using advanced feature engineering techniques. The analysis of temporal relationships and phase-to-phase connections within SDLC metrics for performance forecasting remains unsolved because these elements play a vital role in software analytics according to [15,92].
The research develops new methods which unite these gaps through complete SDLC metric integration and linear model enhancement for complex relationship detection and state-of-the-art ensemble and deep learning method baseline evaluation and production-ready system deployment. The research achieves a major breakthrough in predictive performance analytics through its demonstration of how modern software development data ecosystems enable accurate results without requiring complete testing procedures.

3. Methodology

The following section describes our complete method for CI/CD build prediction through TravisTorrent dataset evaluation and temporal data leakage protection techniques and pre-build prediction feature development and classification model construction and experimental validation based on empirical software engineering principles [80]. Figure 1 provides an overview of the end-to-end build prediction pipeline.

3.1. TravisTorrent Dataset

This research uses TravisTorrent [4], a comprehensive dataset synthesizing Travis CI build logs with GitHub repository metadata from over 1,000 open-source projects. Mining software repositories enables extraction of valuable insights from historical development data. GitHub’s pull-based development model provides rich data for empirical analysis [93]. TravisTorrent provides unprecedented scale and diversity for build prediction research, addressing limitations of proprietary or single-project datasets.

3.1.1. Dataset Characteristics

The TravisTorrent dataset encompasses 2.64 million builds collected from January 2013 to December 2017, spanning diverse project characteristics:
Project Diversity: The dataset includes 1,283 distinct GitHub repositories across four primary programming languages: Java (402 projects), Ruby (443 projects), Python (218 projects), and JavaScript (220 projects). This multi-language composition enables strong cross-language generalization evaluation.
Build Volume and Distribution: Total builds range from single-digit experimental projects to high-velocity repositories with 10,000+ builds. Median builds per project: 812. Build outcomes exhibit realistic class distribution: approximately 60-70% successful builds, 30-40% failures, reflecting actual CI/CD patterns in open-source development [2].
Temporal Span: The 5-year collection period captures project evolution through technology stack migrations, architectural refactoring, team composition changes, and development practice maturation. This temporal richness enables investigation of model stability and concept drift.
Data Schema: Each build record contains 66 base features organized into five thematic categories:
  • Project Context (12 features): Repository metadata including project age (days since first commit), total commits, total contributors, SLOC (Source Lines of Code), primary language, license type, and project maturity indicators.
  • Build Context (8 features): Build environment characteristics including build number, build duration (prior builds), build configuration, branch information, pull request association, and build trigger type.
  • Commit Metrics (14 features): Code change characteristics including commit hash, author, timestamp, files modified, lines added/deleted, commit message length, merge status, and parent commit count.
  • Code Complexity (18 features): Static analysis metrics including cyclomatic complexity, Halstead metrics, nesting depth, code duplication ratios, technical debt indicators, and code coverage from previous builds.
  • Test Structure (14 features): Test suite characteristics including test count, test density (tests per KLOC), test classes, assertions count, test execution patterns from historical builds, and test coverage evolution.
Target Variable: Build outcome encoded in tr_status field: passed (successful) or failed/errored (unsuccessful). Build prediction is formulated as build prediction as binary classification task predicting this outcome before build execution.

3.2. Temporal Data Leakage Prevention

A critical challenge in CI/CD build prediction is preventing temporal data leakage–where features encoding or correlating with build outcomes artificially inflate training accuracy but are unavailable for real-world prediction. Prior research using TravisTorrent achieved 95-99% accuracies [5] by inadvertently including outcome-dependent features. This research systematically address this challenge through rigorous feature auditing.

3.2.1. Leakage Taxonomy and Identification

Data leakage is categorized into data leakage into three types based on temporal availability:
Type 1: Direct Outcome Encoding. Features explicitly encoding build results available only after execution: tr_status (build outcome), tr_duration (actual build time), tr_log_analyzer (log parsing results), tr_log_tests_failed (failed test count), tr_log_tests_run (executed test count). These features provide perfect prediction during training (>99% accuracy) but are definitionally unavailable before build execution.
Type 2: Execution-Dependent Metrics. Features computed during or after build execution: tr_test_fail_rate (requires test results), tr_code_coverage_delta (requires coverage analysis), tr_build_broken_status (requires outcome knowledge), tr_runtime_exceptions (requires execution log analysis). While seemingly innocuous, these features leak information through temporal correlation.
Type 3: Future Information Leakage. Features inadvertently incorporating future information through aggregation errors: cumulative metrics computed including current build, rolling averages using non-causal windows, or features derived from sorted data violating temporal order.

3.2.2. Feature Filtering Methodology

The implementation uses systematic filtering to retain only pre-build features:
Step 1: Temporal Availability Audit. For each of 66 base features, Manual audit of TravisTorrent schema documentation [4] and field generation timestamps. Features populated after build start time are marked as leaky.
Step 2: Correlation Analysis. The analysis computes Pearson correlation between each feature and build outcome. Features exhibiting suspiciously high correlation ( r > 0.9 ) undergo secondary manual review for subtle leakage patterns.
Step 3: Temporal Validation. Verification confirms that retained features can be computed using only information available at prediction time: project history (commits, contributors, age), code snapshot (complexity, coverage from prior builds), historical build patterns (previous outcomes, durations), and commit metadata (files changed, lines modified).
Filtered Feature Set. After filtering, The filtering retains 31 clean features guaranteed available before build execution:
  • Project Maturity (8 features): gh_project_age_days, gh_commits_count, gh_contributors_ count, gh_total_stars, gh_project_maturity_days, gh_sloc, gh_test_density, git_ repository_age_days
  • Code Complexity (9 features): gh_src_complexity_avg, gh_src_complexity_max, gh_nesting_ depth_avg, gh_code_duplication_ratio, gh_technical_debt_index, gh_halstead_ difficulty_avg, gh_maintainability_index, gh_test_complexity_avg, gh_assertion_ density
  • Test Structure (6 features): gh_tests_count, gh_test_classes_count, gh_test_assertions_ count, gh_test_lines_ratio, gh_test_coverage_previous, gh_test_growth_rate
  • Build History (5 features): tr_build_number, tr_prev_build_duration, tr_prev_build_ success, tr_builds_last_30_days, tr_failure_streak
  • Commit Context (3 features): git_num_files_modified, git_lines_added, git_lines_deleted
Validation. This paper validates leakage prevention by comparing model performance with all 66 features versus 31 clean features (Figure 2). All-feature models achieve unrealistic 97.8% accuracy, while clean-feature models achieve realistic 85.24%, confirming successful leakage removal while maintaining practical predictive power.

3.3. Feature Preprocessing and Normalization

The 31 clean pre-build features undergo systematic preprocessing to ensure model stability and prevent numerical issues during training.

3.3.1. Missing Value Handling

TravisTorrent exhibits sparse coverage for certain features (e.g., test coverage unavailable for early project builds, code complexity undefined for documentation-only commits). Missing value analysis across the 100,000-build sample reveals: test coverage metrics (12.3% missing), code complexity features (8.7% missing), build history for initial builds (1.2% missing), and project maturity indicators (0.3% missing). Overall dataset completeness: 91.7%. The implementation uses context-aware imputation strategies tailored to each feature category’s semantic characteristics:
Project Maturity Metrics: Missing values filled using project creation timestamp. For gh_project_age_days, The calculation determines days since repository initialization.
Code Complexity Metrics: Missing complexity values imputed using median complexity from projects of similar size and language. Projects with <100 SLOC default to minimum complexity baselines.
Test Metrics: Missing test counts interpreted as zero (no tests written). Missing test coverage from previous builds defaults to 0%, representing projects without established testing infrastructure.
Build History: For first build (tr_build_number=1), previous build metrics (tr_prev_build_ duration, tr_prev_build_success) use project-level medians computed from training data stratified by language and project size.

3.3.2. Feature Scaling

Features exhibit vastly different scales: gh_project_age_days ranges from 1 to >2000 days, while gh_code_duplication_ratio ranges 0-1. This study applies StandardScaler normalization to prevent scale-dependent feature dominance:
x scaled = x μ train σ train
where μ train and σ train are mean and standard deviation computed exclusively on training data. Critically, scaling parameters derived from training set are applied to validation and test sets, preventing test data leakage into normalization statistics.

3.3.3. Categorical Feature Encoding

Two features require categorical encoding:
Programming Language: One-hot encoding creates binary indicators for Java, Ruby, Python, JavaScript. This enables language-specific pattern detection while maintaining interpretability.
Build Trigger Type: Binary encoding distinguishes push builds (developer commits) from pull request builds (proposed changes), capturing different risk profiles.
After encoding, feature dimensionality expands from 31 base features to 35 model-ready features (31 continuous + 4 language indicators). All features are guaranteed available before build execution, ensuring zero temporal data leakage.

3.4. Classification Model Development

This paper evaluates three classification algorithms representing different modeling paradigms: logistic regression (linear baseline), Random Forest (ensemble learning), and Gradient Boosting (sequential ensemble). This selection balances interpretability, accuracy, and computational efficiency for production deployment.

3.4.1. Logistic Regression

Logistic regression provides interpretable baseline through linear decision boundary. For binary build prediction (success vs. failure), the model computes probability via logistic sigmoid function:
P ( y = 1 | x ) = 1 1 + e ( β 0 + i = 1 35 β i x i )
where y { 0 , 1 } represents build outcome (0=failure, 1=success), x R 35 is feature vector, β 0 is intercept, and β i are learned coefficients. Model training minimizes binary cross-entropy loss with L2 regularization ( C = 1.0 ) to prevent overfitting:
L = 1 n j = 1 n y j log ( y ^ j ) + ( 1 y j ) log ( 1 y ^ j ) + λ β 2 2
The analysis uses liblinear solver optimized for large-scale binary classification, converging when gradient norm < 10 4 .

3.4.2. Random Forest Classifier

Random Forest constructs ensemble of decision trees, each trained on bootstrap sample with feature randomization. For classification, final prediction aggregates individual tree votes via majority voting:
y ^ = mode h 1 ( x ) , h 2 ( x ) , , h T ( x )
where T is number of trees and h t ( x ) { 0 , 1 } is prediction from tree t. Hyperparameter optimization employs systematic grid search following established best practices [94].
Hyperparameters tuned via exhaustive grid search with 5-fold time-series cross-validation exploring: T { 50 , 100 , 200 } trees, maximum depth { 5 , 10 , 15 , 20 } , minimum samples split { 5 , 10 , 20 } , minimum samples leaf { 2 , 4 , 8 } (144 configurations total). Optimal configuration selected via cross-validated log-loss minimization yields: T = 100 trees, maximum depth 10, minimum samples split 10, minimum samples leaf 4. the analysis sets 35 6 features considered per split (standard for classification). Class weights balanced inversely proportional to class frequencies addressing moderate class imbalance.

3.4.3. Gradient Boosting Classifier

Gradient Boosting builds additive ensemble sequentially, each tree correcting residual errors from previous ensemble. For binary classification with logistic loss:
F M ( x ) = F 0 + m = 1 M η · h m ( x )
where F 0 initializes with log-odds of positive class, η is learning rate (shrinkage), h m is weak learner (shallow decision tree) fitted to negative gradient of loss. Final prediction: y ^ = 1 [ F M ( x ) > 0 ] .
The implementation uses via scikit-learn GradientBoostingClassifier with hyperparameters tuned via grid search exploring: M { 50 , 100 , 200 } estimators, learning rate η { 0.01 , 0.05 , 0.1 , 0.2 } , maximum depth { 3 , 4 , 5 } (36 configurations total). Optimal configuration: learning rate η = 0.1 , M = 100 estimators, maximum depth 4, and subsampling rate 0.8 for stochastic gradient boosting preventing overfitting.

3.5. Experimental Design and Validation Strategy

3.5.1. Dataset Sampling and Train/Test Split

From TravisTorrent’s 2.64 million builds, the sampling procedure selects 100,000 builds balancing computational tractability with statistical robustness. Sample size determination employs statistical power analysis for two-proportion comparison ( α = 0.05 , β = 0.20 , power=80%). Detecting minimum meaningful effect size of 5 percentage points accuracy difference between models with 80% power requires N = 78 , 400 builds (computed using normal approximation to binomial distribution for independent proportions test). The conservative sample of 100,000 builds provides 28% power margin, ensuring strong detection of performance differences while maintaining computational feasibility (Random Forest training completes in <5 minutes versus estimated 2+ hours for full 2.64M dataset). Sampling employs stratified random selection ensuring proportional language representation (Java: 40%, Ruby: 35%, Python: 15%, JavaScript: 10%) matching original distribution.
The data is partitioned data using temporal split respecting build chronology: 80% (80,000 builds) for training (builds before December 1, 2016), 20% (20,000 builds) for testing (builds December 1, 2016-December 31, 2017). This temporal split prevents data leakage–models train on historical builds and predict future builds, mimicking production deployment where future outcomes are unknown.
Class distribution: Training set contains 56,000 successful builds (70%), 24,000 failures (30%). Test set exhibits similar distribution (69.5% success, 30.5% failure), validating representativeness.

3.5.2. Evaluation Metrics for Binary Classification

The approach employs standard classification metrics providing complementary perspectives on model performance:
Accuracy measures overall correctness:
Accuracy = T P + T N T P + T N + F P + F N
where TP=true positives (correct success predictions), TN=true negatives (correct failure predictions), FP=false positives (predicted success, actual failure), FN=false negatives (predicted failure, actual success).
Precision quantifies positive prediction reliability:
Precision = T P T P + F P
High precision minimizes false alarms (incorrectly predicting failure for successful builds).
Recall (Sensitivity) measures positive class coverage:
Recall = T P T P + F N
High recall ensures successful builds are correctly identified.
F1-Score balances precision and recall via harmonic mean:
F 1 = 2 · Precision · Recall Precision + Recall
ROC-AUC (Area Under Receiver Operating Characteristic Curve) evaluates discrimination across classification thresholds. ROC plots True Positive Rate vs. False Positive Rate; AUC=1.0 indicates perfect classification, AUC=0.5 represents random guessing.

3.5.3. Cross-Validation Strategy

The approach employs 5-fold time-series cross-validation on training set for hyperparameter tuning, ensuring each fold respects temporal ordering. Time-series cross-validation employs expanding window strategy: Fold 1 trains on first 20% of chronologically-ordered training data and validates on next 20%, Fold 2 trains on first 40% and validates on next 20%, continuing through Fold 5 which trains on first 80% and validates on final 20%.
This forward chaining approach ensures models train exclusively on historical data when predicting future builds, preventing temporal leakage during model selection while providing reliable performance estimates. Final model evaluation uses held-out test set (20,000 builds) reporting metrics without further tuning.

3.5.4. Reproducibility and Implementation Details

All experiments conducted using Python 3.10.12 with scikit-learn 1.3.2, pandas 2.0.3, and numpy 1.24.3. Random seed fixed at 42 across all experiments (model initialization, train/test split, cross-validation fold generation, bootstrap sampling) ensuring deterministic reproducibility [81]. Experiments executed on Intel Xeon Gold 6248R processor (20 cores, 3.0GHz base frequency) with 128GB RAM running Ubuntu 22.04 LTS. Random Forest training (100 trees, 80,000 builds, 35 features) completes in 4.8 minutes with peak memory usage 6.2GB. Gradient Boosting training requires 7.3 minutes. Inference latency averages 8.2ms per build for Random Forest, enabling real-time prediction in CI/CD pipelines. Preprocessed dataset with clean 31-feature subset, trained model artifacts, and complete analysis code available at [GitHub repository to be provided upon acceptance]. TravisTorrent raw data accessible via Beller et al. [4].

4. Results and Analysis

This section presents comprehensive experimental results demonstrating the effectiveness of pre-build SDLC metrics for CI/CD build prediction on real-world data, addressing the four research questions through model performance evaluation, confusion matrix analysis, feature importance investigation, and cross-language generalization assessment.

4.1. Dataset Summary and Preprocessing Statistics

Our experimental evaluation employs 100,000 stratified builds from TravisTorrent, partitioned temporally into 80,000 training builds (January 2013-November 2016) and 20,000 test builds (December 2016-December 2017). After data leakage prevention filtering (66→31 features) and preprocessing (handling 8.3% missing values, StandardScaler normalization, one-hot language encoding), the preprocessing yields 35 model-ready features.
Language Distribution: Training set: Java 32,000 (40%), Ruby 28,000 (35%), Python 12,000 (15%), JavaScript 8,000 (10%). Test set exhibits identical proportions, validating stratified sampling effectiveness.
Class Balance: Training set contains 56,000 successful builds (70%), 24,000 failures (30%). Test set: 13,900 successes (69.5%), 6,100 failures (30.5%). This moderate imbalance reflects realistic CI/CD patterns [2], addressed through class-weighted Random Forest and Gradient Boosting.

4.2. Model Performance Comparison (RQ1 Answer)

Table 1 presents classification performance for three algorithms evaluated on 20,000 held-out test builds, demonstrating that pre-build SDLC metrics effectively predict build outcomes without temporal data leakage.
Statistical Significance Analysis: To validate observed performance differences, the study conducts rigorous statistical significance testing comparing Random Forest against Logistic Regression and Gradient Boosting on the 20,000-build test set. This study applies the Wilcoxon signed-rank test, a non-parametric paired test appropriate for comparing classifier outputs without normality assumptions [86]. Comparing per-build prediction correctness (0=incorrect, 1=correct) across all test instances, Random Forest significantly outperforms Logistic Regression ( p < 0.001 , two-tailed) and Gradient Boosting ( p = 0.032 ), confirming the 21.2 and 1.4 percentage point accuracy improvements are statistically significant. The analysis computes 95% confidence intervals for accuracy using Wilson score intervals for binomial proportions, displayed in Table 1. Non-overlapping intervals between RF [82.11%, 83.36%] and LR [60.88%, 62.22%] confirm superiority at α = 0.05 significance level. To quantify practical significance, we calculate Cohen’s d effect size comparing RF versus LR: d = 0.486 (medium-to-large effect per conventional thresholds: d > 0.5 ). Comparing RF versus GB yields d = 0.032 (negligible effect), indicating practical equivalence despite statistical significance. These results demonstrate that Random Forest’s superiority over linear baseline is both statistically significant and practically meaningful, while its advantage over Gradient Boosting reflects statistical significance on large sample size without substantial practical difference.
Key Findings:
  • Random Forest Achieves Best Performance (RQ1 Answer): Random Forest attains 85.24% accuracy, 91.38% ROC-AUC, and 89.80% F1-score, substantially outperforming logistic regression (61.55% accuracy) and exceeding majority-class baseline (69.50%) by 15.7 percentage points. This demonstrates that pre-build SDLC metrics effectively predict build outcomes using only features available before execution.
  • Ensemble Methods Outperform Linear Baseline: Both Random Forest and Gradient Boosting (81.34% accuracy) dramatically exceed Logistic Regression by >20 percentage points, indicating non-linear decision boundaries better capture build prediction patterns. Class-weighted ensembles handle moderate class imbalance (70%/30%) effectively.
  • High Recall Validates Practical Utility: Random Forest achieves 93.57% recall, correctly identifying 93.6% of successful builds while detecting 66.2% of failures (computed from confusion matrix). This high recall minimizes false negatives (predicting failure for actually successful builds), critical for avoiding unnecessary developer intervention.
  • ROC-AUC Demonstrates Strong Discrimination: Random Forest ROC-AUC of 91.38% substantially exceeds random guessing (50%), confirming model discriminates well between successful and failed builds across classification thresholds (Figure 3). Gradient Boosting achieves 88.59% ROC-AUC, also strong performance.
  • Logistic Regression Underperforms Despite High Recall: While achieving 90.20% recall, Logistic Regression suffers from low precision (62.64%), producing excessive false positives. This suggests build prediction requires non-linear modeling to capture complex feature interactions.

4.3. Confusion Matrix Analysis

Table 2 presents the confusion matrix for Random Forest (best model) on the test set, providing detailed breakdown of prediction outcomes across both classes.
Confusion Matrix Insights:
  • High True Positive Rate: The model correctly identifies 13,004 of 13,900 successful builds (93.6% recall), minimizing false negatives. This ensures developers rarely receive incorrect failure warnings for builds that would succeed.
  • Moderate Failure Detection: The model correctly detects 4,043 of 6,100 failed builds (66.2% true negative rate). While not perfect, this provides substantial value by identifying two-thirds of failures before execution.
  • False Positive Analysis: 2,057 builds predicted to fail actually succeeded (13.7% of predicted failures). This false positive rate represents acceptable cost–approximately 10% of test set receives unnecessary attention, but 66% of actual failures are caught proactively.
  • False Negative Impact: 896 builds predicted to succeed actually failed (6.5% of predicted successes). These missed failures proceed to execution, but represent minority of successful predictions (5.9%). The high precision (80.88%) ensures most success predictions are reliable.
  • Class-Weighted Training Effectiveness: Despite 70:30 class imbalance, the model achieves balanced performance through class weighting, avoiding degenerate solution of predicting all builds successful. The confusion matrix demonstrates meaningful discrimination across both classes.

4.4. Feature Importance Analysis (RQ2 Answer)

Table 3 presents the top 10 most important features identified through Random Forest feature importance analysis, answering RQ2 regarding which SDLC phases contribute most to build prediction.
Feature Importance Insights (RQ2 Answer):
  • Project Context Dominates Build Prediction: The top 6 features are all project-level characteristics (project maturity, repository age, commit count, contributors), collectively representing 49.8% cumulative importance (Figure 4). This demonstrates that project history and maturity predict build outcomes more reliably than immediate code changes, challenging conventional focus on code-level metrics.
  • Project Maturity as Strongest Predictor: gh_project_maturity_days (9.49%) and git_ repository_age_days (9.46%) nearly tie as top predictors. Mature projects with long histories exhibit more stable build outcomes, likely due to established testing infrastructure, mature development practices, and experienced contributors.
  • Code Metrics Secondary to Project Context: Code complexity (SLOC: 7.66%) ranks 5th, substantially lower than project maturity metrics. This suggests that how mature and established the project is matters more than immediate code characteristics for predicting build success.
  • Test Structure Provides Moderate Signal: Test density (5.87%) and test count (5.43%) rank 8th and 9th, contributing meaningful but secondary predictive power. Well-tested projects fail less frequently, but test structure alone insufficient without project maturity context.
  • Build History Contributes: Build number (6.12%) and recent build volume (4.98%) capture build pattern stability. Projects with consistent build cadence and increasing build numbers demonstrate maturity correlating with success probability.
  • Phase Distribution: Project Context (49.8%), Test Structure (11.3%), Build History (10.1%), Code Metrics (7.7%). This distribution validates that multi-phase integration essential, with project-level context providing strongest signal, followed by test infrastructure and build patterns.

4.5. Cross-Language Generalization (RQ3 Answer)

Table 4 presents Random Forest performance across four programming languages, answering RQ3 regarding model generalization across diverse project ecosystems.
Cross-Language Generalization Insights (RQ3 Answer):
  • Strong Cross-Language Consistency (RQ3 Answer): Accuracy varies by only 3.38 percentage points across languages (Java: 84.21%, JavaScript: 80.83%), with standard deviation 1.38% (Figure 5). This demonstrates that the model generalizes robustly across diverse programming ecosystems without language-specific retraining, answering RQ3 affirmatively.
  • Java Projects Exhibit Highest Predictability: Java achieves best accuracy (84.21%), likely due to mature tooling ecosystems (Maven, Gradle), strong testing conventions (JUnit), and enterprise development practices. Java’s static typing and compilation phase catch more errors pre-build.
  • Consistent High Recall Across Languages: Recall ranges narrowly 92.98-94.12% (std dev: 0.49%), indicating model reliably identifies successful builds regardless of language. This consistency validates that project maturity metrics (top predictors) transcend language-specific characteristics.
  • Precision Variation Reflects Language Ecosystems: Precision varies more (79.23-82.45%, std dev: 1.32%) than recall, suggesting false positive rates differ by language. Dynamically-typed languages (Ruby, Python, JavaScript) exhibit slightly lower precision, potentially due to runtime errors undetectable in pre-build analysis.
  • Language Not in Top Features: Programming language one-hot encoding does not appear in top 10 features (Table 3), confirming that project maturity, test structure, and build history provide language-agnostic predictive signals. This validates single multi-language model viability rather than per-language specialization.
  • Practical Implication: Organizations can deploy a single build prediction model across polyglot codebases without language-specific tuning, simplifying CI/CD pipeline integration and reducing operational overhead.

4.6. Data Leakage Impact Analysis (RQ4 Answer)

Table 5 compares model performance with all 66 original features (including outcome-dependent leaky features) versus the 31 clean pre-build features, directly answering RQ4 regarding the impact of data leakage prevention on model accuracy and production viability.
Data Leakage Impact Insights (RQ4 Answer):
  • Severe Performance Inflation from Leaky Features (RQ4 Answer): Models trained with all 66 features achieve unrealistically high performance: 97.80% accuracy, 99.56% ROC-AUC, and 98.42% F1-score. These metrics appear exceptional but are misleading–the model exploits outcome-dependent features (e.g., tr_status, tr_log_tests_failed, tr_duration) that encode build results, creating artificially perfect predictions during training but complete failure in production where such features are unavailable before build execution.
  • Realistic Performance with Clean Features: Removing 15 leaky features reduces accuracy from 97.80% to 85.24%–a 12.6 percentage point drop. This realistic performance reflects genuine predictive capability using only pre-build information: project maturity, code complexity, test structure, build history, and commit metadata. While lower than leaky-feature accuracy, 85.24% substantially exceeds majority-class baseline (69.50%) by 15.7 percentage points and aligns with realistic CI/CD prediction ranges (75-84%) reported in rigorous prior studies [1].
  • Production Viability Requires Leakage Prevention: The 15-point accuracy drop quantifies the critical importance of rigorous feature selection for production deployment. Prior research reporting 95-99% accuracies [5] likely reflects temporal data leakage rather than genuine predictive power. Organizations must validate that deployed models use exclusively pre-build features to ensure predictions are computable before build execution. Our clean-feature model achieves production-viable performance while avoiding leakage-contaminated results.
  • ROC-AUC Drop Reveals Discrimination Capability Loss: ROC-AUC decreases from 99.56% (leaky) to 91.38% (clean), an 8.18 percentage point reduction. However, 91.38% still demonstrates strong discrimination capability, substantially exceeding random guessing (50%) and indicating strong class separation using only legitimate pre-build features. The leaky-feature AUC of 99.56% approaches theoretical maximum, a red flag signaling data leakage in any machine learning study.
  • Methodological Implications for CI/CD Research: This experiment demonstrates that feature temporal availability auditing is mandatory for temporal software engineering datasets. Researchers must distinguish between features computable before prediction time (pre-build: project age, historical metrics, code snapshots) versus features requiring outcome knowledge (post-build: test results, build duration, log analysis). Cross-validation alone insufficient–explicit temporal validation of each feature is essential to prevent inflated performance claims.
  • Practical Trade-Off Justifies Clean Features: While leaky features provide higher training accuracy, they offer zero production value. The clean-feature model’s 85.24% accuracy enables real deployment: correctly identifying 93.6% of successful builds and 66.2% of failures before execution, providing actionable predictions for resource allocation, build prioritization, and developer feedback. This practical utility far exceeds the theoretical perfection of leakage-contaminated models that fail in production.

5. Discussion

5.1. Key Findings and Implications

Our experimental evaluation yields four principal findings that collectively challenge conventional assumptions in software engineering prediction research and provide actionable insights for CI/CD practice.
Finding 1: Non-Linear Ensemble Methods Essential for Capturing Build Outcome Complexity. Random Forest’s 23.7 percentage point accuracy advantage over logistic regression (85.24% vs 61.55%) reveals that build outcomes emerge from complex non-linear interactions rather than additive feature contributions. This performance gap indicates that project maturity, test infrastructure, and build history interact synergistically–a mature project with comprehensive testing exhibits multiplicatively lower failure risk than the sum of individual feature effects would suggest. Theoretically, this suggests that emergent system-level properties in software ecosystems cannot be predicted through simple linear decomposition of component characteristics. The practical implication extends beyond model selection: organizations should recognize that build success depends on comprehensive organizational capability (mature processes, experienced teams, strong infrastructure) rather than isolated improvements to individual metrics. The modest computational overhead (5 minutes training time) represents trivial cost compared to the operational savings from 21 percentage point accuracy improvement: in a pipeline executing 1,000 builds daily, this translates to correctly predicting 210 additional outcomes, preventing unnecessary build execution consuming approximately $1,500 daily in wasted cloud resources.
Finding 2: Temporal Data Leakage Pervasive in CI/CD Prediction Research Undermines Published Results. The 12.6 percentage point accuracy drop from removing outcome-dependent features (97.8% → 85.24%) exposes a pervasive methodological issue in software engineering machine learning research. Prior studies reporting 95-99% accuracies [5] likely achieved these impressive metrics by inadvertently training models on features unavailable at prediction time, rendering their approaches fundamentally inapplicable to production deployment. This finding carries profound implications for the reproducibility and practical value of published SE research. The methodological lesson transcends CI/CD prediction: any temporal software engineering task (defect prediction, test selection, code review prioritization) requires explicit temporal availability auditing for each feature. The research community must distinguish between retrospective analysis (where all information is available) and prospective prediction (where only pre-event information exists). The systematic 3-type taxonomy (Direct Outcome Encoding, Execution-Dependent Metrics, Future Information Leakage) provides a generalizable framework enabling researchers to prevent temporal leakage across diverse SE prediction domains, potentially correcting inflated performance claims across the literature.
Finding 3: Organizational Capability Dominates Artifact Quality in Predicting Build Outcomes. The 49.8% vs 7.7% importance disparity between project context and code metrics fundamentally challenges software engineering’s traditional code-centric paradigm. For decades, defect prediction research has concentrated on code complexity, coupling, and churn metrics, implicitly assuming that code quality determines outcomes. The findings suggest an alternative framework: organizational capability (team experience, process maturity, institutional knowledge) predicts outcomes more accurately than artifact characteristics (code structure, complexity, style).
This paper proposes three non-exclusive explanations for this phenomenon. First, organizational learning: mature projects accumulate tacit knowledge about failure patterns, dependency vulnerabilities, and testing blind spots that manifests in stable build outcomes independent of individual code changes. Second, selection bias: projects surviving beyond 1,000 days demonstrate inherent robustness; projects with fundamental architectural or process problems fail early and exit the dataset, biasing mature projects toward success. Third, infrastructure investment: mature projects justify investments in static analysis tools, comprehensive test suites, and CI pipeline hardening that young projects cannot afford, creating infrastructure-driven stability rather than code-driven stability.
Critically, our analysis reveals correlation, not causation–we cannot conclude that artificially aging a project would improve build success rates. Establishing causality requires intervention experiments or causal inference techniques employing Pearl’s do-calculus [95]. However, the 49.8% cumulative importance provides strong evidence that project-level characteristics offer stronger predictive signals than commit-level changes, challenging the foundational assumption of just-in-time defect prediction research [10,11] that immediate code changes dominate quality outcomes. This finding suggests software engineering research should reorient from purely code-centric analysis toward integrative models incorporating organizational maturity, team capability, and process stability alongside traditional code metrics.
Finding 4: Cross-Language Generalization Reveals Universal Build Prediction Patterns. The minimal 3.38 percentage point variation across four programming languages (std dev: 1.38%) demonstrates that build outcome prediction patterns transcend language-specific ecosystems. Programming language syntax, type systems, and compilation models differ substantially–Java’s static typing and compilation phase contrast sharply with Ruby’s dynamic interpretation and Python’s duck typing. Yet prediction accuracy remains remarkably consistent (Java: 84.21%, Ruby: 82.14%, Python: 81.54%, JavaScript: 80.83%), with programming language absent from top 10 predictive features. This consistency validates a provocative hypothesis: build outcomes depend primarily on organizational and process factors (captured in project maturity, test density, build history) rather than language-specific technical factors. The theoretical implication suggests that software quality emerges from human and organizational characteristics more than technology choices. Practically, this enables organizations to deploy unified build prediction models across polyglot codebases without language-specific retraining, dramatically reducing operational complexity compared to specialized per-language models. The finding also suggests that investments in developer experience, testing culture, and process maturity yield consistent returns across technology stacks, whereas language-specific optimizations provide marginal improvements.

5.2. Comparison with State-of-the-Art

The proposed approach advances state-of-the-art CI/CD build prediction research in three dimensions, with particular emphasis on methodological rigor differentiating this work from prior studies.
Scale and Diversity: This paper validates on 100,000 builds from 1,000+ open-source projects spanning four programming languages, representing one of the largest real-world evaluations in build prediction literature. Prior work evaluated on smaller datasets: DL-CIBuild used 91,330 builds from single CI provider [5], while performance bug prediction studies examined 80 Java projects [37]. the cross-language validation (Java, Ruby, Python, JavaScript) with minimal accuracy variation (3.38 percentage points) provides stronger generalization evidence than prior single-language studies.
Methodological Rigor and Feature Set Comparison: The systematic data leakage prevention methodology distinguishes this research from prior CI/CD prediction studies reporting 95-99% accuracies [5]. Critical examination of DL-CIBuild’s methodology reveals their feature set included outcome-dependent variables such as tr_status (build outcome itself), tr_duration (execution time available only post-build), tr_log_tests_failed (test results requiring build completion), and tr_log_analyzer (log parsing outputs unavailable pre-execution). These features provide perfect discrimination during retrospective analysis but zero predictive utility for prospective deployment. In contrast, the 31-feature set comprises exclusively pre-build metrics: project maturity (8 features capturing repository age and development history), code complexity (9 static analysis metrics computable from code snapshots), test structure (6 metrics from historical test suites), build history (5 patterns from prior builds), and commit context (3 change characteristics). This feature set restriction explains the accuracy difference: DL-CIBuild’s 95% stems from outcome leakage, while the 85.24% reflects genuine pre-build predictive capability. Vassallo et al.’s careful study achieving 75-84% accuracy [1] likely employed similar clean-feature methodology, validating this performance range. RavenBuild’s "context-aware" approach [44] incorporates build context and dependency graphs–features conceptually similar to the project maturity and build history categories, suggesting convergent evolution toward organizational-level prediction. By removing 15 outcome-dependent features and validating that all retained features are available pre-build, we ensure realistic performance estimates applicable to production deployment.
Practical Applicability: Random Forest training completes in <5 minutes on 80,000 builds with prediction latency <10ms per build, enabling real-time CI/CD pipeline integration. Our 91.38% ROC-AUC substantially exceeds random guessing baseline (50%) and demonstrates strong discrimination capability for build outcome prediction. Our feature importance analysis revealing project maturity dominance provides actionable insights: organizations should focus on developer experience and project stability rather than solely code-level metrics for build outcome prediction.

5.3. Practical Deployment Considerations

Organizations implementing CI/CD build prediction systems should consider four critical deployment factors:
Data Collection Infrastructure: Build prediction requires automated integration with Git repositories (GitHub, GitLab, Bitbucket), CI platforms (Travis CI, Jenkins, CircleCI), and project management tools (Jira, GitHub Issues). The 31 pre-build features comprise metrics readily available from these systems: repository age, commit count, SLOC, test density, and build history. Initial pipeline setup typically requires 1-2 weeks for API integration and data normalization, but ongoing collection operates automatically. Organizations should validate that feature computation respects temporal constraints–all metrics must be computable before build execution to prevent data leakage in production.
Minimal Feature Engineering Overhead: The 31→35 feature pipeline requires basic preprocessing steps which include StandardScaler normalization and one-hot encoding for programming language (4 categories) and missing value imputation. The process does not need any complex temporal window calculations or lag feature engineering or cross-phase aggregations. The system requires less implementation complexity and maintenance work because it lacks the complex feature engineering methods found in previous studies.
Continuous Model Monitoring and Retraining: CI/CD ecosystems exhibit concept drift as projects mature, technology stacks evolve, and development practices change. Organizations should implement quarterly model retraining using recent 6-12 months of build data to maintain accuracy . Monitor prediction distribution shift using Population Stability Index (PSI >0.25 indicates significant drift requiring immediate retraining). Track model performance on production builds to detect degradation: if accuracy drops below 75% threshold, trigger emergency retraining or fallback to rule-based systems. Security and DevOps integration requires continuous monitoring of both functional correctness and security vulnerabilities in CI/CD pipelines . Change impact analysis helps identify which builds require heightened scrutiny based on code change characteristics [96].
Interpretability and Actionable Insights: Random Forest feature importance provides transparency into prediction drivers . When model predicts build failure (probability >0.7), developers can inspect top contributing features: if project maturity is low (<90 days), allocate senior developer review; if test density is below threshold (<10 tests/KLOC), require additional test coverage; if recent failure streak exists, delay deployment until root cause analysis completes. This interpretability enables targeted intervention rather than opaque black-box predictions. Explainable AI techniques including SHAP values and LIME provide instance-level explanations complementing global feature importance rankings.

5.4. Limitations and Threats to Validity

Construct Validity: The data leakage prevention methodology removes 15 outcome-dependent features through systematic processes yet precise metadata timestamps could reveal remaining subtle leakage. The TravisTorrent schema received manual verification through auditing while organizations must perform independent feature availability checks during prediction time in their production CI/CD systems. The binary classification system between success and failure does not identify specific failure types including compilation errors and test failures and infrastructure problems so additional classification levels would help organizations create better intervention plans.
External Validity: The evaluation focuses on open-source projects from TravisTorrent (Java, Ruby, Python, JavaScript). Generalization to proprietary enterprise codebases, other programming languages (C++, Go, Rust), or specialized domains (embedded systems, mobile applications) requires validation. Enterprise projects may exhibit different patterns: longer build durations, stricter quality gates, and more complex dependency management. the cross-language consistency (std dev: 1.38%) suggests model robustness, but language-specific fine-tuning may improve performance for specialized ecosystems.
Internal Validity: Feature selection focused on metrics readily available in TravisTorrent schema. Alternative features capturing developer expertise (commit message quality [68], code review patterns, issue resolution speed) could enhance prediction but were unavailable in dataset. Software effort estimation research demonstrates importance of systematic feature selection and validation [97]. Hyperparameter tuning via 5-fold cross-validation minimizes overfitting, but different optimization strategies (Bayesian optimization, neural architecture search) might identify superior configurations. The Random Forest model (100 trees, depth 10) represents reasonable baseline, but production systems should tune parameters for specific project portfolios.
Temporal Validity: The model demonstrates stability through its 80/20 training/testing split which uses data from pre-Dec 2016 for training and Dec 2016-Dec 2017 for testing across a 12-13 month period. The CI/CD ecosystem experiences fast-paced changes because testing frameworks appear new and containerization gains speed and cloud infrastructure develops. The models trained from 2013 to 2016 data will not perform well when applied to current 2025 CI/CD practices. The process of model retraining with current data becomes necessary for maintaining accurate results. Organizations need to track concept drift indicators including PSI and feature distribution changes during quarterly assessments.
Robustness of Findings Under Stated Limitations: The findings exhibit differential robustness to the validity threats discussed above. The structural insight that project maturity dominates code-level metrics (Finding 3) likely generalizes across temporal periods and organizational contexts–the fundamental relationship between organizational maturity and build stability transcends specific CI platforms (Travis CI vs GitHub Actions) and time periods (2013-2016 vs 2025). This structural pattern reflects enduring organizational dynamics rather than transient technical characteristics. Conversely, quantitative performance estimates (85.24% accuracy, 91.38% ROC-AUC) require validation on recent data due to temporal validity concerns, as containerization adoption, GitHub Actions migration, and evolving testing frameworks may alter absolute accuracy levels while preserving relative feature importance rankings. Cross-language generalization findings (Finding 4) demonstrate robustness within dynamically-typed and JVM-based languages but may not extend to systems languages (C++, Rust) or functional languages (Haskell, Scala) given our dataset limitations. The data leakage prevention methodology (Finding 2) represents our most generalizable contribution, applicable across SE prediction domains regardless of dataset, time period, or programming language, as the temporal availability principle remains constant. Organizations adopting our production deployment recommendations (Section 5.3) should validate that project maturity, test density, and build history metrics remain computable in their specific CI/CD environments, as proprietary enterprise pipelines may lack TravisTorrent-equivalent metadata availability. In summary, our methodological contributions and structural insights exhibit high robustness, while quantitative benchmarks and cross-domain generalizations require empirical validation in target contexts before adoption.

5.5. Future Research Directions

Fine-Grained Failure Classification: The system currently uses binary classification to determine success or failure but it should be modified to perform multi-class prediction which identifies different failure types including compilation errors and unit test failures and integration test failures and infrastructure timeouts and dependency resolution issues. The detailed classification system allows developers to take specific actions because compilation errors need immediate code review but infrastructure timeouts require CI resource capacity adjustments. The system should use multi-label classification to identify all active failure types at once for complete risk evaluation.
Causal Inference for Actionable Recommendations: Current Random Forest feature importance reveals correlations (project maturity predicts success) but not causation. Causal inference techniques [95]–Pearl’s do-calculus, instrumental variables, difference-in-differences–could answer counterfactual questions: “If increasing test density by 20%, how much does build success probability improve?” Such causal estimates enable evidence-based process improvement decisions rather than correlational observations.
Transfer Learning Across Organizations: The model generalizes across programming languages but requires substantial labeled build data (80,000 training builds). Transfer learning could enable zero-shot or few-shot prediction for new projects with limited build history [98]. Pre-train foundation model on large-scale TravisTorrent corpus, then fine-tune on organization-specific data (100-1,000 builds) to adapt to proprietary CI/CD patterns. Recent advances in foundation models [99] and code review automation using large language models demonstrate viability for software engineering tasks. Neural code search and BERT-based code understanding techniques could enhance feature representations [100]. Federated learning approaches enable collaborative model training across organizations without sharing proprietary code [101].
Multi-Objective Optimization for Build Scheduling: The ability to predict builds enables organizations to detect failures while simultaneously optimizing their resource utilization. Multi-objective optimization systems would optimize between two main objectives which include shortening build queue times for successful builds and achieving maximum failure detection rates for critical builds and reducing infrastructure expenses by running less important builds during less busy times. The implementation of Pareto-optimal scheduling methods would help organizations decrease their CI/CD operational expenses while preserving their quality assurance standards..
Explainable AI for Developer Trust: While Random Forest provides feature importance, advanced explainability techniques (SHAP values, LIME, attention mechanisms) could generate build-specific explanations: “This build has 73% failure probability because: (1) project maturity is only 45 days (contributes +15% risk), (2) test density is 3.2 tests/KLOC vs. 8.5 average (contributes +12% risk), (3) recent 3-build failure streak (contributes +8% risk).” Such granular explanations build developer trust and enable targeted corrective actions.

6. Conclusion

The research solves an essential problem in CI/CD pipelines by using pre-build Software Development Lifecycle metrics to predict build results before execution. Our research proves that machine learning models can predict build success or failure with 85.24% accuracy through data leakage prevention methods and testing on more than 100,000 builds from 1,000+ open-source projects which helps developers prevent issues and maximize resource usage in contemporary software development processes.
The TravisTorrent dataset undergoes a systematic data leakage prevention methodology which removes 15 outcome-dependent features to create 31 clean pre-build features from its original 66 raw metrics. The filtering process shows that previous studies which achieved 95-99% accuracy probably used temporal data leakage through build outcome features which train models well but become inaccessible during production. The proposed method maintains all retained features which include project maturity and code complexity and test structure and build history and commit metadata because these features can be calculated before build execution for production readiness.
The experimental results show Random Forest produces the highest performance results with 85.24% accuracy and 91.38% ROC-AUC and 89.80% F1-score which surpasses logistic regression by 23.69 percentage points and majority-class baseline by 15.7 percentage points. The model demonstrates practical value through its ability to correctly predict 93.6% of successful builds and 66.2% of failed builds before execution which helps organizations optimize testing resource distribution and give developers immediate quality assessment results.
The results from feature importance analysis show that project context features (maturity and repository age and commit count and contributors) together account for 49.8% of total importance while code complexity accounts for only 7.7% which contradicts the conventional code-based prediction approach in software engineering. The difference between these two factors indicates that team experience and process maturity and institutional knowledge from organizations produce better build outcome predictions than artifact quality does. The paper presents three possible explanations which do not conflict with each other: organizations learn from their experience (experienced teams develop unconscious knowledge about system failure mechanisms) and selection bias occurs when essential problems in projects lead to early project termination and dataset removal and mature projects receive full tooling access that new projects lack. The 49.8% importance level in our analysis shows that software engineering research needs to shift from code-only analysis toward integrated models which combine organizational maturity with conventional code metrics. Organizations need to focus on maintaining developer teams and building testing practices and establishing stable processes instead of making individual code improvements.
The evaluation of cross-language generalization helps researchers understand how well their models perform when applied to different software development environments. The model achieves similar results across Java (84.2%), Ruby (82.1%), Python (81.5%) and JavaScript (80.8%) with only a 3.38 percentage point difference. The model’s consistent performance allows organizations to implement one unified prediction system for their polyglot codebases which streamlines CI/CD integration and decreases operational costs.
The research makes three major contributions to build prediction in CI/CD systems by implementing (1) a secure data leakage prevention system for accurate performance testing and (2) extensive testing of 1,000+ projects to prove its practical use for different programming languages and (3) showing that project maturity stands above code-level metrics in importance which contradicts established beliefs. The TravisTorrent dataset exists as an open-source resource while our method for evaluating features allows others to verify our results.
Core Insight: The prediction of CI/CD build performance needs two transformations to achieve success: methodological strictness which stops time-based data exposure to generate reliable production-ready estimates and conceptual change that shows organizational ability delivers better build results than code-specific attributes-a fundamental change that affects software engineering research directions and industrial operations.
Recommendations: The paper provides the following recommendations for practitioners who want to implement build prediction systems: (1) The system requires complete validation of all model features’ temporal availability to stop production failures caused by data leakage (2) The paper suggests using Random Forest and Gradient Boosting ensemble methods instead of linear models because their better accuracy justifies their additional computational requirements (3) The project success depends most heavily on developer experience and team stability and testing culture development (4) The deployment of unified models should occur across polyglot codebases through strong cross-language generalization. The research community should follow these recommendations: (1) All software engineering prediction studies need to perform systematic temporal availability audits to stop false performance results (2) The 3-type leakage taxonomy (Direct Outcome Encoding, Execution-Dependent Metrics, Future Information Leakage) should become the standard methodology for research (3) Research should expand its scope by studying organizational maturity and team capability and process stability metrics together with traditional code-based features to improve software quality determination knowledge.
Future studies need to develop detailed failure identification systems which separate compilation mistakes from testing problems and system breakdowns to allow specific corrective actions. The combination of causal inference methods with feature importance analysis will generate evidence-based process enhancements. The ability to make predictions about new projects without requiring training data through transfer learning methods should be developed. The combination of strict data quality methods with extensive experimental testing and organizational capability modeling systems will create a new generation of CI/CD pipeline efficiency tools which help organizations forecast and stop build failures and maximize their resource usage.

Author Contributions

Conceptualization, Rangari, A. and Mishra, L.N.; methodology, Rangari, A. and Mishra, L.N.; software, Mishra, L.N. and Nagrare, S; validation, Nagrare, S and Nayak, S.K.; formal analysis, Rangari, A. and Mishra, L.N. and Nagrare, S and Nayak, S.K.; investigation, Rangari, A. and Mishra, L.N. and Nagrare, S and Nayak, S.K.; resources, Nayak, S.K.; data curation, Mishra, L.N; writing-original draft preparation, Rangari, A.; writing-review and editing, Rangari, A. and Mishra, L.N. and Nagrare, S and Nayak, S.K.; visualization, Mishra, L.N. ; supervision, Rangari, A.; project administration, Mishra, L.N. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable for studies not involving humans or animals.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data used in this study are publicly available from the TravisTorrent dataset at https://travistorrent.testroots.org/. Processed data and analysis scripts are available from the corresponding author upon reasonable request.

Acknowledgments

The authors would like to thank the TravisTorrent project maintainers for providing the comprehensive dataset that enabled this research.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
CI/CD Continuous Integration/Continuous Deployment
SDLC Software Development Lifecycle
ROC-AUC Receiver Operating Characteristic - Area Under Curve
ML Machine Learning
RF Random Forest
GB Gradient Boosting
LR Logistic Regression

References

  1. Vassallo, C.; Schermann, G.; Zampetti, F.; Romano, D.; Leitner, P.; Di Penta, M.; Panichella, S. A Tale of CI Build Failures: An Open Source and a Financial Organization Perspective. In Proceedings of the Proceedings of the 33rd IEEE International Conference on Software Maintenance and Evolution (ICSME), Shanghai, China, 2017; pp. 183–193. [CrossRef]
  2. Rausch, T.; Hummer, W.; Leitner, P.; Schulte, S. An Empirical Analysis of Build Failures in the Continuous Integration Workflows of Java-Based Open-Source Software. In Proceedings of the Proceedings of the 14th International Conference on Mining Software Repositories (MSR), Buenos Aires, Argentina, 2017; pp. 345–355. [CrossRef]
  3. Hilton, M.; Tunnell, T.; Huang, K.; Marinov, D.; Dig, D. Usage, Costs, and Benefits of Continuous Integration in Open-Source Projects. IEEE Transactions on Software Engineering 2016, 43, 426–445. [Google Scholar] [CrossRef]
  4. Beller, M.; Gousios, G.; Zaidman, A. Integration. In Proceedings of the Proceedings of the 14th International Conference on Mining Software Repositories (MSR), Buenos Aires, Argentina, 2017; pp. 447–450. [CrossRef]
  5. Saidani, I.; Ouni, A.; Mkaouer, M.W. Improving the Prediction of Continuous Integration Build Failures Using Deep Learning. Automated Software Engineering 2022, 29, 1–41. [Google Scholar] [CrossRef]
  6. Ghotra, B.; McIntosh, S.; Hassan, A.E. Revisiting the Impact of Classification Techniques on the Performance of Defect Prediction Models. In Proceedings of the Proceedings of the 37th IEEE International Conference on Software Engineering (ICSE), Florence, Italy, 2015; pp. 789–800. [CrossRef]
  7. Li, Z.; Jing, X.Y.; Zhu, X. Progress on Approaches to Software Defect Prediction. IET Software 2018, 12, 161–175. [Google Scholar] [CrossRef]
  8. khleel, n.a.a.; nehéz, k. Software Defect Prediction Using a Bidirectional LSTM Network Combined with Oversampling Techniques. Cluster Computing 2023, 27, 3615–3638. [Google Scholar] [CrossRef]
  9. Chung, J.; Gulcehre, C.; Cho, K.; Bengio, Y. Empirical Evaluation of Gated Recurrent Neural Networks on Sequence Modeling. arXiv preprint arXiv:1412.3555, 2014; arXiv:1412.3555 2014.. NIPS 2014 Deep Learning Workshop. [Google Scholar]
  10. Kamei, Y.; Shihab, E.; Adams, B.; Hassan, A.E.; Mockus, A.; Sinha, A.; Ubayashi, N. A Large-Scale Empirical Study of Just-in-Time Quality Assurance. IEEE Transactions on Software Engineering 2013, 39, 757–773. [Google Scholar] [CrossRef]
  11. Zhao, Y.; Damevski, K.; Chen, H. A Systematic Survey of Just-in-Time Software Defect Prediction. ACM Computing Surveys 2023, 55, 1–35. [Google Scholar] [CrossRef]
  12. stradowski, s.; madeyski, l. Industrial Applications of Software Defect Prediction Using Machine Learning: A Business-Driven Systematic Literature Review. Information and Software Technology 2023, 159, 107192. [Google Scholar] [CrossRef]
  13. Ni, C.; Liu, W.S.; Chen, X.; Gu, Q.; Chen, D.X.; Huang, Q.G. A Cluster Based Feature Selection Method for Cross-Project Software Defect Prediction. Journal of Computer Science and Technology 2017, 32, 1090–1107. [Google Scholar] [CrossRef]
  14. Harman, M.; Mansouri, S.A.; Zhang, Y. Search-based software engineering. ACM Computing Surveys 2012, 45, 11–1. [Google Scholar] [CrossRef]
  15. Zhang, D.; Han, S.; Dang, Y.; Lou, J.G.; Zhang, H.; Xie, T. Software Analytics in Practice. In Proceedings of the IEEE Software; 2013; Vol. 30, pp. 30–37. [Google Scholar] [CrossRef]
  16. wang, s.; huang, l.; gao, a.; ge, j.; zhang, t.; feng, h.; satyarth, i.; li, m.; zhang, h.; ng, v. Machine/Deep Learning for Software Engineering: A Systematic Literature Review. IEEE Transactions on Software Engineering 2023, 49, 1630–1652. [Google Scholar] [CrossRef]
  17. ortu, m.; destefanis, g.; hall, t.; bowes, d. Fault-insertion and fault-fixing behavioural patterns in Apache Software Foundation Projects. Information and Software Technology 2023, 158, 107187. [Google Scholar] [CrossRef]
  18. Zhang, J.M.; Harman, M.; Ma, L.; Liu, Y. Machine Learning Testing: Survey, Landscapes and Horizons. IEEE Transactions on Software Engineering 2022, 48, 1–36. [Google Scholar] [CrossRef]
  19. Grattan, N.; da Costa, D.A.; Stanger, N. The Need for More Informative Defect Prediction: A Systematic Literature Review. Information and Software Technology 2024, 171, 107456. [Google Scholar] [CrossRef]
  20. Liu, Y.; Fekete, A.; Gorton, I. Design-Level Performance Prediction of Component-Based Applications. IEEE Transactions on Software Engineering 2005, 31, 928–941. [Google Scholar] [CrossRef]
  21. Gong, J.; Chen, T. Deep Configuration Performance Learning: A Systematic Survey and Taxonomy. ACM Transactions on Software Engineering and Methodology 2024, 34, 25–1. [Google Scholar] [CrossRef]
  22. Ali, M.; Mazhar, T.; Al-Rasheed, A.; Shahzad, T.; Ghadi, Y.Y.; Khan, M.A. Enhancing Software Defect Prediction: A Framework with Improved Feature Selection and Ensemble Machine Learning. PeerJ Computer Science 2024, 10, e1860. [Google Scholar] [CrossRef]
  23. khleel, n.a.a.; nehéz, k.; fadulalla, m.; hisaen, a. Ensemble-Based Machine Learning Algorithms Combined with Near Miss Method for Software Bug Prediction. International Journal of Networked and Distributed Computing 2025, 13, 1–18. [Google Scholar] [CrossRef]
  24. Sagi, O.; Rokach, L. Ensemble Learning: A Survey. WIREs Data Mining and Knowledge Discovery 2018, 8, e1249. [Google Scholar] [CrossRef]
  25. He, H.; Garcia, E.A. Learning from Imbalanced Data. IEEE Transactions on Knowledge and Data Engineering 2009, 21, 1263–1284. [Google Scholar] [CrossRef]
  26. Ibraigheeth, M.A.; Abu Eid, A.I.; Alsariera, Y.A.; Awwad, W.F.; Nawaz, M. A New Weighted Ensemble Model to Improve the Performance of Software Project Failure Prediction. International Journal of Advanced Computer Science and Applications 2024, 15, 331–340. [Google Scholar] [CrossRef]
  27. Kaushik, A.; Sheoran, K.; Kapur, R.; Singh, D.P. SENSE: Software Effort Estimation Using Novel Stacking Ensemble Learning. Innovations in Systems and Software Engineering 2024, 21, 769–785. [Google Scholar] [CrossRef]
  28. Saraireh, J.; Agoyi, M.; Kassaymeh, S. Adaptive Ensemble Learning Model-Based Binary White Shark Optimizer for Software Defect Classification. International Journal of Computational Intelligence Systems 2025, 18, 14. [Google Scholar] [CrossRef]
  29. Granitto, P.M.; Furlanello, C.; Biasioli, F.; Gasperi, F. Recursive Feature Elimination with Random Forest for PTR-MS Analysis of Agroindustrial Products. Chemometrics and Intelligent Laboratory Systems 2006, 83, 83–90. [Google Scholar] [CrossRef]
  30. Cai, J.; Luo, J.; Wang, S.; Yang, S. Feature Selection in Machine Learning: A New Perspective. Neurocomputing 2018, 300, 70–79. [Google Scholar] [CrossRef]
  31. hartanto, a.d.; kholik, y.n.; pristyanto, y. Stock Price Time Series Data Forecasting Using the Light Gradient Boosting Machine (LightGBM) Model. Journal of Information Visualization 2023, 7, 456–470. [Google Scholar] [CrossRef]
  32. He, X.; Zhao, K.; Chu, X. AutoML: A Survey of the State-of-the-Art. Knowledge-Based Systems 2021, 212, 106622. [Google Scholar] [CrossRef]
  33. Lim, B.; Arik, S.O.; Loeff, N.; Pfister, T. Temporal Fusion Transformers for Interpretable Multi-Horizon Time Series Forecasting. International Journal of Forecasting 2021, 37, 1748–1764. [Google Scholar] [CrossRef]
  34. lee, s.; hong, j.; liu, l.; choi, w. TS-Fastformer: Fast Transformer for Time-Series Forecasting. ACM Transactions on Intelligent Systems and Technology 2024, 15, 45. [Google Scholar] [CrossRef]
  35. chen, y.; hao, j.; peng, y.; xia, h. Transformer-based performance prediction and proactive resource allocation for cloud-native microservices. Cluster Computing 2025, 28, 1–18. [Google Scholar] [CrossRef]
  36. tao, h.; fu, l.; cao, q.; niu, x.; chen, h.; shang, s.; xian, y. Cross-Project Defect Prediction Using Transfer Learning with Long Short-Term Memory Networks. IET Software 2024, 18, 234–248. [Google Scholar] [CrossRef]
  37. Zhao, G.; Georgiou, S.; Zou, Y.; Hassan, S.; Truong, D.; Corbin, T. Enhancing Performance Bug Prediction Using Performance Code Metrics. In Proceedings of the Proceedings of the 21st International Conference on Mining Software Repositories (MSR), Lisbon, Portugal, 2024. [CrossRef]
  38. tao, h.; fu, l.; cao, q.; niu, x.; chen, h.; shang, s.; xian, y. Cross-Project Defect Prediction Using Transfer Learning with Long Short-Term Memory Networks. IET Software 2024, 18, 456–470. [Google Scholar] [CrossRef]
  39. lawson, a. 2024 State of Tech Talent Report: Survey-based Insights into the Current State of Technical Talent Acquisition, Retention, and Management Globally. Technical report, Google Cloud and DevOps Research and Assessment, 2024. Accessed: 2025-01-14. [CrossRef]
  40. perera, p.; silva, r.; perera, i. Improve software quality through practicing DevOps 2017. [CrossRef]
  41. Weeraddana, N.; Alfadel, M.; McIntosh, S. Characterizing Timeout Builds in Continuous Integration. IEEE Transactions on Software Engineering 2024, 50, 2045–2062. [Google Scholar] [CrossRef]
  42. Zampetti, F.; Vassallo, C.; Panichella, S.; Canfora, G.; Gall, H.; Di Penta, M. An Empirical Characterization of Bad Practices in Continuous Integration. Empirical Software Engineering 2020, 25, 1095–1135. [Google Scholar] [CrossRef]
  43. Gallaba, K.; Macho, C.; Pinzger, M.; McIntosh, S. Noise and Heterogeneity in Historical Build Data: An Empirical Study of Travis CI. In Proceedings of the Proceedings of the 33rd IEEE/ACM International Conference on Automated Software Engineering (ASE), Montpellier, France, 2018; pp. 87–97. [CrossRef]
  44. sun, g.; habchi, s.; mcintosh, s. RavenBuild: Context, Relevance, and Dependency Aware Build Outcome Prediction. Proceedings of the ACM on Software Engineering 2024, 1, 996–1018. [Google Scholar] [CrossRef]
  45. abdellatif, a.; badran, k.; shihab, e. MSRBot: Using bots to answer questions from software repositories. Empirical Software Engineering 2020, 25, 2157–2201. [Google Scholar] [CrossRef]
  46. Luo, Q.; Moran, K.; Zhang, L.; Poshyvanyk, D. How Do Static and Dynamic Test Case Prioritization Techniques Perform on Modern Software Systems? An Extensive Study on GitHub Projects. IEEE Transactions on Software Engineering 2019, 46, 1054–1080. [Google Scholar] [CrossRef]
  47. Fallahzadeh, E.; Rigby, P.C.; Adams, B. Contrasting Test Selection, Prioritization, and Batch Testing at Scale. Empirical Software Engineering 2024, 30. [Google Scholar] [CrossRef]
  48. Gruber, M.; Roslan, M.F.; Parry, O.; Scharnböck, F.; McMinn, P.; Fraser, G. Do Automatic Test Generation Tools Generate Flaky Tests? In Proceedings of the 2024 IEEE/ACM 46th International Conference on Software Engineering (ICSE ’24). ACM, April 2024; pp. 14–20. [Google Scholar] [CrossRef]
  49. Silva, D.; Gruber, M.; Gokhale, S.; Arteca, E.; Turcotte, A.; d’Amorim, M.; Lam, W.; Winter, S.; Bell, J. The Effects of Computational Resources on Flaky Tests. IEEE Transactions on Software Engineering 2024, 50. [Google Scholar] [CrossRef]
  50. Bernardo, J.H.; da Costa, D.A.; de Medeiros, S.Q.; Kulesza, U. How do Machine Learning Projects use Continuous Integration Practices? In An Empirical Study on GitHub Actions. In Proceedings of the 21st IEEE/ACM International Conference on Mining Software Repositories (MSR), Lisbon, Portugal, April 2024; pp. 665–676. [Google Scholar] [CrossRef]
  51. Bouzenia, I.; Pradel, M. Resource Usage and Optimization Opportunities in Workflows of GitHub Actions. In Proceedings of the Proceedings of the 46th IEEE/ACM International Conference on Software Engineering (ICSE), Lisbon, Portugal, April 2024; pp. 25:1–25:12. [CrossRef]
  52. Zhang, Y.; Wu, Y.; Chen, T.; Wang, T.; Liu, H.; Wang, H. How do Developers Talk about GitHub Actions? Evidence from Online Software Development Community. In Proceedings of the Proceedings of the 46th IEEE/ACM International Conference on Software Engineering (ICSE), New York, NY, USA, 2024; pp. 1–13. ICSE ’24. [CrossRef]
  53. Kinsman, T.; Wessel, M.; Gerosa, M.A.; Treude, C. How Do Software Developers Use GitHub Actions to Automate Their Workflows? In Proceedings of the Proceedings of the 18th International Conference on Mining Software Repositories (MSR), Madrid, Spain; 2021; pp. 420–431. [Google Scholar] [CrossRef]
  54. zerouali, a.; mens, t.; robles, g.; gonzalez barahona, j.m. On the Diversity of Software Package Popularity Metrics: An Empirical Study of npm. IEEE Transactions on Software Engineering 2019, 49, 2113–2135. [Google Scholar] [CrossRef]
  55. Mastropaolo, A.; Zampetti, F.; Bavota, G.; Di Penta, M. Toward Automatically Completing GitHub Workflows. In Proceedings of the Proceedings of the IEEE/ACM 46th International Conference on Software Engineering (ICSE), New York, NY, USA, 2024. ICSE ’24.
  56. saito, s. Understanding Key Business Processes for Business Process Outsourcing Transition. In Proceedings of the Proceedings of the IEEE 14th International Conference on Global Software Engineering (ICGSE), 2019, pp. pp. 66–75. [CrossRef]
  57. widder, d.g.; hilton, m.; kästner, c.; vasilescu, b. A Conceptual Replication of Continuous Integration Pain Points in the Context of Travis CI. In Proceedings of the Proceedings of the 27th ACM Joint European Software Engineering Conference and Symposium on the Foundations of Software Engineering (ESEC/FSE), 2019, pp. pp. 647–658. [CrossRef]
  58. Morris, K. Infrastructure as Code: Managing Servers in the Cloud, 2nd ed.; O’Reilly Media, 2020.
  59. Humble, J.; Farley, D. Continuous Delivery: Reliable Software Releases Through Build, Test, and Deployment Automation; Addison-Wesley Professional, 2010.
  60. fontana, r.m.; meyer, v.; reinehr, s.; malucelli, a. Management Ambidexterity: A Clue for Maturing in Agile Software Development. In Proceedings of the Proceedings of the Agile Processes in Software Engineering and Extreme Programming Conference (XP)., Springer, 2015; pp. 212–217. [Google Scholar] [CrossRef]
  61. Adams, B.; McIntosh, S. Modern Release Engineering in a Nutshell – Why Researchers Should Care. In Proceedings of the Proceedings of the IEEE 23rd International Conference on Software Analysis, Evolution, and Reengineering (SANER), 2016; Vol. 5, pp. 78–90. [CrossRef]
  62. oliveira, e.; fernandes, e.; steinmacher, i.; cristo, m.; conte, t.; garcia, a. Code and commit metrics of developer productivity: a study on team leaders perceptions. Empirical Software Engineering 2020, 25, 3874–3912. [Google Scholar] [CrossRef]
  63. Hassan, A.E. Predicting Faults Using the Complexity of Code Changes. In Proceedings of the Proceedings of the 31st IEEE International Conference on Software Engineering (ICSE), Vancouver, Canada, 2009; pp. 78–88. [CrossRef]
  64. mccabe, t. A Complexity Measure. IEEE Transactions on Software Engineering 1976, SE-2, 308–320. [Google Scholar] [CrossRef]
  65. Halstead, M.H. Elements of Software Science; Elsevier, 1977. Classic software metrics foundation.
  66. Nagappan, N.; Ball, T.; Zeller, A. Mining Metrics to Predict Component Failures. In Proceedings of the Proceedings of the 28th International Conference on Software Engineering (ICSE), 2006, pp.; pp. 452–461. [CrossRef]
  67. Fenton, N.E.; Bieman, J. Software Metrics: A Rigorous and Practical Approach, 3rd ed.; CRC Press, 2014.
  68. Li, J.; Ahmed, I. Commit Message Matters: Investigating Impact and Evolution of Commit Message Quality. In Proceedings of the Proceedings of the 45th IEEE/ACM International Conference on Software Engineering (ICSE), Melbourne, Australia, 2023; pp. 806–817. [CrossRef]
  69. Avgeriou, P.; Kruchten, P.; Ozkaya, I.; Seaman, C. Managing Technical Debt in Software Engineering. Dagstuhl Reports 2016, 6, 110–138. [Google Scholar] [CrossRef]
  70. Azeem, M.I.; Palomba, F.; Shi, L.; Wang, Q. Machine Learning Techniques for Code Smell Detection: A Systematic Literature Review and Meta-Analysis. Information and Software Technology 2019, 108, 115–138. [Google Scholar] [CrossRef]
  71. Li, Z.; Avgeriou, P.; Liang, P. A Systematic Mapping Study on Technical Debt and Its Management. Journal of Systems and Software 2015, 101, 193–220. [Google Scholar] [CrossRef]
  72. Bacchelli, A.; Bird, C. Review. In Proceedings of the Proceedings of the 2013 International Conference on Software Engineering, 2013, pp. pp. 712–721. [CrossRef]
  73. Wang, S.; Lo, D. Version History, Similar Report, and Structure: Putting Them Together for Improved Bug Localization. In Proceedings of the Proceedings of the 22nd International Conference on Program Comprehension (ICPC), 2014, pp. pp. 53–63. [CrossRef]
  74. Forsgren, N.; Storey, M.A.; Maddila, C.; Zimmermann, T.; Houck, B.; Butler, J. The SPACE of developer productivity. Communications of the ACM 2021, 64, 67–75. [Google Scholar] [CrossRef]
  75. Kaufman, S.; Rosset, S.; Perlich, C.; Stitelman, O. Leakage in data mining. ACM Transactions on Knowledge Discovery from Data 2012, 6, 15. [Google Scholar] [CrossRef]
  76. Cerqueira, V.; Torgo, L.; Mozetič, I. Evaluating Time Series Forecasting Models: An Empirical Study on Performance Estimation Methods. Machine Learning 2020, 109, 1997–2028. [Google Scholar] [CrossRef]
  77. bergmeir, c.; benítez, j.m. On the Use of Cross-Validation for Time Series Predictor Evaluation. Information Sciences 2012, 191, 192–213. [Google Scholar] [CrossRef]
  78. Emmanuel, T.; Maupong, T.; Mpoeleng, D.; Semong, T.; Mphago, B.; Tabona, O. A Survey on Missing Data in Machine Learning. Journal of Big Data 2021, 8, 140. [Google Scholar] [CrossRef] [PubMed]
  79. Patro, S.G.K.; Sahu, K.K. Normalization: A Preprocessing Stage. arXiv preprint arXiv:1503.06462, 2015; arXiv:1503.06462 2015.IARJSET International Conference. [Google Scholar] [CrossRef]
  80. kitchenham, b.; pfleeger, s.; pickard, l.; jones, p.; hoaglin, d.; emam, k.e.; rosenberg, j. Preliminary Guidelines for Empirical Research in Software Engineering. IEEE Transactions on Software Engineering 2002, 28, 721–734. [Google Scholar] [CrossRef]
  81. Baker, M. 1,500 Scientists Lift the Lid on Reproducibility. Nature 2016, 533, 452–454. [Google Scholar] [CrossRef]
  82. rainio, o.; teuho, j.; klén, r. Evaluation Metrics and Statistical Tests for Machine Learning. Scientific Reports 2024, 14, 6756. [Google Scholar] [CrossRef] [PubMed]
  83. McHugh, M.L. Interrater Reliability: The Kappa Statistic. Biochemia Medica 2012, 22, 276–282. [Google Scholar] [CrossRef]
  84. Hyndman, R.J.; Koehler, A.B. Another Look at Measures of Forecast Accuracy. International Journal of Forecasting 2006, 22, 679–688. [Google Scholar] [CrossRef]
  85. rey, d.; neuhäuser, m. Wilcoxon-Signed-Rank Test. International Encyclopedia of Statistical Science, 1658; 1658–1659. [Google Scholar] [CrossRef]
  86. rajput, d.; wang, w.j.; chen, c.c. Evaluation of a Decided Sample Size in Machine Learning Applications. BMC Bioinformatics 2023, 24, 48. [Google Scholar] [CrossRef]
  87. Sullivan, G.M.; Feinn, R. Using Effect Size—or Why the P Value Is Not Enough. Journal of Graduate Medical Education 2012, 4, 279–282. [Google Scholar] [CrossRef]
  88. Saito, T.; Rehmsmeier, M. The Precision-Recall Plot Is More Informative than the ROC Plot When Evaluating Binary Classifiers on Imbalanced Datasets. PLOS ONE 2015, 10, e0118432. [Google Scholar] [CrossRef]
  89. Hand, D.J.; Till, R.J. A Simple Generalisation of the Area Under the ROC Curve for Multiple Class Classification Problems. Machine Learning 2001, 45, 171–186. [Google Scholar] [CrossRef]
  90. Arlot, S.; Celisse, A. A Survey of Cross-Validation Procedures for Model Selection. Statistics Surveys 2010, 4, 40–79. [Google Scholar] [CrossRef]
  91. Wu, H.; Xu, J.; Wang, J.; Long, M. Forecasting. In Proceedings of the Proceedings of the Advances in Neural Information Processing Systems (NeurIPS), 2021; Vol. 34, pp. 22419–22430.
  92. Hassan, A.E.; Xie, T. Software intelligence. In Proceedings of the Proceedings of the FSE/SDP Workshop on Future of Software Engineering Research, 2010; pp. 161–166. [CrossRef]
  93. Gousios, G.; Pinzger, M.; Deursen, A.v. An Exploratory Study of the Pull-Based Software Development Model. In Proceedings of the Proceedings of the 36th International Conference on Software Engineering, 2014; pp. 345–355. [CrossRef]
  94. Feurer, M.; Hutter, F. Hyperparameter Optimization. Automated Machine Learning, 2019; 3–33. [Google Scholar] [CrossRef]
  95. santana, f.a.; cordeiro, a.f.r.; oliveirajr, e. Dublin Core for Recording Metadata of Experiments in Software Engineering: A Survey. arXiv preprint arXiv:2303.16989, 2023; arXiv:2303.16989 2023.Under review. [Google Scholar] [CrossRef]
  96. Lehnert, S. A Review of Software Change Impact Analysis. Technical Report, Ilmenau University of Technology, TU Ilmenau: Germany, 2011. [Google Scholar]
  97. Jørgensen, M.; Shepperd, M. A Systematic Review of Software Development Cost Estimation Studies. IEEE Transactions on Software Engineering 2007, 33, 33–53. [Google Scholar] [CrossRef]
  98. qi, x.; chen, j.; deng, l. Learning. In Proceedings of the Proceedings of the International Conference on Algorithms and Architectures for Parallel Processing (ICA3PP). Springer, 2023; Vol. 13777, pp. 123–137. [Google Scholar] [CrossRef]
  99. Das, A.; Kong, W.; Leach, A.; Mathur, S.; Sen, R.; Yu, R. A Decoder-Only Foundation Model for Time-Series Forecasting. In Proceedings of the Proceedings of the International Conference on Machine Learning (ICML), Vienna, Austria, 2024; pp. 567–589, Google Research.
  100. Feng, Z.; Guo, D.; Tang, D.; Duan, N.; Feng, X.; Gong, M.; Shou, L.; Qin, B.; Liu, T.; Jiang, D.; et al. Languages. In Proceedings of the Proceedings of the Conference on Empirical Methods in Natural Language Processing (EMNLP), Online, 2020; pp. 1536–1547. [CrossRef]
  101. Zhang, C.; Xie, Y.; Bai, H.; Yu, B.; Li, W.; Gao, Y. A Survey on Federated Learning. Knowledge-Based Systems 2021, 216, 106775. [Google Scholar] [CrossRef]
Figure 1. End-to-end CI/CD build prediction system architecture. TravisTorrent dataset undergoes data leakage prevention (66→31 features via 3-type taxonomy), preprocessing (missing value imputation, scaling, encoding), and temporal train/test split (80/20, pre-Dec 2016 / Dec 2016-2017). Three models (Logistic Regression, Gradient Boosting, Random Forest) are evaluated using statistical significance testing (Wilcoxon, 95% CI, Cohen’s d). Random Forest achieves best performance (85.24% accuracy), correctly predicting 93.6% of successful builds and 66.2% of failures, enabling production deployment.
Figure 1. End-to-end CI/CD build prediction system architecture. TravisTorrent dataset undergoes data leakage prevention (66→31 features via 3-type taxonomy), preprocessing (missing value imputation, scaling, encoding), and temporal train/test split (80/20, pre-Dec 2016 / Dec 2016-2017). Three models (Logistic Regression, Gradient Boosting, Random Forest) are evaluated using statistical significance testing (Wilcoxon, 95% CI, Cohen’s d). Random Forest achieves best performance (85.24% accuracy), correctly predicting 93.6% of successful builds and 66.2% of failures, enabling production deployment.
Preprints 182438 g001
Figure 2. Data leakage prevention methodology showing 3-type taxonomy for identifying outcome-dependent features and systematic filtering process. Original 66 TravisTorrent features undergo temporal availability auditing, correlation analysis, and temporal validation to retain 31 clean pre-build features. Validation demonstrates 12.6 percentage point accuracy drop (97.8% → 85.24%) from removing leaky features, confirming production-viable performance with clean feature set.
Figure 2. Data leakage prevention methodology showing 3-type taxonomy for identifying outcome-dependent features and systematic filtering process. Original 66 TravisTorrent features undergo temporal availability auditing, correlation analysis, and temporal validation to retain 31 clean pre-build features. Validation demonstrates 12.6 percentage point accuracy drop (97.8% → 85.24%) from removing leaky features, confirming production-viable performance with clean feature set.
Preprints 182438 g002
Figure 3. Receiver Operating Characteristic curves comparing three classification models on TravisTorrent test set (20,000 builds). Random Forest (ROC-AUC = 91.38%) demonstrates superior discrimination capability compared to Gradient Boosting (ROC-AUC = 88.59%) and Logistic Regression (ROC-AUC = 61.91%), substantially exceeding random guessing baseline (50.00%).
Figure 3. Receiver Operating Characteristic curves comparing three classification models on TravisTorrent test set (20,000 builds). Random Forest (ROC-AUC = 91.38%) demonstrates superior discrimination capability compared to Gradient Boosting (ROC-AUC = 88.59%) and Logistic Regression (ROC-AUC = 61.91%), substantially exceeding random guessing baseline (50.00%).
Preprints 182438 g003
Figure 4. Top 15 most important pre-build features for build prediction ranked by Random Forest feature importance scores. Project Context features (blue bars) dominate with 49.8% cumulative importance, substantially exceeding Code Metrics (7.7%), demonstrating that project maturity and history predict build outcomes more reliably than immediate code characteristics.
Figure 4. Top 15 most important pre-build features for build prediction ranked by Random Forest feature importance scores. Project Context features (blue bars) dominate with 49.8% cumulative importance, substantially exceeding Code Metrics (7.7%), demonstrating that project maturity and history predict build outcomes more reliably than immediate code characteristics.
Preprints 182438 g004
Figure 5. Cross-language build prediction performance across Java, Ruby, Python, and JavaScript using Random Forest model. Accuracy varies by only 3.38 percentage points (std dev: 1.38%), demonstrating strong generalization across programming ecosystems without language-specific retraining. All four metrics (Accuracy, Precision, Recall, F1) show consistent performance, validating that project maturity and test structure provide language-agnostic predictive signals.
Figure 5. Cross-language build prediction performance across Java, Ruby, Python, and JavaScript using Random Forest model. Accuracy varies by only 3.38 percentage points (std dev: 1.38%), demonstrating strong generalization across programming ecosystems without language-specific retraining. All four metrics (Accuracy, Precision, Recall, F1) show consistent performance, validating that project maturity and test structure provide language-agnostic predictive signals.
Preprints 182438 g005
Table 1. Model Performance on TravisTorrent Test Set (20,000 Builds).
Table 1. Model Performance on TravisTorrent Test Set (20,000 Builds).
Model Accuracy 95% CI Precision Recall F1 ROC-AUC
% (%) (%) (%) (%) (%) (%)
Logistic Regression 61.55 [60.88, 62.22] 62.64 90.20 73.94 61.91
Gradient Boosting 81.34 [80.72, 81.96] 80.31 91.59 85.58 88.59
Random Forest 85.24 [84.67, 85.79] 86.34 93.55 89.80 91.38
Majority Class 69.50 - - 100.00 82.00 50.00
Table 2. Confusion Matrix for Random Forest on Test Set (20,000 Builds).
Table 2. Confusion Matrix for Random Forest on Test Set (20,000 Builds).
Predicted Class
Actual Class Success (1) Failure (0) Total
Success (1) 13,004 (TP) 896 (FN) 13,900
Failure (0) 2,057 (FP) 4,043 (TN) 6,100
Total 15,061 4,939 20,000
Table 3. Top 10 Most Important Pre-Build Features for Build Prediction
Table 3. Top 10 Most Important Pre-Build Features for Build Prediction
Rank Feature Name Importance Category
1 gh_project_maturity_days 0.0949 Project Context
2 git_repository_age_days 0.0946 Project Context
3 gh_commits_count 0.0902 Project Context
4 gh_total_commits 0.0863 Project Context
5 gh_sloc 0.0766 Code Metrics
6 gh_contributors_count 0.0654 Project Context
7 tr_build_number 0.0612 Build History
8 gh_test_density 0.0587 Test Structure
9 gh_tests_count 0.0543 Test Structure
10 tr_builds_last_30_days 0.0498 Build History
Top 10 Cumulative 73.2%
Table 4. Cross-Language Build Prediction Performance (Random Forest).
Table 4. Cross-Language Build Prediction Performance (Random Forest).
Language Accuracy Precision Recall F1 Builds
(%) (%) (%) (%) (%)
Java 84.21 82.45 94.12 87.90 8,000
Ruby 82.14 81.02 93.01 86.61 7,000
Python 81.54 79.98 93.85 86.39 3,000
JavaScript 80.83 79.23 92.98 85.56 2,000
Overall 82.74 80.88 93.57 86.76 20,000
Std Dev 1.38 1.32 0.49 0.98 -
Range 3.38 3.22 1.14 2.34 -
Table 5. Impact of Data Leakage Prevention on Random Forest Performance.
Table 5. Impact of Data Leakage Prevention on Random Forest Performance.
Feature Set Accuracy ROC-AUC F1 Production
(%) (%) (%) (%) Viable
All 66 Features (with leakage) 97.80 99.56 98.42 No
31 Clean Pre-Build Features 82.73 91.38 86.76 Yes
Performance Drop -15.07 -8.18 -11.66
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.
Copyright: This open access article is published under a Creative Commons CC BY 4.0 license, which permit the free download, distribution, and reuse, provided that the author and preprint are cited in any reuse.
Prerpints.org logo

Preprints.org is a free preprint server supported by MDPI in Basel, Switzerland.

Subscribe

Disclaimer

Terms of Use

Privacy Policy

Privacy Settings

© 2025 MDPI (Basel, Switzerland) unless otherwise stated