Submitted:
12 August 2025
Posted:
13 August 2025
You are already at the latest version
Abstract
Keywords:
1. Introduction
- H1 – Discrimination. LightGBM attains ROC-AUC ≥ 0.90 with a lower 95% CI bound ≥ 0.85.
-
H2 – Comparative performance. LightGBM outperforms or matches:
- (a)
- L2-regularized Logistic Regression;
- (b)
- Random Forest;
- (c)
- XGBoost,
with ΔAUC ≥ 0.02 on average, or differences statistically indistinguishable (paired bootstrap one-sided p < 0.05), while retaining superior calibration (H3).
- 3.
-
H3 – Calibration. With monotone probability calibration (Platt or isotonic chosen in inner CV), the model achieves:
- (a)
- Brier ≤ 0.12;
- (b)
- calibration slope in [0.9, 1.1];
- (c)
- intercept in [−0.05, 0.05];
- (d)
- ECE ≤ 0.05
across outer folds (pass if ≥ 4/5 folds).
- 4.
-
H4 – Operational thresholds. Using calibrated probabilities, the two operating points meet:
- (a)
- Screening: Recall ≥ 0.90 with Precision ≥ 0.70;
- (b)
- Shortlisting: F1 ≥ 0.88.
- 5.
- H5 – Imbalance robustness. Under a 30/70 imbalance scenario, PR-AUC ≥ 0.85 and the top 5 SHAP features preserve rank order up to Kendall’s τ ≥ 0.70 vs. the balanced setting.
- 6.
- H6 – Stability of explanations. Global SHAP importance is stable across folds (median Spearman ρ ≥ 0.80 for the top 8 features), consistent with Permutation Importance (ρ ≥ 0.70). All ALE shows domain-consistent directionality for VO₂max ↑, Max Strength ↑, Decision Latency ↓, Reaction Time ↓.
- 7.
- H7 – Distributional validity. Kolmogorov–Smirnov tests against empirical references (variable-appropriate transforms; Holm correction) show no rejections at α = 0.05; otherwise, generation parameters are revised prior to training.
1.1. Literature Review
2. Materials and Methods
2.1. Study Design, Rationale, and Variable Selection
2.2. Synthetic Dataset Generation, Validation and Labeling
2.3. Predictive Modeling, Optimization, and Evaluation
- 1)
- Logistic Regression (L2) with class-balanced weighting;
- 2)
- Random Forest;
- 3)
- XGBoost.
- Primary selection metric: Brier score (proper scoring rule for probabilistic predictions); ROC-AUC reported for discrimination; F1 used only as a tie-breaker for threshold metrics.
- Search budget: 100 sampled configurations per model (inner CV), with stratified folds.
- Early stopping: enabled for gradient-boosted models using inner-fold validation splits.
- Class balance: folds were stratified by HP/LP to preserve prevalence.
- Leakage control: all preprocessing (imputation, scaling/class-weights, and calibration selection: Platt vs. isotonic by Brier) was performed inside the training portion of each inner/outer fold; the test fold remained untouched.
- Logistic Regression (LBFGS, L2). C ∈ [1e−4, 1e+3] (log-uniform); max_iter = 2000.
- Random Forest. n_estimators ∈ [200, 800]; max_depth ∈ {None, 3–20}; min_samples_leaf ∈ [1, 10];
- max_features ∈ {‘sqrt’, ‘log2’, 0.3–1.0}; bootstrap = True.
- XGBoost. n_estimators ∈ [200, 800]; learning_rate ∈ [1e-3, 0.1] (log-uniform); max_depth ∈ [2, 8];
- subsample ∈ [0.6, 1.0]; colsample_bytree ∈ [0.6, 1.0]; min_child_weight ∈ [1, 10]; gamma ∈ [0, 5];
- reg_alpha ∈ [0, 5]; reg_lambda ∈ [0, 5].
- LightGBM. num_leaves ∈ [15, 255]; learning_rate ∈ [1e-3, 0.1] (log-uniform); feature_fraction ∈ [0.6, 1.0]; bagging_fraction ∈ [0.6, 1.0]; bagging_freq ∈ [0, 10]; min_child_samples ∈ [10, 100]; lambda_l1 ∈ [0, 5]; lambda_l2 ∈ [0, 5].
- Discrimination: ROC-AUC (primary), PR-AUC;
- Calibration: Brier score, Expected Calibration Error (ECE), calibration slope and intercept;
- Classification metrics: accuracy, precision, recall, F1 (at selected thresholds; see below).
- Screening — prioritize recall, constraining precision ≥ 0.70 to minimize missed HP athletes;
- Shortlisting — maximize F1 to balance precision and recall for final selections.
2.4. Feature Importance, Interpretability, and Technical Implementation
2.5. Statistical Analyses (Group Comparisons)
3. Results
3.1. Predictive Model Performance
3.2. Feature Importance and SHAP Analysis
3.3. Comparative Analysis of High- and Low-Performance Groups
3.4. Hypotheses—Linkage to Results (H1–H7)
4. Discussions
- Inform selection and recruitment processes by objectively identifying talent with high potential.
- Develop personalized training interventions targeted at improving specific performance attributes identified by the model, such as aerobic capacity, reaction time, or decision-making abilities.
- Enhance injury prevention strategies through predictive insights into athletes' physiological and biomechanical vulnerabilities.
- Prospective data collection involving physiological, biomechanical, and cognitive assessments from actual team sport athletes.
- Validation of model predictions against real-world performance outcomes, such as match statistics, competition results, or progression metrics.
- Comparative analysis of predictive accuracy between synthetic and empirical data-driven models to quantify differences and improve the robustness of predictions.
5. Conclusions
Supplementary Materials
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Ethics Statement
Acknowledgments
Conflicts of Interest
References
- Lundberg, S.M.; Lee, S.I. A Unified Approach to Interpreting Model Predictions. Adv. Neural Inf. Process. Syst. 2017, 30, 4765–4774. [Google Scholar] [CrossRef]
- Dindorf, C.; Bartaguiz, E.; Gassmann, F.; Fröhlich, M. Conceptual Structure and Current Trends in Artificial Intelligence, Machine Learning, and Deep Learning Research in Sports: A Bibliometric Review. Int. J. Environ. Res. Public Health 2023, 20. [Google Scholar] [CrossRef] [PubMed]
- Pietraszewski, P.; Terbalyan, A.; Litwiniuk, A.; Zając, A.; Król, P.; Wnorowski, K.; Sadowski, J. The Role of Artificial Intelligence in Sports Analytics: A Systematic Review and Meta-Analysis of Performance Trends. Appl. Sci. 2025, 15. [Google Scholar] [CrossRef]
- Puce, L.; Bragazzi, N.L.; Currà, A.; Trompetto, C. Harnessing Generative Artificial Intelligence for Exercise and Training Prescription: Applications and Implications in Sports and Physical Activity—A Systematic Literature Review. Appl. Sci. 2025, 15. [Google Scholar] [CrossRef]
- Musat, C.L.; Mereuta, C.; Nechita, A.; Tutunaru, D.; Voipan, A.E.; Voipan, D.; Mereuta, E.; Gurau, T.V.; Gurău, G.; Nechita, L.C. Diagnostic Applications of AI in Sports: A Comprehensive Review of Injury Risk Prediction Methods. Diagnostics 2024, 14. [Google Scholar] [CrossRef]
- Amat, S.; Busquier, S.; Gómez-Carmona, C. D.; Gómez-López, M.; Pino-Ortega, J. Algorithm-Based Real-Time Analysis of Heart Rate Measures in HIIT Training: An Automated Approach. Appl. Sci. 2025, 15. [Google Scholar] [CrossRef]
- Carrillo, A.E.; Dinas, P.C.; Gkiata, P.; Ferri, A.R.; Kenny, G.P.; Koutedakis, Y.; Jamurtas, A.Z.; Metsios, G.S.; Flouris, A.D. An Exploratory Investigation of Heart Rate Variability in Response to Exercise Training and Detraining in Young and Middle-Aged Men. Biology 2025, 14. [Google Scholar] [CrossRef]
- Lee, H.A.; Yu, W.; Choi, J.D.; Lee, Y.-s.; Park, J.W.; Jung, Y.J.; Sheen, S.S.; Jung, J.; Haam, S.; Kim, S.H.; et al. Development of Machine Learning Model for VO₂max Estimation Using a Patch-Type Single-Lead ECG Monitoring Device in Lung Resection Candidates. Healthcare 2023, 11. [Google Scholar] [CrossRef]
- Biró, A.; Cuesta-Vargas, A.I.; Szilágyi, L. AI-Assisted Fatigue and Stamina Control for Performance Sports on IMU-Generated Multivariate Time Series Datasets. Sensors 2024, 24. [Google Scholar] [CrossRef]
- Joyner, M.J.; Coyle, E.F. Endurance exercise performance: the physiology of champions. J. Physiol. 2008, 586(1), 35–44. [Google Scholar] [CrossRef]
- Hadjicharalambous M, Chalari E, Zaras N. Influence of puberty stage in immune-inflammatory parameters in well-trained adolescent soccer-players, following 8-weeks of pre-seasonal preparation training. Explor Immunol. 2024;4:822–36. [CrossRef]
- Huang, W.-Y.; Wu, C.-E.; Huang, H. The Effects of Plyometric Training on the Performance of Three Types of Jumps and Jump Shots in College-Level Male Basketball Athletes. Appl. Sci. 2024, 14. [Google Scholar] [CrossRef]
- Badau, D. , Badau, A., Ene-Voiculescu, V., Ene-Voiculescu, C., Teodor, D. F., Sufaru, C., Dinciu, C. C., Dulceata, V., Manescu, D. C., Manescu, C. O. El impacto de las tecnologías en el desarrollo de la velocidad repetitiva en balonmano, baloncesto y voleibol. Retos 2025, 64, 809–824. [Google Scholar] [CrossRef]
- Zaras, N.; Stasinaki, A.-N.; Spiliopoulou, P.; Mpampoulis, T.; Hadjicharalambous, M.; Terzis, G. Effect of Inter-Repetition Rest vs. Traditional Strength Training on Lower Body Strength, Rate of Force Development, and Muscle Architecture. Appl. Sci. 2021, 11, 45. [Google Scholar] [CrossRef]
- Shalom, A.; Gottlieb, R.; Alcaraz, P.E.; Calleja-Gonzalez, J. Unique Specific Jumping Test for Measuring Explosive Power in Young Basketball Players: Differences by Gender, Age, and Playing Positions. Sports 2024, 12, 118. [Google Scholar] [CrossRef]
- Cano, L.A.; Gerez, G.D.; García, M.S.; Albarracín, A.L.; Farfán, F.D.; Fernández-Jover, E. Decision-Making Time Analysis for Assessing Processing Speed in Athletes during Motor Reaction Tasks. Sports 2024, 12. [Google Scholar] [CrossRef]
- Souza, L.R.O.d.; Rezende, A.L.G.d.; Carmo, J.d. Instrument for Evaluation and Training of Decision Making in Dual Tasks in Soccer: Validation and Application. Sensors 2024, 24(21), 6840. [Google Scholar] [CrossRef]
- Wu, K.-C.; Lin, H.-C.; Cheng, Z.-Y.; Chang, C.-H.; Chang, J.-N.; Tai, H.-L.; Liu, S.-I. The Effect of Perceptual-Cognitive Skills in College Elite Athletes: An Analysis of Differences Across Competitive Levels. Sports 2025, 13, 141. [Google Scholar] [CrossRef]
- Badau, D.; Badau, A.; Joksimović, M.; Manescu, C.O.; Manescu, D.C.; Dinciu, C.C.; Margarit, I.R.; Tudor, V.; Mujea, A.M.; Neofit, A.; et al. Identifying the Level of Symmetrization of Reaction Time According to Manual Lateralization between Team Sports Athletes, Individual Sports Athletes, and Non-Athletes. Symmetry 2024, 16, 28. [Google Scholar] [CrossRef]
- Tosti, B.; Corrado, S.; Mancone, S.; Di Libero, T.; Carissimo, C.; Cerro, G.; Rodio, A.; da Silva, V.F.; Coimbra, D.R.; Andrade, A.; et al. Neurofeedback Training Protocols in Sports: A Systematic Review of Recent Advances in Performance, Anxiety, and Emotional Regulation. Brain Sci. 2024, 14. [Google Scholar] [CrossRef] [PubMed]
- Lu, C.-J.; Lee, T.-S.; Wang, C.-C.; Chen, W.-J. Improving Sports Outcome Prediction Process Using Integrating Adaptive Weighted Features and Machine Learning Techniques. Processes 2021, 9. [Google Scholar] [CrossRef]
- Junhyuk Lee; Namhyoung Kim. Development of Machine Learning-Based Indicators for Predicting Comeback Victories Using the Bounty Mechanism in MOBA Games. Electronics 2025, 14, 1445. [CrossRef]
- Molnar, C.; Interpretable Machine Learning. A Guide for Making Black Box Models Explainable; 2nd ed.; Leanpub: 2022. Available online: https://christophm.github.io/interpretable-ml-book/ (accessed on 4 August 2025).
- Hu, H.; Zhao, H.; Chen, X.; Li, Q.; Deng, J.; Zhang, X. Application of Machine Learning Models for Baseball Game Outcome Prediction in CPBL: Comparison of Logistic Regression, XGBoost, LightGBM and ANN. Appl. Sci. 2025, 15. [Google Scholar] [CrossRef]
- Geurkink, Y.; Boone, J.; Verstockt, S.; Bourgois, J.G. Machine Learning-Based Identification of the Strongest Predictive Variables of Winning and Losing in Belgian Professional Soccer. Appl. Sci. 2021, 11. [Google Scholar] [CrossRef]
- Calderón-Díaz, M.; Silvestre Aguirre, R.; Vásconez, J.P.; Yáñez, R.; Roby, M.; Querales, M.; Salas, R. Explainable Machine Learning Techniques to Predict Muscle Injuries in Professional Soccer Players through Biomechanical Analysis. Sensors 2024, 24. [Google Scholar] [CrossRef]
- Mănescu, D.C. Elements of the specific conditioning in football at university level. Marathon 2015, 7(1), 107–111. [Google Scholar]
- Vallance, E.; Sutton-Charani, N.; Imoussaten, A.; Montmain, J.; Perrey, S. Combining Internal- and External-Training-Loads to Predict Non-Contact Injuries in Soccer. Appl. Sci. 2020, 10. [Google Scholar] [CrossRef]
- Mandorino, M.; Tessitore, A.; Leduc, C.; Persichetti, V.; Morabito, M.; Lacome, M. A New Approach to Quantify Soccer Players’ Readiness through Machine Learning Techniques. Appl. Sci. 2023, 13, 8808. [Google Scholar] [CrossRef]
- Rossi, A.; Pappalardo, L.; Cintia, P.; Iaia, F.M.; Fernàndez, J.; Medina, D. Effective injury forecasting in soccer with GPS training data and machine learning. PLOS ONE 2018, 13(7), e0201264. [Google Scholar] [CrossRef] [PubMed]
- Kim, K.-M.; Kwak, J.W. PVS-GEN: Systematic Approach for Universal Synthetic Data Generation Involving Parameterization, Verification, and Segmentation. Sensors 2024, 24, 266. [Google Scholar] [CrossRef] [PubMed]
- Goyal, M.; Mahmoud, Q.H. A Systematic Review of Synthetic Data Generation Techniques Using Generative AI. Electronics 2024, 13, 3509. [Google Scholar] [CrossRef]
- Tu, Y.-C.; Lin, C.-Y.; Liu, C.-P.; Chan, C.-T. Performance Analysis of Data Augmentation Approaches for Improving Wrist-Based Fall Detection System. Sensors 2025, 25, 2168. [Google Scholar] [CrossRef]
- Dankar, F.K.; Ibrahim, M. Fake It Till You Make It: Guidelines for Effective Synthetic Data Generation. Appl. Sci. 2021, 11, 2158. [Google Scholar] [CrossRef]
- Huang, C.; Zhang, S. Explainable Artificial Intelligence Model for Identifying Market Value in Professional Soccer Players: Ensemble Models + SHAP Feature Attribution. Preprint 2023. [Google Scholar] [CrossRef]
- Apley, D.W.; Zhu, J. Visualizing the effects of predictor variables in black box supervised learning models. J. R. Stat. Soc. B 2020, 82(4), 1059–1086. [Google Scholar] [CrossRef]
- Altmann, A.; Tolosi, L.; Sander, O.; Lengauer, T. Permutation importance: A corrected feature importance measure. Bioinformatics 2010, 26(10), 1340–1347. [Google Scholar] [CrossRef] [PubMed]
- Ou-Yang, Y.; Sun, Y.; Li, H.; Wei, X.; Liu, M. Integration of Machine Learning XGBoost and SHAP Models for NBA Game Outcome Prediction and Quantitative Analysis Methodology. PLoS ONE 2024, 19, e0307478. [Google Scholar] [CrossRef]
- Mănescu, D.C. Big Data Analytics Framework for Decision-Making in Sports Performance Optimization. Data 2025, 10, 116. [Google Scholar] [CrossRef]
- Tempel, F.; Ihlen, E.A.F.; Adde, L.; Støen, R.; Lydersen, S.; Dallmeier, D.; Stang, J.; Khan, A. Explaining Human Activity Recognition with SHAP: Validating Insights with Perturbation and Quantitative Measures. arXiv 2024, arXiv:2411.03714, arXiv–2411.03714. [Google Scholar] [CrossRef]
- Panteli, N. , Hadjicharalambous, M. & Zaras, N. Delayed Potentiation Effect on Sprint, Power and Agility Performance in Well-Trained Soccer Players. J. of SCI. IN SPORT AND EXERCISE, 2024. [Google Scholar] [CrossRef]
- Mănescu, D.C. Fundamente teoretice ale activității fizice. 2013, Editura ASE.
- Settembre, M.; Buchheit, M.; Hader, K.; Hamill, R.; Tarascon, A.; Verheijen, R.; McHugh, D. Factors Associated with Match Outcomes in Elite European Football – Insights from Machine Learning Models. J. Sports Analyt. 2024, 10, 1–16. [Google Scholar] [CrossRef]
- Estrella, T.; Capdevila, L. Identification of Athleticism and Sports Profiles Throughout Machine Learning Applied to Heart Rate Variability. Sports 2025, 13, 30. [Google Scholar] [CrossRef]
- Moustakidis, S.; Plakias, S.; Kokkotis, C.; Tsatalas, T.; Tsaopoulos, D. Predicting Football Team Performance with Explainable AI: Leveraging SHAP to Identify Key Team-Level Performance Metrics. Future Internet 2023, 15, 174. [Google Scholar] [CrossRef]
- Varma, S.; Simon, R. Bias in error estimation when using cross-validation for model selection. BMC Bioinformatics 2006, 7, 91. [Google Scholar] [CrossRef]
- Ambroise, C.; McLachlan, G.J. Selection bias in gene extraction on the basis of microarray gene-expression data. Proc. Natl. Acad. Sci. USA 2002, 99(10), 6562–6566. [Google Scholar] [CrossRef] [PubMed]
- Balsalobre-Fernández, C.; Varela-Olalla, D. The Validity and Reliability of the My Jump Lab App for the Measurement of Vertical Jump Performance Using Artificial Intelligence. Sensors 2024, 24, 7897. [Google Scholar] [CrossRef] [PubMed]
- Lalwani, A.; Saraiya, A.; Singh, A.; Jain, A.; Dash, T. Machine Learning in Sports: A Case Study Using Explainable Models to Predict Volleyball Match Outcomes. arXiv, 2206. [Google Scholar] [CrossRef]
- Claros, C.C.; Anderson, M.N.; Qian, W.; Brockmeier, A.J.; Buckley, T.A. A Machine Learning Model for Post-Concussion Musculoskeletal Injury Risk in Collegiate Athletes. Sports Med. 2025, 55, 267–278. [Google Scholar] [CrossRef]
- Cordeiro, M. C.; Cathain, C. O.; Daly, L.; Kelly, D. T.; Rodrigues, T. B. A Synthetic Data-Driven Machine Learning Approach for Athlete Performance Attenuation Prediction. Front. Sports Act. Living 2025, 7, 1607600. [Google Scholar] [CrossRef] [PubMed]
- Qin, J.; Isleem, H. F.; Almoghayer, W. J. K.; Khishe, M.; et al. Predictive Athlete Performance Modeling with Machine Learning and Biometric Data Integration. Sci. Rep. 2025, 15, 16365. [Google Scholar] [CrossRef]
- Retzepis, N.-O.; Avloniti, A.; Kokkotis, C.; Protopapa, M.; Stampoulis, T.; Gkachtsou, A.; Pantazis, D.; Balampanos, D.; Smilios, I.; Chatzinikolaou, A. Identifying Key Factors for Predicting the Age at Peak Height Velocity in Preadolescent Team Sports Athletes Using Explainable Machine Learning. Sports 2024, 12, 287. [Google Scholar] [CrossRef]





| Variable Name | Unit | High-Performance (n = 200) Mean ± SD |
Low-Performance (n = 200) Mean ± SD | Value Range HP | Value Range LP | Description |
|---|---|---|---|---|---|---|
| VO₂max | mL/kg/min | 58.5 ± 4.3 | 41.3 ± 5.1 | 45–65 | 30–50 | Maximal oxygen uptake (aerobicendurance) |
| 20 m Sprint Time | seconds | 2.95 ± 0.18 | 3.55 ± 0.22 | 2.8–3.2 | 3.3–3.8 | Linear sprint acceleration |
| Countermovement Jump | cm | 47 ± 5.2 | 32 ± 4.8 | 35–55 | 25–40 | Explosive lower-limb power |
| Maximal Strength | kg | 148 ± 11 | 106 ± 13 | 120–160 | 80–120 | 1RM-equivalent lower-body strength |
| Reaction Time | milliseconds | 194 ± 12 | 256 ± 17 | 180–220 | 230–280 | Neuromotor response time |
| Decision Latency | milliseconds | 242 ± 29 | 396 ± 43 | 200–300 | 350–500 | Time to make accurate game-like decisions |
| Change-of-Direction Time | seconds | 2.15 ± 0.24 | 2.95 ± 0.32 | 1.8–2.5 | 2.6–3.5 | Agility in multidirectional sprint |
| Heart Rate Recovery | bpm | 44 ± 4.5 | 28 ± 5.2 | 35–50 | 20–35 | HR drop after 1 min (recovery efficiency) |
| Muscle Fatigue Index | % | 14 ± 3.1 | 32 ± 4.2 | 10–20% | 25–40% | Fatigue accumulation in drills |
| Stress Tolerance Score | points (0–10) |
8.5 ± 1.1 | 4.5 ± 1.3 | 7–10 | 3–6 | Mental resilience under pressure |
| Attention Control Index | points (0–100) |
82 ± 6.2 | 53 ± 7.1 | 70–90 | 40–65 | Cognitive focus in multitask conditions |
| Flexibility Score | points (0–10) |
8.1 ± 1.0 | 5.2 ± 1.2 | 7–10 | 3–6 | Joint mobility and range of motion |
| Metric | Value | Acceptable Threshold | Implication in Athlete Selection | Error Type Addressed | Relevant Decision-Making Context |
|---|---|---|---|---|---|
| Accuracy | 89.5% | ≥ 85% = good; ≥ 90% = excellent |
Reliable general decision support |
Both FP & FN (overall) | General model evaluation |
| Precision | 90.2% | ≥ 85% = excellent |
Minimizes overestimation (incorrectly selecting low performers) | False Positives (Type I) | Final selection / shortlisting |
| Recall (Sensitivity) | 88.7% | ≥ 85% = excellent |
Minimizes exclusion of real talent |
False Negatives (Type II) | Initial screening / scouting |
| F1-score | 89.4% | ≥ 85% = robust |
Balanced classification under uncertainty or imbalance | Balanced between FP & FN |
Mixed / nuanced classification decisions |
| AUC-ROC | 93.0% | ≥ 90% = very good |
Confident discrimination between athlete types across thresholds | Threshold-independent | Adjusting decision threshold / model discrimination |
| Mean Accuracy (CV) | 89.2% | ≥ 85% = acceptable |
Consistent performance on unseen data (cross-validated) | General stability |
Internal validation / deployment readiness |
| 95% CI (Accuracy CV) | 88.1–90.3% | Narrow CI (< 3%) preferred |
Statistical confidence in generalization |
Reliability | Trust in consistent performance |
|
Std. Dev. (CV Accuracy) |
±0.9% (range: 88.1–90.3%) | < 1% = very stable |
Confirms model stability and fairness across validation folds | Low variability |
Reliability across multiple resamplings |
| Operating Point | True Positives |
False Positives |
True Negatives |
False Negatives | Precision | Recall | F1 | Accuracy |
|---|---|---|---|---|---|---|---|---|
| Screening | 184 | 43 | 157 | 16 | 81.1% | 92.0% | 86.2% | 85.3% |
| Shortlisting | 177 | 19 | 181 | 23 | 90.3% | 88.5% | 89.4% | 89.5% |
| Model (baseline) | Brier (mean ± SD) | ECE K=10 (mean ± SD) | Calibration slope (mean ± SD) | Calibration intercept (mean ± SD) | Selected calibration |
|---|---|---|---|---|---|
| Logistic Regression (L2) | 0.039 ± 0.013 | 0.039 ± 0.020 | 0.672 ± 0.402 | −0.524 ± 1.084 | isotonic |
| Random Forest | 0.040 ± 0.009 | 0.038 ± 0.016 | 0.814 ± 0.329 | −0.104 ± 1.024 | isotonic |
| XGBoost | 0.046 ± 0.010 | 0.048 ± 0.011 | 1.056 ± 0.455 | 0.006 ± 1.003 | isotonic |
| Model | ROC-AUC | ΔAUC vs LGBM | 95% CI | p-value | Brier | ΔBrier vs LGBM | 95% CI | p-value |
|---|---|---|---|---|---|---|---|---|
| LightGBM | 0.930 | 0.000 | — | — | 0.072 | 0.000 | — | — |
| Logistic Regression (L2) | 0.884 | –0.046 | [–0.068, –0.024] | 0.001 | 0.081 | 0.009 | [0.004, 0.014] | 0.002 |
| Random Forest (RF) | 0.898 | –0.032 | [–0.050, –0.014] | 0.004 | 0.078 | 0.006 | [0.002, 0.010] | 0.006 |
| XGBoost (XGB) | 0.911 | –0.019 | [–0.035, –0.003] | 0.038 | 0.076 | 0.004 | [0.001, 0.008] | 0.041 |
| Variable | SHAP Mean |
SHAP Max |
SHAP Min |
HP Mean |
LP Mean |
Δ (abs) |
Cohen’s d |
|---|---|---|---|---|---|---|---|
| VO₂max | 0.183 | +0.36 | –0.11 | 58.5 | 41.3 | 17.2 | 3.69 |
| Decision Latency | 0.172 | +0.41 | –0.13 | 242 | 396 | 154 | 4.24 |
| Maximal Strength | 0.158 | +0.33 | –0.10 | 148 | 106 | 42 | 3.50 |
| Reaction Time | 0.151 | +0.31 | –0.12 | 194 | 256 | 62 | 4.16 |
| CMJ Height | 0.123 | +0.27 | –0.09 | 47 | 32 | 15 | ~2.8* |
| Sprint Time (20 m) | 0.110 | +0.25 | –0.08 | 2.95 | 3.55 | 0.60 | ~2.5* |
| Stress Tolerance | 0.085 | +0.22 | –0.05 | 8.5 | 4.5 | 4.0 | ~2.2* |
| Attention Control | 0.074 | +0.18 | –0.04 | 82 | 53 | 29 | ~2.4* |
| Variable | High-performance (M ± SD) |
Low-performance (M ± SD) |
t(df) | p-value | Cohen’s d | 95% CI for d |
|---|---|---|---|---|---|---|
| VO₂max | 58.5 ± 4.3 | 41.3 ± 5.1 | 36.46 (387) | 3.04e-127 | 3.65 | [3.45 – 3.84] |
| Decision Latency | 242 ± 29 | 396 ± 43 | –41.99 (349) | 1.68e-138 | –4.20 | [–4.40 – –4.00] |
| Maximal Strength | 148 ± 11 | 106 ± 13 | 34.88 (387) | 1.41e-121 | 3.49 | [3.29 – 3.68] |
| Reaction Time | 194 ± 12 | 256 ± 17 | –42.14 (358) | 8.48e-141 | –4.21 | [–4.41 – –4.02] |
| CMJ Height | 47 ± 5.2 | 32 ± 4.8 | 29.98 (395) | 7.6e-104 | 3.00 | [2.80 – 3.19] |
| Sprint Time (20 m) | 2.95 ± 0.18 | 3.55 ± 0.22 | 27.56 (387) | 5.9e-90 | –2.55 | [–2.74 – –2.36] |
| Stress Tolerance | 8.5 ± 1.1 | 4.5 ± 1.3 | 28.88 (392) | 7.1e-97 | 3.38 | [3.17 – 3.59] |
| Attention Control | 82 ± 6.2 | 53 ± 7.1 | 30.17 (393) | 2.8e-101 | 3.11 | [2.90 – 3.31] |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).