1. Introduction
1.1. Brief Summary of the Study
This study addresses the critical challenge of trust in Autonomous Cyber Defense Systems (ACDS) by investigating the application of Explainable AI (XAI) frameworks. As cyber attacks evolve in scale and complexity, AI-driven autonomous responses become essential. However, the "blackbox" nature of advanced AI models creates significant risks, including unexplained false positives/negatives, adversarial manipulation, and a lack of operational accountability. This research systematically examines how XAI methodologies can be integrated into the ACDS pipeline to provide transparency, justify autonomous actions, and foster trust among security operators. The study evaluates current XAI techniques, measures their impact on system performance, and proposes a tailored framework for deploying trustworthy ACDS.
1.2. Purpose of the Research
The primary purpose of this research is threefold: (1) To analyze and categorize existing XAI frameworks for their suitability in dynamic, high-stakes cyber defense environments. (2) To empirically evaluate the trade-offs between explanation fidelity, computational cost, and defensive efficacy in a simulated ACDS context. (3) To synthesize a set of design principles and a conceptual architecture for an XAI-integrated ACDS that enhances trust without compromising defensive capabilities.
1.3. Methodology
A mixed-methods research design was employed. First, a systematic literature review was conducted to establish the state-of-the-art in XAI and its applications in cybersecurity. Second, a quantitative experimental study was performed using the CIC-IDS2017 dataset. Multiple ML models (Decision Tree, Random Forest, Deep Neural Network) were trained for intrusion detection and subsequently explained using post-hoc techniques (SHAP, LIME, and a surrogate model). Metrics for explanation accuracy, stability, and computational overhead were collected alongside standard performance metrics (accuracy, F1-score). The study adhered to ethical guidelines for the use of public cybersecurity datasets.
2. Literature Review
2.1. Background and Context
The cybersecurity landscape is defined by a chronic shortage of skilled personnel and an overwhelming volume of threats. ACDS promise to alleviate this burden by automating threat detection, analysis, and response (Töpfer et al., 2022). AI, especially deep learning, excels at identifying complex patterns in network traffic, malware, and user behavior. However, this capability comes at the cost of interpretability. The "right to explanation" embedded in regulations like the GDPR, and operational needs in military/enterprise contexts (e.g., the need for a Commander's understanding of an autonomous action), make explainability a paramount concern (Arrieta et al., 2020). Trust, as defined by recent NIST guidelines, encompasses reliability, resilience, safety, and transparency – all of which are impacted by explainability (NIST, 2023).
2.2. Research Questions
This study is guided by the following research questions:
RQ1: Which categories of XAI frameworks (intrinsic, post-hoc, model-agnostic, model-specific) are most suited for integration into different stages (detection, diagnosis, response) of an ACDS pipeline?
RQ2: What are the measurable performance trade-offs (e.g., latency, accuracy, resource consumption) when integrating post-hoc XAI techniques into a real-time ACDS? RQ3: How do explanations generated by XAI frameworks affect the trust and decision-making efficacy of human security analysts in a simulated incident response loop?
2.3. Significance of the Study
This research contributes to the emerging field of trustworthy AI in cybersecurity. It moves beyond mere technical performance of AI models to address the socio-technical challenge of trust. The findings are significant for: (1) Security Operations Center (SOC) Managers, by providing evidence-based guidance on deployable, explainable AI tools; (2) ACDS Developers, by offering a framework for designing transparent systems from the ground up; and (3) Policy Makers, by highlighting the technical feasibility and necessities of explainability in regulating autonomous cyber systems. The proposed hybrid XAI framework aims to serve as a blueprint for nextgeneration, accountable cyber defense.
3. Methodology
3.1. Research Design
A sequential mixed-methods design was used. Phase 1 was qualitative: a systematic narrative review of academic literature (2018-2024) from IEEE Xplore, ACM Digital Library, and
SpringerLink using keywords "XAI," "explainable AI," "cybersecurity," "autonomous defense," and "trustworthy AI." Phase 2 was quantitative: an experimental study measuring the impact of XAI on ML-based intrusion detection systems.
3.2. Participants or Datasets
The primary dataset for the quantitative phase was the CIC-IDS2017 dataset (Sharafaldin et al., 2018). It contains benign traffic and common up-to-date attacks, with 80 network flow-based features. It was selected for its realism and widespread use as a benchmark in intrusion detection research.
3.3. Data Collection Methods
Literature Review: 78 relevant papers were identified, screened, and synthesized to map XAI techniques to cyber defense tasks.
Experimental Data: The CIC-IDS2017 dataset was preprocessed (handling missing values, normalization, label encoding). It was split into 70% training, 15% validation, and 15% testing sets. Three model architectures were implemented: a Decision Tree (DT - intrinsically interpretable), a Random Forest (RF - ensemble black-box), and a Deep Neural Network (DNN - deep black-box).
3.4. Data Analysis Procedures
Model Training & Baseline Performance: All models were trained, and baseline metrics (Accuracy, Precision, Recall, F1-Score, Inference Time) were recorded.
-
XAI Application: Post-hoc explanations were generated for a stratified sample of 1000 test instances (including true positives, false positives, true negatives, false negatives).
o SHAP (SHapley Additive exPlanations): KernelExplainer for DT and RF; DeepExplainer for DNN.
o LIME (Local Interpretable Model-agnostic Explanations): Applied to all models.
o Surrogate Model: A global logistic regression model was trained on the predictions of the RF and DNN models.
-
Evaluation Metrics:
o Explanation Fidelity: For LIME and the surrogate model, we measured how well the explanation model approximated the black-box model's predictions (using R² for regression tasks and accuracy for classification tasks on the explained instances).
o Stability: Measured by applying LIME multiple times to the same instance and calculating the Jaccard similarity between the top-k features identified.
o Computational Overhead: The time added to the inference pipeline to generate an explanation was measured.
Human-Trust Simulation: A simplified survey was designed where 5 expert security analysts were presented with model predictions for 10 incident scenarios, both with and without SHAP-generated feature attribution plots. They rated their confidence in the system's decision on a Likert scale (1-5).
3.5. Ethical Considerations
As the study used a publicly available, anonymized dataset (CIC-IDS2017), no personal data was involved. The simulated human trust component involved voluntary participation of professionals, who were informed about the study's purpose and assured of the anonymity of their responses. The research acknowledges the dual-use nature of cybersecurity AI and emphasizes its defensive application.
4. Results
4.1. Model Performance Baseline
Table 1.
Baseline Performance of ML Models on CIC-IDS2017 Test Set.
Table 1.
Baseline Performance of ML Models on CIC-IDS2017 Test Set.
| Model |
Accuracy |
F1-Score (Weighted) |
Avg. Inference Time (ms) |
| Decision Tree (DT) |
96.2% |
0.959 |
0.8 |
| Random Forest (RF) |
99.1% |
0.990 |
4.2 |
| Deep Neural Net (DNN) |
98.7% |
0.987 |
3.1 (GPU) |
4.2. XAI Framework Performance
Figure 1.
Trade-off between Model Predictive Performance (F1-Score) and Explanation Fidelity (as measured by SHAP R²) for different model classes.
Figure 1.
Trade-off between Model Predictive Performance (F1-Score) and Explanation Fidelity (as measured by SHAP R²) for different model classes.
Table 2.
Evaluation of Post-hoc XAI Techniques (Averaged over Black-box Models RF & DNN).
Table 2.
Evaluation of Post-hoc XAI Techniques (Averaged over Black-box Models RF & DNN).
| XAI Technique |
Avg. Explanation Fidelity (R²)
|
Avg. Stability (Jaccard@5) |
Avg. Overhead per Explanation (ms) |
| SHAP |
0.89 |
0.92 |
320 |
| LIME |
0.76 |
0.71 |
45 |
| Surrogate Model |
0.82 |
1.00 |
5 (after training) |
4.3. Impact on Human Trust
When presented with XAI explanations (SHAP force plots), the average confidence rating of security analysts in the system's decision increased from 2.8 (SD=0.9) to 4.1 (SD=0.6). The increase was most pronounced for false positive and false negative cases.
5. Discussion
5.1. Interpretation of Results
The results confirm a clear tension between model complexity and explainability. While the Random Forest model achieved the highest detection performance (F1=0.99), explaining its decisions required computationally expensive post-hoc methods like SHAP, which added significant latency (~320ms). The intrinsically interpretable Decision Tree, while fast and transparent, suffered from lower accuracy, making it unsuitable as a standalone solution for sophisticated threats.
The high stability and fidelity of SHAP make it a robust choice for post-incident forensic analysis within an ACDS, where understanding the root cause of a detection is critical. LIME, while faster, showed lower fidelity and instability, potentially leading to misleading explanations. The surrogate model offered a fast global overview but failed to accurately explain complex, nonlinear local decisions.
The trust survey, though preliminary, provides strong indicative evidence that XAI outputs significantly enhance human operator confidence, a prerequisite for effective human-AI teaming in SOCs.
5.2. Comparison with Existing Literature
Our findings align with Arrieta et al. (2020) on the taxonomy and trade-offs of XAI methods. The identified computational overhead of SHAP corroborates concerns raised by Vilone and Longo (2021) regarding the deployability of high-fidelity XAI in real-time systems. However, this study extends prior work by specifically quantifying these trade-offs in a cyber defense context, where latency constraints are stringent (Spring et al., 2020). The proposed need for a hybrid approach echoes suggestions by Töpfer et al. (2022) for "multi-modal" explanations in cybersecurity.
5.3. Implications
The implications are both technical and operational. Technically, ACDS architects must move beyond a single-model approach. A proposed hybrid XAI-ACDS architecture could involve: (1) A fast, interpretable model (e.g., a rule-based system or shallow tree) for high-frequency, lowrisk decisions with inherent explanations. (2) A high-performance black-box ensemble for complex anomaly detection. (3) An on-demand XAI module (using optimized SHAP or a faithful surrogate) that is triggered for high-severity alerts, slow-time analysis, or when human operator intervention is required. This balances performance with on-demand transparency. Operationally, SOC workflows must be redesigned to incorporate explanation interfaces. Training must upskill analysts to interpret XAI outputs critically.
5.4. Limitations
This study has several limitations. The CIC-IDS2017 dataset, while valuable, does not represent all modern attack vectors (e.g., advanced persistent threats). The human trust evaluation involved a small, non-representative sample and a simplified simulation. The study did not explore adversarial attacks on the XAI methods themselves, a known vulnerability (Slack et al., 2020). Furthermore, the computational overhead was measured in a research environment and may vary in production systems.
5.5. Directions for Future Research
Future work should: (1) Develop and benchmark resource-efficient XAI techniques specifically for real-time streaming cyber data. (2) Conduct comprehensive human-in-the-loop experiments in live SOC environments to validate the impact of XAI on decision speed and accuracy. (3) Investigate explanation security to make XAI frameworks robust against manipulation. (4) Establish standardized evaluation metrics and benchmarks for XAI in cybersecurity, similar to those for model performance. (5) Explore the role of XAI in autonomous response justification, creating audit trails for automated actions like blocking IPs or isolating hosts.
6. Conclusions
The transition to Autonomous Cyber Defense Systems is inevitable, but its success is contingent on trust. This research demonstrates that Explainable AI is the keystone for building that trust. No single XAI method is a panacea; rather, a principled, hybrid approach is required. By strategically combining intrinsically interpretable models for routine decisions with high-performance blackbox models coupled with on-demand, high-fidelity explanations for critical events, we can construct ACDS that are both effective and accountable. The quantified trade-offs between performance, explainability, and speed provide a practical guide for system designers. Ultimately, trustworthy ACDS will not replace human analysts but will empower them, creating a synergistic partnership where AI handles scale and speed, while humans provide oversight, contextual wisdom, and ethical judgment, guided by clear explanations. The journey towards fully trustworthy autonomous cyber defense begins with making the black box transparent.
References
- KM, Z.; Akhtaruzzaman, K.; Tanvir Rahman, A. Building trust in autonomous cyber decision infrastructure through explainable AI. International Journal of Economy and Innovation 2022, 29, 405–428. [Google Scholar]
- Kumar, V.; Kaware, P.; Singh, P.; Sonkusare, R.; Kumar, S. Extraction of information from bill receipts using optical character recognition. 2020 International Conference on Smart Electronics and Communication (ICOSEC); LOCATION OF CONFERENCE, IndiaDATE OF CONFERENCE; pp. 72–77.
- Kumar, V.; Kumar, S.; Sreekar, L.; Singh, P.; Pai, P.; Nimbre, S.; Rathod, S.S. AI Powered Smart Traffic Control System for Emergency Vehicles; CONFERENCE NAME, LOCATION OF CONFERENCE, COUNTRYDATE OF CONFERENCE; pp. 651–663.
- Arrieta, A.B.; Díaz-Rodríguez, N.; Del Ser, J.; Bennetot, A.; Tabik, S.; Barbado, A.; Garcia, S.; Gil-Lopez, S.; Molina, D.; Benjamins, R.; et al. Explainable Artificial Intelligence (XAI): Concepts, taxonomies, opportunities and challenges toward responsible AI. Inf. Fusion 2020, 58, 82–115. [Google Scholar] [CrossRef]
- Bhatt, U.; Xiang, A.; Sharma, S.; Weller, A.; Taly, A.; Jia, Y.; Eckersley, P. Explainable machine learning in deployment. In Proceedings of the 2020 Conference on Fairness, Accountability, and Transparency, 2020; pp. 648–657. [Google Scholar]
- Gunning, D.; Stefik, M.; Choi, J.; Miller, T.; Stumpf, S.; Yang, G. Z. XAI— Explainable artificial intelligence. Science Robotics 2019, *4*(37), eaay7120. [Google Scholar] [CrossRef] [PubMed]
- Hoffman, R. R.; Mueller, S. T.; Klein, G.; Litman, J. Metrics for explainable AI. 2018. [Google Scholar]
- Challenges and prospects. arXiv arXiv:1812.04608.
- Lundberg, S. M.; Lee, S. I. A unified approach to interpreting model predictions. In Advances in Neural Information Processing Systems; 2017; pp. 4765–4774. [Google Scholar]
- Miller, T. Explanation in artificial intelligence: Insights from the social sciences. Artif. Intell. 2019, 267, 1–38. [Google Scholar] [CrossRef]
- National Institute of Standards and Technology (NIST). AI Risk Management Framework (AI RMF 1.0); U.S. Department of Commerce, 2023. [Google Scholar]
- Ribeiro, M.; Singh, S.; Guestrin, C. “Why Should I Trust You?”: Explaining the Predictions of Any Classifier. In Proceedings of the 2016 Conference of the North American Chapter of the Association for Computational Linguistics: Demonstrations, LOCATION OF CONFERENCE, United StatesDATE OF CONFERENCE.
- Rudin, C. Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead. Nat. Mach. Intell. 2019, 1, 206–215. [Google Scholar] [CrossRef] [PubMed]
- Sharafaldin, I.; Lashkari, A.H.; Ghorbani, A.A. Toward generating a new intrusion detection dataset and intrusion traffic characterization. In Proceedings of the International Conference on Information Systems Security and Privacy, Funchal, Portugal, 22–24 January 2018; pp. 108–116. [Google Scholar]
- Slack, D.; Hilgard, S.; Jia, E.; Singh, S.; Lakkaraju, H. Fooling LIME and SHAP: Adversarial attacks on post hoc explanation methods. In Proceedings of the AAAI/ACM Conference on AI, Ethics, and Society, 2020; pp. 180–186. [Google Scholar]
- Spring, J.; Hatleback, E.; Householder, A.; Manion, A. The difficulty of proving a negative: The challenge of evaluating autonomous cyber defense systems; Carnegie Mellon University, Software Engineering Institute, 2020. [Google Scholar]
- Töpfer, M.; Endres, T.; Paschke, A. Explainable artificial intelligence for cybersecurity: A survey. IEEE Access 2022, *10*, 123700–123714. [Google Scholar]
- Vilone, G.; Longo, L. Notions of explainability and evaluation approaches for explainable artificial intelligence. Inf. Fusion 2021, 76, 89–106. [Google Scholar] [CrossRef]
|
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |