Submitted:
04 May 2025
Posted:
06 May 2025
You are already at the latest version
Abstract
Keywords:
1. Introduction
2. Literature Review
- Resampling techniques, such as Synthetic Minority Over-sampling Technique (SMOTE), which synthetically generates minority class examples.
- Cost-sensitive learning, which penalizes misclassification of minority instances more heavily.
- Anomaly detection frameworks, which treat fraud as a deviation from normal transaction behavior rather than a supervised learning problem.
3. Methodology
3.1. Data Generation and Simulation
- Features: 20 independent numerical features, simulating various transactional and account-based metrics.
- Informative Features: 15 features contributed directly to class separation.
- Redundant Features: 5 were linearly dependent to mimic correlated financial attributes.
- Class Distribution: A highly imbalanced ratio of 95:5 was used to represent legitimate versus fraudulent transactions, respectively.
3.2. Data Preprocessing
- Train-Test Split: The dataset was divided into training (64%), validation (16%), and test (20%) subsets.
- Standardization: Features were normalized using standard scaler to have zero mean and unit variance, which accelerates convergence in gradient-based models.
- Label Encoding: Binary class labels were encoded as 0 (legitimate) and 1 (fraudulent).
3.3. Neural Network Architecture
- Input Layer: Accepts 20 features.
- Hidden Layer 1: 32 neurons with ReLU activation, followed by a 30% dropout layer to prevent overfitting.
- Hidden Layer 2: 16 neurons with ReLU activation, to further capture non-linear feature interactions.
- Output Layer: A single neuron with sigmoid activation for binary classification.
- Loss Function: Binary cross-entropy, suitable for probabilistic binary output.
- Optimizer: Adam optimizer, chosen for its robustness and adaptive learning rates.
- Metrics: Accuracy and AUC (Area Under the Curve) were tracked during training.
3.4. Model Training

3.5. Model Evaluation and Visualization
- Accuracy measures the overall proportion of correctly predicted instances out of all instances. It is calculated as:
- Precision measures the accuracy of positive predictions. It tells you what proportion of predicted positive cases were actually positive.
- Recall measures how well the model identifies actual positive cases. It tells you what proportion of actual positive cases were correctly identified.
- F1-Score is the harmonic mean of precision and recall, providing a single metric that balances both. It’s especially useful when you need to balance false positives and false negatives.
3.6. Integration into AFRAMES
4. Results
4.1. Classification Performance
4.2. Confusion Matrix and Insights
- True Negatives (TN): The model correctly identified the vast majority of legitimate transactions.
- True Positives (TP): Most fraudulent transactions were correctly flagged.
- False Positives (FP): Minimal false alarms, indicating robust generalization.
- False Negatives (FN): A small number of missed fraud cases, which may be mitigated with threshold tuning or post-processing strategies.
4.3. ROC Curve and Threshold Analysis
4.4. Dimensionality Reduction and Visualization

4.5. Relevance to AFRAMES
- Credit risk scoring
- Customer behavior analysis
- Anti-money laundering (AML) systems
5. Conclusions
References
- Khang, A., 2025. Shaping cutting-edge technologies and applications for digital banking and financial services.
- Bahnsen, A.C., Aouada, D., Stojanovic, A., Ottersten, B.: Feature engineering strategies for credit card fraud detection. Expert Systems with Applications, 51, 134–142 (2016). [CrossRef]
- Carcillo, F., Le Borgne, Y.-A., Caelen, O., Bontempi, G.: Streaming active learning strategies for real-life credit card fraud detection: assessment and visualization. International Journal of Data Science and Analytics 5(4), 285–300 (2018). [CrossRef]
- Jurgovsky, J., Granitzer, M., Ziegler, K., Calabretto, S., Portier, P.-E., He-Guelton, L., Caelen, O.: Sequence classification for credit-card fraud detection. Expert Systems with Applications 100, 234–245 (2018). [CrossRef]
- LeCun, Y., Bengio, Y., Hinton, G.: Deep learning. Nature 521, 436–444 (2015). [CrossRef]
- Ngai, E.W.T., Hu, Y., Wong, Y.H., Chen, Y., Sun, X.: The application of data mining techniques in financial fraud detection: A classification framework and an academic review of literature. Decision Support Systems 50(3), 559–569 (2011). [CrossRef]
- Sahin, Y., Duman, E.: Detecting credit card fraud by ANN and logistic regression. In: Innovations in Intelligent Systems and Applications (INISTA), pp. 315–319. IEEE (2011). [CrossRef]
- Xu, X., Wang, X.: Financial fraud detection based on SVM and BP neural network. In: International Symposium on Computational Intelligence and Design (ISCID), vol. 1, pp. 361–364. IEEE (2009). [CrossRef]
- Goodfellow, I., Bengio, Y., Courville, A.: Deep Learning. MIT Press (2016). http://www.deeplearningbook.org.
- Zhang, Y., Zhang, Y.: Research on credit card fraud detection model based on improved BP neural network. Procedia Computer Science 166, 600–605 (2019).
- Daah, Clement, Amna Qureshi, Irfan Awan, and Savas Konur. "Enhancing zero trust models in the financial industry through blockchain integration: A proposed framework." Electronics 13, no. 5 (2024): 865. 10.3390/electronics13050865.
- Dakalbab, Fatima, Manar Abu Talib, Qassim Nasir, and Tracy Saroufil. "Artificial intelligence techniques in financial trading: A systematic literature review." Journal of King Saud University-Computer and Information Sciences 36, no. 3 (2024): 102015. [CrossRef]
- Biju, Ajitha Kumari Vijayappan Nair, Ann Susan Thomas, and J. Thasneem. "Examining the research taxonomy of artificial intelligence, deep learning & machine learning in the financial sphere—a bibliometric analysis." Quality & Quantity 58, no. 1 (2024): 849-878. [CrossRef]
- Mohsen, Sara Ebrahim, Allam Hamdan, and Haneen Mohammad Shoaib. "Digital transformation and integration of artificial intelligence in financial institutions." Journal of Financial Reporting and Accounting 23, no. 2 (2025): 680-699. [CrossRef]

| Metric | Value |
|---|---|
| Accuracy | 98% |
| Precision | 95% |
| Recall | 88% |
| F1-Score | 91% |
| AUC-ROC | 98% |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the author. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).