1. Introduction
The proliferation of digital financial services has intensified the threat of new account fraud, where malicious users exploit onboarding vulnerabilities to commit illicit activities. Unlike transaction fraud, this task is complicated by missing information, limited user history, and heterogeneous data types, making traditional detection methods insufficiently robust. Existing models often fail to handle incomplete inputs effectively, lack dedicated mechanisms for multi-view feature extraction, and struggle to adapt to evolving fraud patterns.
To address these limitations, we propose NAFNet, a unified neural architecture designed for new account fraud detection. The model includes a dynamic imputation module that reconstructs missing features using both group-level and global statistical signals; a multi-view feature encoder that separately processes demographic, behavioral, and transactional views with bilinear fusion; and an attention-augmented DNN backbone that captures global dependencies via multi-head attention and residual learning. Coupled with an adaptive fine-tuning strategy, NAFNet effectively improves representation quality, enhances model generalization, and remains resilient in the face of noisy, sparse, and rapidly changing financial data.
2. Related Work
Graph-centric models capture relational structures among accounts: homogeneous graphs for inductive link prediction in banking fraud [
1] and scalable graph-powered architectures for large-scale detection [
2]. Liu [
3] introduces HKNR, which combines LLaMA-2-7B embedding based candidate recall, a temporal GNN for user encoding, and knowledge augmented multi task ranking. These LLM derived entity and topic signals and temporal graph modeling inform NAFNet’s bilinear cross view fusion and progressive fine tuning to better track shifting fraud patterns.Optimization-driven designs integrate swarm intelligence with deep networks to improve convergence but neglect incomplete or multi-view features [
4].
Sun et al. [
5] introduce TransTARec, a time-adaptive translating embedding; integrating its time-aware translation vectors into NAFNet’s behavioral view via bilinear fusion can better capture temporal drift. Luo et al. [
6] present Gemini-GraphQA, integrating a large language model with a graph encoder and retrieval-augmented generation for executable graph reasoning. For NAFNet, this motivates exploiting graph-structured account relations with LLM-guided retrieval to inject structural semantics into the bilinear cross-view encoder and strengthen fraud signal extraction; real-time adaptive learning stresses responsiveness without robust semantic integration [
7]. Yu [
8] proposes MFTCoder++, a stable multilingual code-generation fine-tuning framework with adaptive task scheduling, attention-guided optimization, adversarial regularization, and gated fusion that decouples semantic logic from syntax. For NAFNet, analogous decoupled fusion across demographic/behavioral/transactional views and curriculum-style adaptive scheduling during progressive fine-tuning can reduce optimization bias under heterogeneous features and improve robustness.
For imbalanced credit risk, combining LightGBM, XGBoost, and TabNet with SMOTEENN boosts classification but assumes complete features [
9]; standard supervised baselines for credit card fraud rely heavily on preprocessing [
10]. Guo [
11] introduces MHST-GB, a multi-modal hierarchical spatio-temporal ensemble that fuses modality-specific neural encoders via correlation-guided attention and integrates LightGBM with feedback that adjusts attention by tree-based feature importance. For NAFNet, adopting this neural+LightGBM dual path and feedback-driven fusion can bolster bilinear cross-view encoding under sparse categorical signals and temporal drift. Deployable and behavior-aware real-time systems advance monitoring but underuse attention and fine-grained view separation [
12]; AI-driven behavioral modeling shows similar limits [
13]. Remaining gaps include handling missing data, multi-view fusion, and adaptive representation. Our NAFNet integrates dynamic feature imputation, view-specific encoding, and attention-based representation learning in a unified, scalable architecture.
3. Methodology
In this section, we proposes
NAFNet, a novel deep learning framework specifically designed for detecting new account fraud (NAF) in banking systems, where feature incompleteness and data imbalance pose significant challenges. NAFNet integrates a dynamic Feature Imputation Module to handle missing attributes, a Multi-View Feature Encoder that extracts cross-perspective embeddings, and an Attention-Augmented Deep Neural Network (DNN) backbone to capture complex fraud patterns. To ensure adaptability, we introduce a Progressive Fine-Tuning Strategy that incrementally updates the model using real-time new account data. Extensive mathematical modeling is incorporated to rigorously define each component, demonstrating the framework’s ability to achieve state-of-the-art performance in synthetic and real-world scenarios. The pipline of NAFNet in bank system data is shown in
Figure 1
3.1. Feature Imputation Module
The Feature Imputation Module in NAFNet is designed to tackle the pervasive issue of incomplete feature profiles in new bank accounts, which is a critical challenge in fraud detection. Unlike conventional imputation that merely fills missing values with static statistics, our module integrates a learning-based dynamic imputation mechanism combined with feature augmentation derived from historical data of old users.
First, we exploit the complete profiles of historical (old) users to construct enriched auxiliary features. Specifically, for each missing feature
in the new user data, we group old users into demographic and behavioral clusters
(e.g., based on age brackets, gender, or account type) and compute statistical aggregates such as:
These aggregates are used not only for imputation but also to
construct new derived features, such as:
where
is a small constant to prevent division by zero. This transformation captures how much a new user’s imputed value deviates from the typical range in their group, enabling the model to better distinguish anomalous behaviors.
For the imputation itself, we implement a
Feature Fusion Layer that learns an optimal combination of group-level and global statistics. The imputed value is dynamically computed as:
where
is the overall mean for feature
j, and
is a learnable parameter updated via backpropagation. This layer is implemented as a small neural network with one fully connected (FC) layer followed by a sigmoid activation:
where
is a concatenation of group and global statistics for feature
j. This design allows the network to adaptively determine the importance of group-specific versus global patterns for each feature. The pipline of Feature Imputation Module is shown in
Figure 2
3.1.1. Network Architecture
We implement three compact sub-layers.
Statistical Encoding embeds
:
Weight Generation maps the encoding to the gate:
Imputation & Augmentation computes and constructs . The encoder and generator are small FC networks with hidden size and ReLU activations; . For interpretability we compare to global mean, KNN impute, and VAE reconstruction; ablations include NAFNet-NoImp.
3.1.2. LightGBM Feature Selection
After augmentation, LightGBM ranks features by cumulative split gain:
We select the top-k features using early stopping and regularization (e.g., max depth, minimum child weight) to reduce overfitting. The selected subset is fed to the downstream Multi-View Encoder, yielding cleaner inputs and improved prediction efficiency.
3.2. Multi-View Feature Encoder
We encode bank-account behavior via three independent two-layer view encoders for Demographic (
), Behavioral (
), and Transactional (
) features. For view
v:
Encoders do not share weights; output dimension is d. A parameter-sharing variant can add a shared base with view-specific projections.
Cross-view dependencies use bilinear interactions:
with optional low-rank parameterization:
Information flows through gated fusion:
The final embedding concatenates views and pairwise interactions:
Regularization and stabilization:
3.3. Attention-Augmented DNN Backbone
Multi-head self-attention models global cross-feature and cross-view relations on
Z:
The attention-enhanced embedding enters a 3-layer DNN:
with LeakyReLU
and dropout after each layer. Prediction uses a sigmoid head:
Ablation confirms the role of attention (NAFNet-Full vs NAFNet-NoAtt): removing attention reduces AUC and F1 (
Table 2). Residual connections and weight decay mitigate overfitting.
3.4. Progressive Fine-Tuning Strategy
We adapt to evolving data with incremental updates from pretrained
:
and a decaying learning rate:
This schedule stabilizes updates while tracking new fraud patterns. Illustrations are shown in
Figure 3.
4. Evaluation Metrics
We evaluate the performance of our model using the following metrics to ensure comprehensive assessment:
Accuracy measures the overall correctness of predictions:
Precision focuses on the correctness of positive predictions:
Recall assesses the ability to detect positive cases:
F1-Score provides a harmonic mean of Precision and Recall:
In addition, we include the Area Under the ROC Curve (AUC), which measures the model’s ability to distinguish between classes:
where TPR is the true positive rate and FPR is the false positive rate.
5. Experiment Results
Table 1 shows the main performance comparison between NAFNet and several baseline models, including tree-based and deep learning methods.
Per reviewer request, we added ablation experiments that quantify the contributions of attention, cross-view interaction, and the imputation module. Results are summarized in
Table 2.
To analyze the contribution of each key module in NAFNet, we conduct ablation studies.
Table 2 summarizes the impact of removing the attention layer, cross-view interaction, and imputation module.
Analysis:
Table 2 shows that removing attention (NAFNet-NoAtt) lowers AUC from 0.978 to 0.969 and F1 from 0.927 to 0.909, demonstrating empirical benefit of attention. Removing the imputation module (NAFNet-NoImp) also reduces performance, supporting our learnable imputation design. Removing cross-view interactions (NAFNet-NoXView) degrades results as well, validating bilinear interactions and gated fusion.
We also compare NAFNet with common ensemble strategies.
Table 3 reports the results of stacking, voting, and hybrid ensemble models, demonstrating that while ensemble methods are strong, NAFNet-Full still achieves superior performance. And the changes in model training indicators are shown in
Figure 4.
6. Conclusions
This study proposed NAFNet, a novel deep learning framework for detecting new account fraud in banking systems. By integrating a multi-view encoder, an attention-augmented DNN, and a progressive fine-tuning strategy, NAFNet effectively addresses the challenges posed by incomplete feature profiles and evolving fraud patterns. Experimental results confirm that each module contributes to the model’s strong performance, providing a scalable and robust solution for real-world fraud detection.
Limitations and future work: Despite good empirical results, NAFNet has limitations. First, the learnable imputer depends on representative historical samples; when historical data are non-representative or extremely sparse, priors can mislead imputation and harm downstream performance. Second, bilinear interactions and multi-head attention increase compute and memory costs; production deployment may require low-rank approximations or pruning. Future work includes adding uncertainty estimates for imputed values to allow safe fallback strategies, integrating VAE/self-supervised pretraining to improve robustness to distribution shift, and exploring lighter-weight interaction approximations (e.g., hashing or factorized bilinear terms).
References
- Bukhori, H.A.; Munir, R. Inductive link prediction banking fraud detection system using homogeneous graph-based machine learning model. In Proceedings of the 2023 IEEE 13th Annual Computing and Communication Workshop and Conference (CCWC). IEEE; 2023; pp. 0246–0251. [Google Scholar]
- Li, Z.; Wang, B.; Huang, J.; Jin, Y.; Xu, Z.; Zhang, J.; Gao, J. A graph-powered large-scale fraud detection system. International Journal of Machine Learning and Cybernetics 2024, 15, 115–128. [Google Scholar] [CrossRef]
- Liu, J. Knowledge-Augmented News Recommendation via LLM Recall, Temporal GNN Encoding, and Multi-Task Ranking. In Proceedings of the 2025 6th International Conference on Big Data &, 2025, Artificial Intelligence & Software Engineering (ICBASE). IEEE; pp. 141–144.
- Karthikeyan, T.; Govindarajan, M.; Vijayakumar, V. An effective fraud detection using competitive swarm optimization based deep neural network. Measurement: Sensors 2023, 27, 100793. [Google Scholar] [CrossRef]
- Sun, Y. Transtarec: Time-adaptive translating embedding model for next poi recommendation. In Proceedings of the 2024 5th International Conference on Computer Engineering and Application (ICCEA). IEEE; 2024; pp. 647–651. [Google Scholar]
- Luo, X.; Wang, E.; Guo, Y. Gemini-GraphQA: Integrating Language Models and Graph Encoders for Executable Graph Reasoning. Preprints 2025. [Google Scholar] [CrossRef]
- Bello, H.O.; Ige, A.B.; Ameyaw, M.N. Adaptive machine learning models: concepts for real-time financial fraud prevention in dynamic environments. World Journal of Advanced Engineering Technology and Sciences 2024, 12, 021–034. [Google Scholar] [CrossRef]
- Yu, H. Hybrid Modal Decoupled Fusion for Stable Multilingual Code Generation. Preprints 2025. [Google Scholar] [CrossRef]
- Yu, C.; Jin, Y.; Xing, Q.; Zhang, Y.; Guo, S.; Meng, S. Advanced user credit risk prediction model using lightgbm, xgboost and tabnet with smoteenn. In Proceedings of the 2024 IEEE 6th International Conference on Power, Intelligent Computing and Systems (ICPICS). IEEE; 2024; pp. 876–883. [Google Scholar]
- Afriyie, J.K.; Tawiah, K.; Pels, W.A.; Addai-Henne, S.; Dwamena, H.A.; Owiredu, E.O.; Ayeh, S.A.; Eshun, J. A supervised machine learning algorithm for detecting and predicting fraud in credit card transactions. Decision Analytics Journal 2023, 6, 100163. [Google Scholar] [CrossRef]
- Guo, R. Multi-Modal Hierarchical Spatio-Temporal Network with Gradient-Boosting Integration for Cloud Resource Prediction. Preprints 2025. [Google Scholar] [CrossRef]
- Manoharan, G.; Dharmaraj, A.; Sheela, S.C.; Naidu, K.; Chavva, M.; Chaudhary, J.K. Machine learning-based real-time fraud detection in financial transactions. In Proceedings of the 2024 International Conference on Advances in Computing, Communication and Applied Informatics (ACCAI). IEEE; 2024; pp. 1–6. [Google Scholar]
- Xu, J.; Yang, T.; Zhuang, S.; Li, H.; Lu, W. AI-based financial transaction monitoring and fraud prevention with behaviour prediction. Applied and Computational Engineering 2024, 77, 218–224. [Google Scholar] [CrossRef]
|
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).