Preprint
Article

This version is not peer-reviewed.

CSSA: A Cross-Modal Semantic-Structural Alignment Framework via LLMs and Graph Contrastive Learning for Fraud Detection of Online Payment

Submitted:

01 March 2026

Posted:

03 March 2026

You are already at the latest version

Abstract
Graph Neural Networks (GNNs) have demonstrated exceptional performance in modeling structural dependencies within networked data. However, in complex decision-making environments, structural information alone often fails to capture the latent semantic logic and domain-specific heuristics. While Large Language Models (LLMs) excel in semantic reasoning, their integration with graph-structured data remains loosely coupled in existing literature. This paper proposes CSSA, a novel Cross-modal Semantic-Structural Alignment framework that synergizes the zero-shot reasoning of LLMs with the topological aggregation of GNNs through a contrastive learning objective. Specifically, we treat node attributes as semantic prompts for LLMs to distill high-level "risk indicators," while a GNN branch encodes the local neighborhood topology. A cross-modal alignment layer is then introduced to minimize the representational gap between semantic intent and structural behavior. We evaluate CSSA on a massive dataset of 2.84 million online transaction records. Experimental results demonstrate that CSSA achieves a superior F1-score and AUC compared to state-of-the-art GNNs, particularly in scenarios characterized by extreme class imbalance and covert adversarial patterns.
Keywords: 
;  ;  ;  ;  

I. Introduction

One of the main challenges of modern learning models on relational data lies in reconciling local topological structures and global semantic reasoning, while GNN has emerged as a de facto solution to process unstructured data. Their main strategy is to combine nodes with close neighbors, a method which assumes that proximity implies functionality similarity[1].However, for some crucial applications it is insufficient to only consider the topology structure. One shortcoming of GNNs lies in that they cannot perform complicated analysis and processing operations over nodes’ features.reducing complex concepts to simplistic numbers [2].
Meanwhile, recent breakthroughs on large scale LLMs have significantly expanded the horizon of NLP tasks. Large-scale LLMs are capable to perform general reasoning beyond their training data as well as domain-specific expertises,which allows them to understand human intent, which is not trivial for standard networks.Nevertheless, they do not possess any kind of internal structure perception since they treat the interdependent data like a sequence of elements without considering intricate network patterns and dynamic structural properties that exist in big networks [3].
This restriction is problematic for researchers, as they have to give up either knowledge about the structure of their network and hence lose semantic information, or they have to give up analysis and therefore loose syntactic information[4]. This has serious implications especially when it comes to critical applications such as online bankingwhere adversaries intentionally modify the graph connectivity in order to look benign and have contradictory behavioral properties.
In this work, we propose CSSA (Cross-modal Semantic-Structural Alignment) which overcomes the aforementioned limitation of learning multi-modal latent alignment. We argue that the representation of each node cannot be only determined by its graph structure and features butbut rather from their interplay:The technical foundation of CSSA is a two path architecture and one of the branches includes the Semantic Reasoner with a finetuned large language model through LORA training,while the second one includes a Structural Encoder based on an attention-based graph convolutional neural network.
The main contribution of our method lies in the proposed CCA module. Unlike existing “featurefusion” methods, which simply concatenate feature vectors together to generate a fused representation,contrastive learning, which is used in our CCA model for representation disagreement alignment.This new module aligns LM inference with the topology of a graph,thus facilitating the identification of “semantic-structural inconsistencies”—cases when the node has normal connections, but its logical behavior is inconsistent with that.or even vice versa. We experimentally evaluate our approach using an extensive e-commerce payment dataset comprising of 2.84 million transactions [5].By showcasing how novel aspects of this system enable us to demonstrate that our CSSA framework is a powerful tool for processing text-attributed graphs which need both network structure and semantics-aware inference capability.

III. Methodology

The CSSA framework aims to learn a unified representation Z i for each node i by aligning its semantic representation Z i s e m and structural representation Z i s t r [11].

A. Semantic Logic Distillation (LLM Branch)

We model the semantics for each node (i.e., context) as an embedding that consists in a combination of features related to the node itself and past interactions containing such a node. We don’t employ standard embeddings,Instead, we turn to an LLM for a Reasoning Vector [12] and with a LoRA-extended version of our model train the embedding module:
Z i s e m =   MLP ( LLM ( P i ) )
The prompt explicitly tells the LLM to determine whether or not the facts about a given node are “logically consistent” with what it knows about that topic[13].

B. Structural Feature Encoding (GNN Branch)

Meanwhile, another GNN is applied to learn on the graph G =(V,E) . To avoid being distracted by useless edges,we propose a dynamic message passing scheme which learns the importance of edges in relation to nodes’ importance[14]:
h i l + 1   = o ( j N ( i ) { i } i j W l h j l )
where B is the mini-batch and τ is a temperature parameter that forces the network to find a joint embedding space in which semantically inconsistent examples are structurally inconsistent[15].

C. Joint Optimization

The output of the classifier is made through an ensemble decision:
y i = S o f t m a x ( M L P ( Z i s e m ) Z i s t r )
The final loss function is given by, L = , where is cross entropy loss.

IV. Data

A. Dataset Synthesis and Pre-processing

In order to evaluate the ability of our CSSA architecture to capture more complicated outliers and abnormalities, we use large scale time series data obtained through one of the largest online marketplaces operating on a worldwide basis; this provides detailed information about the sales transactions made by almost 2.84 million unique customers during a fortnight period.We represent our dataset by a multi-layered dynamic network denoted by G = (N,E,T), where N (~= 2030 nodes) corresponds to agents,consisting of individual users as well as organizations. The transaction relation E represents transactions between these two parties with the following feature matrix for each transaction: Transaction Features : Times when transactions took place to detect fast frauds.Description: Sector code, address history. Quantitative Variables: Price, frequency of fraud label. Response variable is the fraud label which takes value Y ∈ {0,1}), with the label y = 1 meaning that the transaction was proven to be fraudulent.
The dataset is also contaminated by frauds and presents an imbalance problem since the fraud samples are relatively rare (~0.22% of the data), which requires that the model be able to learn from few positives while being drown into many negatives:which is given in ref.[16].

B. Implementation and Training Protocol

We implement our CSSA model with PyTorch Geometric and HuggingFace Transformers libraries[17,18,19]. The overall structure of our model has two main modules:
Semantic Reasoning Layer: We choose ChatGLM3-6B to be our base reasoner[20,21,22]. In order to improve the financial background without losing too much ability due to full fine-tuning,we utilize LoRA where r = 16, α = 32; i.e., extracting “semantic risk indicator” information from natural-language descriptions about transactions.
Graph-based Embedding Layer: We utilize a three-layer GCNs with the size of 256 for each layer as graph representation module[23,24,25,26], in which we use batch norm and set drop rate as 0.3 to avoid over-fitting certain popular classes.
We train our joint model for 150 epochs with four V100-80G GPUs on a distributed computer system and we are careful not to over-fit because of scarce data while doing the contrastive learning.
In the latent space, to avoid overfitting of the alignment loss on the majority legitimate class category, we use an oversampling ratio of 5:1 (minority) [27,28,29,30].

V. Experiments

A. Comparative Performance Analysis

To comprehensively verify the superiority of the proposed CSSA model, we compare it with following baselines:
(1)
Standard GCN: a standard graph convolution method with only structure feature aggregation.
(2)
Graph attention network (GAT). A state-of-the-art structural modeling approach, which is able to learn the weight of a neighbor’s feature via self-attention mechanism.
(3)
Transformer on Tabular: Sequential model in which we consider the series of transactions as a set of features, learning their complicated interrelationships with multi-head attention.
(4)
LLM Zero-shot: The zero-shot LLM is a pure text-reasoning and non-structural-aware model with no finetune.
In Table 1, we summarize the results of analyzing our dataset to understand better how different approaches perform for crossmodal learning. First, it is clear that GCN and GAT have very low recalls:indicating that purely structural features may be insufficient to distinguish between complex fraud cases from legitimate large purchases.While Tabular Transformer performs better than GCN at learning feature correlations, it does not take into account the networkbased context of graphs.
CSSA achieves the highest values in terms of Macro-F1 (0.639) and PR-AUC (0.712), which are considerable improvements over our baseline model based on GAT. It also outperforms it significantly by a large margin in terms of Precision (0.953).suggesting that the semantic-structural fusion system acts like a very good filter.By rigorous mutual consensus between LLM’s reasoning capability and GN’s structure,We show that CSSA is able to effectively reduce the number of wrong classifications which often affect attention based methods when large amounts of information are present and there is a high amount of noise in the background.

B. Ablation Study

To better understand the impact of different modules in the proposed CSSA architecture and compare the contributions from semantical information with that from modality alignment, we further systematically remove each module for an ablation study, as shown in Table 2.
When we remove the contrastive alignment loss function () and keep both branches of models, i.e., “Concatenationonly”, the performance on Macro-F1 drops by 18.7%. It indicates that simple concatenation is not effective,and the model requires an explicit module for mapping between semantics and structure.When replacing the distilled LR of the LLM by ordinary number embedding (no LLM Reasoning),we see the worst performance degradation (-30.7%) which suggests, for the e-commerce fraud detection task,the context-based reasoning ability of the LLM is the most important feature element, which can be derived as follows:
Removing the GNN module (no GNN Structure), causes a sharp drop in Precision scores [20], which suggests that although LLMs are very good at detecting possibly fake stories,graph neural networks are crucial to test these suspicions with real transaction network data.

C. Explainability and Case Study

The CSSA method has a clear advantage in terms of explainability; that is, when predicting results, the explanation path will be provided by the language model part to each abnormal node found.For instance, for “Seller Coordination”, we found that the graph component identified a spike in interactions while the NLP component returned “mismatched activity times and location information from the seller.” Such evidence provides strong grounds on which one can rely as they manually verify the results.

VI. Conclusions

In this paper, we propose CSSA to bridge the gap between context-sensitive knowledge in large-scale LMs and topology-aware capabilities of GNs by casting AD as a cross-modal retrieval problem.our model avoids limitations of classic feature learning approaches, as well as rigid graph diffusion strategies.
Specifically, we mainly introduce a new concept named Cross-modal Contrastive Alignment (CCA), which aims at aligning semantic “inference routes” encoded in LLMs and topological structures learned through GNNs in terms of mathematical representation.Large-scale empirical evaluation conducted over real-world large e-commerce dataset shows that our proposed CSSA model can achieve very good trade-offbetween precision and recall,especially for the difficult cases with only 0.22% fraud transaction rates, we are able to detect them at an accuracy level of as high as 95.3% which proves that combining the semantic equivalence with relational structure provides much stronger indication about entities’ functioning than each of these sources alone.

References

  1. Dahiphale, D.; Madiraju, N.; Lin, J.; Karve, R.; Agrawal, M.; Modwal, A.; Merchant, A. Enhancing Trust and Safety in Digital Payments: An LLM-Powered Approach. 2024 IEEE International Conference on Big Data (BigData), 2024, December; IEEE; pp. 4854–4863. [Google Scholar]
  2. Ouyang, K.; Ke, Z.; Fu, S.; Liu, L.; Zhao, P.; Hu, D. Learn from global correlations: Enhancing evolutionary algorithm via spectral gnn. arXiv 2024, arXiv:2412.17629. [Google Scholar]
  3. Hacini, A. D.; Benabdelouahad, M.; Abassi, I.; Houhou, S.; Boulmerka, A.; Farhi, N. LLM-Assisted Financial Fraud Detection with Reinforcement Learning. Algorithms 2025, 18(12), 792. [Google Scholar] [CrossRef]
  4. Kanikanti, V. S. N.; Mula, K.; Muthukumarasamy, K.; Kubam, C. S.; Goswami, B.; Gadam, H. Streaming analytics pipelines for LLM-based financial anomaly detection in Real-Time retail transaction flows. International Journal of Information Technology 2026, 1–6. [Google Scholar]
  5. Luo, R.; Wang, N.; Zhu, X. Fraud detection and risk assessment of online payment transactions on e-commerce platforms based on llm and gcn frameworks. arXiv 2025, arXiv:2509.09928. [Google Scholar] [CrossRef]
  6. Sui, Mingxiu; Su, Yiyun; Shen, Jiaqing; et al. Intelligent Anti-Money Laundering on Cryptocurrency: A CNN-GNN Fusion Approach. Authorea 2026. [Google Scholar] [CrossRef]
  7. Bisht, K. S. Conversational Finance: LLM-Powered Payment Assistant Architecture. European Journal of Computer Science and Information Technology 2025, 13(27), 116–130. [Google Scholar] [CrossRef]
  8. Chen, Y.; Liu, L.; Fang, L. An Enhanced Credit Risk Evaluation by Incorporating Related Party Transaction in Blockchain Firms of China. Mathematics 2024, 12(17), 2673. [Google Scholar] [CrossRef]
  9. Chen, Lizi; Zou, Yue; Pan, Pengfei; et al. Cascading Credit Risk Assessment in Multiplex Supply Chain Networks. Authorea 2026. [Google Scholar] [CrossRef]
  10. Ke, Z.; Cao, Y.; Chen, Z.; Yin, Y.; He, S.; Cheng, Y. Early warning of cryptocurrency reversal risks via multi-source data. Finance Research Letters 2025, 107890. [Google Scholar] [CrossRef]
  11. Malingu, C. J.; Kabwama, C. A.; Businge, P.; Agaba, I. A.; Ankunda, I. A.; Mugalu, B.; Musinguzi, D. Application of LLMs to Fraud Detection. World J. Adv. Res. Rev 2025, 26, 178–183. [Google Scholar] [CrossRef]
  12. O'Neill, O.; Ramanayake, R.; Mandal, A.; Pawar, U.; Flanagan, W.; Chatbri, H.; Martin, C. A Practical Taxonomy for Finance-Specific LLM Risk Detection and Monitoring. NeurIPS 2025 Workshop: Generative AI in Finance.
  13. Wang, Qin; Huang, Bolin; Liu, Qianying. Deep Learning-Based Design Framework for Circular Economy Supply Chain Networks: A Sustainability Perspective. In Proceedings of the 2025 2nd International Conference on Digital Economy and Computer Science (DECS '25), Association for Computing Machinery, New York, NY, USA, 2026; pp. 836–840. [Google Scholar] [CrossRef]
  14. Artola Velasco, A.; Tsirtsis, E.; Okati, N.; Gomez Rodriguez, M. Is Your LLM Overcharging You? Tokenization, Transparency, and Incentives. arXiv 2025, arXiv:2505.21627. [Google Scholar] [CrossRef]
  15. Trozelli, P.; Andersson Holm, A. Comparing GPT and Traditional Machine Learning in Fraud Detection; 2024. [Google Scholar]
  16. Zheng, Haoran; Lin, Yuqing; He, Qi; et al. Blockchain Payment Fraud Detection with a Hybrid CNN-GNN-LSTM Model. Authorea 2026. [Google Scholar] [CrossRef]
  17. Bhatt, S.; Garg, G. NLP for Fraud Detection and Security in Financial Documents. In Transformative Natural Language Processing: Bridging Ambiguity in Healthcare, Legal, and Financial Applications; Springer Nature Switzerland: Cham, 2025; pp. 131–155. [Google Scholar]
  18. Yuan, Keyu; Lin, Yuqing; Wu, Wenjun; et al. Detection of Blockchain Online Payment Fraud Via CNN-LSTM. Authorea 2026. [Google Scholar] [CrossRef]
  19. Rollinson, N.; Polatidis, N. LLM-Generated Samples for Android Malware Detection. Digital 2026, 6(1), 5. [Google Scholar] [CrossRef]
  20. Li, Z.; Ke, Z. RepoLLM: A multi-modal foundation model for drug repurposing via alignment of molecules, EHRs, and knowledge graphs. ICML 2025 Workshop on Multi-modal Foundation Models and Large Language Models for Life Sciences. 2025. Available online: https://openreview.net/forum?id=adweKg5NZa.
  21. Yao, Z.; Zhu, Q. A Two-Stage Wiener Process-Based Method for Bearing Remaining Useful Life Prediction Considering Measurement Uncertainty. In Measurement Science and Technology; 2026. [Google Scholar] [CrossRef]
  22. Li, Long; Hao, Jiaran; Klein Liu, Jason; Zhou, Zhijian; Miao, Yanting; Pang, Wei; Tan, Xiaoyu; et al. "The choice of divergence: A neglected key to mitigating diversity collapse in reinforcement learning with verifiable reward. arXiv 2025, arXiv:2509.07430. [Google Scholar]
  23. Wang, Chengkai; Wu, Di; Liao, Yunsheng; Zheng, Wenyao; Zeng, Ziyi; Gao, Xurong; Wu, Hemmings; et al. "NeuroCLIP: A Multimodal Contrastive Learning Method for rTMS-treated Methamphetamine Addiction Analysis.". arXiv 2025, arXiv:2507.20189. [Google Scholar]
  24. Liu, S.; Zhu, M. In-trajectory inverse reinforcement learning: Learn incrementally before an ongoing trajectory terminates. Advances in Neural Information Processing Systems 2024, 37, 117164–117209. [Google Scholar]
  25. Shi, Ge; Sun, Lin; An, Quan; Tang, Lei; Shi, Jiantao; Chen, Chuang; Feng, Lihang; Ma, Hongyang. Quantifying Urban Park Cooling Effects and Tri-Factor Synergistic Mechanisms: A Case Study of Nanjing’s Central Districts. Systems 2026, 14(no. 2), 130. [Google Scholar] [CrossRef]
  26. Wei, D.; Wang, Z.; Kang, H.; Sha, X.; Xie, Y.; Dai, A.; Ouyang, K. A comprehensive analysis of digital inclusive finance’s influence on high quality enterprise development through fixed effects and deep learning frameworks. Scientific Reports 2025, 15(1), 30095. [Google Scholar] [CrossRef] [PubMed]
  27. Li, F.; Xu, Q.; Bao, S.; Han, B.; Yang, Z.; Huang, Q. Hybrid generative fusion for efficient and privacy-preserving face recognition dataset generation. In Proceedings of the IEEE/CVF International Conference on Computer Vision, 2025; pp. 495–501. [Google Scholar]
  28. Wu, S.; Zhang, J. Spatiotemporal multi-view continual dictionary learning with graph diffusion. Knowledge-Based Systems 2025, 316, 113388. [Google Scholar] [CrossRef]
  29. Min, M.; Duan, J.; Liu, M.; Wang, N.; Zhao, P.; Zhang, H.; Han, Z. Task Offloading with Differential Privacy in Multi-Access Edge Computing: An A3C-Based Approach. In IEEE Transactions on Cognitive Communications and Networking; 2026. [Google Scholar]
  30. Liu, S.; Zhu, M. Learning multi-agent behaviors from distributed and streaming demonstrations. Advances in Neural Information Processing Systems 2023, 36, 53552–53564. [Google Scholar]
  31. Li, F.; Xu, Q.; Bao, S.; Yang, Z.; Cao, X.; Huang, Q. One image is worth a thousand words: A usability preservable text-image collaborative erasing framework. arXiv 2025, arXiv:2505.11131. [Google Scholar] [CrossRef]
  32. Wang, Zhi; Gong, Tao; Nian, Yingpu; Yi, Bo; Wang, Xingwei; Zhou, Xinhao; Lv, Jianhui; Min, Geyong; Li, Keqin. SFTRAP: Satisfying Fidelity Threshold Routing and Adaptive Purification for Throughput Maximum in Quantum Network. IEEE Transactions on Communications 2026, 74, 3952–3967. [Google Scholar]
  33. Guan, Z.; Cao, H.; Zhong, M.; Yang, E.; Ai, L.; Ni, Y.; Shi, B. Symphony-Coord: Emergent Coordination in Decentralized Agent Systems. arXiv 2026, arXiv:2602.00966. [Google Scholar] [CrossRef]
  34. Tian, J.; Wang, Z. DLRREC: Denoising Latent Representations via Multi-Modal Knowledge Fusion in Deep Recommender Systems. arXiv 2025, arXiv:2512.00596. [Google Scholar] [CrossRef]
  35. Ning, Qirui; Zhang, Jinlai; Xie, Yuhang; Liu, Kaifeng; Gao, Kai; Chen, Bin; Chen, Gengbiao; Fan, Qing; Liu, Hui; Du, Ronghua. Multi-Resolution Context Augmentation and Dual Channel Attention for 3D lane detection. IEEE Internet of Things Journal, 2025. [Google Scholar]
  36. Wang, J.; Lin, C.; Nie, L.; Liao, K.; Shao, S.; Zhao, Y. Digging into contrastive learning for robust depth estimation with diffusion models. In Proceedings of the 32nd ACM International Conference on Multimedia, 2024, October; pp. 4129–4137. [Google Scholar]
  37. Wang, Chengkai; Zhang, Yifan; Wu, Chengyu; Liu, Jun; Huang, Xingliang; Wu, Liuxi; Wang, Yitong; Feng, Xiang; Lu, Yiting; Wang, Yaqi. MMDental-A multimodal dataset of tooth CBCT images with expert medical records. Scientific Data 2025, 12(no. 1), 1172. [Google Scholar] [CrossRef] [PubMed]
  38. Wang, Ke; Zhao, Zishuo; Song, Xinyuan; Shi, Bill; Xia, Libin; Tong, Chris; Ai, Lynn; Qu, Felix; Yang, Eric. VeriLLM: A Lightweight Framework for Publicly Verifiable Decentralized Inference. arXiv 2025, arXiv:2509.24257. [Google Scholar] [CrossRef]
  39. Li, M.; Zhang, D.; Dong, Q.; Xie, X.; Qin, K. Adaptive dataset quantization. Proceedings of the AAAI Conference on Artificial Intelligence 2025, Vol. 39(No. 11), 12093–12101. [Google Scholar] [CrossRef]
  40. Yu, Q.; Ke, Z.; Xiong, G.; Cheng, Y.; Guo, X. Identifying money laundering risks in digital asset transactions based on ai algorithms. 2024 4th International Conference on Electronic Information Engineering and Computer Communication (EIECC), 2024, December; IEEE; pp. 1081–1085. [Google Scholar]
  41. Zheng, Y.; Zhong, B.; Liang, Q.; Li, N.; Song, S. Decoupled spatio-temporal consistency learning for self-supervised tracking. Proceedings of the AAAI Conference on Artificial Intelligence 2025, Vol. 39(No. 10), 10635–10643. [Google Scholar] [CrossRef]
  42. Li, T.; Sun, W. MLP-SLAM: Multilayer Perceptron-Based Simultaneous Localization and Mapping. arXiv 2024, arXiv:2410.10669. [Google Scholar]
  43. Shao, Z.; Zhou, Y.; Li, F.; Zhu, H.; Liu, B. Joint facial action unit recognition and self-supervised optical flow estimation. Pattern Recognition Letters 2024, 181, 70–76. [Google Scholar] [CrossRef]
  44. Peng, Yong; Gu, Shaowei; Wu, Guohua; Liang, Yunbin; Ouyang, Kaichen; Liang, Xifeng; Wang, Kui; Fan, Chaojie. A novel plug-and-play meta-black-box optimization module based on video streams for non-contact physiological signal extraction. Swarm and Evolutionary Computation 2026, 102, 102336. [Google Scholar] [CrossRef]
  45. Zhang, Ye; Chen, Qi; Gao, Longsen; Liu, Rui; Chu, Linyue; Mo, Kangtong; Kang, Zhengjian; Huang, Wenyou; Zhang, Xingyu. Invertible liquid neural network-based learning of inverse kinematics and dynamics for robotic manipulators. Scientific Reports 2025, 15(no. 1), 42311. [Google Scholar] [CrossRef] [PubMed]
  46. Liu, S.; Xu, S.; Qiu, W.; Zhang, H.; Zhu, M. Explainable reinforcement learning from human feedback to improve alignment. arXiv 2025, arXiv:2512.13837. [Google Scholar] [CrossRef]
  47. Wang, Jiyuan; Lin, Chunyu; Guan, Cheng; Nie, Lang; He, Jing; Li, Haodong; Liao, Kang; Zhao, Yao. "Jasmine: Harnessing diffusion prior for self-supervised depth estimation.". arXiv 2025, arXiv:2503.15905. [Google Scholar] [CrossRef]
  48. Shao, Z.; Li, F.; Zhou, Y.; Chen, H.; Zhu, H.; Yao, R. Identity-invariant representation and transformer-style relation for micro-expression recognition. Applied Intelligence 2023, 53(17), 19860–19871. [Google Scholar] [CrossRef]
  49. Feng, Tongtong; Wang, Xin; Zhou, Zekai; Wang, Ren; Zhan, Yuwei; Li, Guangyao; Li, Qing; Zhu, Wenwu. Evoagent: Agent autonomous evolution with continual world model for long-horizon tasks. arXiv 2025, e-prints, arXiv–2502. [Google Scholar]
  50. Tian; Xuxiang, Aaron; Zhang, Ruofan; Tang, Jiayao; Cho, Young Min; Li, Xueqian; Yi, Qiang; Wang, Ji; et al. "Beyond the Strongest LLM: Multi-Turn Multi-Agent Orchestration vs. Single LLMs on Benchmarks.". arXiv 2025, arXiv:2509.23537. [Google Scholar]
  51. Jiang, Jiantong; Yang, Peiyu; Zhang, Rui; et al. Towards Efficient Large Language Model Serving: A Survey on System-Aware KV Cache Optimization. TechRxiv 2025. [Google Scholar] [CrossRef]
  52. Wu, Jiachun; Zhang, Jinlai; Zhu, Jihong; Duan, Yijian; Fang, Youyang; Zhu, Jingyu; Yin, Lairong; et al. "Multi-scale convolution and dynamic task interaction detection head for efficient lightweight plum detection. Food and Bioproducts Processing 2025, 149, 353–367. [Google Scholar] [CrossRef]
  53. Li, M.; Gou, H.; Zhang, D.; Liang, S.; Xie, X.; Ouyang, D.; Qin, K. Beyond Random: Automatic Inner-loop Optimization in Dataset Distillation. arXiv 2025, arXiv:2510.04838. [Google Scholar] [CrossRef]
  54. Pan, P.; Chen, L.; He, Q.; Yuan, K.; Wang, H.; Zhang, W. FinSCRA: An LLM-Powered Multi-Chain Reasoning Framework for Interpretable Node Classification on Text-Attributed Graphs. Preprints 2026. [Google Scholar] [CrossRef]
  55. Zhang, Y.; Liu, R.; Chen, Q.; Mo, K.; Chu, L.; Kang, Z. SDARL: Safe Deep Adaptive Representation Learning for High-DoF Non-Linear System Manipulation in Space. In IEEE Access; 2025. [Google Scholar]
  56. Ke, Zong; Shen, Jiaqing; Zhao, Xuanyi; Fu, Xinghao; Wang, Yang; Li, Zichao; Liu, Lingjie; Mu, Huailing. A stable technical feature with GRU-CNN-GA fusion. Applied Soft Computing 2025, 114302. [Google Scholar] [CrossRef]
  57. Wang, Q.; Tsai, W. T.; Shi, T.; Tang, W.; Du, B. Hide and seek in transaction networks: a multi-agent framework for simulating and detecting money laundering activities. Complex & Intelligent Systems 2025, 11(6), 271. [Google Scholar] [CrossRef]
  58. Xu, D. M.; Hu, X. X.; Wang, W. C.; Wang, J.; Shi, C. C.; Zang, H. F. Innovative daily runoff prediction model integrating black-winged kite algorithm and Mamba2–Transformer architecture. Ecological Informatics 2025, 103565. [Google Scholar] [CrossRef]
  59. Qu, Y.; Panariello, M.; Todisco, M.; Evans, N. Reference-free adversarial sex obfuscation in speech. 2025 Asia Pacific Signal and Information Processing Association Annual Summit and Conference (APSIPA ASC), 2025, October; IEEE; pp. 2128–2133. [Google Scholar]
  60. Liu, S.; Zhu, M. Distributed inverse constrained reinforcement learning for multi-agent systems. Advances in Neural Information Processing Systems 2022, 35, 33444–33456. [Google Scholar]
  61. Wang, T. L.; Gu, S. W.; Liu, R. J.; Chen, L. Q.; Wang, Z.; Zeng, Z. Q. Cuckoo catfish optimizer: a new meta-heuristic optimization algorithm. Artificial Intelligence Review 2025, 58(10), 326. [Google Scholar] [CrossRef]
  62. Liu, Z.; Qian, S.; Cao, S.; Shi, T. Mitigating age-related bias in large language models: Strategies for responsible artificial intelligence development. INFORMS Journal on Computing 2025. [Google Scholar] [CrossRef]
  63. Qu, Y.; Fu, D.; Fan, J. Subject Information Extraction for Novelty Detection with Domain Shifts. arXiv 2025, arXiv:2504.21247. [Google Scholar] [CrossRef]
  64. Li, M.; Zhang, D.; He, T.; Xie, X.; Li, Y. F.; Qin, K. Towards effective data-free knowledge distillation via diverse diffusion augmentation. In Proceedings of the 32nd ACM International Conference on Multimedia, 2024, October; pp. 4416–4425. [Google Scholar]
  65. Wang, Junqiao; Zhang, Zeng; He, Yangfan; Zhang, Zihao; Song, Xinyuan; Song, Yuyang; Shi, Tianyu; et al. Enhancing code llms with reinforcement learning in code generation: A survey. arXiv 2024, arXiv:2412.20367. [Google Scholar]
  66. Liu, S.; Zhu, M. Meta inverse constrained reinforcement learning: Convergence guarantee and generalization analysis. The Twelfth International Conference on Learning Representations, 2023. [Google Scholar]
  67. Zhang, P. Boosting Learning Efficiency in Few-Shot Tasks With Layer-Adaptive PID Control. IEEE Transactions on Pattern Analysis and Machine Intelligence. [CrossRef]
  68. Zheng, Y.; Zhong, B.; Liang, Q.; Mo, Z.; Zhang, S.; Li, X. Odtrack: Online dense temporal token learning for visual tracking. Proceedings of the AAAI conference on artificial intelligence 2024, Vol. 38(No. 7), 7588–7596. [Google Scholar] [CrossRef]
  69. Tian, J.; Wang, Z.; Zhao, J.; Ding, Z. Mmrec: Llm based multi-modal recommender system. 2024 19th International Workshop on Semantic and Social Media Adaptation & Personalization (SMAP), 2024, November; IEEE; pp. 105–110. [Google Scholar]
  70. Wang, JiYuan; Lin, Chunyu; Sun, Lei; Liu, Rongying; Nie, Lang; Li, Mingxing; Liao, Kang; Chu, Xiangxiang; Zhao, Yao. From editor to dense geometry estimator. arXiv 2025, arXiv:2509.04338. [Google Scholar] [CrossRef]
  71. Yi, Qiang; He, Yangfan; Wang, Jianhui; Song, Xinyuan; Qian, Shiyao; Yuan, Xinhang; Xin, Yi; et al. "Score: Story coherence and retrieval enhancement for ai narratives.". arXiv 2025, arXiv:2503.23512. [Google Scholar] [CrossRef]
  72. Zhao, Peng; Liu, Xiaoyu; Su, Xuqi; Wu, Di; Li, Zi; Kang, Kai; Li, Keqin; Zhu, Armando. "Probabilistic Contingent Planning Based on Hierarchical Task Network for High-Quality Plans.". Algorithms 2025, 18(no. 4), 214. [Google Scholar] [CrossRef]
  73. Tang, J.; Zhang, W.; Liu, H.; Yang, M.; Jiang, B.; Hu, G.; Bai, X. Few could be better than all: Feature sampling and grouping for scene text detection. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 2022; pp. 4563–4572. [Google Scholar]
  74. Zhuang, J.; Jin, H.; Zhang, Y.; Kang, Z.; Zhang, W.; Dagher, G. G.; Wang, H. Exploring the vulnerability of the content moderation guardrail in large language models via intent manipulation. arXiv 2025, arXiv:2505.18556. [Google Scholar] [CrossRef]
  75. Kang, Zhaolu; Gong, Junhao; Yan, Jiaxu; Xia, Wanke; Wang, Yian; Wang, Ziwen; Ding, Huaxuan; et al. "Hssbench: Benchmarking humanities and social sciences ability for multimodal large language models.". arXiv 2025, arXiv:2506.03922. [Google Scholar] [CrossRef]
  76. Liu, S.; Zhu, M. Utility: Utilizing explainable reinforcement learning to improve reinforcement learning. The Thirteenth International Conference on Learning Representations, 2025. [Google Scholar]
  77. Wan, W.; Zhou, F.; Liu, L.; Fang, L.; Chen, X. Ownership structure and R&D: The role of regional governance environment. International Review of Economics & Finance 2021, 72, 45–58. [Google Scholar] [CrossRef]
  78. Peng, Yong; Gu, Shaowei; Liang, Yunbin; Ouyang, Kaichen; Li, Yingli; Wang, Kui; Wu, Guohua; Fan, Chaojie. Wave Optics Optimizer: A novel meta-heuristic algorithm for engineering optimization. Communications in Nonlinear Science and Numerical Simulation 2025, 109337. [Google Scholar] [CrossRef]
  79. Lu, Yao; Yang, Wen; Zhang, Yunzhe; Chen, Zuohui; Chen, Jinyin; Xuan, Qi; Wang, Zhen; Yang, Xiaoniu. Understanding the dynamics of dnns using graph modularity. European Conference on Computer Vision, 2022; Springer Nature Switzerland: Cham; pp. 225–242. [Google Scholar]
  80. Zheng, Y.; Zhong, B.; Liang, Q.; Li, G.; Ji, R.; Li, X. Toward unified token learning for vision-language tracking. IEEE Transactions on Circuits and Systems for Video Technology 2023, 34(4), 2125–2135. [Google Scholar] [CrossRef]
  81. Xiong, Jing; Li, Zixuan; Zheng, Chuanyang; Guo, Zhijiang; Yin, Yichun; Xie, Enze; Yang, Zhicheng; et al. Dq-lore: Dual queries with low rank approximation re-ranking for in-context learning. arXiv 2023, arXiv:2310.02954. [Google Scholar]
  82. Xiao, C.; Hou, L.; Fu, L.; Chen, W. Diffusion-based self-supervised imitation learning from imperfect visual servoing demonstrations for robotic glass installation. 2025 IEEE International Conference on Robotics and Automation (ICRA), 2025, May; IEEE; pp. 10401–10407. [Google Scholar]
  83. Shi, Chuancheng; Li, Shangze; Guo, Shiming; Xie, Simiao; Wu, Wenhua; Dou, Jingtong; Wu, Chao; et al. "Where Culture Fades: Revealing the Cultural Gap in Text-to-Image Generation.". arXiv 2025, arXiv:2511.17282. [Google Scholar]
  84. Chen, X.; Xiao, C.; Cao, W.; Zhang, W.; Liu, Y. Framework and pathway for the construction of a unified data-element market in china. Strategic Study of Chinese Academy of Engineering 2025, 27(1), 40–50. [Google Scholar]
  85. Yao, J.; Li, C.; Xiao, C. Swift sampler: Efficient learning of sampler by 10 parameters. Advances in Neural Information Processing Systems 2024, 37, 59030–59053. [Google Scholar]
  86. Xu, Shixiong; Chen, Naxi; Pan, Pengfei; et al. Financial Anomaly Transaction Detection Using Autoencoder-Based Models; Authorea, 23 February 2026. [Google Scholar] [CrossRef]
Table 1. Performance comparison of CSSA against others.
Table 1. Performance comparison of CSSA against others.
Model Architecture Precision Recall Macro-F1 PR-AUC
Vanilla GCN 0.654 0.122 0.206 0.315
GAT (Self-Attention) 0.712 0.185 0.294 0.388
Tabular Transformer 0.685 0.254 0.371 0.402
LLM-only (ChatGLM3) 0.451 0.382 0.414 0.42
CSSA (Ours) 0.953 0.481 0.639 0.712
Table 2. Ablation study results demonstrating the impact of LLM reasoning and Cross-modal Alignment.
Table 2. Ablation study results demonstrating the impact of LLM reasoning and Cross-modal Alignment.
Configuration Precision Recall Macro-F1 Δ F1
Full CSSA Framework 0.953 0.481 0.639 -
w/o Cross-modal Alignment (L_align) 0.824 0.312 0.452 -18.70%
w/o LLM Reasoning (Raw Embeddings) 0.745 0.214 0.332 -30.70%
w/o GNN Structure (LLM-only) 0.451 0.382 0.414 -22.50%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.
Copyright: This open access article is published under a Creative Commons CC BY 4.0 license, which permit the free download, distribution, and reuse, provided that the author and preprint are cited in any reuse.
Prerpints.org logo

Preprints.org is a free preprint server supported by MDPI in Basel, Switzerland.

Subscribe

Disclaimer

Terms of Use

Privacy Policy

Privacy Settings

© 2026 MDPI (Basel, Switzerland) unless otherwise stated