Submitted:
22 July 2025
Posted:
23 July 2025
You are already at the latest version
Abstract
Keywords:
I. Introduction
II. Related Work and Foundation
III. Method
VI. Experimental Results
A. Dataset
B. Experimental Results
V. Conclusion
References
- Y. Gu, L. Dong, F. Wei, et al., "MiniLLM: knowledge distillation of large language models," arXiv preprint, arXiv:2306.08543, 2023.
- X. Xu, M. Li, C. Tao, et al., "A survey on knowledge distillation of large language models," arXiv preprint, arXiv:2402.13116, 2024.
- C. Yang, Y. Zhu, W. Lu, et al., "Survey on knowledge distillation for large language models: methods, evaluation, and application," ACM Transactions on Intelligent Systems and Technology, 2024.
- S. Muralidharan, S. T. Sreenivas, R. Joshi, et al., "Compact language models via pruning and knowledge distillation," Advances in Neural Information Processing Systems, vol. 37, pp. 41076–41102, 2024.
- J. Liu, C. Zhang, J. Guo, et al., "DDK: distilling domain knowledge for efficient large language models," Advances in Neural Information Processing Systems, vol. 37, pp. 98297–98319, 2024.
- C. Liu, C. Tao, J. Feng, et al., "Multi granularity structural knowledge distillation for language model compression," Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pp. 1001–1011, 2022.
- L. Li, Y. Lin, S. Ren, et al., "Dynamic knowledge distillation for pre trained language models," arXiv preprint, arXiv:2109.11295, 2021.
- A. Kai, L. Zhu, and J. Gong, "Efficient compression of large language models with distillation and fine tuning," Journal of Computer Science and Software Applications, vol. 3, no. 4, pp. 30–38, 2023.
- H. Zheng, Y. Wang, R. Pan, G. Liu, B. Zhu, and H. Zhang, "Structured gradient guidance for few-shot adaptation in large language models," arXiv preprint, arXiv:2506.00726, 2025.
- H. Zheng, Y. Ma, Y. Wang, G. Liu, Z. Qi, and X. Yan, "Structuring low-rank adaptation with semantic guidance for model fine-tuning," 2025.
- X. Liu, Y. Qin, Q. Xu, Z. Liu, X. Guo, and W. Xu, "Integrating knowledge graph reasoning with pretrained language models for structured anomaly detection," 2025.
- Y. Xing, T. Yang, Y. Qi, M. Wei, Y. Cheng, and H. Xin, "Structured memory mechanisms for stable context representation in large language models," arXiv preprint, arXiv:2505.22921, 2025.
- Y. Peng, "Structured knowledge integration and memory modeling in large language systems," Transactions on Computational and Scientific Methods, vol. 4, no. 10, 2024.
- X. Wang, "Time-aware and multi-source feature fusion for transformer-based medical text analysis," Transactions on Computational and Scientific Methods, vol. 4, no. 7, 2024.
- L. Zhu, F. Guo, G. Cai, and Y. Ma, "Structured preference modeling for reinforcement learning-based fine-tuning of large models," Journal of Computer Technology and Software, vol. 4, no. 4, 2025.
- H. Zhang, Y. Ma, S. Wang, G. Liu, and B. Zhu, "Graph-based spectral decomposition for parameter coordination in language model fine-tuning," arXiv preprint, arXiv:2504.19583, 2025.
- J. He, G. Liu, B. Zhu, H. Zhang, H. Zheng, and X. Wang, "Context-guided dynamic retrieval for improving generation quality in RAG models," arXiv preprint, arXiv:2504.19436, 2025.
- Y. Peng, "Context-aligned and evidence-based detection of hallucinations in large language model outputs," Transactions on Computational and Scientific Methods, vol. 5, no. 6, 2025.
- Y. Deng, "Transfer methods for large language models in low-resource text generation tasks," Journal of Computer Science and Software Applications, vol. 4, no. 6, 2024.
- R. Wang, "Joint semantic detection and dissemination control of phishing attacks on social media via LLama-based modeling," arXiv preprint, arXiv:2504.00282, 2025.
- F. Guo, L. Zhu, Y. Wang, and G. Cai, "Perception-guided structural framework for large language model design," Journal of Computer Technology and Software, vol. 4, no. 5, 2025.
- Z. Fang, "A deep learning-based predictive framework for backend latency using AI-augmented structured modeling," Journal of Computer Technology and Software, vol. 3, no. 7, 2024.
- X. Jiao, Y. Yin, L. Shang, et al., "TinyBERT: distilling BERT for natural language understanding," arXiv preprint, arXiv:1909.10351, 2019.
- Z. Sun, H. Yu, X. Song, et al., "MobileBERT: a compact task-agnostic BERT for resource-limited devices," arXiv preprint, arXiv:2004.02984, 2020.
- W. Wang, F. Wei, L. Dong, et al., "MiniLM: deep self-attention distillation for task-agnostic compression of pre-trained transformers," Advances in Neural Information Processing Systems, vol. 33, pp. 5776–5788, 2020.
- B. Zhao, Q. Cui, R. Song, et al., "Decoupled knowledge distillation," Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 11953–11962, 2022.



Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).