Submitted:
01 June 2025
Posted:
03 June 2025
You are already at the latest version
Abstract
Keywords:
1. Introduction
2. Related Work
3. Methodology
3.1. Reinforcement Learning Component
3.1.1. State and Action Definition
3.1.2. Reward Function
3.1.3. Deep Q-Network (DQN) Architecture
3.1.4. Training Process for the RL Agent
- Randomly initialize Q-network weights.
- Observe state , select and execute action .
- Record next state and reward .
- Store the transition in the replay buffer.
- Sample mini-batches and compute the loss.
- Apply gradient descent to update .
- Periodically sync the target network for stability.
3.2. Prediction Network for Load Forecasting
3.2.1. LSTM Encoder
3.2.2. Transformer Encoder
3.2.3. Output of the Prediction Network
3.3. Loss Function
3.3.1. Reinforcement Learning Loss
3.3.2. Prediction Loss
3.3.3. Total Loss Function
3.4. Data Preprocessing
3.4.1. Normalization of Task Data
3.4.2. Time-Series Transformation
3.4.3. Feature Engineering for Resource Utilization
4. Evaluation Metrics
4.1. Evaluation Metrics Description
5. Experiment Results
6. Conclusion
References
- Wang, L.; Wu, J.; Gao, Y.; Zhang, J. Deep reinforcement learning based resource allocation for cloud native wireless network. arXiv 2023, arXiv:2305.06249. [Google Scholar] [CrossRef]
- Rossi, A.; Visentin, A.; Carraro, D.; Prestwich, S.; Brown, K.N. Forecasting workload in cloud computing: towards uncertainty-aware predictions and transfer learning. Cluster Computing 2025, 28, 258. [Google Scholar] [CrossRef]
- Arbat, S.; Jayakumar, V.K.; Lee, J.; Wang, W.; Kim, I.K. Wasserstein adversarial transformer for cloud workload prediction. In Proceedings of the Proceedings of the AAAI Conference on Artificial Intelligence; 2022; Volume 36, pp. 12433–12439. [Google Scholar]
- Jin, T. Integrated machine learning for enhanced supply chain risk prediction. In Proceedings of the Proceedings of the 2024 8th International Conference on Electronic Information Technology and Computer Engineering; 2024; pp. 1254–1259. [Google Scholar]
- Wang, E. Hybrid FM-GCN-Attention Model for Personalized Recommendation. In Proceedings of the 2025 International Conference on Electrical Automation and Artificial Intelligence (ICEAAI). IEEE; 2025; pp. 1307–1310. [Google Scholar]
- Chen, X. Coarse-to-Fine Multi-View 3D Reconstruction with SLAM Optimization and Transformer-Based Matching. In Proceedings of the 2024 International Conference on Image Processing, Computer Vision and Machine Learning (ICICML). IEEE; 2024; pp. 855–859. [Google Scholar]
- Zhou, G.; Tian, W.; Buyya, R.; Xue, R.; Song, L. Deep reinforcement learning-based methods for resource scheduling in cloud computing: A review and future directions. Artificial Intelligence Review 2024, 57, 124. [Google Scholar] [CrossRef]
- Gu, Y.; Liu, Z.; Dai, S.; Liu, C.; Wang, Y.; Wang, S.; Theodoropoulos, G.; Cheng, L. Deep Reinforcement Learning for Job Scheduling and Resource Management in Cloud Computing: An Algorithm-Level Review. arXiv 2025, arXiv:2501.01007. [Google Scholar]
- Zhao, X.; Wu, C. Large-scale machine learning cluster scheduling via multi-agent graph reinforcement learning. IEEE Transactions on Network and Service Management 2021, 19, 4962–4974. [Google Scholar] [CrossRef]



| Model | Std. Dev. of Load | Avg. Response Timeout | Avg. Efficiency | Resource Utilization |
|---|---|---|---|---|
| FCFS | 15.2 | 35.4 | 0.78 | 70% |
| RR | 12.5 | 28.3 | 0.82 | 75% |
| Min-Min | 10.1 | 22.1 | 0.86 | 80% |
| DynaSched-Net | 8.2 | 19.7 | 0.92 | 90% |
| RL only | 9.3 | 22.5 | 0.85 | 85% |
| Prediction only | 11.8 | 27.3 | 0.80 | 78% |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).