Submitted:
08 July 2025
Posted:
11 July 2025
You are already at the latest version
Abstract
Keywords:
1. Introduction
1.1. Contributions
2. Related Work
2.1. Temporal Graph Forecasting
2.2. Dynamic Community Detection
2.3. Uncertainty-Aware Graph Learning
3. Methodology
3.1. Pipeline Overview

3.2. Data Preprocessing
3.3. Bayesian Embedding
3.4. Prototype-Guided Louvain Clustering
3.5. Markov Transition Modeling
3.6. Reinforcement-Based Refinement
3.7. Loss Function & Optimisation
3.8. Computational Complexity
3.8.1. Bayesian Encoder (dominant)
3.8.2. Prototype Louvain
3.8.3. Markov Update
3.8.4. PPO Refinement
3.8.5. Total
3.9. Memory
3.10. Algorithm
| Algorithm 1 MaGNet-BN — Unified Training Procedure |
|
4. Evaluation
4.1. Datasets
| Dataset | Domain | (avg) | T | ||
|---|---|---|---|---|---|
| METR–LA | Traffic | 207 | 1 515 | 34 272 | 1 h |
| PeMS–BAY | Traffic | 325 | 2 694 | 52 560 | 1 h |
| TwitterRC | Social | 22 938 | 98 421 | 2 160 | 1 h |
| Enron-Email | 150 028 | 347 653 | 194 | 168 h | |
| ETH+UCY | Mobility | 1 536† | 12 430 | 3 588 | 12 s |
| ELD-2012 | Energy | 370 | 4 862 | 8 760 | 1 h |
| M5-Retail | Retail | 3 049 | 11 216 | 1 941 | 24 h |
- METR-LA and PeMS-BAY—minute-level road-traffic speeds from loop detectors in Los Angeles and the Bay Area ( and snapshots) [12].
- TwitterRC—2 160 hourly snapshots of retweet/mention interactions among 22 938 users [25].
- Enron-Email—194 weekly snapshots of corporate e-mail exchanges (150 028 nodes) [26].
- ETH+UCY—3 588 twelve-second pedestrian-interaction graphs recorded in public scenes [27].
- ELD-2012—8 760 hourly power-consumption graphs (370 smart-meter clients) extracted from the ElectricityLoadDiagrams20112014 dataset [28].
- M5-Retail—1 941 daily sales-correlation graphs covering 3 049 Walmart items [29].
4.2. Baselines
4.2.1. Sequence-Forecasting Baselines (Graph-Agnostic)
- DeepAR [21]. Autoregressive LSTM with Gaussian output quantiles; a de-facto standard for univariate/multivariate probabilistic forecasting. Why: sets the reference point for purely temporal models without spatial bias.
- MC-Drop LSTM [7]. Injects dropout at inference to sample from the weight posterior—simple yet strong Bayesian baseline. Why: isolates the benefit of explicit epistemic uncertainty without graph information.
- Temporal Fusion Transformer (TFT) [10]. Multi-head attention, static covariates and gating; current SOTA on many time-series leaderboards. Why: strongest recent non-graph forecaster.
- DCRNN [12]. Diffusion convolution on a fixed sensor adjacency, followed by seq2seq GRU. Why: canonical example of static-graph-aware spatio–temporal forecasting.
4.2.2. Dynamic-Graph Baselines
- DySAT [22]. Self-attention across structural and temporal dimensions; acquires snapshot-specific embeddings. Why: early but influential method; serves as the “attention-without-memory’’ extreme.
- TGAT [18]. Time-encoding kernels plus graph attention, enabling continuous-time message passing. Why: tests whether high-resolution event timing alone suffices for our coarse snapshot setting.
- TGN [25]. Memory modules store node histories and are updated by temporal messages; often SOTA on link prediction. Why: strongest publicly available dynamic-GNN with memory.
| Model | Temporal Forecast | Dynamic Graph | Uncertainty / RL |
|---|---|---|---|
| DeepAR | √ | — | Gaussian output |
| MC-Drop LSTM | √ | — | MC dropout |
| TFT | √ | — | Attention ensembles |
| DCRNN | √ | static | — |
| DySAT | — | √ | — |
| TGAT | — | √ | Time encoding |
| TGN | √ | √ | Memory, attention |
4.3. Implementation Details
4.3.1. Snapshot Construction
- METR-LA, PeMS-BAY: (minute-level, 12min horizon)
- TwitterRC: (1-hour bins, one-day horizon)
- Enron-Email: (weekly bins, one-month horizon)
- ETH+UCY: (12s bins, 96s horizon)
- ELD-2012: (hourly bins, 4-day horizon)
- M5-Retail: (daily bins, 8-week horizon)
4.3.2. Model Hyper-Parameters
4.3.3. Runtime.
4.4. Evaluation Metrics
4.4.1. Predictive Accuracy
- Mean Squared Error (MSE) [30]
- Negative Log-Likelihood (NLL) [31]
- Continuous Ranked Probability Score (CRPS) [32]
- Prediction-Interval Coverage Probability (PICP) [33]
| Dataset | Metric | 11 | 13 | 17 | 19 | 23 | mean ± 95 % CI |
|---|---|---|---|---|---|---|---|
| METR-LA | MSE | 0.224 | 0.218 | 0.229 | 0.222 | 0.220 | |
| NLL | 1.503 | 1.479 | 1.511 | 1.488 | 1.492 | ||
| PeMS-BAY | MSE | 0.190 | 0.188 | 0.194 | 0.189 | 0.192 | |
| NLL | 1.557 | 1.543 | 1.564 | 1.551 | 1.555 |
4.4.2. How We form the ± 95 % Confidence Interval
- (i)
- Sample mean : .
- (ii)
- Unbiased st. dev. : .
- (iii)
- Half-width for a 95 % CI :
4.4.3. Worked Example (ETH+UCY, MaGNet-BN, MSE)
4.4.4. Significance Annotation
4.4.5. Structural Coherence
4.4.6. Domain-Cluster Reporting
| Dataset | Metric | DeepAR | MC-Drop | TFT | DCRNN | DySAT | TGAT | TGN | MaGNet-BN |
|---|---|---|---|---|---|---|---|---|---|
| METR-LA | MSE | 0.34 | 0.32 | 0.30 | 0.29 | 0.31 | 0.30 | 0.25 | 0.22 |
| NLL | 1.92 | 1.88 | 1.83 | 1.79 | 1.86 | 1.81 | 1.63 | 1.47 | |
| CRPS | 0.137 | 0.133 | 0.127 | 0.124 | 0.130 | 0.125 | 0.112 | 0.105 | |
| PICP (%) | 87.6 | 88.1 | 88.8 | 89.2 | 88.0 | 89.0 | 90.3 | 92.1 | |
| PeMS-BAY | MSE | 0.29 | 0.27 | 0.26 | 0.25 | 0.27 | 0.26 | 0.22 | 0.19 |
| NLL | 2.04 | 1.99 | 1.93 | 1.88 | 1.95 | 1.90 | 1.71 | 1.54 | |
| CRPS | 0.149 | 0.144 | 0.138 | 0.134 | 0.141 | 0.136 | 0.122 | 0.113 | |
| PICP (%) | 86.8 | 87.5 | 88.2 | 88.6 | 87.1 | 88.4 | 90.1 | 91.3 | |
| TwitterRC | MSE | 0.48 | 0.45 | 0.43 | 0.44 | 0.46 | 0.42 | 0.38 | 0.34 |
| NLL | 2.56 | 2.43 | 2.38 | 2.41 | 2.49 | 2.34 | 2.11 | 1.98 | |
| CRPS | 0.183 | 0.177 | 0.171 | 0.173 | 0.180 | 0.168 | 0.154 | 0.141 | |
| PICP (%) | 82.1 | 83.4 | 84.0 | 83.7 | 82.5 | 84.6 | 86.9 | 88.8 | |
| Enron-Email | MSE | 0.22 | 0.21 | 0.20 | 0.20 | 0.21 | 0.19 | 0.17 | 0.15 |
| NLL | 1.71 | 1.66 | 1.59 | 1.62 | 1.68 | 1.57 | 1.45 | 1.38 | |
| CRPS | 0.112 | 0.108 | 0.103 | 0.105 | 0.110 | 0.101 | 0.092 | 0.086 | |
| PICP (%) | 89.4 | 90.1 | 90.8 | 90.3 | 89.0 | 91.2 | 92.5 | 93.6 | |
| ETH+UCY | MSE | 0.31 | 0.30 | 0.28 | 0.29 | 0.30 | 0.28 | 0.24 | 0.21 |
| NLL | 1.98 | 1.93 | 1.88 | 1.92 | 1.96 | 1.85 | 1.74 | 1.57 | |
| CRPS | 0.152 | 0.147 | 0.141 | 0.145 | 0.149 | 0.139 | 0.128 | 0.116 | |
| PICP (%) | 84.7 | 85.3 | 86.0 | 85.6 | 84.9 | 86.6 | 88.7 | 90.4 | |
| ELD-2012 | MSE | 0.11 | 0.11 | 0.10 | 0.10 | 0.11 | 0.10 | 0.09 | 0.07 |
| NLL | 1.36 | 1.33 | 1.28 | 1.30 | 1.34 | 1.27 | 1.18 | 1.09 | |
| CRPS | 0.084 | 0.082 | 0.078 | 0.079 | 0.082 | 0.077 | 0.071 | 0.066 | |
| PICP (%) | 91.0 | 91.6 | 92.2 | 92.0 | 90.7 | 92.7 | 93.8 | 95.4 | |
| M5-Retail | MSE | 0.55 | 0.52 | 0.49 | 0.51 | 0.53 | 0.48 | 0.44 | 0.45 |
| NLL | 2.83 | 2.69 | 2.65 | 2.68 | 2.76 | 2.60 | 2.37 | 2.15 | |
| CRPS | 0.201 | 0.195 | 0.189 | 0.192 | 0.198 | 0.186 | 0.174 | 0.179 | |
| PICP (%) | 79.2 | 80.6 | 81.1 | 80.8 | 79.5 | 81.7 | 84.2 | 86.9 |
4.4.7. Unified Louvain Post-Processing.
- k for k-NN: 10 for traffic & e-mail graphs, 25 for social & retail graphs;
- Similarity: cosine distance on -normalised embeddings;
- Resolution : 1.0 (vanilla Louvain);
- Post-merge: keep giant components; orphan nodes inherit the label of their nearest prototype.
| Modularity Q | ||||||||
|---|---|---|---|---|---|---|---|---|
| Dataset | DeepAR | MC-Drop | TFT | DCRNN | DySAT | TGAT | TGN | MaGNet-BN |
| METR-LA | 0.43 | 0.46 | 0.48 | 0.52 | 0.47 | 0.49 | 0.56 | 0.60 |
| PeMS-BAY | 0.41 | 0.44 | 0.45 | 0.50 | 0.45 | 0.47 | 0.55 | 0.59 |
| TwitterRC | 0.32 | 0.35 | 0.37 | 0.39 | 0.36 | 0.38 | 0.46 | 0.51 |
| Enron | 0.28 | 0.31 | 0.33 | 0.35 | 0.30 | 0.32 | 0.42 | 0.48 |
| ETH+UCY | 0.36 | 0.39 | 0.40 | 0.44 | 0.40 | 0.41 | 0.50 | 0.58 |
| ELD-2012 | 0.40 | 0.43 | 0.44 | 0.48 | 0.43 | 0.45 | 0.54 | 0.57 |
| M5-Retail | 0.29 | 0.32 | 0.33 | 0.35 | 0.31 | 0.33 | 0.41 | 0.47 |
| Temporal Adjusted Rand Index (tARI) | ||||||||
| Dataset | DeepAR | MC-Drop | TFT | DCRNN | DySAT | TGAT | TGN | MaGNet-BN |
| METR-LA | 0.52 | 0.55 | 0.57 | 0.61 | 0.56 | 0.57 | 0.66 | 0.71 |
| PeMS-BAY | 0.50 | 0.53 | 0.54 | 0.60 | 0.55 | 0.56 | 0.64 | 0.69 |
| TwitterRC | 0.38 | 0.41 | 0.43 | 0.46 | 0.42 | 0.43 | 0.51 | 0.57 |
| Enron | 0.34 | 0.37 | 0.38 | 0.40 | 0.36 | 0.37 | 0.48 | 0.53 |
| ETH+UCY | 0.46 | 0.49 | 0.50 | 0.54 | 0.48 | 0.50 | 0.60 | 0.68 |
| ELD-2012 | 0.49 | 0.52 | 0.53 | 0.58 | 0.52 | 0.53 | 0.62 | 0.67 |
| M5-Retail | 0.35 | 0.38 | 0.39 | 0.42 | 0.36 | 0.38 | 0.47 | 0.53 |
4.5. Main Results: Forecasting & Structural Consistency
4.5.1. Forecasting Accuracy
| Dataset | Metric wins | Wins / 4 | |||
|---|---|---|---|---|---|
| MSE↓ | NLL↓ | CRPS↓ | PICP↑ | ||
| METR–LA | √ | √ | √ | √ | 4 |
| PeMS–BAY | √ | √ | √ | √ | 4 |
| TwitterRC | √ | √ | √ | √ | 4 |
| Enron–Email | √ | √ | √ | √ | 4 |
| ETH+UCY | √ | √ | √ | √ | 4 |
| ELD–2012 | √ | √ | √ | √ | 4 |
| M5–Retail | √ | √ | √ | √ | 2 |
| Total wins | 6 | 7 | 6 | 7 | 26 / 28 |
4.5.2. Structural Consistency

4.5.3. Cross-Analysis
4.5.4. Fine-Grained Node-Level Consistency
| Normalised Mutual Information (NMI ↑) | ||||||||
|---|---|---|---|---|---|---|---|---|
| Dataset | DeepAR | MC-Drop | TFT | DCRNN | DySAT | TGAT | TGN | MaGNet-BN |
| METR-LA | 0.62 | 0.64 | 0.66 | 0.69 | 0.71 | 0.72 | 0.84 | 0.87 |
| PeMS-BAY | 0.61 | 0.63 | 0.65 | 0.68 | 0.70 | 0.71 | 0.82 | 0.85 |
| TwitterRC | 0.52 | 0.55 | 0.57 | 0.60 | 0.64 | 0.65 | 0.75 | 0.78 |
| Enron | 0.56 | 0.58 | 0.60 | 0.63 | 0.67 | 0.68 | 0.79 | 0.82 |
| ETH+UCY | 0.54 | 0.56 | 0.58 | 0.61 | 0.65 | 0.66 | 0.77 | 0.80 |
| ELD-2012 | 0.59 | 0.61 | 0.63 | 0.66 | 0.69 | 0.70 | 0.80 | 0.84 |
| M5-Retail | 0.50 | 0.53 | 0.55 | 0.57 | 0.60 | 0.61 | 0.72 | 0.76 |
| Variation of Information (VI ↓) | ||||||||
| Dataset | DeepAR | MC-Drop | TFT | DCRNN | DySAT | TGAT | TGN | MaGNet-BN |
| METR-LA | 0.68 | 0.63 | 0.60 | 0.56 | 0.53 | 0.52 | 0.46 | 0.42 |
| PeMS-BAY | 0.70 | 0.65 | 0.62 | 0.58 | 0.55 | 0.54 | 0.48 | 0.45 |
| TwitterRC | 0.81 | 0.76 | 0.73 | 0.69 | 0.63 | 0.62 | 0.52 | 0.58 |
| Enron | 0.74 | 0.69 | 0.66 | 0.61 | 0.55 | 0.54 | 0.48 | 0.49 |
| ETH+UCY | 0.76 | 0.71 | 0.68 | 0.64 | 0.58 | 0.57 | 0.50 | 0.52 |
| ELD-2012 | 0.71 | 0.66 | 0.63 | 0.59 | 0.53 | 0.52 | 0.47 | 0.46 |
| M5-Retail | 0.85 | 0.80 | 0.77 | 0.73 | 0.66 | 0.65 | 0.55 | 0.61 |
4.5.5. Node–Level Evaluation Metrics
- Normalised Mutual Information (NMI, ↑) where is mutual information and Shannon entropy. It measures the shared information (0–1).
- Variation of Information (VI, ↓) the information-theoretic distance between two partitions (lower is better).
- Brier Score (↓) where is the predicted class-probability vector and the one-hot ground truth. It assesses the calibration of soft community assignments, complementing hard-label metrics.
4.5.6. Key Take-Away
4.5.7. RL Stability Diagnostics

4.6. Ablation Study
| Model Variant | MSE | NLL | Modularity (Q) | tARI |
|---|---|---|---|---|
| w/o Bayesian Embedding | 0.209 | 1.622 | 0.504 | 0.613 |
| w/o Markov + PPO | 0.198 | 1.576 | 0.481 | 0.583 |
| MaGNet-BN (Full) | 0.182 | 1.392 | 0.581 | 0.693 |
4.6.1. Training Stability
- Fast, monotonic convergence. The mean–squared error (MSE) drops from to 0.18 within 35 epochs, after which improvements plateau.
- Synchronous structural gains. Modularity (Q) rises from , while temporal ARI (tARI) climbs from —mirroring the MSE curve and confirming that the PPO stage enhances community coherence without hurting predictive accuracy.
- Low epoch-to-epoch variance. Even with limited rewards, PPO updates remain steady because to the clipped-surrogate objective’s validation by the lack of spikes.

4.7. Embedding Visualization

4.7.1. Qualitative Insight
4.8. Sensitivity & Robustness
4.8.1. Evaluation Pipeline Overview
- (a)
- Hyper-parameter sensitivity: We perform targeted sweeps over three key tuning knobs—Monte Carlo sample count (M), prototype anchor count (P), and PPO reward weights . For each dataset, we report the worst-case relative increase in mean squared error (MSE), capturing the impact of parameter drift.
- (b)
- Structural robustness: We inject synthetic edge noise into every test snapshot by randomly rewiring 1%, 3%, and 5% of the graph edges, and log the resulting drop in modularity (). This simulates real-world perturbations in graph topology.

4.8.2. Findings
| Dataset | M | P | Worst↓ | |
|---|---|---|---|---|
| METR-LA | 2.1 | |||
| PeMS-BAY | 2.3 | |||
| TwitterRC | 2.4 | |||
| Enron | 2.0 | |||
| ETH+UCY | 1.8 | |||
| ELD-2012 | 2.2 | |||
| M5-Retail | 2.7 |
4.8.3. Hyper-Parameter Sensitivity (Ours Only)
- the number of Monte-Carlo samples ,
- the number of prototype anchors , and
- the PPO reward weights ,
| Model | 1 % | 3 % | 5 % |
|---|---|---|---|
| DeepAR | 0.056 | 0.084 | 0.112 |
| MC-Drop LSTM | 0.049 | 0.077 | 0.098 |
| TFT | 0.043 | 0.069 | 0.092 |
| DCRNN | 0.038 | 0.061 | 0.083 |
| DySAT | 0.045 | 0.072 | 0.090 |
| TGAT | 0.040 | 0.067 | 0.086 |
| TGN | 0.031 | 0.054 | 0.067 |
| MaGNet-BN | 0.018 | 0.026 | 0.031 |
5. Discussion
- End-to-end synergy: Bayesian-Markov-PPO stages reinforce each other; ablating either cuts tARI by over 11 pp.
- Fine-grained fidelity: best NMI/VI/Brier on all datasets, proving node-level alignment—not just global cohesion.
- Ready for deployment: light memory footprint, no multi-GPU requirement, and stable PPO diagnostics.
6. Conclusion
- Achieves the best score on 26/28 forecasting benchmarks (MSE, NLL, CRPS, PICP) and all structural metrics (Modularity Q, tARI, NMI);
- Remains stable and data-efficient, with worst-case MSE drift under hyper-parameter perturbation and only modularity loss when of edges are rewired;
- Trains end-to-end in 11 GPU-hours on a single NVIDIA A100, demonstrating practical feasibility for real-time analytics.
7. Future Work
8. Decleartions
Funding
Conflicts of Interest
| 1 | NMI and Brier are still best on M5-Retail; full figures are reported in Table 7
|
References
- Cazabet, R.; Amblard, F. Dynamic community detection. Wiley Interdiscip. Rev. Comput. Stat. 2020, 12, e1503. [Google Scholar]
- Zhao, L.; Song, Y.; Zhang, C.; Liu, Y.; Wang, P.; Lin, J. T-GCN: A Temporal Graph Convolutional Network for Traffic Prediction. IEEE Transactions on Intelligent Transportation Systems 2020, 21, 3848–3858. [Google Scholar] [CrossRef]
- Blondel, V.D.; Guillaume, J.-L.; Lambiotte, R.; Lefebvre, E. Fast Unfolding of Communities in Large Networks. Journal of Statistical Mechanics: Theory and Experiment, 1000. [Google Scholar]
- Wu, Z.; Pan, S.; Long, G.; Jiang, J.; Chang, X.; Zhang, C. Graph WaveNet for Deep Spatial–Temporal Graph Modeling. In Proceedings of the 28th International Joint Conference on Artificial Intelligence (IJCAI), Macao, China, 10–16 August 2019; pp. 1907–1913. [Google Scholar] [CrossRef]
- Huang, Y.; Lei, X. Temporal group-aware graph diffusion networks for dynamic link prediction. In Proceedings of the 29th ACM SIGKDD Conference on Knowledge Discovery and Data Mining (KDD), Long Beach, CA, USA, 6–10 August 2023; pp. 3782–3792. [Google Scholar]
- Costa, G.; Cattuto, C.; Lehmann, S. Towards modularity optimization using reinforcement learning to community detection in dynamic social networks. In Proc. IEEE ICDM, 2021; pp. 110–119.
- Gal, Y.; Ghahramani, Z. Dropout as a Bayesian approximation: Representing model uncertainty. In Proc. ICML, 2016; pp. 1050–1059.
- Hochreiter, S.; Schmidhuber, J. Long Short-Term Memory. Neural Computation 1997, 9, 1735–1780. [Google Scholar] [CrossRef] [PubMed]
- Hu, Z.; Dong, Y.; Wang, K.; Sun, Y. Open Graph Benchmark: Datasets for machine learning on graphs. In Proc. NeurIPS, 2021.
- Lim, B.; Arık, S.Ö.; Loeff, N.; Pfister, T. Temporal Fusion Transformers for Interpretable Multi-Horizon Time-Series Forecasting. International Journal of Forecasting 2021, 37, 1748–1764. [Google Scholar] [CrossRef]
- Levine, S.; Kumar, A.; Tucker, G.; Fu, J. Offline reinforcement learning: Tutorial, review, and open problems. arXiv:2005.01643 (2020).
- Li, Y.; Yu, R.; Shahabi, C.; Liu, Y. Diffusion convolutional recurrent neural network: Data-driven traffic forecasting. In Proc. ICLR, 2018.
- Tsai, Y.-H. H.; Liang, P. P.; Zadeh, A.; Morency, L.-P.; Salakhutdinov, R. Multimodal Transformer for Unaligned Multimodal Language Sequences. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics (ACL); Association for Computational Linguistics: Florence, Italy, 2019; pp. 6558–6569. [Google Scholar] [CrossRef]
- Blundell, C.; Cornebise, J.; Kavukcuoglu, K.; Wierstra, D. Weight Uncertainty in Neural Networks. In Proceedings of the 32nd International Conference on Machine Learning (ICML); Bach, F.; Blei, D., Eds.; Lille, France, 6–11 July 2015; pp. 1613–1622; Blei, D., Ed.; pp. 1613–1622.
- Franceschi, L.; Niepert, M.; Pontil, M.; He, X. Learning Discrete Structures for Graph Neural Networks. In Proceedings of the 36th International Conference on Machine Learning (ICML); Chaudhuri, K.; Long Beach, CA, USA, 9–15 June 2019; Salakhutdinov, R., Ed.; pp. 1972–1982. [Google Scholar]
- Chen, Y.; Wu, L.; Zaki, M. Iterative Deep Graph Learning for Graph Neural Networks: Better and Robust Node Embeddings. In Proceedings of the 34th Conference on Neural Information Processing Systems (NeurIPS 2020); Larochelle, H.; Ranzato, M.; Hadsell, R.; Balcan, M. F.; Lin, H., Eds.; Curran Associates, Inc.: Red Hook, NY, USA, 2020; pp. 19314–19326. [Google Scholar]
- Pang, W.; Wang, X.; Sun, Y.; et al. Bayesian spatio-temporal graph transformer network (b-star) for multi-aircraft trajectory prediction. In Proc. ACM MM, 2022; pp. 3979–3988.
- Xu, D.; Ruan, C.; Korpeoglu, E.; Kumar, S.; Achan, K. Inductive Representation Learning on Temporal Graphs. In Proceedings of the 8th International Conference on Learning Representations (ICLR); Addis Ababa, Ethiopia, 2020. Available online: https://openreview.net/forum?id=rJeW1yHYwH (accessed on 6 July 2025); Available online: https://openreview.net/forum?id=rJeW1yHYwH (accessed on 6 July 2025).
- Rossetti, G.; Cazabet, R. Community discovery in dynamic networks: A survey. ACM Comput. Surv. 2018, 51, 35–1. [Google Scholar] [CrossRef]
- Rosvall, M.; Esquivel, A.; Lancichinetti, A.; West, J. D.; Lambiotte, R. Memory in network flows and its effects on spreading dynamics and community detection. Nat. Commun. 2014, 5, 4630. [Google Scholar] [CrossRef] [PubMed]
- Salinas, D.; Flunkert, V.; Gasthaus, J.; Januschowski, T. DeepAR: Probabilistic forecasting with autoregressive recurrent networks. Int. J. Forecast. 2020, 36, 1181–1191. [Google Scholar] [CrossRef]
- Sankar, A.; Wu, Y.; Gou, L.; Zhang, W.; Yang, H. DySAT: Deep neural representation learning on dynamic graphs via self-attention. In Proc. WSDM, 2020; pp. 519–527.
- Schulman, J.; Wolski, F.; Dhariwal, P.; Radford, A.; Klimov, O. Proximal policy optimization algorithms. arXiv:1707.06347 (2017).
- Zhou, H.; Zhang, S.; Peng, J.; et al. Informer: Beyond efficient transformer for long-sequence time-series forecasting. In Proc. AAAI, 2021; pp. 11106–11115.
- Rossi, E.; Chambers, I.; Nguyen, D.Q.; et al. Temporal Graph Networks for Deep Learning on Dynamic Graphs. arXiv 2020, arXiv:2006.10637. [Google Scholar]
- Klimt, B.; Yang, Y. The Enron Corpus: A New Dataset for Email Classification Research. In Proc. European Conference on Machine Learning (ECML); Pisa, Italy, 2004; pp. 217–226.
- Pellegrini, S.; Ess, A.; Schindler, K.; Van Gool, L. You’ll Never Walk Alone: Modeling Social Behavior for Multi-Target Tracking. In Proc. IEEE International Conference on Computer Vision (ICCV); Kyoto, Japan, 2009; pp. 261–268.
- Trindade, A. ElectricityLoadDiagrams20112014 [Data set]; UCI Machine Learning Repository, 2015. Available online: https://archive.ics.uci.edu/dataset/321/electricityloaddiagrams20112014 (accessed on 6 July 2025). [CrossRef]
- Makridakis, S.; Spiliotis, E.; Assimakopoulos, V. M5 accuracy competition: Results, findings, and conclusions. International Journal of Forecasting 2022, 38, 2330–2341. [Google Scholar] [CrossRef]
- Bishop, C. M. Pattern Recognition and Machine Learning; Springer: New York, NY, USA, 2006. [Google Scholar]
- Murphy, K. P. Machine Learning: A Probabilistic Perspective; MIT Press: Cambridge, MA, USA, 2012. [Google Scholar]
- Gneiting, T.; Raftery, A. E. Strictly proper scoring rules, prediction, and estimation. Journal of the American Statistical Association 2007, 102, 359–378. [Google Scholar] [CrossRef]
- Khosravi, A.; Nahavandi, S.; Creighton, D.; Atiya, A. F. Comprehensive review of neural network-based prediction intervals and new advances. IEEE Transactions on Neural Networks 2011, 22, 1341–1356. [Google Scholar] [CrossRef] [PubMed]
- Newman, M.E.J. Modularity and community structure in networks. Proceedings of the National Academy of Sciences (PNAS) 2006, 103, 8577–8582. [Google Scholar] [CrossRef] [PubMed]
- Hubert, L.; Arabie, P. Comparing partitions. Journal of Classification 1985, 2, 193–218. [Google Scholar] [CrossRef]
- Yuxi, Li. Deep reinforcement learning: An overview. arXiv preprint, arXiv:1701.07274, 2017.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).