Submitted:
10 October 2024
Posted:
11 October 2024
You are already at the latest version
Abstract
Keywords:
1. Introduction
2. Electricity Demand Forecasting
2.1. Electricity Demand
2.2. Selection of Forecasting Methodologies
2.2.1. Forecasting Methodologies Based on Time Series
2.2.2. Deep learning Techniques
2.3. External factors
2.4. Modeling Methodology
2.4.1. Pipeline for Building Electricity Forecasting Models
2.4.2. Key Aspects of Forecasting Methodology
3. Implementation and Results
3.1. Dataset
3.2. Case study
- o
- Among the variable selection algorithms tested here, the most notable ones were CFS subset evaluation, classifier attribute evaluation, and relief. These algorithms selected the five or six variables that were most strongly correlated with the target variable, yielding predictions with error rates similar to or lower than those obtained when all variables in the database were considered (as indicated in the last line of Table 3). This underscores the relevance of the selected variables.
- o
- Regarding the performance of models with different prediction algorithms, LSTM stands out as having the highest computational cost. Its processing time was 10 times higher than for the DNN, five times more than for the CNN, and 2.5 times more than for the combined CNN+LSTM model. However, the models based on the LSTM and a combination of CNN+LSTM achieved the highest accuracy.
- o
- From comparing the performance of shallow and deep models, it was observed that the latter incurred a higher computational cost. Nevertheless, in most scenarios, they demonstrated superior accuracy compared to shallow models.
- o
- The distinguishing factor between the tables is the prediction horizon. It is evident that increasing the prediction horizon leads to a reduction in accuracy.
- o
- Processing involving only the target variable (Sl), with or without external factors, gave the biggest errors. This implies that deep learning algorithms face limitations in terms of their generalization capacity when operating with a very limited number of input variables, resulting in elevated error rates.
- o
- From a comparison of the predictions based on only the target variable (Sl), the selected variables (Fs), and all variables in the dataset (Av), it is notable that errors in the Fs and Av columns are very similar. This suggests that despite the reduction of the number of input variables from 23 to six or seven in the variable selection step, this did not lead to an increase in the prediction error.
- o
- Multivariate and multistep models offer flexibility by permitting variations in the number of input variables and prediction intervals, making them appealing for electricity consumption forecasts.
- o
- It has been demonstrated that incorporating external factors enhances the accuracy of the model, with increases of up to 60%.
- o
- Variable selection is a necessary measure when the number of input variables is large. This enables a reduction in the dimensionality of the problem while still yielding a model with good accuracy. This reduction in dimensionality promotes the application of deep learning techniques and the utilization of deep models.
- o
- In deep learning models, defining the architecture is a crucial step. The simulation results demonstrated that it is unnecessary to incorporate more than three intermediate layers or to introduce an excessive number of neurons in these layers. Adhering to the number of neurons defined in Eq. (8) yields satisfactory results.
- o
- From the perspective of accuracy, we note that the lowest error rate achieved was 0.15% for the CNN+LSTM deep learning technique, which included both variable selection and the representation of external factors. In [37], the lowest error achieved using a CNN was 0.8%, while in [38], employing LSTM, it was 1.44%. In addition, the authors of [5] suggest that errors ranging from 1% to 5% are typically expected for aggregate consumption.
- o
- Table 4 and Table 5 show simulation results conducted under various input conditions and forecast horizons, where the lowest error rates are shown in bold. It is evident that the CNN models and composite models (CNN+LSTM) consistently gave the lowest error rates across most scenarios. This is due to the advantages of the CNN in terms of its feature extraction capability from input variables. The presence of the CNN in both models significantly contributed to the reductions in the error rate. The proficiency of the LSTM in handling time series data further enhanced the composite model's performance, allowing it to outperform other models in certain simulation scenarios.
4. Conclusions and Future Research
Author Contributions
Data Availability Statement
Acknowledgments
Conflicts of Interest
Code Availability
Appendix A. Model Architectures
![]() |
![]() |
![]() |
![]() |
Appendix B. Hyperparameter Selection
| Hyperparameter | Interval | Selection |
|---|---|---|
| Number of hidden layers | [1..100] | 1 to 6 |
| Number of neurons | [1..100] | 5 |
| Learning Rating | [0.0001..0.01] | 0.01 |
| Dropout rate | [0.1.. 0.5] | 0.1 |
| Activation function | Relu, Sigmoid, Tanh | Relu |
| Optimizer | Adam,RMSprop,SGD | Adam |
| Batch Size | [16..512] | 128 |
| Loss function | Mse, mae, mape | Mse |
| Kernel | [2..5] | 2 |
| Number of epochs | [10..500] | 30 |
References
- Merce, R.A.; Grover-Silva, E.; Le Conte, J. Load and Demand Side Flexibility Forecasting. In Proceedings of the ENERGY Tenth Int. Conf. Smart Grids, Green Commun. IT Energy-aware Technol, Lisbon, Portugal; 2020; pp. 1–6. [Google Scholar]
- Khan, K.A.; Quamar, M.M.; Al-Qahtani, F.H.; Asif, M.; Alqahtani, M.; Khalid, M. Smart Grid Infrastructure and Renewable Energy Deployment: A Conceptual Review of Saudi Arabia. Energy Strateg. Rev. 2023, 50, 101247. [Google Scholar] [CrossRef]
- Mystakidis, A.; Koukaras, P.; Tsalikidis, N.; Ioannidis, D.; Tjortjis, C. Energy Forecasting: A Comprehensive Review of Techniques and Technologies. Energies 2024, 17. [Google Scholar] [CrossRef]
- Peñaloza, A.K.A.; Balbinot, A.; Leborgne, R.C. Review of Deep Learning Application for Short-Term Household Load Forecasting. In Proceedings of the 2020 IEEE PES Transmission & Distribution Conference and Exhibition - Latin America (T&D LA); 2020; pp. 1–6. [Google Scholar]
- Aybar-Mejía, M.; Villanueva, J.; Mariano-Hernández, D.; Santos, F.; Molina-García, A. A Review of Low-Voltage Renewable Microgrids: Generation Forecasting and Demand-Side Management Strategies. Electronics 2021, 10. [Google Scholar] [CrossRef]
- Bunn, D.W. Short-Term Forecasting: A Review of Procedures in the Electricity Supply Industry. J. Oper. Res. Soc. 1982, 33, 533–545. [Google Scholar] [CrossRef]
- Min, H.; Hong, S.; Song, J.; Son, B.; Noh, B.; Moon, J. SolarFlux Predictor: A Novel Deep Learning Approach for Photovoltaic Power Forecasting in South Korea. Electronics 2024, 13. [Google Scholar] [CrossRef]
- Manzolini, G.; Fusco, A.; Gioffrè, D.; Matrone, S.; Ramaschi, R.; Saleptsis, M.; Simonetti, R.; Sobic, F.; Wood, M.J.; Ogliari, E.; et al. Impact of PV and EV Forecasting in the Operation of a Microgrid. Forecasting 2024, 6, 591–615. [Google Scholar] [CrossRef]
- Sengupta, S.; Basak, S.; Saikia, P.; Paul, S.; Tsalavoutis, V.; Atiah, F.; Ravi, V.; Peters, A. A Review of Deep Learning with Special Emphasis on Architectures, Applications and Recent Trends. Knowledge-Based Syst. 2020, 194, 105596. [Google Scholar] [CrossRef]
- Hahn, H.; Meyer-Nieberg, S.; Pickl, S. Electric Load Forecasting Methods: Tools for Decision Making. Eur. J. Oper. Res. 2009, 199, 902–907. [Google Scholar] [CrossRef]
- Park, H.-Y.; Lee, B.-H.; Son, J.-H.; Ahn, H.-S. A Comparison of Neural Network-Based Methods for Load Forecasting with Selected Input Candidates. In Proceedings of the 2017 IEEE International Conference on Industrial Technology (ICIT); 2017; pp. 1100–1105. [Google Scholar]
- Kong, W.; Dong, Z.Y.; Jia, Y.; Hill, D.J.; Xu, Y.; Zhang, Y. Short-Term Residential Load Forecasting Based on LSTM Recurrent Neural Network. IEEE Trans. Smart Grid 2019, 10, 841–851. [Google Scholar] [CrossRef]
- Mannini, R.; Darure, T.; Eynard, J.; Grieu, S. Predictive Energy Management of a Building-Integrated Microgrid: A Case Study. Energies 2024, 17. [Google Scholar] [CrossRef]
- Elsworth, S.; Güttel, S. Time Series Forecasting Using LSTM Networks: A Symbolic Approach 2020.
- Glavan, M.; Gradišar, D.; Moscariello, S.; Juričić, Đ.; Vrančić, D. Demand-Side Improvement of Short-Term Load Forecasting Using a Proactive Load Management – a Supermarket Use Case. Energy Build. 2019, 186, 186–194. [Google Scholar] [CrossRef]
- Alzubaidi, L.; Zhang, J.; Humaidi, A.J.; Al-Dujaili, A.; Duan, Y.; Al-Shamma, O.; Santamaría, J.; Fadhel, M.A.; Al-Amidie, M.; Farhan, L. Review of Deep Learning: Concepts, CNN Architectures, Challenges, Applications, Future Directions. J. Big Data 2021, 8, 53. [Google Scholar] [CrossRef] [PubMed]
- Ungureanu, S.; Topa, V.; Cziker, A.C. Deep Learning for Short-Term Load Forecasting—Industrial Consumer Case Study. Appl. Sci. 2021, 11. [Google Scholar] [CrossRef]
- Jiang, W. Deep Learning Based Short-Term Load Forecasting Incorporating Calendar and Weather Information. Internet Technol. Lett. 2022, 5, e383. [Google Scholar] [CrossRef]
- Alom, M.Z.; Taha, T.M.; Yakopcic, C.; Westberg, S.; Sidike, P.; Nasrin, M.S.; Hasan, M.; Van Essen, B.C.; Awwal, A.A.S.; Asari, V.K. A State-of-the-Art Survey on Deep Learning Theory and Architectures. Electronics 2019, 8. [Google Scholar] [CrossRef]
- Son, N. Comparison of the Deep Learning Performance for Short-Term Power Load Forecasting. Sustainability 2021, 13. [Google Scholar] [CrossRef]
- Rumelhart, D.E.; Hinton, G.E.; Williams, R.J. Learning Internal Representations by Error Propagation. In Parallel Distributed Processing: Explorations in the Microstructure of Cognition, {V}olume 1: {F}oundations; Rumelhart, D.E., Mcclelland, J.L., Eds.; MIT Press: Cambridge, MA, 1986; pp. 318–362. [Google Scholar]
- Parmezan, A.R.S.; Souza, V.M.A.; Batista, G.E.A.P.A. Evaluation of Statistical and Machine Learning Models for Time Series Prediction: Identifying the State-of-the-Art and the Best Conditions for the Use of Each Model. Inf. Sci. (Ny). 2019, 484, 302–337. [Google Scholar] [CrossRef]
- Yalcinoz, T.; Eminoglu, U. Short Term and Medium Term Power Distribution Load Forecasting by Neural Networks. Energy Convers. Manag. 2005, 46, 1393–1405. [Google Scholar] [CrossRef]
- Santos Amaral, L.; Medeiros de Araújo, G.; Reinaldo de Moraes, R.A. Analysis of the Factors That Influence the Performance of an Energy Demand Forecasting Model. Adv. Notes Inf. Sci. 2022, 2, 92–102. [Google Scholar] [CrossRef]
- Amaral, L.S.; de Araújo, G.M.; Moraes, R.; de Oliveira Villela, P.M. An Expanded Study of the Application of Deep Learning Models in Energy Consumption Prediction BT - Data and Information in Online Environments.; Pinto, A.L., Arencibia-Jorge, R., Eds.; Springer Nature Switzerland: Cham, 2022; pp. 150–162. [Google Scholar]
- Kuo, P.-H.; Huang, C.-J. A High Precision Artificial Neural Networks Model for Short-Term Energy Load Forecasting. Energies 2018, 11. [Google Scholar] [CrossRef]
- Dudek, G. Multilayer Perceptron for Short-Term Load Forecasting: From Global to Local Approach. Neural Comput. Appl. 2020, 32, 3695–3707. [Google Scholar] [CrossRef]
- Buildings. Available online: https://www.iea.org/energy-system/buildings.
- KawBamura, T. Distribution of Discomfort Index in Japan in Summer Season. J Meteorol. Res. 1965, 17, 460–466. [Google Scholar]
- Ono, H.-S.P.; Kawamura, T. Sensible Climates in Monsoon Asia. Int. J. Biometeorol. 1991, 35, 39–47. [Google Scholar] [CrossRef] [PubMed]
- Suganthi, L.; Samuel, A.A. Energy Models for Demand Forecasting—A Review. Renew. Sustain. Energy Rev. 2012, 16, 1223–1240. [Google Scholar] [CrossRef]
- Curry, B.; Morgan, P.H. Model Selection in Neural Networks: Some Difficulties. Eur. J. Oper. Res. 2006, 170, 567–577. [Google Scholar] [CrossRef]
- Zhang, G.; Eddy Patuwo, B.; Y. Hu, M. Forecasting with Artificial Neural Networks:: The State of the Art. Int. J. Forecast. 1998, 14, 35–62. [Google Scholar] [CrossRef]
- Petropoulos, F.; Apiletti, D.; Assimakopoulos, V.; Babai, M.Z.; Barrow, D.K.; Ben Taieb, S.; Bergmeir, C.; Bessa, R.J.; Bijak, J.; Boylan, J.E.; et al. Forecasting: Theory and Practice. Int. J. Forecast. 2022, 38, 705–871. [Google Scholar] [CrossRef]
- Cogollo, M.R.; Velasquez, J.D. Methodological Advances in Artificial Neural Networks for Time Series Forecasting. IEEE Lat. Am. Trans. 2014, 12, 764–771. [Google Scholar] [CrossRef]
- Lippmann, R. An Introduction to Computing with Neural Nets. IEEE ASSP Mag. 1987, 4, 4–22. [Google Scholar] [CrossRef]
- Sheela, K.G.; Deepa, S.N. Review on Methods to Fix Number of Hidden Neurons in Neural Networks. Math. Probl. Eng. 2013, 2013, 425740. [Google Scholar] [CrossRef]
- Chollet, F. Keras: The Python Deep Learning API. Available online: https://keras.io/.
- Ribeiro, A.M.N.C.; do Carmo, P.R.X.; Endo, P.T.; Rosati, P.; Lynn, T. Short- and Very Short-Term Firm-Level Load Forecasting for Warehouses: A Comparison of Machine Learning and Deep Learning Models. Energies 2022, 15. [Google Scholar] [CrossRef]
- Gnanambal, S.; Thangaraj, M.; Meenatchi, V.; Gayathri, V. Classification Algorithms with Attribute Selection: An Evaluation Study Using WEKA. Int. J. Adv. Netw. Appl 2018, 9, 3640–3644. [Google Scholar]
- Nespoli, A.; Ogliari, E.; Pretto, S.; Gavazzeni, M.; Vigani, S.; Paccanelli, F. Electrical Load Forecast by Means of LSTM: The Impact of Data Quality. Forecasting 2021, 3, 91–101. [Google Scholar] [CrossRef]
- Hobbs, B.F.; Helman, U.; Jitprapaikulsarn, S.; Konda, S.; Maratukulam, D. Artificial Neural Networks for Short-Term Energy Forecasting: Accuracy and Economic Value. Neurocomputing 1998, 23, 71–84. [Google Scholar] [CrossRef]










| Interval | Condition |
|---|---|
| Id > 80 | Heat stress |
| 75 < Id < 80 | Uncomfortable due to heat |
| 60 < Id < 75 | Comfortable |
| 55 < Id < 60 | Uncomfortable due to cold |
| Id < 55 | Stress due to cold |
| Interval | Condition | Value Range |
|---|---|---|
| Holiday | Holiday or not | [0,1] |
| Workday | Workday or not | [0,0.5,1] |
| Working | Working or not | [0,0.5,1] |
| Technique |
Source | Selected variables |
Magnitude | DNN | CNN | LSTM | CNN+LSTM | ||||
|---|---|---|---|---|---|---|---|---|---|---|---|
| Shallow | Deep | Shallow | Deep | Shallow | Deep | Shallow | Deep | ||||
| CFS subset evaluation | WEKA |
RT demand, DACC, DA MLC, RT MLC, MIN_5MIN_RSP, MAX_5MIN_RSP |
MAPE | 0.22 | 0.18 | 1.8 | 0.74 | 0.37 | 0.24 | 0.21 | 0.18 |
| t(s) | 152 | 185 | 276 | 340 | 887 | 1798 | 739 | 848 | |||
|
Classifier attribute evaluation |
MAX_5MIN_RSP, DA EC, DA CC, RT demand, DA LMP | MAPE | 0.27 | 0.23 | 1.6 | 0.48 | 0.21 | 0.16 | 0.26 | 0.24 | |
| t(s) | 153 | 187 | 342 | 384 | 994 | 1813 | 761 | 790 | |||
| Principal components | RT LMP, RT EC, DA LMP, DA EC, RT MLC | MAPE | 7.8 | 6.84 | 9.67 | 6.5 | 10.95 | 9.79 | 8.06 | 7.57 | |
| t(s) | 87 | 130 | 150 | 219 | 594 | 912 | 996 | 386 | |||
| Relief | RT demand, DA demand, DA EC, DA LMC, reg. service price | MAPE | 0.52 | 0.33 | 3.82 | 0.22 | 0.23 | 0.16 | 0.16 | 0.17 | |
| t(s) | 112.8 | 123 | 274 | 339 | 906 | 1792 | 692 | 823 | |||
| Mutual information | Python | DA MLC, DA LMP, MIN_5MIN_RSP, DA EC, dew point | MAPE | 5.9 | 5.76 | 8.15 | 5.03 | 10.21 | 7.82 | 7.23 | 6.11 |
| t(s) | 155 | 186 | 270 | 339 | 581 | 1827 | 725 | 792 | |||
| - | - | All | MAPE | 0.86 | 0.38 | 2.27 | 1.05 | 0.29 | 0.2 | 0.25 | 0.20 |
| t(s) | 118 | 123 | 158 | 153 | 591 | 920 | 695 | 879 | |||
| Technique | Model | Sl (1) |
Sl+ W (2) |
Sl+ S+ C+ Id (3) |
(%) (4) |
Fs (5) |
Fs+ S+ C+ Id (6) |
(%) (7) |
Av (8) |
Av+ S+ C+ Id (9) |
(%) (10) |
|---|---|---|---|---|---|---|---|---|---|---|---|
| DNN | Sh | 11.3 | 6.76 | 5.38 | 52.4 | 0.29 | 0.19 | 34.5 | 0.19 | 0.18 | 5.3 |
| D | 10.85 | 5.98 | 8.57 | 44.9 | 0.37 | 0.17 | 54.1 | 0.22 | 0.17 | 22.7 | |
| CNN | Sh | 12.13 | 8.46 | 8.0 | 34 | 2.85 | 2.23 | 21.8 | 3.40 | 2.26 | 33.5 |
| D | 11.04 | 4.95 | 4.29 | 61.1 | 0.18 | 0.18 | 0 | 0.26 | 0.16 | 38.5 | |
| LSTM | Sh | 12.37 | 8.31 | 7.86 | 36.5 | 0.65 | 0.25 | 61.5 | 0.47 | 0.17 | 63.8 |
| D | 11.74 | 7.59 | 9.0 | 23.3 | 14.1 | 7.07 | 49.8 | 14.96 | 10.45 | 30.1 | |
| CNN+LSTM | Sh | 10.38 | 6.56 | 5.67 | 45.4 | 0.31 | 0.21 | 32.2 | 0.42 | 0.19 | 54.8 |
| D | 11.11 | 5.09 | 7.60 | 31.6 | 0.25 | 0.15 | 40 | 0.25 | 0.24 | 4 |
| Technique | Model | Sl (1) |
Sl+ W (2) |
Sl+ S+ C+ Id (3) |
(%) (4) |
Fs (5) |
Fs+ S+ C+ Id (6) |
(%) (7) |
Av (8) |
Av+ S+ C+ Id (9) |
(%) (10) |
|---|---|---|---|---|---|---|---|---|---|---|---|
| DNN | Sh | 33.2 | 23.44 | 27.98 | 15.7 | 36.58 | 29.77 | 18.6 | 29.8 | 29.76 | 0.13 |
| D | 48.1 | 27.63 | 37.01 | 23.1 | 45.03 | 32.4 | 28.1 | 44.0 | 40.58 | 7.77 | |
| CNN | Sh | 12.01 | 15.37 | 4.92 | 59.0 | 4.91 | 4.83 | 1.6 | 4.92 | 4.68 | 4.9 |
| D | 10.6 | 9.19 | 3.77 | 64.4 | 4.79 | 3.96 | 17.3 | 4.39 | 3.63 | 17.7 | |
| LSTM | Sh | 13.00 | 14.36 | 9.82 | 24.5 | 4.39 | 4.30 | 2.0 | 4.29 | 3.66 | 14.7 |
| D | 15.58 | 15.65 | 15.27 | 2.0 | 15.8 | 11.34 | 4.46 | 15.4 | 15.15 | 1.62 | |
| CNN+LSTM | Sh | 15.00 | 8.04 | 4.58 | 69.5 | 4.42 | 3.93 | 11.0 | 4.12 | 4.09 | 0.7 |
| D | 11.24 | 8.65 | 10.7 | 4.8 | 15.94 | 10.82 | 32.1 | 15.6 | 15.36 | 1.5 |
| DNN | CNN | LSTM | CNN+LSTM | ||||||||||
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
| Number of Neurons |
Number of neurons |
Number of neurons |
Number of neurons |
||||||||||
| 6 | 15 | 32 | 6 | 15 | 32 | 6 | 15 | 32 | 6 | 15 | 32 | ||
| Number of intermediate layers | 2 | 0.24 | 0.15 | 0.41 | 0.18 | 0.15 | 0.15 | 0.23 | 0.16 | 0.19 | 0.3 | 0.15 | 0.25 |
| 3 | 0.3 | 0.54 | 0.16 | 0.15 | 0.26 | 0.29 | 0.52 | 0.15 | 0.17 | 0.18 | 0.43 | 0.52 | |
| 4 | 0.17 | 0.44 | 0.2 | 0.19 | 0.18 | 0.17 | 2.34 | 0.46 | 0.81 | 0.18 | 0.17 | 0.17 | |
| 6 | 0.24 | 0.15 | 0.16 | 0.3 | 0.18 | 0.41 | 10.0 | 15.4 | 0.40 | 15.53 | 0.40 | 0.35 | |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).



