Sensors 22 04062
Sensors 22 04062
Sensors 22 04062
Article
Predicting Energy Consumption Using LSTM, Multi-Layer
GRU and Drop-GRU Neural Networks
Sameh Mahjoub 1,2, * , Larbi Chrifi-Alaoui 1 , Bruno Marhic 1 and Laurent Delahoche 1
1 Laboratory of Innovative Technology (LTI, UR 3899), University of Picardie Jules Verne, 80000 Amiens, France;
larbi.alaoui@u-picardie.fr (L.C.-A.); bruno.marhic@u-picardie.fr (B.M.); laurent.delahoche@u-picardie.fr (L.D.)
2 Control & Energy Management Laboratory (CEMLab), University of Sfax, Sfax 3029, Tunisia
* Correspondence: sameh.mahjoub@u-picardie.fr
Abstract: With the steep rise in the development of smart grids and the current advancement in
developing measuring infrastructure, short term power consumption forecasting has recently gained
increasing attention. In fact, the prediction of future power loads turns out to be a key issue to
avoid energy wastage and to build effective power management strategies. Furthermore, energy
consumption information can be considered historical time series data that are required to extract all
meaningful knowledge and then forecast the future consumption. In this work, we aim to model
and to compare three different machine learning algorithms in making a time series power forecast.
The proposed models are the Long Short-Term Memory (LSTM), the Gated Recurrent Unit (GRU)
and the Drop-GRU. We are going to use the power consumption data as our time series dataset and
make predictions accordingly. The LSTM neural network has been favored in this work to predict the
future load consumption and prevent consumption peaks. To provide a comprehensive evaluation of
this method, we have performed several experiments using real data power consumption in some
French cities. Experimental results on various time horizons show that the LSTM model produces
a better result than the GRU and the Drop-GRU forecasting methods. There are fewer prediction
errors and its precision is finer. Therefore, these predictions based on the LSTM method will allow us
Citation: Mahjoub, S.; Chrifi-Alaoui,
to make decisions in advance and trigger load shedding in cases where consumption exceeds the
L.; Marhic, B.; Delahoche, L.
Predicting Energy Consumption
authorized threshold. This will have a significant impact on planning the power quality and the
Using LSTM, Multi-Layer GRU and maintenance of power equipment.
Drop-GRU Neural Networks. Sensors
2022, 22, 4062. https://doi.org/ Keywords: neural networks; time series; LSTM; GRU; Drop-GRU; energy consumption prediction;
10.3390/s22114062 load shedding
autoregressive integrated moving average (ARIMA) [9]. The machine learning models
consist of three methodologies: Decision Tree (DT), Support Vector Machine (SVM) and
Artificial Neural Network (ANN) [10]. Finally, as a breakthrough in artificial intelligence,
deep learning approaches automatically perform in treating highly nonlinear features via
a cascade of multiple layers. Recurrent Neural Network (RNN) and Back-Propagation
Neural Network (BPNN) are two popular algorithms used for predicting time series [11].
So, a forecasting system based BPNN is widely used by researchers, due to its strong
adaptability and computing ability. However, it also has some shortcomings, such as the
final training result which can easily fall into a local extremum [12]. RNN can remember the
historical information and choose whether to retain this information for the current output.
Nevertheless, it fails to maintain the long temporal dependence because of the serious
vanishing/exploding gradient problem [13]. To alleviate these problems, an improved
version of RNN, named long short-term memory network (LSTM) is proposed. The LSTM
network is based on the introduction of a cell memory in the RNN which presents a promis-
ing solution to inhibit the gradient disappearance [14]. Their contribution is particularly
manifest in the case of long sequences of events. It has also demonstrated a significant
improvement in terms of forecasting stability. The LSTM can give more stable forecasting
power in time-series prediction compared to traditional RNN [15]. In this paper, an LSTM
network based algorithm for forecasting power consumption is presented. Indeed, LSTM
models are currently widely used as the most robust approach to dealing with sequential
data. Their contribution is especially obvious in the case of rather long sequences of events.
Time series analysis and forecasting are currently at the beginning of their potential. It
should be noted that, in the case of the construction of global models for the time series,
the LSTM models allow the creation of high-performance models, both for point forecasts
and long-term forecasts [16]. The fundamental question is then how powerful and accurate
these newly introduced techniques are when compared with traditional approaches. In this
work, a dynamic model based on time series analysis and the LSTM is proposed to forecast
the power consumption, detect the power peaks, and then shed the load. The aim of this
prediction is to maintain the power balance between producers and consumers to ensure
the security of the electrical grid. In order to make full use of the various data in the power
consumption and achieve accurate predictions, different deep learning-based algorithms
were proposed, which comprise the LSTM, the GRU, and the Drop-GRU models. In the
proposed model, the GRU module is exploited to model dynamic changes in historical
power consumption data for better learning potential characteristics in time sequence
data, and the dropout process is used as a weight regularization technique for reducing
overfitting. In an analogous way, the LSTM method is chosen due to its ability to preserve
and train the characteristics of the provided data for a long period of time. Therefore, this
technique has achieved a significant success in recent years. Indeed, the paper provides
in-depth guidance on data processing and training of LSTM models for a set of power
consumption time series data. Moreover, a comparison study between GRU, Drop-GRU,
and LSTM models concerning their performance in reducing error rates is performed.
The article is organized as follows. Section 2 presents the context of this work. Section 3
explains the methodology adopted in this study. Firstly, it provides the general framework
of the proposed predictive models. Then it gives, step by step, the implementation process
of the proposed algorithms; it covers the dataset processing description, the parameters
and the evaluation indices. Section 4 shows the experimental detail, the results analysis
and the performance evaluation of the LSTM networks in comparison to the GRU and the
Drop-GRU models. Finally, Section 5 concludes this work.
this subject is part of a vast VERTPOM research project which wants to develop positive
energy territories through the design of smart electricity grids. The VERTPOM project will
maintain an optimized balance between the energy available from production regarding
uses by applying a set of algorithms for forecasting and simulating the levels of power
production and consumption on the various distribution systems. It treats an actual
database. The use of artificial intelligence networks, such as machine and deep learning, is
preferred. The key challenge of this project is to simulate all possible scenarios allowing
a positive production/consumption balance. Hence, the main objectives of VERTPOM
are anticipating the power requirement according to all the parameters available in the
area such as climatology, scheduled events and consumption in order to make the best
decision regarding consumption priorities among renewable and conventional energies.
Moreover, it should optimize the peak consumption and interact intelligently with the
consumer-player. So the aim of all previous objectives is to develop a set of digital tools for
power management, to design adequate smartgrids for energy suppliers and consumers
and to guarantee the safety of electrical infrastructures.
Energy storage
Photovoltaic sources
Industrial consumers
Public buildings
Traditional sources
of electricity Smart meter
control
Wind sources
Public lighting
Smart meters,
control,
data concentrator
Electric vehicles
Consumers
Data pre-processing
Handle missing values and normalization
Feature selection
Set parameters, train-test split
In fact, these data show that each area’s power consumption has various distribution
characteristics as illustrated in Figures 3 and 4. The presented profiles are of different types:
cyclical, irregular, stationary and with trends, etc.
12
10
8
Power (KW)
0
0 1 2 3 4
Time (year)
0.6
0.5
0.4
Power (KW)
0.3
0.2
0.1
0
0 1 2 3 4
Time (year)
2
where the smoothing constant is equal to n+ 1
with n as the number of time periods.
The EMA formula is based on the previous day0 s EMA value. Since it has to start the
computations somewhere, the initial value for the first EMA calculation will actually
be an SMA. It is calculated by taking the arithmetic mean of a given set of values over
a specified period of time.
The formula for computing the SMA is presented as:
A1 + A2 + · · · + A n
SMA = , (2)
n
where
A denotes the average in period n.
• Consumption profiles have different types of data: the process of reduced centered
standardization is applied. The normalisation is performed by ZScore and the formula
is given as:
x − xmean
ZScore = , (3)
xσ
where: s
n
1
n − 1 i∑
xσ = ( xi − xmean )2 (4)
=1
n
1
xmean =
n ∑ xi . (5)
i =1
The goal of data standardization is to use a common scale, without loss of information.
The idea is to push all the values of the variable to be between −1 and 1, while keeping the
distances between the values. These two preliminary preprocessing steps allow the LSTM
network to work properly.
maintenance and the update of information comprised in cell status. Figure 5 shows the
structure of an LSTM cell. The computation process can be given in the following as [17]:
f t = σ ( w f [ h t −1 , X t ] + b f ) (6)
i t = σ ( w i [ h t − 1 , X t ] + bi ) (7)
o t = σ ( w o [ h t − 1 , X t ] + bo ) (8)
at = tanh(wa [ht−1 , Xt ] + ba ) (9)
c t = f t ∗ c t −1 + i t ∗ a t (10)
ht = ot ∗ tanh(ct ), (11)
where σ is the sigmoid activation function and it can be defined as:
σ ( x ) = (1 + e − x ) −1 . (12)
The notations f t , it and ot are the output values of the forget, the input and the output
gates, respectively. ct refers to the memory cell and at is the update and the activation of
the current cell status. Xt is the input vector and ht represents the output vector result at
time t. Finally, W f ,i,a,o are the weights matrices and b f ,i,a,o the bias vectors.
Cell state ht
Ct−1 X + Ct
F orget gate
X tanh
σ σ tanh σ X
ht−1 ht
Input gate Output gate
Figure 6 illustrates the forecast strategy framework with the LSTM model. This model
can be divided into three big parts [18]:
• The input layer is mainly used for preprocessing the original data;
• The hidden layer is used to optimize the parameters and training the data;
• The output layer is used to predict the data according to the model trained in the
hidden layer.
Raw data
Input
Data scaling and partitioning layer
Adam
optimization X X
Loss
calculation
Model prediction
4np 4pp 4p mp m
The weights between Recursive weights The bias of The weights between The bias of
the input layer in the hidden layer. the hidden layer. the hidden layer the output layer.
and the hidden layer. and the output layer.
Selecting optimal parameters for a neural network architecture can often mean the
difference between poor and peak performances. However, there is little information in
the literature on the choice of different parameters, n, m and p, of the neural network; it
involves the experience of experts.
ht−1 X + ht
rt
X 1-
X
zt at
σ σ tanh
Xt
Figure 7. The internal structure of the GRU model.
z t = σ ( w z [ h t − 1 , X t ] + bz ) (14)
E ( u ) = E ( u1 , u2 , . . . , u n ). (18)
∂E ∂E
u i +1 = u i − α ⇒ 4 ui = − α , (20)
∂ui ∂ui
where α represents the learning rate. It is a positive parameter. ui gives the value of
variables in the ith iteration of running the algorithm. In this work, the Adam algorithm
is chosen as the suitable optimizer that can update the network weights and improve the
Sensors 2022, 22, 4062 9 of 20
performances of our model. This algorithm needs less memory and is well adapted to
solving problems that implicate the learning of complex and large datasets.
3.6.2. Dropout
The deep learning neural network has a powerful memory. However, the network
tends to learn the features of data that cannot be generalized, resulting in overfitting.
Dropout is one of the most popular regularization techniques that was proposed to solve
this problem. It returns the output of a proportion of the hidden units to zero randomly
according to the Dropout in order to reduce the neural network complexity [22]. The
dropout layer deactivates some of neurons in the training process. In this work, we have
integrated a dropout layer between the two GRU layers to facilitate and accelerate the
training step.
History Data
4. Experimental Results
In this work, we present a prediction of the power consumption of a survey area for
different prediction horizons of “one day”, “three days”, “one week” and “two weeks”.
In the proposed methodology, we have implemented three models, namely LSTM, GRU
and Drop-GRU. For the several architecture structures of the various models, the network
performance is dissimilar. The internal architecture of the proposed models is predefined
and unchangeable; each topology admits a vector input of n values which are the current
power consumption at time t = 0 and the previous consumptions. We construct four neural
networks of different architectures; each one is adapted to its prediction horizon. These
networks are able to predict consumption after 30 min. Through the repetition process and
the right choice of parameters for these networks, we can predict the entire period of the
desired horizon. The parameters of each network (number of inputs, number of neurons
in the hidden layer, number of iterations, number of outputs, . . . ) are determined by the
process of training on real data. The choice of the number of data for training is set at
four times greater than the prediction horizon. The choice of these parameters is illustrated
in Table 2.
As shown in this table, the neural network architecture depends on the number of
days to predict. Indeed, the number of inputs, the number of units in the hidden layer and
the size of training data are proportional to the prediction horizon.
0.7
Real data
LSTM validation
0.6
GRU validaion
Drop-GRU validation
0.5
0.3
0.2
0.1
0
0 1 2 3 4 5 6 7 8 9
Time (day)
0.7
Real data
LSTM validation
0.6
GRU validaion
Drop-GRU validation
0.5
Power (pu)
0.4
0.3
0.2
0.1
0
0 1 2 3 4 5 6 7 8 9 10
Time (day)
0.7
Real data
LSTM validation
0.6
GRU validaion
Drop-GRU validation
0.5
Power (pu)
0.4
0.3
0.2
0.1
0
0 5 10 15 20 25
Time (day)
0.7
Real data
LSTM validation
0.6
GRU validaion
Drop-GRU validation
0.5
0.3
0.2
0.1
0
0 5 10 15 20 25 30 35 40
Time (day)
0.7
Real data
LSTM
0.6
GRU
Drop-GRU
0.5
Power (pu)
0.4
0.3
0.2
0.1
0
0 1 2 3 4 5 6
Time (day)
0.7
Real data
LSTM
0.6
GRU
Drop-GRU
0.5
Power (pu)
0.4
0.3
0.2
0.1
0
0 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15
Time (day)
Figure 14. Prediction results of 3 days power consumption.
0.7
Real data
LSTM
0.6
GRU
Drop-GRU
0.5
Power (pu)
0.4
0.3
0.2
0.1
0
12 13 14 15
Time (day)
0.7
Real data
LSTM
0.6
GRU
Drop-GRU
0.5
Power (pu)
0.4
0.3
0.2
0.1
0
0 3 6 9 12 15 18 21 24 27 30 33
Time (day)
Figure 16. Prediction results of 7 days power consumption for the first area.
0.7
Real data
LSTM
0.6
GRU
Drop-GRU
0.5
Power (pu)
0.4
0.3
0.2
0.1
0
30 33
Time (day)
Figure 17. Zoom version Prediction of 7 days power consumption for the first area.
0.7
Real data
LSTM
0.6
GRU
Drop-GRU
0.5
Power (pu)
0.4
0.3
0.2
0.1
0
5 10 15 20 25 30 35 40 45 50 55 60 65 70 75
Time (day)
Figure 18. Prediction of 15 days power consumption for the first area.
0.7
Real data
LSTM
0.6
GRU
Drop-GRU
0.5
Power (pu)
0.4
0.3
0.2
0.1
0
70 Time (day) 75
Figure 19. Zoom version Prediction of 15 days power consumption for the first area.
LSTM
0.7
Predicted power consumption
Average power consumption
0.6
Threshold of power consumption
Load shedding
0.5
0.3
0.2
0.1
0
15
Time (day)
Figure 20. Detection of power consumption peaks during 1 day using the LSTM model.
GRU
0.7
Predicted power consumption
Average power consumption
0.6
Threshold of power consumption
Load shedding
0.5
Power (pu)
0.4
0.3
0.2
0.1
0
15 Time (day)
Figure 21. Detection of power consumption peaks during 1 days using the GRU model.
Drop-GRU
0.6
Predicted power consumption
Average power consumption
Threshold of power consumption
0.5
Load shedding
0.4
Power (pu)
0.3
0.2
0.1
0
15
Time (day)
Figure 22. Detection of power consumption peaks during 1 days using the Drop-GRU model.
Sensors 2022, 22, 4062 17 of 20
LSTM
0.7
Predicted power consumption
Average power consumption
0.6 Threshold of power consumption
Load shedding
0.5
0.3
0.2
0.1
0
35
Time (day)
Figure 23. Detection of power consumption peaks during 3 days using the LSTM model.
GRU
0.7
Predicted power consumption
Average power consumption
0.6 Threshold of power consumption
Load shedding
0.5
Power (pu)
0.4
0.3
0.2
0.1
0
35
Time (day)
Figure 24. Detection of power consumption peaks during 3 days using the GRU model.
Drop-GRU
0.7
Predicted power consumption
Average power consumption
0.6 Threshold of power consumption
Load shedding
0.5
Power (pu)
0.4
0.3
0.2
0.1
0
35
Time (day)
Figure 25. Detection of power consumption peaks during 3 days using the Drop-GRU model.
As shown from these figures, the three models are able to predict the peaks of the
power consumption with a very low shift of time compared to the real time. The power
Sensors 2022, 22, 4062 18 of 20
consumption forecasted by the GRU and Drop-GRU are very similar to the true data but
the GRU model is better at the prediction of high values.
In this table, it is observed that the uncertainty of the predicted results of the three
models and the execution times increase with the increasing of the prediction horizon.
The statistics of simulation results of the LSTM, GRU and Drop-GRU models for the testing
dataset show that the four RMSE and the four MAE values of LSTM models are larger
than those of the GRU and GRU-Drop models. In contrast, the four correlation values
R are smaller than those of GRU and Drop-GRU. Three of four correlation values of the
Drop-GRU models are also larger than those of the GRU models.
The results show that the Drop-GRU neural network is more efficient and performs
better than the GRU and the LSTM models. The Drop-GRU algorithm produces better
results in terms of accuracy and prediction speed compared to the LSTM and GRU models.
The GRU performed better than the Drop-GRU in the detection of high values of power con-
sumption but it is less fast. We can deduce that the Drop-GRU architecture produces very
satisfactory results and the prediction results are precise and reliable. The two performance
indices, RMSE and MAE, have low values and R has perfect values (near +1), as shown in
Table 3 and in the different prediction curves (Figures 13–23). The consumption forecasts
are close and representative of the actual consumption. We emphasize that the learning
time depends on the approach used for forecasting. This learning time is proportional to
the size of the network and to the prediction horizon.
To conclude, the three models are able to simulate power prediction processes accu-
rately by employing time sequences input. The experimental results show the effectiveness
of the proposed prediction algorithms for medium-short time period power load fore-
casting. GRU and LSTM models predict power consumption with good accuracy. Given
that GRU has simpler structures and fewer parameters, and requires less time for model
training, it may be the preferred method for short term prediction and it can be improved
by hybridization with other techniques such as the dropout.
5. Conclusions
This paper presents an energy management strategy based on the forecasting pro-
cess and proposes three deep neural networks: LSTM, GRU and Drop-GRU. The main
Sensors 2022, 22, 4062 19 of 20
objective of these approaches is to forecast and control the load consumption. The power
consumption prediction methods firstly treated the input data, performed effective fea-
ture extraction and then built the appropriate network structure to optimize the ability
prediction. Finally, a comparative study of the proposed algorithms is performed. These
three techniques were implemented and tested on a set of power load data and the results
indicated that the Drop-GRU was superior to the GRU and the LSTM. More specifically,
the GRU approach is very suitable for our project to forecast the energy consumption over
a defined horizon based on previous consumption readings, which will allow us to predict
consumption peaks and predict in advance an optimal decision-making scenario for load
shedding. The future direction of the research is to develop hybrid models with an even
higher accuracy and even higher speeds and we can further improve these results by taking
into account other external factors such as meteorological information and information on
holidays. So, from the prediction results and the external data on the production capac-
ity, it is easy to detect high consumption points that exceed the authorized consumption
threshold and then protect the electrical grid.
Author Contributions: Conceptualization, L.C.-A. and B.M.; methodology, S.M., B.M. and L.D.;
software, S.M.; validation, S.M., L.C.-A.; formal analysis, L.D. and L.C.-A.; investigation, S.M. and
L.D.; resources, B.M. and L.D.; data curation, B.M. and L.D.; writing—original draft preparation, S.M.;
writing—review and editing, L.C.-A. and B.M.; visualization, S.M.; supervision, L.C.-A. and L.D.;
project administration, B.M., L.D. and L.C.-A.; funding acquisition, B.M. All authors have read and
agreed to the published version of the manuscript.
Funding: This research received no external funding.
Institutional Review Board Statement: Not applicable.
Informed Consent Statement: Not applicable.
Data Availability Statement: Not applicable.
Conflicts of Interest: The authors declare no conflict of interest.
References
1. Yan, K.; Wang, X.; Du, Y.; Jin, N.; Huang, H.; Zhou, H. Multi-step short-term power consumption forecasting with a hybrid deep
learning strategy. Energies 2018, 11, 3089. [CrossRef]
2. Siami-Namini, S.; Tavakoli, N.; Namin, A.S. A comparison of ARIMA and LSTM in forecasting time series. In Proceedings of the
2018 17th IEEE International Conference on Machine Learning and Applications (ICMLA), Orlando, FL, USA, 17–20 December
2018; pp. 1394–1401.
3. Chen, K. APSO-LSTM: An improved LSTM neural network model based on APSO algorithm. In Journal of Physics: Conference
Series; IOP Publishing: Bristol, UK, 2020; p. 012151.
4. Karevan, Z.; Suykens, J.A. Transductive LSTM for time-series prediction: An application to weather forecasting. Neural Netw.
2020, 125, 1–9. [CrossRef] [PubMed]
5. Tsai, Y.T.; Zeng, Y.R.; Chang, Y.S. Air pollution forecasting using RNN with LSTM. In Proceedings of the 2018 IEEE 16th
International Conference on Dependable, Autonomic and Secure Computing, 16th International Conference on Pervasive
Intelligence and Computing, 4th International Conference on Big Data Intelligence and Computing and Cyber Science and
Technology Congress (DASC/PiCom/DataCom/CyberSciTech), Athens, Greece, 12–15 August 2018; pp. 1074–1079.
6. Jin, Z.; Yang, Y.; Liu, Y. Stock closing price prediction based on sentiment analysis and LSTM. Neural Comput. Appl. 2019, 32,
9713–9729. [CrossRef]
7. Shakya, A.; Michael, S.; Saunders, C.; Armstrong, D.; Pandey, P.; Chalise, S.; Tonkoski, R. Solar irradiance forecasting in remote
microgrids using markov switching model. IEEE Trans. Sustain. Energy 2016, 8, 895–905. [CrossRef]
8. Rendon-Sanchez, J.F.; de Menezes, L.M. Structural combination of seasonal exponential smoothing forecasts applied to load
forecasting. Eur. J. Oper. Res. 2019, 275, 916–924. [CrossRef]
9. Cao, T.D.; Delahoche, L.; Marhic, B.; Masson, J.B. Occupancy Forecasting using two ARIMA Strategies. In Proceedings of the
ITISE 2019: International Conference on Time Series and Forecasting, Granada, Spain, 25–27 September 2019; Volume 2.
10. Wang, K.; Qi, X.; Liu, H. Photovoltaic power forecasting based LSTM-Convolutional Network. Energy 2019, 189, 116225.
[CrossRef]
11. Wang, J.Q.; Du, Y.; Wang, J. LSTM based long-term energy consumption prediction with periodicity. Energy 2020, 197, 117197.
[CrossRef]
Sensors 2022, 22, 4062 20 of 20
12. Wang, L.; Wang, Z.; Qu, H.; Liu, S. Optimal forecast combination based on neural networks for time series forecasting. Appl. Soft
Comput. 2018, 66, 1–17. [CrossRef]
13. Yang, B.; Yin, K.; Lacasse, S.; Liu, Z. Time series analysis and long short-term memory neural network to predict landslide
displacement. Landslides 2019, 16, 677–694. [CrossRef]
14. Lin, C.; Zhang, Y.; Ivy, J.; Capan, M.; Arnold, R.; Huddleston, J.M.; Chi, M. Early diagnosis and prediction of sepsis shock by
combining static and dynamic information using convolutional-LSTM. In Proceedings of the 2018 IEEE International Conference
on Healthcare Informatics (ICHI), New York, NY, USA, 4–7 June 2018; pp. 219–228.
15. Kim, K.; Kim, D.K.; Noh, J.; Kim, M. Stable forecasting of environmental time series via long short term memory recurrent neural
network. IEEE Access 2018, 6, 75216–75228. [CrossRef]
16. Heidari, A.; Khovalyg, D. Short-term energy use prediction of solar-assisted water heating system: Application case of combined
attention-based LSTM and time-series decomposition. Sol. Energy 2020, 207, 626–639. [CrossRef]
17. Tovar, M.; Robles, M.; Rashid, F. PV Power Prediction, Using CNN-LSTM Hybrid Neural Network Model. Case of Study:
Temixco-Morelos, México. Energies 2020, 13, 6512. [CrossRef]
18. Zhang, T.; Song, S.; Li, S.; Ma, L.; Pan, S.; Han, L. Research on gas concentration prediction models based on LSTM multidimen-
sional time series. Energies 2019, 12, 161. [CrossRef]
19. Wu, L.; Kong, C.; Hao, X.; Chen, W. A short-term load forecasting method based on GRU-CNN hybrid neural network model.
Math. Probl. Eng. 2020, 2020, 1428104. [CrossRef]
20. Shiang, E.P.L.; Chien, W.C.; Lai, C.F.; Chao, H.C. Gated recurrent unit network-based cellular trafile prediction. In Proceedings of
the 2020 International Conference on Information Networking (ICOIN), Barcelona, Spain, 7–10 January 2020; pp. 471–476.
21. Ghaziasgar, M.; Naeini, A.T. Neural Network for Routing in a Directed and Weighted Graph. In Proceedings of the 2008 Eighth
International Conference on Intelligent Systems Design and Applications, Kaohsuing, Taiwan, 26–28 November 2008; pp. 631–636.
22. Wen, L.; Zhou, K.; Yang, S. Load demand forecasting of residential buildings using a deep learning model. Electr. Power Syst. Res.
2020, 179, 106073. [CrossRef]
23. Nguyen, Q.H.; Ly, H.B.; Ho, L.S.; Al-Ansari, N.; Le, H.V.; Tran, V.Q.; Pham, B.T. Influence of data splitting on performance of
machine learning models in prediction of shear strength of soil. Math. Probl. Eng. 2021, 2021. [CrossRef]
24. Li, Y.; Zhu, Z.; Kong, D.; Han, H.; Zhao, Y. EA-LSTM: Evolutionary attention-based LSTM for time series prediction. Knowl.-Based
Syst. 2019, 181, 104785. [CrossRef]
25. Sommer, F.; et Stuke, M. An efficient and fast method to calculate integral experimental correlation coefficients–S2Cor. Ann. Nucl.
Energy 2021, 157, 108209. [CrossRef]