Energies 17 05181 v2
Energies 17 05181 v2
Energies 17 05181 v2
Article
Multiple Load Forecasting of Integrated Renewable Energy
System Based on TCN-FECAM-Informer
Mingxiang Li 1 , Tianyi Zhang 2 , Haizhu Yang 1 and Kun Liu 3, *
1 School of Electrical Engineering and Automation, Henan Polytechnic University, Jiaozuo 454003, China;
[email protected] (M.L.); [email protected] (H.Y.)
2 School of Electrical and Information Engineering, Changsha University of Science and Technology,
Changsha 410114, China; [email protected]
3 Tianjin Eco-Environmental Monitoring Center, Tianjin 300191, China
* Correspondence: [email protected]; Tel.: +86-156-2078-4296
Abstract: In order to solve the problem of complex coupling characteristics between multivariate
load sequences and the difficulty in accurate multiple load forecasting for integrated renewable
energy systems (IRESs), which include low-carbon emission renewable energy sources, in this paper,
the TCN-FECAM-Informer multivariate load forecasting model is proposed. First, the maximum
information coefficient (MIC) is used to correlate the multivariate loads with the weather factors to
filter the appropriate features. Then, effective information of the screened features is extracted and
the frequency sequence is constructed using the frequency-enhanced channel attention mechanism
(FECAM)-improved temporal convolutional network (TCN). Finally, the processed feature sequences
are sent to the Informer network for multivariate load forecasting. Experiments are conducted with
measured load data from the IRES of Arizona State University, and the experimental results show
that the TCN and FECAM can greatly improve the multivariate load prediction accuracy and, at
the same time, demonstrate the superiority of the Informer network, which is dominated by the
attentional mechanism, compared with recurrent neural networks in multivariate load prediction.
IRES multivariate load forecasting, and Ref. [4] analyzes a variety of regression models by
comparing them, and ultimately recommends the use of the Gaussian process regression
model for power load forecasting.
Deep learning methods use multi-layer neural network stacking to construct models
for load forecasting. Ref. [5] developed a complex neural network (ComNN) for multivari-
ate load forecasting. Ref. [6] utilized a long short-term neural network (LSTM) combined
with genetic algorithms (GAs) for pre-training, and then fine-tuned it using transfer learning
to build a multivariate load forecasting model GA-LSTM. Ref. [7] proposed a multivariate
load prediction model combining the transformer and Gaussian process.
In order to further improve the overall performance of the model and fully exploit the
effective information of relevant features and historical load data, signal decomposition
techniques and convolutional neural networks are applied in load forecasting.
For signal decomposition techniques, Ref. [8] utilized the backpropagation (BP) neural
network for the initial load prediction and utilized improved complete ensemble empirical
mode decomposition with adaptive noise (ICEEMDAN) to decompose and reconstruct the
results, before utilizing the gated recurrent unit (GRU) algorithm for the further prediction
of the reconstructed components and combining it with the prediction results of the BP
neural network. Ref. [9] proposed a two-stage decomposition hybrid forecasting model
for the multivariate load forecasting of IRESs by combining modal decomposition with
the bidirectional long short-term memory neural network (Bi-LSTM). Ref. [10] utilized
variational mode decomposition (VMD) with the k-means clustering algorithm (FK) com-
bined with the SecureBoost model for the forecasting of electricity load, and Ref. [11] used
a combination of complete ensemble empirical mode decomposition with adaptive noise
(CEEMDAN) and VMD to form a secondary mode decomposition. CEEMDAN decom-
poses the load sequence into multiple subsequences, while VMD further decomposes the
high-frequency components for multivariate load prediction. Ref. [12] combines ICEEM-
DAN and successive variational mode decomposition (SVMD) to form an aggregated
hybrid mode decomposition (AHMD) for modal decomposition, and after that, the multi-
variate linear regression (MLR) model is used, respectively, with the temporal convolutional
network-bidirectional gated recurrent unit (TCN-BiGRU) fusion neural network for the
prediction of low-frequency components and medium–high-frequency components, respec-
tively, and the prediction results are superimposed for IES load prediction. In Ref. [13],
modal decomposition is combined with the multitask learning model; the middle- and
high-frequency sequences of multivariate loads after modal decomposition are input into
the multitask transformer model for prediction; the low-frequency sequences are input into
the multitask bi-directional gated logic recurrent unit for prediction; and the results of the
two are ultimately superimposed to obtain the final prediction results.
In terms of the convolutional neural network, Ref. [14] performed feature extraction
by fusing the convolutional neural network (CNN) and GRU, followed by a multitask
learning approach to build a multivariate load prediction model. Ref. [15] utilized the
multi-channel convolutional neural network (MCNN) to extract high-dimensional spatial
features and combined it with the LSTM network for IRES load prediction. Ref. [16], on the
other hand, used a three-layer CNN to extract features, followed by load prediction using
the transformer model.
The analysis of the previous related research revealed the following aspects that can
be improved in the research related to multivariate load forecasting:
(1) Most of the different feature extraction methods use traditional modal decomposition
and convolutional neural networks, and more advanced networks can be further
adopted and improved to enhance the model’s feature analysis capability and improve
the model’s performance.
(2) Most of the deep prediction models in the research use traditional recurrent neural
networks, and attempts can be made to use other types of network structures to
capture correlations between sequences and to improve the ability of deep prediction
Energies 2024, 17, 5181 3 of 15
1.2. Contributions
(1) The CAM and DCT are combined with TCN for feature analysis. The CAM and DCT
enhance the extraction of key features, and the TCN is responsible for the modeling
of time series to lay the foundation for the next prediction.
(2) The Informer network combined with the improved TCN is applied to the prediction
of multivariate loads, and the feasibility of the network model of the encoder–decoder
framework, which is dominated by the attention structure in the deep prediction
model, is verified for the multivariate load prediction.
(3) This paper uses the MIC method to analyze the correlation between multivariate
loads and other factors such as meteorology mainly in order to analyze the correlation
between multivariate loads and other factors, and to make the selection of features for
the input model.
The next part of this paper focuses on the following: the second part introduces
the TCN model, the improvement module, and the subsequent Informer network model
used for predicting, and proposes an overall modeling framework for improving the TCN
combined with the Informer network. In the third part, the effectiveness of the proposed
model is analyzed using practical examples. Finally, conclusions are drawn and future
plans are outlined.
F is the convolution operation, F (t) is the output value at time point t, k is the size of the
convolution kernel, f (i ) is the i element in the convolution kernel, and xt−d·i is the input
value corresponding to shifting from the t moment to the past d · i moment.
The expansion coefficient of inflated causal convolution grows exponentially with
the increase in the number of layers, which in turn expands the receptive fields, enabling
the model to capture local features on different time scales, thus improving the modeling
capability for time-series data. And local features are the patterns and regularities captured
is the input value corresponding to shifting from the t moment to the past d ⋅ i mo-
ment.
The expansion coefficient of inflated causal convolution grows exponentially with
the increase in the number of layers, which in turn expands the receptive fields, enabling
Energies 2024, 17, 5181
the model to capture local features on different time scales, thus improving the modeling
4 of 15
capability for time-series data. And local features are the patterns and regularities cap-
tured within these limited receptive fields. However, compared to ordinary convolutional
networks, TCNs have a reduced number of layers, but they are able to obtain a larger
within these limited receptive fields. However, compared to ordinary convolutional net-
receptive field, thus avoiding the repeated extraction of information and being able to
works, TCNs have a reduced number of layers, but they are able to obtain a larger receptive
fully analyze the features of the data.
field, thus avoiding the repeated extraction of information and being able to fully analyze
As the number of layers increases, the expansion coefficient increases exponentially,
the features of the data.
which in turn expands the receptive fields, allowing the model to capture local features
As the number of layers increases, the expansion coefficient increases exponentially,
on different time scales. And local features are the patterns and regularities captured
which in turn expands the receptive fields, allowing the model to capture local features
within these limited
on different receptive
time scales. Andfields.
localHowever,
features arecompared with the
the patterns andordinary convolutional
regularities captured
network, the limited
within these TCN isreceptive
reduced in the number
fields. However, ofcompared
layers, butwith
it isthe
able to obtain
ordinary a larger re-
convolutional
ceptive
network, field,
thethus
TCNavoiding
is reducedthe repeated
in the numberextraction of information
of layers, and being
but it is able able atolarger
to obtain fully
analyze the characteristics of the data.
receptive field, thus avoiding the repeated extraction of information and being able to fully
In addition,
analyze traditionalofneural
the characteristics the data.networks are prone to gradient explosion and network
degradation when the number of layers
In addition, traditional neural networks is gradually
are pronedeepened,
to gradientso the residual
explosion andconnectiv-
network
ity module [18]
degradation whenis applied in theofoutput
the number layers layer of the TCN.
is gradually In Figure
deepened, 1, residual
so the the residual connec-
connectivity
tivity module structure is shown. The module consists of two identical internal
module [18] is applied in the output layer of the TCN. In Figure 1, the residual connectivity units and
amodule
one-dimensional convolutional network, where the internal units include
structure is shown. The module consists of two identical internal units and a one- the expansion
causal convolutional
dimensional layer, the
convolutional normalization
network, where thelayer, the units
internal activation function,
include and the causal
the expansion drop-
out layer.
convolutional layer, the normalization layer, the activation function, and the dropout layer.
Figure 1.
Figure Residual connectivity
1. Residual connectivity module
module structure.
structure.
2.2. Frequency-Enhanced
2.2. Frequency-Enhanced Channel
Channel Attention
Attention Mechanism
Mechanism (FECAM)
(FECAM)
The FECAM is an improvement in CAM. The CAM has been applied in CNNs. In the
The FECAM is an improvement in CAM. The CAM has been applied in CNNs. In the
squeeze-and-excitation (SE) module [19], the relationships between different channels of
squeeze-and-excitation (SE) module [19], the relationships between different channels of
the feature map are modeled using global information, and the feature map is re-modified
to improve its representational ability. However, the frequency domain information of the
data extracted by the Fourier transform in channel attention can introduce high-frequency
components due to the period problem, thus affecting the feature extraction. The FECAM
uses the DCT [20] instead of the Fourier transform to avoid the above problem. The steps
of the FECAM are as follows:
First, the FECAM splits each input feature map along the channel dimension into n
subgroups as V i = {v0 , v1 , . . . , vn−1 }. Then, the DCT is performed on all the components of
Energies 2024, 17, 5181 5 of 15
each subgroup, where each single channel is subjected to the same frequency of components
to obtain Equation (2):
j= L j
Freqi = DCTj (V i ) = ∑ j=0 s (V i l ) Bl (2)
j πl 1
Bl = cos( ( j + )) (3)
Ls 2
where l ∈ {0, 1, . . . , Ls − 1}, i ∈ {0, 1, . . . , n − 1}, and j ∈ {0, 1, . . . , LS − 1}, in which
j are the frequency component 1D indices corresponding to V i , and Freqi ∈ R L is the
L-dimensional vector after the discrete cosine transformation.
Then, the whole frequency channel vector can be obtained by stack operation.
1 LS
LS i∑
Zc = GAP( xc ) = x c (i ) (7)
=1
In Equation (4), σ and δ are RELU and Sigmoid activation functions, and W1 and
W2 are the two fully connected layers. In Equation (5), att is the learned attention vector
that can be scaled for each channel of data, Z consists of Zc from different channels; in
Equation (6), Zc is obtained by scaling the inputs x in the temporal dimension, and LS is
the length of the time-series sequence.
Through the frequency-enhanced channel attention mechanism, each channel feature
interacts with each frequency component to comprehensively obtain important temporal
information from the frequency domain and enhance the diversity in information of features
extracted by the network.
QK T
A( Q, K, V ) = So f tmax( √ )V (8)
d
where Q is the improved query matrix, K is the attention content, and QK T calculates the
attention weight of Q on V.
The distillation operation assigns higher weights to dominant features with dominant
attention. From the j layer to j + 1, the distillation operation is as follows:
h i
X jt+1 = MaxPool ( ELU (Conv1d( X jt ))) (9)
AB
[·] AB is the attention module, and the one-dimensional convolution on the time series is
denoted by Conv1d[·]. This is followed by maximum pooling after passing the activation
function ELU.
The decoder is mainly responsible for processing the feature information obtained
from the encoder, and finally outputting the result through the fully connected layer.
Energies 2024, 17, 5181 6 of 15
where p( ai , b j ) is the joint probability density of variables ai and b j , and p( ai ) and p(b j ) are
the marginal probability densities of variables ai and b j , respectively.
Second, define a grid, denoted as G ( x, y); divide ai and b j in the dataset D = ( ai , b j )
into two grids, x and y, respectively; and compute to obtain the maximum MI value of D
under the division of the grid G:
f M1 ( D, x, y) = max f M1 ( D | G ) (11)
f M1 ( D, x, y)
M ( D ) x,y = (12)
log2 min{ x, y}
2.5.
2.5. TCN-FECAM-Informer
TCN-FECAM-Informer
Combined with the
Combined with theanalysis
analysisofofthe
theabove-related
above-related model
model structure,
structure, thisthis paper
paper pro-
proposes
poses a multivariate
a multivariate load load pre-model
pre-model fusingfusing the TCN,
the TCN, frequency-enhanced
frequency-enhanced channel
channel atten-
attention
tion mechanism, and Informer network, and it is shown
mechanism, and Informer network, and it is shown in Figure 2. in Figure 2.
Figure 2.
Figure Thewhole
2. The wholeframework
framework of
of TCN-FECAM-Informer.
TCN-FECAM-Informer.
Firstly, the features with strong correlation with the load screened by the MIC
method are fed into the TCN to extract the local information in it, such as load fluctuation
due to rainfall. Secondly, the output of the TCN is processed through the FECAM to en-
hance the diversity in information of the extracted information, and to generate attention
vectors with weights to increase the influence of important features. Then, the processed
Energies 2024, 17, 5181 7 of 15
Firstly, the features with strong correlation with the load screened by the MIC method
are fed into the TCN to extract the local information in it, such as load fluctuation due to
rainfall. Secondly, the output of the TCN is processed through the FECAM to enhance the
diversity in information of the extracted information, and to generate attention vectors
with weights to increase the influence of important features. Then, the processed feature
data are input to the Informer network for prediction to obtain the final prediction result.
3. Example Analysis
3.1. Experimental Environment
The experiments in this paper were realized through Python version 3.10.11 with torch
version 2.0.1 as the learning framework. The computer configuration was an Intel Core
i7-12700H 2.30 GHz CPU, NVIDIA GeForce RTX 3060 laptop GPU (Produced by NVIDIA,
a company headquartered in Santa Clara, CA, USA), and a Windows 11 operating system.
Definition Equation
1 m
MAE Mean Absolute Error m ∑i =1 (Yi − Ŷi )
MAE = q
RMSE Root Mean Square Error 2
RMSE = m1 ∑im=1 (Yi − Ŷi )
MAPE Mean Absolute Percentage Error 100% m Ŷi − Yi
MAPE = ∑
m i =1 Yi
Yi is the true value, Ŷi is the predicted value, and m is the number of samples in test set.
Figure3.
Figure Characterizationof
3.Characterization ofthe
themain
maineffects
effectsof
ofmultiple
multiple loads.
loads.
The multivariate load prediction considers the mutual coupling characteristics among
The multivariate load prediction considers the mutual coupling characteristics
electric, heat, and cold loads, which reduces the errors of electric, heat, and cold loads by
among electric, heat, and cold loads, which reduces the errors of electric, heat, and cold
24.5%, 2.90%, and 22.42%, respectively, indicating that there is a correlation among the
loads by 24.5%, 2.90%, and 22.42%, respectively, indicating that there is a correlation
multivariate loads. In the comprehensive energy system of the campus, the correlation
among the multivariate loads. In the comprehensive energy system of the campus, the
correlation between electric load, cooling load, and other loads is strong, and the correla-
tion between heat load and other loads is weak.
between electric load, cooling load, and other loads is strong, and the correlation between
heat load and other loads is weak.
Electric Load/kW
MAPE/% MAE RMSE
Informer 2.64 405.56 583.27
TCN-Informer 1.76 264.08 368.45
TCN-FECAM-Informer 1.36 205.01 287.46
Heat Load/MMBTU
MAPE/% MAE RMSE
Informer 4.27 74.56 96.88
TCN-Informer 2.96 51.70 65.03
TCN-FECAM-Informer 2.03 35.21 45.10
Cold Load/Tons
MAPE/% MAE RMSE
Informer 4.91 688.43 918.01
TCN-Informer 4.57 609.26 814.59
TCN-FECAM-Informer 4.00 545.49 730.27
Analyzing the prediction results, the following conclusions can be drawn: firstly,
in multivariate load prediction, the prediction error of the TCN-Informer network is
lower than that of the Informer model, which indicates that the TCN is able to carry out
preliminary extraction of the local features of the factors affecting the load, and thus the
overall prediction effect of the combined model is superior to that of the single network
model. In contrast, the model proposed in this paper combines the FECAM to analyze
the frequency domain of the processed data from TCNs, filtering the high-frequency noise
to further enhance the effectiveness of the extracted features, which further improves the
effectiveness in dealing with multivariate load forecasting. The prediction error of the
TCN-FECAM-Informer model is lower than that of the TCN-Informer model, in which
the MAPE of the electric, heat, and cooling loads decreases by 22.7%, 31.4%, and 12.5%,
respectively. In summary, the combined forecasting model proposed in this paper has good
forecasting results and is feasible in multivariate load forecasting. The multivariate load
forecasting results of different models are shown in Figures 4–6 below.
Energies 2024, 17, x FOR PEER REVIEW 11 of 16
Transformer model and the LSTM model, which is a network that uses LSTM for both the
encoder and decoder. The results are shown in Tables 7–9.
Electric Load/kW
MAPE/% MAE RMSE
LSTM(Seq2seq) 5.10 750.99 1006.56
TCN-FECAM-Transformer 2.26 319.77 400.34
TCN-FECAM-Informer 1.36 205.01 287.46
Heat Load/MMBTU
MAPE/% MAE RMSE
LSTM(Seq2seq) 8.80 145.01 178.03
TCN-FECAM-Transformer 3.69 65.27 83.42
TCN-FECAM-Informer 2.03 35.21 45.10
Cold Load/Tons
MAPE/% MAE RMSE
LSTM(Seq2seq) 29.90 3639.03 3984.12
TCN-FECAM-
7.93 966.05 1175.56
Transformer
TCN-FECAM-
4.00 545.49 730.27
Informer
The prediction effect of the LSTM model is not as good as those of the other two
models. The transformer model and Informer model are encoder–decoder architecture
models based on the attention mechanism, which are unsupervised learning models. In
this paper, supervised training is used for training, and the attention mechanisms of the
transformer network and Informer network are utilized to calculate the correlation between
the data at any position of the input sequence, to explore more effective information of the
input sequence, to overcome the problem of long dependence on the data that cannot be
extracted by the traditional recurrent neural network, and to improve the overall prediction
performance of the model. The Informer network, on the other hand, improves on the
transformer network and further enhances the predictive performance of the model, with
the MAPE of electric, heat, and cold loads reduced by 39.8%, 45.0%, and 49.6%, respectively.
The comparison results are shown in Figures 7–9 below.
Energies 2024, 17, x FOR PEER REVIEW 13 of 16
Energies 2024, 17, x FOR PEER REVIEW 13 of 16
Energies 2024, 17, x FOR PEER REVIEW 13 of 16
Figure
Figure8.
8. Comparison of TCN-FECAM-Informer,
Comparison of TCN-FECAM-Informer,TCN-FECAM-Transformer,
TCN-FECAM-Transformer,
andand LSTM
LSTM in heat
in heat load
Figure
load 8. Comparison of TCN-FECAM-Informer, TCN-FECAM-Transformer, and LSTM in heat
forecasting.
forecasting.
Figure 8. Comparison of TCN-FECAM-Informer, TCN-FECAM-Transformer, and LSTM in heat
load forecasting.
load forecasting.
4. Conclusions
In this paper, in order to deal with the coupling relationship between multivariate loads
for IRESs, which include low-carbon-emission renewable energy sources, the correlation
analysis and feature selection of different influencing factors of multivariate loads are
carried out by using the MIC; the basic principles of the FECAM, TCN, and Informer
network are analyzed; and the TCN-FECAM-Informer multivariate load forecasting model
is proposed.
The TCN-FECAM-Informer multivariate load combination prediction model proposed
in this paper gives full play to the feature information extraction capability of the TCN,
FECAM, and Informer network, thus improving the prediction accuracy of multivariate
loads, which has certain application prospects in the field of multivariate load prediction
and provides ideas for the subsequent research.
In future research, methods to enhance the performance of IRES multivariate load
forecasting under more uncertainties will be further analyzed.
Author Contributions: Conceptualization, M.L. and H.Y.; methodology, M.L. and T.Z.; investigation,
M.L.; resources, K.L.; data curation M.L.; writing—original draft preparation, M.L.; writing—review
and editing, K.L. All authors have read and agreed to the published version of the manuscript.
Funding: This research received no external funding.
Data Availability Statement: The data used in this study are indicated in the text, and specific
enquiry information can be found in references [21,22] or contact the first author by email.
Conflicts of Interest: The authors declare no conflicts of interest.
Nomenclature
Abbreviations
IRES Integrated renewable energy system
TCN Temporal convolutional network
SVM Support vector machine
LSSVM Least-squares support vector machine
CNN Convolutional neural network
ComNN Complex neural network
MCNN Multi-channel convolutional neural network
LSTM Long short-term neural network
BiLSTM Bidirectional long short-term memory neural network
GA Genetic algorithm
BP Backpropagation
GRU Gated recurrent unit
TCN-BiGRU Temporal convolutional network–bidirectional gated recurrent unit
MBiGRU Multitask bi-directional gated recurrent unit
MultiTr Multitask transformer
MLR Multivariate linear regression
VMD Variational mode decomposition
SVMD Successive variational mode decomposition
FK Federated k-means clustering algorithm
CEEMDAN Complete ensemble empirical mode decomposition with adaptive noise
Improved complete ensemble empirical mode decomposition with
ICEEMDAN
adaptive noise
AHMD Aggregated hybrid modal decomposition
CAM Channel attention mechanism
DCT Discrete cosine transform
Energies 2024, 17, 5181 14 of 15
References
1. Jin, Y.; Xiong, X.; Zhang, J.; Pan, L.; Li, Y.; Wu, X.; Shen, J. A prospective review on the design and operation of integrated energy
system: The spotlight cast on dynamic characteristics. Appl. Therm. Eng. 2024, 253, 123751. [CrossRef]
2. Tan, Z.; De, G.; Li, M.; Lin, H.; Yang, S.; Huang, L.; Tan, Q. Combined electricity-heat-cooling-gas load forecasting model for
integrated energy system based on multi-task learning and least square support vector machine. J. Clean. Prod. 2020, 248, 119252.
[CrossRef]
3. Li, G.; Li, Y.; Roozitalab, F. Midterm Load Forecasting: A Multistep Approach Based on Phase Space Reconstruction and Support
Vector Machine. IEEE Syst. J. 2020, 14, 4967–4977. [CrossRef]
4. Madhukumar, M.; Sebastian, A.; Liang, X.; Jamil, M.; Shabbir, M.N.S.K. Regression Model-Based Short-Term Load Forecasting for
University Campus Load. IEEE Access 2022, 10, 8891–8905. [CrossRef]
5. Zhao, P.; Cao, D.; Hu, W.; Huang, Y.; Hao, M.; Huang, Q.; Chen, Z. Geometric Loss-Enabled Complex Neural Network for
Multi-Energy Load Forecasting in Integrated Energy Systems. IEEE Trans. Power Syst. 2024, 39, 5659–5671. [CrossRef]
6. Wang, Y.; Wang, H.; Meng, X.; Dong, H.; Chen, X.; Xiang, H.; Xing, J. Considering the dual endogenous-exogenous uncertainty
integrated energy multiple load short-term forecast. Energy 2023, 285, 129387. [CrossRef]
7. Hu, J.; Hu, W.; Cao, D.; Sun, X.; Chen, J.; Huang, Y.; Chen, Z.; Blaabjerg, F. Probabilistic net load forecasting based on transformer
network and Gaussian process-enabled residual modeling learning method. Renew. Energy 2024, 225, 120253. [CrossRef]
8. Xu, P.; Song, Y.; Du, J.; Zhang, F. Town gas daily load forecasting based on machine learning combinatorial algorithms: A case
study in North China. Chin. J. Chem. Eng. 2024, in press. [CrossRef]
9. Shi, J.; Teh, J.; Alharbi, B.; Lai, C.-M. Load forecasting for regional integrated energy system based on two-phase decomposition
and mixture prediction model. Energy 2024, 297, 131236. [CrossRef]
Energies 2024, 17, 5181 15 of 15
10. Yang, Y.; Wang, Z.; Zhao, S.; Wu, J. An integrated federated learning algorithm for short-term load forecasting. Electr. Power Syst.
Res. 2023, 214, 108830. [CrossRef]
11. Chen, J.; Hu, Z.; Chen, W.; Gao, M.; Du, Y.; Lin, M. Load prediction of integrated energy system based on combination of quadratic
modal decomposition and deep bidirectional long short-term memory and multiple linear regression. Autom. Electr. Power Syst.
2021, 45, 85–94.
12. Chen, H.; Huang, H.; Zheng, Y.; Yang, B. A load forecasting approach for integrated energy systems based on aggregation hybrid
modal decomposition and combined model. Appl. Energy 2024, 375, 124166. [CrossRef]
13. Zhang, Y.; Sun, M.; Ji, X.; Ye, P.; Yang, M.; Cai, F. Short-term load forecasting of integrated energy system based on modal
decomposition and multi-task learning model. High Volt. Eng. 2024, 1–18. [CrossRef]
14. Xuan, W.; Shouxiang, W.; Qianyu, Z.; Shaomin, W.; Liwei, F. A multi-energy load prediction model based on deep multi-task
learning and ensemble approach for regional integrated energy systems. Int. J. Electr. Power Energy Syst. 2021, 126, 106583.
[CrossRef]
15. Li, R.; Sun, F.; Ding, X.; Han, Y.; Liu, Y.; Yan, J. Ultra-short-term load forecasting method of user-level integrated energy system
considering multi-energy space-time coupling. Power Syst. Technol. 2020, 44, 4121–4131.
16. Tian, Z.; Liu, W.; Jiang, W.; Wu, C. CNNs-Transformer based day-ahead probabilistic load forecasting for weekends with limited
data availability. Energy 2024, 293, 130666. [CrossRef]
17. Limouni, T.; Yaagoubi, R.; Bouziane, K.; Guissi, K.; Baali, E.H. Accurate one step and multistep forecasting of very short-term PV
power using LSTM-TCN model. Renew. Energy 2023, 205, 1010–1024. [CrossRef]
18. Wu, X.; Fan, B.; Wang, J.; Hu, Q. Life prediction of lithium battery based on VMD-TCN-Attention. Chin. J. Power Sources 2023, 47,
1319–1325.
19. Hu, J.; Shen, L.; Sun, G. Squeeze-and-Excitation Networks. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision
and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 7132–7141.
20. Jiang, M.; Zeng, P.; Wang, K.; Liu, H.; Chen, W.; Liu, H. FECAM: Frequency enhanced channel attention mechanism for time
series forecasting. Adv. Eng. Inform. 2023, 58, 102158. [CrossRef]
21. Aus. Campus Metabolism [DB/OL] America. Available online: http://cm.asu.edu/ (accessed on 31 July 2023).
22. National Centers for Environmental Information [DB/OL] America. Available online: https://www.ncei.noaa.gov/ (accessed on
31 July 2023).
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual
author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to
people or property resulting from any ideas, methods, instructions or products referred to in the content.