1. Introduction
Energy consumption refers to the amount of energy used over a certain period of time, usually measured in units such as kilowatt-hours (kWh) or British thermal units (BTUs). It is an important metric for assessing energy usage and efficiency, and for understanding the overall energy needs of a region or country [
1]. According to the International Energy Agency (IEA) [
2], global energy consumption is expected to continue to increase in the coming decades, driven by population growth, urbanization, and industrialization in developing countries.
However, most of the current research has focused on forecasting in countries or regions [
3,
4]. However, only a limited number of studies exist that specifically focus on forecasting for individual buildings. Energy consumption in buildings encompasses the precise measurement of energy utilized for specific purposes such as heating, cooling, lighting, and other essential functions within residential, commercial, and institutional structures. In China and India, buildings account for a 37% [
5] and 35% [
6] of global energy consumption, making it an important area for energy efficiency and sustainability efforts [
7].
In this study, we propose the forecasting model for energy consumption for various commercial buildings, such as Hawthorn Campus - ATC Building, Hawthorn Campus - AMDC Building, Wantirna Campus - KIOSC Building. In the study, the proposed model, along with its trained data preprocessing method, demonstrates superior performance compared to other popular models, particularly when there is a sufficient amount of training data available or when there is a lack of training data. The results indicate that the proposed method and model can be used to accurately predict energy consumption in commercial buildings, which is crucial for energy management and conservation. In addition, the proposed method can be easily applied to other commercial buildings with similar energy consumption patterns, providing a practical solution for energy management in the commercial building sector.
The rest of this study is organized as follows.
Section 2 delivers the background of the research.
Section 4 provides the details of the proposed model.
Section 5 describes some popular bench-marking models.
Section 6 presents the experiment to evaluate the efficiency of the models in different datasets. Finally, the conclusion of this study is given in
Section 7.
2. Background
2.1. Forecasting Models
A forecasting model is a mathematical algorithm or statistical tool used to predict future trends, values, or events based on historical data and patterns. A forecasting model M that analyzes the historical values from the current time to return the predicted future value at time (denoted as ) is built. The objective of the forecasting model is to minimize the discrepancy between the estimated value and the actual value by seeking the closest approximation. To achieve this in a temporal context where data points are indexed in time order, a specific type of forecasting model is often used.
A time series forecasting model is a predictive algorithm that utilizes historical time-series data to anticipate future trends or patterns in data over time. Two techniques are available for building the model
M to obtain this objective, i.e., (i) univariate and (ii) multivariate time-series (TS) [
8,
9,
10]. In univariate TS, only a 1D sequence of energy consumption value
is utilized to produce the estimated value
, where
k is a period of time from the current time
[
11,
12]. By contrast, for multivariate TS, we could employ one or more other historical features in addition to the energy consumption for training model
M [
13]. They can be time field or other specific sources. Therefore, the input for multivariate TS is a multi-dimensional sequence
, with
is a vector of dimension
n). In forecasting new values, the model could be enhanced if related available features are taken into account. The additional information could help to model capture the dependencies or correlations between features and target variable. Therefore, the model could better understand the context, mitigate the impact of missing values, and make more precise predictions. Therefore, multivariate TS are frequently employed for building the forecasting model recently [
14].
2.2. Energy Consumption Forecasting
Energy consumption forecasting refers to predicting a particular region’s future energy consumption based on historical consumption and other relevant factors [
1]. Accurate energy consumption forecasting is essential for proper energy planning, pricing, and management. It plays a significant role in the transition toward a more sustainable energy future. There have been numerous studies on energy consumption forecasting, and various models have been proposed for this purpose [
15]. As shown in
Figure 1, in the literature, forecasting models often fall into two categories: (
i) Conventional Models and (
ii) Artificial Intelligence (AI) Models.
Table 1 shows different forecasting models for commercial buildings. It shows the name of forecasting techniques, location, and performance evaluation indexes with the best accuracy.
Conventional models used in energy consumption forecasting commonly include Stochastic time series (TS) models [
16], regression models (RMs) [
17], and gray models (GMs) [
18,
19]. These models typically require historical energy consumption data as input and use various statistical and mathematical techniques to make future energy consumption predictions. However, these models may not be able to capture complex nonlinear relationships and may require manual feature engineering, making them less efficient and scalable compared to AI-based models. AI-based models have become increasingly popular in the field of energy forecasting due to their ability to learn patterns and relationships in complex data [
32,
33,
34]. The use of AI in energy forecasting has the potential to reduce energy costs, optimize energy production, and enhance energy security. Especially, LSTM models are able to capture both short-term and long-term dependencies in TS data. They are also capable of handling non-linear relationships between input and output variables, which is important in energy forecasting where the relationships may be complex. Finally, LSTMs could process sequential data of varying lengths, which is useful for handling variable-length TS data in energy forecasting.
3. Data Used in this Study
3.1. Data Collection
Data collected from three buildings in different regions are employed to evaluate the models, i.e Hawthorn Campus - ATC Building (denoted as ), Hawthorn Campus - AMDC Building (denoted as ), and Wantirna Campus - KIOSC Building (denoted as ) is incredibly valuable as it provides real-time insights into the building’s performance, energy consumption, and operational efficiencies, allowing for the swift response to potential issues. This real-time data not only enhance decision-making capabilities for building management, maintenance, and optimization but also provides a basis for developing more accurate forecasting models. Furthermore, it can guide strategic energy management, potentially leading to significant cost savings, improved sustainability, and increased occupant comfort over time. The two datasets and contain the energy consumption from 2017 to 2019 and the dataset contains the energy consumption from 2018 to 2019. The prediction value is the difference between the previous and the intermediate next time in using energy, or the cumulation of energy. The historical value is 96 data in every 15 minutes to predict the next value.
Figure 2.
Location of Hawthorn Campus and Wantirna Campus in Metropolitan Melbourne, Victoria, Australia.
Figure 2.
Location of Hawthorn Campus and Wantirna Campus in Metropolitan Melbourne, Victoria, Australia.
Figure 3.
Electricity accumulation in every 15 minutes at Hawthorn Campus - ATC Building.
Figure 3.
Electricity accumulation in every 15 minutes at Hawthorn Campus - ATC Building.
Figure 4.
Average electricity accumulation in every 1 hour at Hawthorn Campus - ATC Building in 2018. As can be seen, electricity consumption peaks between 8 a.m. and 21 p.m.
Figure 4.
Average electricity accumulation in every 1 hour at Hawthorn Campus - ATC Building in 2018. As can be seen, electricity consumption peaks between 8 a.m. and 21 p.m.
3.2. Data Setup
Dataset. In this experiment, three datasets from three buildings in different regions are employed to evaluate the models, i.e Hawthorn Campus - ATC Building (denoted as ), Hawthorn Campus - AMDC Building (denoted as ), and Wantirna Campus - KIOSC Building (denoted as ). The two datasets and contain the energy consumption from 2017 to 2019 and the dataset contains the energy consumption from 2018 to 2019. The prediction value is the difference between the previous and the intermediate next time in using energy, or the cumulation of energy. The historical value is 96 data in every 15 minutes to predict the next value.
Configuration of proposed model. The proposed model,
, consists of a succession of one input layer, two LSTM layers, and one dense layer at the end. The input layer contains two input types as described in
Section 4.1. In the following, the first LSTM layer wraps eight LSTM units and the second wraps four units. The last dense layer has one unit for predicting energy consumption.
Configuration of bench-marking models. As mentioned earlier, three competitive models are used for comparison: LSTM, Bi-LSTM, and LR. The LSTM model consists of one single layer with one unit, followed by a Dense layer for prediction. The Bi-LSTM consists of one single Bi-Directional LSTM layer of one unit, followed by a Dense layer for prediction as the LSTM model. The LR model trains with one dense layer.
Training Configuration. Both , LSTM, Bi-LSTM, LR models are trained using the same training set, and evaluated on the same test set. In and , the models are trained on the data in 2017 and 2018. In , the models are trained on the data in 2018 to demonstrate the ability of models with a lack of training data.
4. Methodology
In this section, we propose the forecasting model (Multivariate Multilayered LSTM), which is referred to as
. The overview of the proposed method is illustrated in
Figure 5. There are three phases, i.e. data preprocessing, model training and evaluation.
4.1. Data Preprocessing
Two major techniques are used in the data preprocessing phase, i.e. (i) data differencing and (ii) time labelling. Additionally, there are two other techniques, i.t (iii) concatenation and (iv) window slicing.
In regard to data differencing, due to a large amount of energy consumption and the value is not commonly stationary, the difference between the data value and its previous data value is taken into account. By removing these patterns through data differencing, the resulting stationary TS can be more easily modelled and forecasted. This technique is often required for many forecasting models [
35].
In time labelling, the amount of energy consumption during peak time points is typically greater than that during non-peak time. Consequently, the labelling convention assigns a value of 1 to peak time and 0 to non-peak time. [
36] These labelled time periods serve as valuable features for training the forecasting model. In the concatenation phase, the data from data differencing and labelled time are concatenated and then sliced into vectors with a length of
k during the window-slicing phase.
4.2. Forecasting Model - Multivariate Multilayered LSTM
is an extension of the LSTM model, which is a type of recurrent neural network (RNN) architecture used for sequential data processing tasks such as TS forecasting. There are h sub-layers In the hidden layer. The model consists of multiple LSTM layers, with each layer having its own set of neurons that process the input variables independently. The output of each layer is then fed into the next layer, allowing the model to capture more complex and abstract relationships between the input variables. There are some additional layers such as the dropout layer, normalization layer, etc in the hidden part for training the model efficiently.
As shown in
Figure 6, the memory cell is responsible for storing information about the long-term dependencies and patterns in the input sequence, while the gates control the flow of information into and out of the cell. The gates are composed of sigmoid neural network layers and a point-wise multiplication operation, that allows the network to learn which information to keep or discard. In particular, each
cell at the layer
m, denoted as
, has three inputs, i.e the hidden state
, cell state
of the previous cell, and the hidden state
of the cell
i at the previous layer
. The cell
has two recurrent features, i.e., hidden state
, and cell state
.
is a mathematical function as Equation
1, that takes three inputs and returns two outputs. Both outputs leave the cell at time
i and are fed into that same cell at time
, and the input sequence
is also fed into the cell. In the first layer (
), the hidden state
is the input
.
Inside the
cell, the previous hidden state
and input vector
are fed into three gates. i.e input gate (
), forget gate (
), output gate (
). They are
functions (
), each of which produces a scalar value as described in Equations
2,
3,
4 respectively.
where
and
denote weight, which is the parameter that should be updated during the training of the cell. Another scalar function, called the update function (denoted as
) has a
activation function as described in Equation
5.
where
and
are further weighted to be learned. The returned cell state (
) and hidden state (
) are formulated in Equations
6,
7 respectively.
In energy forecasting, loss optimization is an important step to improve the accuracy of the model [
37]. One common technique for loss optimization is using the Adam optimizer, which is a stochastic gradient descent optimizer that uses moving averages of the parameters to adapt the learning rate. The Adam optimizer computes individual adaptive learning rates for different parameters from estimates of the first and second moments of the gradients. This makes it suitable for optimizing the loss function in models where there are a large number of parameters. By using the Adam optimizer, the model can efficiently learn and update the weights of the neurons in each time step, resulting in better prediction accuracy.
5. Bench-Marking Models
In this study, we compare the proposed model with three well-known models, i.e., linear regression (LR), long-short-term memory (LSTM), and bidirectional long-short-term memory (Bi-LSTM).
5.1. Linear Regression
LR allows knowing the relationship between the response variable (energy consumption) and the return variables (the other variables). As a causative technique, regression analysis predicts energy demand from one or more reasons (independent variables), which might include things like the day of the week, energy prices, the availability of housing, or other variables. When there is a clear pattern in the previous forecast data, the LR method is applied. Due to this, its simple application has been used in numerous works related to the prediction of electricity consumption. Bianco V et al. (2009) used a LR model to conduct a study on the projection of Italy’s electricity consumption [
38] while Saab C et al. (2001) looked into various univariate modelling approaches to project Lebanon’s monthly electric energy usage [
39]. With the help of our statistical model, this research has produced fantastic outcomes.
The LR model works by fitting a line to a set of data points, with the goal of minimizing the sum of the squared differences between the predicted and actual values of the dependent variable. The slope of the line represents the relationship between the dependent and independent variables, while the intercept represents the value of the dependent variable when the independent variable is equal to zero. The LR model describes the linear relationship between the previous values
and the estimated future value
, formulated as follow:
5.2. LSTM
The LSTM technique is a type of Recurrent Neural Network (RNN). The RNNs [
40] are capable of processing data sequences, or data that must be read together in a precise order to have meaning, in contrast to standard neural networks. This ability is made possible by the RNNs’ architectural design, which enables them to receive input specific to each instant of time in addition to the value of the activation from the previous instant. Given their ability to preserve data from earlier actions, these earlier temporal instants provide for a certain amount of “memory”. Consequently, they possess a memory cell, which maintains the state throughout time [
41].
Figure 7.
Overview of the LSTM model.
Figure 7.
Overview of the LSTM model.
As noted in
Section 4, LSTM [
42] model have ability to remove or add information to decide what information need to go through the network from the cell state [
42]. Different from
model, LSTM model has just one LSTM layer with the input is the input sequence
. Therefore, the hidden state (
h) and cell state (
c) for the
LSTM cell are calculated as Equation
9.
In the experiment, we compare the performance of the proposed model with the univariate and multivariate LSTM model. The univariate LSTM takes the first input (
i) described in
Section 4.1 and the multivariate LSTM takes both those inputs.
5.3. Bidirectional LSTM
Bi-LSTM is also an RNN. It utilizes information in both the previous and the following directions in training phase [
43]. The fundamental principle of the Bi-LSTM model is that it examines a specific sequence from both the front and back. Which uses one LSTM layer for forward processing and the other for backward processing. The network would be able to record the evolution of energy that would power both its history and its future [
44]. This bidirectional processing is achieved by duplicating the hidden layers of the LSTM, where one set of layers processes the input sequence in the forward direction and another set of layers processes the input sequence in the reverse direction. As illustrated in
Figure 8, the hidden state (
) and the cell state (
) in the
forward LSTM cell are calculated as similar as the Equation
9. On the contrary, each LSTM cell in the backward LSTM takes the following hidden state (
), and following cell state (
), and
as input. Therefore, the hidden state (
) and cell state (
) of the
backward LSTM cell are calculated as Equation
10.
After the calculations of both forward and backward LSTM cells, the hidden states of two directions could be concatenated or combined in some way to obtain the output. The common combination is the sigmoid function as noted in
Figure 8. The output is fed into the Dense layer to obtain the final prediction. Similar to the LSTM model, we also compare the proposed model with univariate and multivariate Bi-LSTM models.
6. Experiment
6.1. Metric
To better evaluate the performance, a model is tested by making a set of predictions and then comparing it with a set of known actual values , where D is the size of the test set. Three common metrics are used to compare the overall distance of these two sets, i.e. mean average error (), root mean square error (), and mean absolute percentage error ().
MAE As shown in Equation
11,
MAE is calculated by taking the absolute difference between the predicted and actual values and averaging them. This calculation results in a single number that represents the average magnitude of the errors in the predictions made by the model.
RMSE As shown in Equation
12,
RMSE is calculated by taking the square root of the mean of the squared differences between the predicted and actual values. This calculation results in a single number that represents the typical magnitude of the errors in the predictions made by the model.
MAPE As shown in Equation
13,
MAPE is calculated by taking the absolute difference between the predicted and actual values, dividing it by the actual value, and then taking the average of these values over the entire dataset. This calculation results in a single number that represents the average percentage difference between the predicted and actual values.
6.2. Result and Discussion
This study aims to experimentally address the effectiveness of the proposed model by answering the following research inquiries: the general performance of training the proposed model and the comparative performance analysis against other competitive models.
6.2.1. General Performance
In the first part of the experiments,
is trained and evaluated with two types of data preprocessing strategies, i.e., with a labelled time field and without a labelled time field.
Figure 9 shows the results in the test set in
,
, and
. In this question, there are two results, (
i) result in the sufficient training set and (
ii) result in the lack of training set.
For the first result (
i), the model is sufficiently trained with data in 2017, and 2018 and evaluated in 2019 from
and
.
Figure 9 shows that the model achieves better performance in all three metrics with the labelled time field. The results are similar under the same setting for the other models. The details are provided in
Table 2 and the line plot in
Figure 10. Therefore, the models can learn and extract more valued features if they are trained with the appropriate data preprocessing strategy.
For the second result (
ii), the model is only trained with data in 2018 and evaluated in 2019 from
.
Figure 10 shows that the model performed well in predicting and matching the actual values, as evidenced by its superior fit line compared with the other models in
Figure 9 and
Figure 10. These findings suggest that the proposed preprocessing method is effective, particularly in situations with a limited amount of training data available for model training.
Figure 10.
Comparison of with labelled time () with without labelled time, and other models in case of lack of training data ( in 2019). The time step is 7 days.
Figure 10.
Comparison of with labelled time () with without labelled time, and other models in case of lack of training data ( in 2019). The time step is 7 days.
6.2.2. Experience Different Models
In the second part of the experiments, we compare the performance evaluation results of
to other competitive models (LSTM, Bi-LSTM, LR models) on three datasets in terms of three metrics. Two sets of performance metrics are presented, one set includes time label information (
MAE_t,
RMSE_t, and
MAPE_t), whereas the other set does not include time label information (
MAE,
RMSE, and
MAPE).
Table 2 presents the performance evaluation results of
and LSTM, Bi-LSTM, and LR models on three datasets in terms of three metrics.
Figure 11 shows the MAPE error of models with and without labelled time.
Table 2.
Comparison of with competitive models with and without labelled time. MAE_t, RMSE_t and MPAE_t are denoted metrics for model trained with labelled time. MAE, RMSE and MPAE are denoted metrics for model trained without labelled time. MPAE_t and MPAE are rescale to the range from 0 to 1. Better values are marked in bold.
Table 2.
Comparison of with competitive models with and without labelled time. MAE_t, RMSE_t and MPAE_t are denoted metrics for model trained with labelled time. MAE, RMSE and MPAE are denoted metrics for model trained without labelled time. MPAE_t and MPAE are rescale to the range from 0 to 1. Better values are marked in bold.
|
with labelled time |
without labelled time |
Dataset |
Model |
MAE_t |
RMSE_t |
MAPE_t |
MAE |
RMSE |
MAPE |
|
|
0.75 |
1.15 |
0.16 |
1.08 |
1.37 |
0.26 |
LSTM |
1.30 |
1.47 |
0.32 |
1.35 |
1.58 |
0.34 |
Bi-LSTM |
1.46 |
1.85 |
0.24 |
1.53 |
1.63 |
0.31 |
Linear Regression |
1.40 |
1.53 |
0.29 |
1.40 |
1.50 |
0.27 |
|
|
0.49 |
0.74 |
0.14 |
0.84 |
0.96 |
0.18 |
LSTM |
1.47 |
1.66 |
0.39 |
1.89 |
2.00 |
0.50 |
Bi-LSTM |
1.45 |
1.16 |
0.22 |
1.89 |
2.06 |
0.48 |
Linear Regression |
0.60 |
0.75 |
0.17 |
1.21 |
1.52 |
0.34 |
|
|
0.66 |
0.85 |
0.07 |
0.69 |
0.87 |
0.10 |
LSTM |
0.79 |
0.95 |
0.18 |
0.67 |
0.94 |
0.43 |
Bi-LSTM |
1.61 |
2.15 |
0.40 |
1.23 |
1.92 |
0.81 |
Linear Regression |
0.84 |
1.18 |
0.27 |
0.73 |
0.90 |
0.73 |
Figure 11.
Bart chart in the comparison of with labelled time () with without labelled time, and other models with and without labelled time in case of lack of training data ( in 2019).
Figure 11.
Bart chart in the comparison of with labelled time () with without labelled time, and other models with and without labelled time in case of lack of training data ( in 2019).
In general, the results indicate that models using labelled time information tend to perform better than those that do not use labelled time information in the same setting. For example, on , the four models get the lower values, while the same models have higher error values. On the other hand, among the different models, the model using the labelled time information tends to perform the best overall with the lowest MAE_t, RMSE_t, and MAPE_t in most cases. To summarize, using the proposed preprocessing method with time information tends to improve the performance of the models. The model using time label information performs the best in general. The labelled time field provides useful information for predicting energy consumption in peak and non-peak periods. These findings suggest that considering time information can help in accurately predicting the target variable in the studied datasets.
7. Conclusions
In conclusion, this work presents a method for pre-processing data and a model for accurately predicting energy consumption in commercial buildings, specifically focusing on buildings in the Hawthorn and Wantirna campuses. The proposed pre-processing method effectively improves the accuracy of energy consumption prediction, even when training data are limited. The results demonstrate the applicability of the proposed method and model for accurately predicting energy consumption in various commercial buildings.
The proposed model, denoted as , achieved the lowest MAE values of 0.75, 0.49, and 0.66 for , , and , respectively. This achievement is crucial for effective energy management and conservation in commercial buildings. The practicality of this approach extends to other commercial buildings with similar energy consumption patterns, making it a viable solution for energy management in the commercial building sector. Visualizations were also provided to aid in understanding data patterns and trends in the model predictions.
Considering the impact of the COVID-19 pandemic on energy consumption patterns, it is essential to develop forecasting models that account for these changes in demand across sectors. Additionally, further research can explore the effectiveness of the proposed pre-processing method and models in predicting energy consumption for different types of buildings or larger datasets. Exploring alternative techniques, such as seasonal decomposition or time series analysis, for incorporating time information into the models could also yield valuable insights. These advancements in energy consumption forecasting contribute to significant cost savings and environmental benefits in commercial buildings.
References
- Khalil, M.; McGough, A.S.; Pourmirza, Z.; Pazhoohesh, M.; Walker, S. Machine learning, deep learning and statistical analysis for forecasting building energy consumption — a systematic review. Engineering Applications of Artificial Intelligence. 2022, 115, 105287. [Google Scholar] [CrossRef]
- IEA. Clean energy transitions in emerging and developing economies. 2021. [Google Scholar]
- Shin, S.-Y.; Woo, H.-G. Energy consumption forecasting in korea using machine learning algorithms. Energies 2022, 15, 4880. [Google Scholar] [CrossRef]
- Özbay, H.; Dalcalı, A. Effects of covid-19 on electric energy consumption in turkey and ann-based short-term forecasting. Turkish Journal of Electrical Engineering Computer Sciences 2021, 29, 78–97. [Google Scholar] [CrossRef]
- Ji, Y.; Lomas, K.J.; Cook, M.J. Hybrid ventilation for low energy building design in south China. Building and Environment 2009, 44, 2245–2255. [Google Scholar] [CrossRef]
- Manu, S.; Shukla, Y.; Rawal, R.; Thomas, L.E.; De Dear, R. Field studies of thermal comfort across multiple climate zones for the subcontinent: India Model for Adaptive Comfort (IMAC). Building and Environment 2016, 98, 55–70. [Google Scholar] [CrossRef]
- Delzendeh, E.; Wu, S.; Lee, A.; Zhou, Y. The impact of occupants’ behaviours on building energy analysis: A research review. Renewable and Sustainable Energy Reviews 2017, 80, 1061–1071. [Google Scholar] [CrossRef]
- Itzhak, N.; Tal, S.; Cohen, H.; Daniel, O.; Kopylov, R.; Moskovitch, R. Classification of univariate time series via temporal abstraction and deep learning. In Proceedings of the 2022 IEEE International Conference on Big Data (Big Data); IEEE, 2022; pp. 1260–1265. [Google Scholar]
- Ibrahim, M.; Badran, K.M.; Hussien, A.E. Artificial intelligence-based approach for univariate time-series anomaly detection using hybrid cnn-bilstm model. In Proceedings of the 2022 13th International Conference on Electrical Engineering (ICEENG); IEEE, 2022; pp. 129–133. [Google Scholar]
- Hu, M.; Ji, Z.; Yan, K.; Guo, Y.; Feng, X.; Gong, J.; Zhao, X.; Dong, L. Detecting anomalies in time series data via a meta-feature based approach. IEEE Access 2018, 6, 27760–27776. [Google Scholar] [CrossRef]
- Niu, Z.; Yu, K.; Wu, X. Lstm-based vae-gan for time-series anomaly detection. Sensors 2020, 20, 3738. [Google Scholar] [CrossRef]
- Warrick, P.; Homsi, M.N. Cardiac arrhythmia detection from ecg combining convolutional and long short-term memory networks. In 2017 Computing in Cardiology (CinC); IEEE, 2017; pp. 1–4. [Google Scholar]
- Karim, F.; Majumdar, S.; Darabi, H.; Harford, S. Multivariate lstm-fcns for time series classification. Neural networks 2019, 116, 237–245. [Google Scholar] [CrossRef]
- Gasparin, A.; Lukovic, S.; Alippi, C. Deep learning for time series forecasting: The electric load case. CAAI Transactions on Intelligence Technology 2021, 7. [Google Scholar] [CrossRef]
- Wei, N.; Li, C.; Peng, X.; Zeng, F.; Lu, X. Conventional models and artificial intelligence-based models for energy consumption forecasting: A review. Journal of Petroleum Science and Engineering 2019, 181, 106187. [Google Scholar] [CrossRef]
- Divina, F.; Garcia Torres, M.; Gomez Vela, F.A.; Vazquez Noguera, J.L. A comparative study of time series forecasting methods for short term electric energy consumption prediction in smart buildings. Energies 2019, 12, 1934. [Google Scholar] [CrossRef]
- Johannesen, N.J.; Kolhe, M.; Goodwin, M. Relative evaluation of regression tools for urban area electrical energy demand forecasting. Journal of cleaner production 2019, 218, 555–564. [Google Scholar] [CrossRef]
- Singhal, R.; Choudhary, N.; Singh, N. Short-Term Load Forecasting Using Hybrid ARIMA and Artificial Neural Network Model. 2020, 01, 935–947. [Google Scholar]
- Li, K.; Zhang, T. Forecasting electricity consumption using an improved grey prediction model. Information 2018, 9, 204. [Google Scholar] [CrossRef]
- Kim, Y.; Son, H.G.; Kim, S. Short term electricity load forecasting for institutional buildings. Energy Reports 2019, 5, 1270–1280. [Google Scholar] [CrossRef]
- Chitalia, G.; Pipattanasomporn, M.; Garg, V.; Rahman, S. Robust short-term electrical load forecasting framework for commercial buildings using deep recurrent neural networks. Applied Energy 2020, 278, 115410. [Google Scholar] [CrossRef]
- Pinto, T.; Praça, I.; Vale, Z.; Silva, J. Ensemble learning for electricity consumption forecasting in office buildings. Neurocomputing 2021, 423, 747–755. [Google Scholar] [CrossRef]
- Pallonetto, F.; Jin, C.; Mangina, E. Forecast electricity demand in commercial building with machine learning models to enable demand response programs. Energy and AI 2022, 7, 100121. [Google Scholar] [CrossRef]
- Skomski, E.; Lee, J.Y.; Kim, W.; Chandan, V.; Katipamula, S.; Hutchinson, B. Sequence-to-sequence neural networks for short-term electrical load forecasting in commercial office buildings. Energy and Buildings 2020, 226, 110350. [Google Scholar] [CrossRef]
- Dagdougui, H.; Bagheri, F.; Le, H.; Dessaint, L. Neural network model for short-term and very-short-term load forecasting in district buildings. Energy and Buildings 2019, 203, 109408. [Google Scholar] [CrossRef]
- Khan, Z.A.; Hussain, T.; Ullah, A.; Rho, S.; Lee, M.; Baik, S.W. Towards efficient electricity forecasting in residential and commercial buildings: A novel hybrid CNN with a LSTM-AE based framework. Sensors 2020, 20, 1399. [Google Scholar] [CrossRef]
- Karijadi, I.; Chou, S.Y. A hybrid RF-LSTM based on CEEMDAN for improving the accuracy of building energy consumption prediction. Energy and Buildings 2022, 259, 111908. [Google Scholar] [CrossRef]
- Hwang, J.; Suh, D.; Otto, M.O. Forecasting electricity consumption in commercial buildings using a machine learning approach. Energies 2020, 13, 5885. [Google Scholar] [CrossRef]
- Fernández-Martínez, D.; Jaramillo-Morán, M.A. Multi-Step Hourly Power Consumption Forecasting in a Healthcare Building with Recurrent Neural Networks and Empirical Mode Decomposition. Sensors 2022, 22, 3664. [Google Scholar] [CrossRef] [PubMed]
- Jozi, A.; Pinto, T.; Marreiros, G.; Vale, Z. Electricity consumption forecasting in office buildings: an artificial intelligence approach. In 2019 IEEE Milan PowerTech; 2019; pp. 1–6. [Google Scholar]
- Mariano-Hernández, D.; Hernández-Callejo, L.; Solís, M.; Zorita-Lamadrid, A.; Duque-Pérez, O.; Gonzalez-Morales, L.; Alonso-Gómez, V.; Jaramillo-Duque, A.; Santos García, F. Comparative study of continuous hourly energy consumption forecasting strategies with small data sets to support demand management decisions in buildings. Energy Science & Engineering 2022, 10, 4694–4707. [Google Scholar]
- del Real, A.J.; Dorado, F.; Duran, J. Energy demand forecasting using deep learning: Applications for the french grid. Energies 2020, 13, 2242. [Google Scholar] [CrossRef]
- Fathi, S.; Srinivasan, R.S.; Kibert, C.J.; Steiner, R.L.; Demirezen, E. AI-based campus energy use prediction for assessing the effects of climate change. Sustainability 2020, 12, 3223. [Google Scholar] [CrossRef]
- Khan, S.U.; Khan, N.; Ullah, F.U.M.; Kim, M.J.; Lee, M.Y.; Baik, S.W. Towards intelligent building energy management: AI-based framework for power consumption and generation forecasting. Energy and Buildings 2023, 279, 112705. [Google Scholar] [CrossRef]
- Athiyarath, S.; Paul, M.; Krishnaswamy, S. A comparative study and analysis of time series forecasting techniques. SN Computer Science 2020, 1. [Google Scholar] [CrossRef]
- Noor, R.M.; Yik, N.S.; Kolandaisamy, R.; Ahmedy, I.; Hossain, M.A.; Yau, K.L.A.; Shah, W.M.; Nandy, T. Predict Arrival Time by Using Machine Learning Algorithm to Promote Utilization of Urban Smart Bus. 2020. [Google Scholar]
- Ciampiconi, L.; Elwood, A.; Leonardi, M.; Mohamed, A.; Rozza, A. A survey and taxonomy of loss functions in machine learning. 2023. [Google Scholar]
- Bianco, V.; Manca, O.; Nardini, S. Electricity consumption forecasting in italy using linear regression models. Energy 2009, 34, 1413–1421. [Google Scholar] [CrossRef]
- Saab, S.; Badr, E.; Nasr, G. Univariate modeling and forecasting of energy consumption: the case of electricity in lebanon. Energy 2001, 26, 1–14. [Google Scholar] [CrossRef]
- Yuan, X.; Li, L.; Wang, Y. Nonlinear dynamic soft sensor modeling with supervised long short-term memory network. IEEE transactions on industrial informatics 2019, 16, 3168–3176. [Google Scholar] [CrossRef]
- Durand, D.; Aguilar, J.; R-Moreno, M.D. An analysis of the energy consumption forecasting problem in smart buildings using lstm. Sustainability 2022, 14, 13358. [Google Scholar] [CrossRef]
- Hochreiter, S.; Schmidhuber, J. Long short-term memory. Neural computation 1997, 9, 1735–1780. [Google Scholar] [CrossRef]
- Graves, A.; Schmidhuber, J. Framewise phoneme classification with bidirectional lstm and other neural network architectures. Neural networks 2005, 18, 602–610. [Google Scholar] [CrossRef] [PubMed]
- Le, T.; Vo, M.T.; Vo, B.; Hwang, E.; Rho, S.; Baik, S.W. Improving electric energy consumption prediction using cnn and bi-lstm. Applied Sciences 2019, 9, 4237. [Google Scholar] [CrossRef]
|
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2023 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).