Next Article in Journal
Analysis of a Clapping Vibration Energy Harvesting System in a Rotating Magnetic Field
Previous Article in Journal
Observability Decomposition-Based Decentralized Kalman Filter and Its Application to Resilient State Estimation under Sensor Attacks
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Improving the Efficiency of Multistep Short-Term Electricity Load Forecasting via R-CNN with ML-LSTM

by
Mohammed F. Alsharekh
1,2,*,
Shabana Habib
2,3,
Deshinta Arrova Dewi
4,
Waleed Albattah
2,3,
Muhammad Islam
2,5 and
Saleh Albahli
2,3
1
Department of Electrical Engineering, Unaizah College of Engineering, Qassim University, Unaizah 56452, Saudi Arabia
2
System Control Processing Research Group, Qassim University, Buraydah 51452, Saudi Arabia
3
Department of Information Technology, College of Computer, Qassim University, Buraydah 51452, Saudi Arabia
4
Faculty of Data Science and Information Technology, INTI International University, Nilai 71800, Negeri Sembilan, Malaysia
5
Department of Electrical Engineering, College of Engineering and Information Technology, Onaizah Colleges, Onaizah 56447, Saudi Arabia
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(18), 6913; https://doi.org/10.3390/s22186913
Submission received: 15 August 2022 / Revised: 27 August 2022 / Accepted: 29 August 2022 / Published: 13 September 2022
(This article belongs to the Section Industrial Sensors)

Abstract

:
Multistep power consumption forecasting is smart grid electricity management’s most decisive problem. Moreover, it is vital to develop operational strategies for electricity management systems in smart cities for commercial and residential users. However, an efficient electricity load forecasting model is required for accurate electric power management in an intelligent grid, leading to customer financial benefits. In this article, we develop an innovative framework for short-term electricity load forecasting, which includes two significant phases: data cleaning and a Residual Convolutional Neural Network (R-CNN) with multilayered Long Short-Term Memory (ML-LSTM) architecture. Data preprocessing strategies are applied in the first phase over raw data. A deep R-CNN architecture is developed in the second phase to extract essential features from the refined electricity consumption data. The output of R-CNN layers is fed into the ML-LSTM network to learn the sequence information, and finally, fully connected layers are used for the forecasting. The proposed model is evaluated over residential IHEPC and commercial PJM datasets and extensively decreases the error rates compared to baseline models.

1. Introduction

Electricity load forecasting predicts future load based on single or multiple features or parameters. Features could be of multiple types, such as an ongoing month, hour, weather situation, electricity costs, economic conditions, geographical circumstances, etc. [1]. Electricity load forecasting is significantly increasing due to the development and extension of the energy market, as they endorsed trading electricity for each hour. Profitable market interactions could be enabled by accurate load forecasting [2] and leakage current prediction [3], which helps power firms guarantee electricity stability and decrease electricity wastage [4]. The electricity load prediction is handled through short-term electricity load forecasting [5], which is particularly significant due to smart grid development [6].
The United States Energy Information Administration stated that from 2015–40 the increase in power consumption would be boosted up to 28% [7], while the International Energy Agency stated that buildings and building construction account for approximately 36% of the world’s total energy consumption. Stimulating building energy efficiency is vital in the low carbon economy [8,9]. Accurate energy consumption forecasting is indispensable for buildings’ energy-saving design and renovation. The scrutiny of dissimilarity among the energy consumption forecasting and the experimented data can also provide a foundation for building operation monitoring. It can also provide a source for energy peak regulation for large buildings [10,11]. In addition, some large-scale buildings are essential to secure public resource supply in an area. Thus, accurate energy consumption forecasting can sustain the resident energy allocation sectors.
The following few important points could establish accurate forecasting models of consumption. First, it is crucial to precisely recognize the parameters that contain robust effects on a state’s consumption and add these indicators to the prediction model. Furthermore, choosing an appropriate modeling procedure is also a significant point. The input and output variables contain such a nonlinear relationship that it is very challenging to express them mathematically. Principles to advance the prediction performance develop conspicuously nowadays, instead of hypothetical principles in the model selection. In the end, the vital point is that the methodology must be capable of producing effective forecasting results.
Reviewing the literature in recent years, machine learning models are increasingly used by scholars to forecast the short-term energy consumption of buildings. Additionally, most scholars are adopting hybrid models due to their effective performance compared to single models. However, energy consumption in buildings is exaggerated by various factors and is extremely nonlinear. Thus, this area is still worthy and commendable to be studied and explored further. This research work contributes the following points to the current literature.
  • The collected benchmark datasets contain a lot of missing values and outliers, which occur due to defaulted meters, weather conditions, and abnormal customer consumption. These abnormalities and redundancies in datasets lead the forecasting network to ambiguous predictions. To resolve this problem, we performed data preprocessing strategies, including outlier removal via the three sigma rules of thumb algorithm, missing value via NAN interpolation method, and the normalization of the data using the MinMax scaler.
  • We present a deep R-CNN integrated with ML-LSTM for power forecasting using real power consumption data. The motivation behind R-CNN with ML-LSTM is to extract patterns and time-varied information from the input data for effective forecasting.
  • The proposed model results in the lowest error rates of MAE, MSE, RMSE, and MAPE and the highest R2 compared to recent literature. For the IHEPC dataset, the proposed model achieved 0.0447, 0.0132, 0.002, 0.9759, and 1.024 for RMSE, MAE, MSE, R2, and MAPE, respectively, over the hourly IHEPC dataset while these values are 0.0447, 0.0132, 0.002, 0.9759, and 1.024 over the IHEPC daily dataset. For the PJM dataset, the proposed model achieved 0.0223, 0.0163, 0.0005, 0.9907, 0.5504 for RMSE, MAE, MSE, R2, and MAPE, respectively. The lowest error metrics indicated the supremacy of the proposed model over state-of-the-art methods.

2. Literature Review

Short-term load forecasting is a current research area, and numerous studies have been conducted in the literature. These studies are mainly divided into four categories based on the learning algorithm: physical, persistence, artificial intelligence (AI), and statistical. The persistence model can predict future time series data behavior like electricity consumption or forecasting but failed for several hour ahead predictions [12]. Therefore, persistence models are not decisive for electricity forecasting. Physical models are based on mathematical expressions that consider meteorological and historical data. N. Mohan et al. [13] present a dynamic empirical model for short-term electricity forecasting based on a physical model. These models are also unreliable for electricity forecasting due to the high memory, and computational space required [12]. As compared to physical models, statistical models are less computationally expensive [14] and are typically based on autoregressive methods, i.e., GARCH [15], ARIMA [16], and linear regression methods. These models are based on linear data, while electricity consumption prediction or load forecasting is a nonlinear and complex problem. As presented in [12], the GARCH model can capture uncertainty but has limited capability to capture non-stationary and nonlinear characteristics of electricity consumption data. Several studies based on linear regression have also been developed in previous studies. For instance, N. Fumo et al. proposed a linear and multiple regression model for electricity load forecasting [17]. Similarly, [18,19] developed multi-regression-based electricity load forecasting models. Statistical-based models cannot capture uncertainty patterns directly but use other techniques, as presented in [20], to reduce uncertainty. AI methods can learn nonlinear complex data, divided into shallow and deep structure methods (DSM). The shallow-based methods like SVM [21], ANN [22], wavelet neural networks [23], random forest [24], and ELM [25] performed poorly, except for feature mining. Thus, improving the performance of these models additionally needs feature extraction and selection methods, remaining a challenging problem [20,26].
The AI methods are further categorized into machine learning and deep learning. In machine learning, several studies have been conducted in the literature for electricity load forecasting. Chen et al. used SVR for power consumption forecasting using electricity consumption and temperature data [27]. Similarly, in [28], an SVR-based hybrid electricity forecasting model was developed. In this study, the SVR model was integrated with adjusting the seasonal index and optimization algorithm fruity fly for better performance. Zhong et al. [29] proposed a vector field base SVR model for energy prediction and transformed the nonlinearity of the feature space into linearity to address the concern of the nonlinearity of the model to the input data. C. Li et al. conducted a study based on Random Forest Regression (RFR). First, the features are extracted in the frequency domain using fast Fourier transformation and then used RFR model for simulation and prediction [30]. Similarly, several deep learning models are developed for short-term electricity load forecasting. W. Kong et al. used LSTM recurrent network for electricity forecasting [31]. Another study [32] proposed bidirectional LSTM with an attention mechanism and rolling update technique for electricity forecasting. A recurrent neural network (RNN), based on a pooling mechanism, is developed in [20] to address the concern of overfitting.
Features extraction methods include different regular patterns, spectral analysis, and noise. However, these methods reduce the accuracy for meter level load [20] and these methods reduce the accuracy due to the low proportion of regular patterns in electricity data. DSM, also called deep neural networks (DNNs), addressed the challenges in shallow-based methods. DNN-based models are based on multi-layer processing and learn hierarchical features from input data. For sequence and pattern learning, LSTM [20,26] and CNN [33] are the most powerful architecture recently proposed. The long-tailed dependencies in raw time series data LSTM cannot capture it [34].
Similarly, CNN networks cannot learn temporal features of power consumption. Therefore, hybrid models are developed to forecast power consumption effectively. T. Y. Kim et al. combine CNN with LSTM for power consumption prediction [35], while another study presented in [36] used CNN with bidirectional LSTM. ZA Khan et al. developed a CNN-LSTM-autoencoder-based model to predict short-term load forecasting [37]. Furthermore, CNN with a GRU-based model is developed in [38]. These models address the problems in DNN networks, but the prediction accuracy is unreliable for real-time implementation. Therefore, we developed a two-stage framework for short-term electricity load forecasting in this work. In the first stage, the raw electricity consumption data collected from a residential house is preprocessed to remove missing values, outliers, etc. This refined data is then fed to our newly proposed R-CNN with ML-LSTM architecture to address the concerns of DNNs and improve the forecasting performance for effective power management.

3. Proposed Method

The overall architecture of R-CNN with ML-LSTM is shown in Figure 1 for short-term electricity load forecasting. A two-stage framework is presented, which includes data preprocessing and proposed R-CNN with ML-LSTM architecture. Data preprocessing includes filling missing values removing outliers, and normalizing the data for efficient training. The second step comprises R-CNN with ML-LSTM architecture, where R-CNN is employed for pattern learning. At the same time, the ML-LSTM layers are incorporated to learn the sequential information of electricity consumption data. Each step of the proposed framework is further explained in the following sub-sections.

3.1. Data Preprocessing

Smart meter sensors-based data generation contains outliers and missing values for several reasons, such as meter faults, weather conditions, unmanageable supply, storage issues, etc. [39], and must be preprocessed before training. Herein, we apply a unique preprocessing step. For evaluating the proposed method, we used IHEC Dataset, which includes the above-mentioned erroneous values. In addition, the performance is evaluated over the PJM benchmark dataset. To remove the outlier values in the dataset, we used three sigma rules [40] of thumb according to Equation (1).
Ƒ D i = A V G d + 2 . S T D d , i f   d i   > a v g d + s t d d d i , o t h e r w i s e
where d i is a vector or superset of A representing a value for power consumption in duration, i.e., minute, hour, day, etc. At the same time, A V G d is the average of D , and S T D D represents the standard deviation of the D . A recovering interpolation method is used as presented in Equation (2) for missing values.
Ƒ D i = d i 1 + d i + 1 2 , d i n a n , d i 1 , d i + 1 n a n 0 , d i n a n , d i 1   o r   d i + 1 n a n d i , d i n a n
If d i is missing or null, we placed it as a N A N . The IHEPC dataset was recorded in a one minute resolution, while the PJM dataset was recorded in a one hour resolution. The IHEPC dataset is down sampled for daily load forecasting into hourly resolutions. The input datasets are resampled into low resolution (from minutes to hours) in downsampling. The IHEPC dataset includes 2,075,259 recodes, down-sampled into 34,588 records for daily load forecasting, as shown in Figure 2.
After data cleaning, we apply the data transformation technique to transform the cleaned data into a specific format more suitable for effective training. First, we use the power transformation technique to remove shifts and transform the data into more Gaussian-like. The power transformation includes Box–Cox [41] and Yeo–Johnson [42]. The Box–Cox is sensitive to negative values, while Yeo–Johnson can support both negative and positive values. In this work, we used the Box–Cox technique for power transformation to remove a shift from the electricity data distribution. This work uses univariate electricity load forecasting datasets, so the Box–Cox transformation for a single parameter is shown in Equation (3).
d i λ = d i λ 1 λ , i f   D 0 I n d , i f   D = 0
Finally, the min–max data normalization technique converts the data into a specific range because deep learning networks are sensitive to diverse data. The equation of min–max normalization is shown in Equation (4).
D i = d d m i n d m a x d m i n
where d is the actual data, while d m i n is the minimum and d m a x is the maximum values in the dataset.

3.2. R-CNN with ML-LSTM

The proposed architecture integrates R-CNN with ML-LSTM for power load forecasting. R-CNN and ML-LSTM can store the complex fluctuating trends and extract complicated features for electricity load forecasting. First, the R-CNN layers extract patterns, which are then passed to ML-LSTM as input for learning. CNN is a well know deep learning architecture consisting of four types of layers: convolutional, pooling, fully connected, and regression [43,44]. The convolutional layers include multiple convolution filters, which perform a convolutional operation between convolutional neuron weights and the input volume-connected region, which generates a feature map [45,46]. The basic equation of the convolutional layer operation is shown in Equation (5).
𝒞 m , n , o ȴ = Ƒ ( ( W o ȴ ) t   X m , n ȴ + Ƅ k ȴ
where Ƅ k ȴ is the bias of oth convolution filter in the lth layer and X m , n ȴ demonstrates the location and activation function Ƒ ( ) . In the convolution operation, the weights W o ȴ must be shared with the overall input region, known as weight sharing. During model building, the weight sharing significantly decreases the cost of calculation time and training parameter numbers. After convolution, the pooling operation is performed. The pooling layer reduces feature map resolution for input feature aggregation [47,48]. The output of the pooling layer is shown in Equation (6).
Ƥ m , n , o ȴ = Pool p i , j , o ȴ
where i , j ε P m , n , ε P m , n is the region of location of i , j . CNN has three types of pooling layers: max, min, and average poling. The CNN’s general network comprises several convolutional and pooling layers. Before the regression, the fully connected layers are typically set where every neuron in the previous layer is connected to every other in the next layer. The main purpose of the fully connected layer is to represent the learned feature distribution to a single space for high-level reasoning. The regression layer is the final output of the CNN model.
Due to the strong feature extraction ability, CNN architectures are extensively applied for image classification, video classification, time series, etc. Similarly, in time-series forecasting, these models are used for traffic [49,50], renewables [51], election prediction [52], and power forecasting [53]. Recent studies of image classification show the crucial performance of CNNs. As the network depth increases to a certain level, the degradation problem occurs in which the model performance is saturated. The experimentation shows that saturation is an optimization problem that is not caused by overfitting. To address the degradation concern, R-CNN architecture has been developed [54]. The conventional CNN learns the data in a linear mechanism, i.e., a direct function Ƒ a , but the R-CNN learns it differently, defined as H a = Ƒ a + a .
The ResNet solves the degradation problem and performs satisfactory results over image recognition data, but electricity consumption is time series sequential data. The CNN architecture cannot learn the sequential features of power consumption data. Therefore, the R-CNN with ML-LSTM architecture is developed in this research study for future electricity load forecasting. The R-CNN layers extract spatial information from electricity consumption data. The extracted features of R-CNN are then fed to ML-LSTM as input for temporal learning.
The output of R-CNN is then forwarded to ML-LSTM architecture, that is responsible for storing time information. The ML-LSTM maintains long-term memory by merging its units to update the earlier hidden state, aiming to understand temporal relationships in the sequence. The three gates unit’s mechanism is incorporated to determine each memory unit state through multiplication operations. The input gate, output gate, and forget gate represent each gate unit in the LSTM. The memory cells are updated with an activation. The operation of each gate in the LSTM can be shown in Equations (7)–(9), and the output of each gate is represented by i , f , and o notation, while is the activation function, w represents the weight, and b is the bias.
i t = ( W i h 𝓽 1 ,   𝔁 𝓽 + b i
f t = ( W f h 𝓽 1 ,   𝔁 𝓽 + b f
o t = ( W o h 𝓽 1 ,   𝔁 𝓽 + b o

3.3. Architecture Design

The proposed R-CNN with ML-LSTM is based on three types of layers: R-CNN, ML-LSTM, and fully connected layers. The kernel size, filter numbers, and strides are adjustable in R-CNN layers according to the model’s performance. Many learning speeds, changes, and performance can happen by adjusting these parameters varying on the input data [55]. We can confirm the performance change by increasing or decreasing these parameters. We used a different kernel size in each layer to minimize the loss of temporal information. The data pass through the residual R-CNN layer, followed by the pooling layer for pattern learning. The output is then fed to the ML-LSTM for sequence learning, which is then forwarded to fully connected (FC) layers for final forecasting. Table 1 shows the layer type, the kernel’s size, and R-CNN parameters with the ML-LSTM network.

4. Results

The experimental setup, evaluation metrics, dataset, performance assessment over hourly data, and performance assessment over daily data of R-CNN with the ML-LSTM model are briefly discussed in the following section.

4.1. Experimental Setup

To validate the effectiveness of the proposed approach, the IHEPC dataset is used to implement comprehensive experiments. The R-CNN with ML-LSTM is trained over an Intel-Core-i7 CPU having 32GB RAM and GEFORCE-GTX-2060-GPU in Windows 10. The implementation was performed in Python 3.5 using the Keras framework.

4.2. Evaluation Metrics

The model performance is evaluated on mean square error (MSE), mean absolute error (MAE), root mean square error (RMSE), coefficient of determination ( R 2 ) , and mean absolute percentage error (MAPE) metrics. MAE computes the closeness between actual and forecasted values, MSE calculates square error, RMSE is the square root of MSE, R 2 exhibits model fitting effect ranging from 0 to 1 where closer to 1 indicates best prediction performance, and MAPE is the absolute ratio error for all samples. The mathematical equations of each metric are demonstrated in Equations (10)–(12) Where y i is the actual power consumption value and y ^ i is the forecasted value.
M A E = i = 1 n y i y ^ i n
M S E = i = 1 n y i y ^ i 2 n
R M S E = i = 1 n y i y ^ i 2 n
R 2 = 1 i = 1 n y i y ^ i 2 i = 1 n y i y i 2
M A P E = 1 n i = 1 n y i y ^ i y i

4.3. Datasets

The R-CNN with ML-LSTM model is evaluated over the UCI repository’s IHEPC and PJM datasets. The IHEPC comprises nine attributes: date and time variables, active and reactive power, voltage, intensity, and three submetering variables. Description of IHEPC attributes and their unit are shown in Table 2. The IHEPC dataset was collected from a residential building in France between the period of 2006 to 2010. PJM (Pennsylvania-New Jersey-Maryland) is a regional transmission society that operates the eastern electricity grid of the US. The PJM transmits the electricity to several US regions, including Maryland, Michigan, Delaware, etc. The power consumption data are stored on PJM’s official website, which is recorded in one-hour resolution in megawatts.

4.4. Comparative Analysis

The performance of R-CNN with the ML-LSTM model is compared to other models over residential electricity consumption IHEPC and regional electricity forecasting PJM datasets. The performance of the proposed model over these datasets and its comparison with other models are clarified in the subsequent sections.

5. Evaluation of IHEPC Dataset

The supremacy of the proposed model over the IHEPC dataset is evaluated on hourly and daily load forecasting, whereas the R-CNN with ML-LSTM achieved satisfactory results. The performance of the test dataset for daily and hourly load forecasting is shown in Figure 3a,b. The comparison with other baseline methods for hourly load forecasting is presented in Table 3. For hourly load forecasting, the results are compared with linear regression [35], ANN [56], CNN [37], CNN-LSTM [35], CNN-BDLSTM [36], CNNLSTM-autoencoder [37], SE-AE [57], GRU [58], FCRBM [59], CNNESN [60], residual CNN stacked LSTM [61], CNN-BiGRU [62], CNN-GRU [63], STLF-Net [64], residual GRU [65] and Khan et al. [66]. Compared to all these studies, R-CNN with ML-LSTM achieved lower error rates of 0.0325, 0.0144, 0.0011, 0.9841, and 1.024 for RMSE, MAE, MSE, R2, and MAPE, respectively.
The performance of the proposed model over the hourly resolution of data also secured the lowest error rate compared to the baseline model. For a precedent, the performance over the daily resolution of data is compared with regression [35], CNN [37], LSTM [35], CNN-LSTM [35], and FCRBM [59], where the details results of each study are given in Table 4. Comparatively, the R-CNN with ML-LSTM model also reduces the error rates over the daily dataset and achieved 0.0447, 0.0132, 0.002, 0.9759, and 2.457 for RMSE, MAE, MSE, R2, and MAPE, respectively, for daily load forecasting.

6. Evaluation of the PJM Dataset

The superiority of R-CNN with ML-LSTM is also evaluated over several PJM datasets for daily load forecasting. The PJM benchmark includes 14 datasets of electricity load forecasting, while for the experimentation, we chose 10 datasets from the PJM as selected by [61]. In the literature, we found the comparison for 10 regions dataset, which is demonstrated in Table 5, where the proposed model acquired the lowest error rate for each dataset. The performance of R-CNN with ML-LSTM is compared with Mujeeb et al. [67], Gao et al. [68], Chou et al. [69], Khan et al. [61], and Han et al. [58]. The R-CNN with ML-LSTM secures the lower error metrics for all datasets, where the details are given in Table 5, while the prediction results for all datasets in the PJM region are shown in Figure 4.

7. Conclusions

A two-phase framework is proposed in this work for power load forecasting. Data cleaning is the first phase of our framework where data preprocessing strategies are applied over raw data to make it clean for effective training. Secondly, a deep R-CNN with ML-LSTM architecture is developed where the R-CNN learns patterns from electricity data, and the outputs are fed to ML-LSTM layers. Electricity consumption comprises time series data that include spatial and temporal features. The R-CNN layers extract spatial features in this work, while the ML-LSTM architecture is incorporated for sequence learning. The proposed model was tested over residential and commercial benchmark datasets and conducted with satisfactory results. For residential power consumption forecasting, IHEPC data was used, while the PJM dataset was used for commercial evaluation. The experiments are performed for daily and hourly power consumption forecasting and extensively decrease the error rates. In the future, the proposed model will be tested over medium-term and long-term electricity load forecasting. In addition, we will integrate environmental sensor data that help to predict future electricity consumption. Furthermore, we also intend to investigate the performance of the R-CNN and ML-LSTM in other prediction domains such as fault prediction, renewable power generation prediction, and traffic flow prediction.

Author Contributions

Formal analysis, S.H.; Funding acquisition, M.I.; Investigation, W.A.; Methodology, S.H.; Project administration, M.F.A.; Resources, S.A.; Software, M.I.; Supervision, S.A.; Validation, W.A.; Visualization, D.A.D.; Writing—original draft, M.F.A. and S.H.; Writing—review & editing, D.A.D., W.A., M.F.A., S.A. and M.I. All authors have read and agreed to the published version of the manuscript.

Funding

This research work was funded by the Deputyship for Research & Innovation, Ministry of Education, Saudi Arabia through the project number (QU-IF-04-02-28647).

Acknowledgments

The authors extend their appreciation to the Deputyship for Research & Innovation, Ministry of Education, Saudi Arabia for funding this research work through the project number (QU-IF-04-02-28647). The authors also thank to Qassim University for technical support.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Yar, H.; Imran, A.S.; Khan, Z.A.; Sajjad, M.; Kastrati, Z. Towards smart home automation using IoT-enabled edge-computing paradigm. Sensors 2021, 21, 4932. [Google Scholar] [CrossRef] [PubMed]
  2. Vrablecová, P.; Ezzeddine, A.B.; Rozinajová, V.; Šárik, S.; Sangaiah, A.K. Smart grid load forecasting using online support vector regression. Comput. Electr. Eng. 2018, 65, 102–117. [Google Scholar] [CrossRef]
  3. Sopelsa Neto, N.F.; Stefenon, S.F.; Meyer, L.H.; Ovejero, R.G.; Leithardt, V.R.Q. Fault Prediction Based on Leakage Current in Contaminated Insulators Using Enhanced Time Series Forecasting Models. Sensors 2022, 22, 6121. [Google Scholar] [CrossRef] [PubMed]
  4. He, W. Load forecasting via deep neural networks. Procedia Comput. Sci. 2017, 122, 308–314. [Google Scholar] [CrossRef]
  5. Huang, N.; Lu, G.; Xu, D. A permutation importance-based feature selection method for short-term electricity load forecasting using random forest. Energies 2016, 9, 767. [Google Scholar] [CrossRef]
  6. Dang-Ha, T.-H.; Bianchi, F.M.; Olsson, R. Local short term electricity load forecasting: Automatic approaches. In Proceedings of the 2017 International Joint Conference on Neural Networks (IJCNN), Anchorage, AK, USA, 14–19 May 2017; pp. 4267–4274. [Google Scholar]
  7. Sieminski, A. International energy outlook. Energy Inf. Adm. (EIA) 2014, 18, 2. [Google Scholar]
  8. Lam, J.C.; Wan, K.K.; Tsang, C.; Yang, L. Building energy efficiency in different climates. Energy Convers. Manag. 2008, 49, 2354–2366. [Google Scholar] [CrossRef]
  9. Allouhi, A.; El Fouih, Y.; Kousksou, T.; Jamil, A.; Zeraouli, Y.; Mourad, Y. Energy consumption and efficiency in buildings: Current status and future trends. J. Clean. Prod. 2015, 109, 118–130. [Google Scholar] [CrossRef]
  10. Zhao, L.; Zhang, J.-l.; Liang, R.-b. Development of an energy monitoring system for large public buildings. Energy Build. 2013, 66, 41–48. [Google Scholar] [CrossRef]
  11. Ma, X.; Cui, R.; Sun, Y.; Peng, C.; Wu, Z. Supervisory and Energy Management System of large public buildings. In Proceedings of the 2010 IEEE International Conference on Mechatronics and Automation, Xi’an, China, 4–7 August 2010; pp. 928–933. [Google Scholar]
  12. Khodayar, M.; Kaynak, O.; Khodayar, M.E. Rough deep neural architecture for short-term wind speed forecasting. IEEE Trans. Ind. Inform. 2017, 13, 2770–2779. [Google Scholar] [CrossRef]
  13. Mohan, N.; Soman, K.; Kumar, S.S. A data-driven strategy for short-term electric load forecasting using dynamic mode decomposition model. Appl. Energy 2018, 232, 229–244. [Google Scholar] [CrossRef]
  14. Shi, Z.; Liang, H.; Dinavahi, V. Direct interval forecast of uncertain wind power based on recurrent neural networks. IEEE Trans. Sustain. Energy 2017, 9, 1177–1187. [Google Scholar] [CrossRef]
  15. Bikcora, C.; Verheijen, L.; Weiland, S. Density forecasting of daily electricity demand with ARMA-GARCH, CAViaR, and CARE econometric models. Sustain. Energy Grids Netw. 2018, 13, 148–156. [Google Scholar] [CrossRef]
  16. Boroojeni, K.G.; Amini, M.H.; Bahrami, S.; Iyengar, S.; Sarwat, A.I.; Karabasoglu, O. A novel multi-time-scale modeling for electric power demand forecasting: From short-term to medium-term horizon. Electr. Power Syst. Res. 2017, 142, 58–73. [Google Scholar] [CrossRef]
  17. Fumo, N.; Biswas, M.R. Regression analysis for prediction of residential energy consumption. Renew. Sustain. Energy Rev. 2015, 47, 332–343. [Google Scholar] [CrossRef]
  18. Vu, D.H.; Muttaqi, K.M.; Agalgaonkar, A. A variance inflation factor and backward elimination based robust regression model for forecasting monthly electricity demand using climatic variables. Appl. Energy 2015, 140, 385–394. [Google Scholar] [CrossRef]
  19. Braun, M.; Altan, H.; Beck, S. Using regression analysis to predict the future energy consumption of a supermarket in the UK. Appl. Energy 2014, 130, 305–313. [Google Scholar] [CrossRef]
  20. Shi, H.; Xu, M.; Li, R. Deep learning for household load forecasting—A novel pooling deep RNN. IEEE Trans. Smart Grid 2017, 9, 5271–5280. [Google Scholar] [CrossRef]
  21. Wang, Y.; Xia, Q.; Kang, C. Secondary forecasting based on deviation analysis for short-term load forecasting. IEEE Trans. Power Syst. 2010, 26, 500–507. [Google Scholar] [CrossRef]
  22. Tsekouras, G.J.; Hatziargyriou, N.D.; Dialynas, E.N. An optimized adaptive neural network for annual midterm energy forecasting. IEEE Trans. Power Syst. 2006, 21, 385–391. [Google Scholar] [CrossRef]
  23. Chen, Y.; Luh, P.B.; Guan, C.; Zhao, Y.; Michel, L.D.; Coolbeth, M.A.; Friedland, P.B.; Rourke, S.J. Short-term load forecasting: Similar day-based wavelet neural networks. IEEE Trans. Power Syst. 2009, 25, 322–330. [Google Scholar] [CrossRef]
  24. Lahouar, A.; Slama, J.B.H. Day-ahead load forecast using random forest and expert input selection. Energy Convers. Manag. 2015, 103, 1040–1051. [Google Scholar] [CrossRef]
  25. Li, S.; Wang, P.; Goel, L. A novel wavelet-based ensemble method for short-term load forecasting with hybrid neural networks and feature selection. IEEE Trans. Power Syst. 2015, 31, 1788–1798. [Google Scholar] [CrossRef]
  26. Kong, W.; Dong, Z.Y.; Hill, D.J.; Luo, F.; Xu, Y. Short-term residential load forecasting based on resident behaviour learning. IEEE Trans. Power Syst. 2017, 33, 1087–1088. [Google Scholar] [CrossRef]
  27. Chen, Y.; Xu, P.; Chu, Y.; Li, W.; Wu, Y.; Ni, L.; Bao, Y.; Wang, K. Short-term electrical load forecasting using the Support Vector Regression (SVR) model to calculate the demand response baseline for office buildings. Appl. Energy 2017, 195, 659–670. [Google Scholar] [CrossRef]
  28. Cao, G.; Wu, L. Support vector regression with fruit fly optimization algorithm for seasonal electricity consumption forecasting. Energy 2016, 115, 734–745. [Google Scholar] [CrossRef]
  29. Zhong, H.; Wang, J.; Jia, H.; Mu, Y.; Lv, S. Vector field-based support vector regression for building energy consumption prediction. Appl. Energy 2019, 242, 403–414. [Google Scholar] [CrossRef]
  30. Li, C.; Tao, Y.; Ao, W.; Yang, S.; Bai, Y. Improving forecasting accuracy of daily enterprise electricity consumption using a random forest based on ensemble empirical mode decomposition. Energy 2018, 165, 1220–1227. [Google Scholar] [CrossRef]
  31. Kong, W.; Dong, Z.Y.; Jia, Y.; Hill, D.J.; Xu, Y.; Zhang, Y. Short-term residential load forecasting based on LSTM recurrent neural network. IEEE Trans. Smart Grid 2017, 10, 841–851. [Google Scholar] [CrossRef]
  32. Wang, S.; Wang, X.; Wang, S.; Wang, D. Bi-directional long short-term memory method based on attention mechanism and rolling update for short-term load forecasting. Int. J. Electr. Power Energy Syst. 2019, 109, 470–479. [Google Scholar] [CrossRef]
  33. Raza, M.Q.; Mithulananthan, N.; Li, J.; Lee, K.Y. Multivariate ensemble forecast framework for demand prediction of anomalous days. IEEE Trans. Sustain. Energy 2018, 11, 27–36. [Google Scholar] [CrossRef] [Green Version]
  34. Alsanea, M.; Dukyil, A.S.; Riaz, B.; Alebeisat, F.; Islam, M.; Habib, S. To Assist Oncologists: An Efficient Machine Learning-Based Approach for Anti-Cancer Peptides Classification. Sensors 2022, 22, 4005. [Google Scholar] [CrossRef]
  35. Kim, T.-Y.; Cho, S.-B. Predicting residential energy consumption using CNN-LSTM neural networks. Energy 2019, 182, 72–81. [Google Scholar] [CrossRef]
  36. Ullah, F.U.M.; Ullah, A.; Haq, I.U.; Rho, S.; Baik, S.W.J.I.A. Short-Term Prediction of Residential Power Energy Consumption via CNN and Multilayer Bi-directional LSTM Networks. IEEE Access 2019, 8, 123369–123380. [Google Scholar] [CrossRef]
  37. Khan, Z.A.; Hussain, T.; Ullah, A.; Rho, S.; Lee, M.; Baik, S.W. Towards Efficient Electricity Forecasting in Residential and Commercial Buildings: A Novel Hybrid CNN with a LSTM-AE based Framework. Sensors 2020, 20, 1399. [Google Scholar] [CrossRef]
  38. Afrasiabi, M.; Mohammadi, M.; Rastegar, M.; Stankovic, L.; Afrasiabi, S.; Khazaei, M. Deep-based conditional probability density function forecasting of residential loads. IEEE Trans. Smart Grid 2020, 11, 3646–3657. [Google Scholar] [CrossRef]
  39. Genes, C.; Esnaola, I.; Perlaza, S.M.; Ochoa, L.F.; Coca, D. Recovering missing data via matrix completion in electricity distribution systems. In Proceedings of the 2016 IEEE 17th International Workshop on Signal Processing Advances in Wireless Communications (SPAWC), Edinburgh, UK, 3–6 July 2016; pp. 1–6. [Google Scholar]
  40. Habib, S.; Alsanea, M.; Aloraini, M.; Al-Rawashdeh, H.S.; Islam, M.; Khan, S. An Efficient and Effective Deep Learning-Based Model for Real-Time Face Mask Detection. Sensors 2022, 22, 2602. [Google Scholar] [CrossRef]
  41. Box, G.E.; Cox, D.R. An analysis of transformations. J. R. Stat. Soc. Ser. B 1964, 26, 211–243. [Google Scholar] [CrossRef]
  42. Yeo, I.K.; Johnson, R.A. A new family of power transformations to improve normality or symmetry. Biometrika 2000, 87, 954–959. [Google Scholar] [CrossRef]
  43. Zhao, X.; Wei, H.; Wang, H.; Zhu, T.; Zhang, K. 3D-CNN-based feature extraction of ground-based cloud images for direct normal irradiance prediction. Sol. Energy 2019, 181, 510–518. [Google Scholar] [CrossRef]
  44. Yar, H.; Hussain, T.; Khan, Z.A.; Koundal, D.; Lee, M.Y.; Baik, S.W. Vision sensor-based real-time fire detection in resource-constrained IoT environments. Comput. Intell. Neurosci. 2021, 2021, 5195508. [Google Scholar] [CrossRef] [PubMed]
  45. Wang, F.; Li, K.; Duić, N.; Mi, Z.; Hodge, B.-M.; Shafie-khah, M.; Catalão, J.P. Association rule mining based quantitative analysis approach of household characteristics impacts on residential electricity consumption patterns. Energy Convers. Manag. 2018, 171, 839–854. [Google Scholar] [CrossRef]
  46. Ullah, W.; Hussain, T.; Khan, Z.A.; Haroon, U.; Baik, S.W. Intelligent dual stream CNN and echo state network for anomaly detection. Knowl.-Based Syst. 2022, 253, 109456. [Google Scholar] [CrossRef]
  47. Yar, H.; Hussain, T.; Khan, Z.A.; Lee, M.Y.; Baik, S.W. Fire Detection via Effective Vision Transformers. J. Korean Inst. Next Gener. Comput. 2021, 17, 21–30. [Google Scholar]
  48. Aladhadh, S.; Alsanea, M.; Aloraini, M.; Khan, T.; Habib, S.; Islam, M. An Effective Skin Cancer Classification Mechanism via Medical Vision Transformer. Sensors 2022, 22, 4008. [Google Scholar] [CrossRef]
  49. Ma, X.; Dai, Z.; He, Z.; Ma, J.; Wang, Y.; Wang, Y. Learning traffic as images: A deep convolutional neural network for large-scale transportation network speed prediction. Sensors 2017, 17, 818. [Google Scholar] [CrossRef]
  50. Khan, Z.A.; Ullah, W.; Ullah, A.; Rho, S.; Lee, M.Y.; Baik, S.W. An Adaptive Filtering Technique for Segmentation of Tuberculosis in Microscopic Images; Association for Computing Machinery: New York, NY, USA, 2020; pp. 184–187. [Google Scholar]
  51. Wang, H.; Yi, H.; Peng, J.; Wang, G.; Liu, Y.; Jiang, H.; Liu, W. Deterministic and probabilistic forecasting of photovoltaic power based on deep convolutional neural network. Energy Convers. Manag. 2017, 153, 409–422. [Google Scholar] [CrossRef]
  52. Ali, H.; Farman, H.; Yar, H.; Khan, Z.; Habib, S.; Ammar, A. Deep learning-based election results prediction using Twitter activity. Soft Comput. 2021, 26, 7535–7543. [Google Scholar] [CrossRef]
  53. Chen, K.; Chen, K.; Wang, Q.; He, Z.; Hu, J.; He, J. Short-term load forecasting with deep residual networks. IEEE Trans. Smart Grid 2018, 10, 3943–3952. [Google Scholar] [CrossRef]
  54. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  55. He, K.; Sun, J. Convolutional neural networks at constrained time cost. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 5353–5360. [Google Scholar]
  56. Rajabi, R.; Estebsari, A. Deep Learning Based Forecasting of Individual Residential Loads Using Recurrence Plots. In Proceedings of the 2019 IEEE Milan PowerTech, Milan, Italy, 23–27 June 2019; pp. 1–5. [Google Scholar]
  57. Kim, J.-Y.; Cho, S.-B. Electric energy consumption prediction by deep learning with state explainable autoencoder. Energies 2019, 12, 739. [Google Scholar] [CrossRef]
  58. Han, T.; Muhammad, K.; Hussain, T.; Lloret, J.; Baik, S.W. An efficient deep learning framework for intelligent energy management in IoT networks. IEEE Internet Things J. 2020, 8, 3170–3179. [Google Scholar] [CrossRef]
  59. Mocanu, E.; Nguyen, P.H.; Gibescu, M.; Kling, W.L. Deep learning for estimating building energy consumption. Sustain. Energy Grids Netw. 2016, 6, 91–99. [Google Scholar] [CrossRef]
  60. Khan, Z.A.; Hussain, T.; Baik, S.W. Boosting energy harvesting via deep learning-based renewable power generation prediction. J. King Saud Univ.-Sci. 2022, 34, 101815. [Google Scholar] [CrossRef]
  61. Khan, Z.A.; Ullah, A.; Haq, I.U.; Hamdy, M.; Maurod, G.M.; Muhammad, K.; Hijji, M.; Baik, S.W. Efficient Short-Term Electricity Load Forecasting for Effective Energy Management. Sustain. Energy Technol. Assess. 2022, 53, 102337. [Google Scholar] [CrossRef]
  62. Khan, Z.A.; Ullah, A.; Ullah, W.; Rho, S.; Lee, M.; Baik, S.W. Electrical energy prediction in residential buildings for short-term horizons using hybrid deep learning strategy. Appl. Sci. 2020, 10, 8634. [Google Scholar] [CrossRef]
  63. Sajjad, M.; Khan, Z.A.; Ullah, A.; Hussain, T.; Ullah, W.; Lee, M.Y.; Baik, S.W. A novel CNN-GRU-based hybrid approach for short-term residential load forecasting. IEEE Access 2020, 8, 143759–143768. [Google Scholar] [CrossRef]
  64. Abdel-Basset, M.; Hawash, H.; Sallam, K.; Askar, S.S.; Abouhawwash, M. STLF-Net: Two-stream deep network for short-term load forecasting in residential buildings. J. King Saud Univ.-Comput. Inf. Sci. 2022, 34, 4296–4311. [Google Scholar] [CrossRef]
  65. Khan, S.U.; Haq, I.U.; Khan, Z.A.; Khan, N.; Lee, M.Y.; Baik, S.W. Atrous Convolutions and Residual GRU Based Architecture for Matching Power Demand with Supply. Sensors 2021, 21, 7191. [Google Scholar] [CrossRef]
  66. Khan, Z.A.; Hussain, T.; Haq, I.U.; Ullah, F.U.M.; Baik, S.W. Towards efficient and effective renewable energy prediction via deep learning. Energy Rep. 2022, 8, 10230–10243. [Google Scholar] [CrossRef]
  67. Mujeeb, S.; Javaid, N. ESAENARX and DE-RELM: Novel schemes for big data predictive analytics of electricity load and price. Sustain. Cities Soc. 2019, 51, 101642. [Google Scholar] [CrossRef]
  68. Gao, W.; Darvishan, A.; Toghani, M.; Mohammadi, M.; Abedinia, O.; Ghadimi, N. Different states of multi-block based forecast engine for price and load prediction. Int. J. Electr. Power Energy Syst. 2019, 104, 423–435. [Google Scholar] [CrossRef]
  69. Chou, J.S.; Truong, D.N. Multistep energy consumption forecasting by metaheuristic optimization of time-series analysis and machine learning. Int. J. Energy Res. 2021, 45, 4581–4612. [Google Scholar] [CrossRef]
Figure 1. Proposed framework for electricity consumption forecasting.
Figure 1. Proposed framework for electricity consumption forecasting.
Sensors 22 06913 g001
Figure 2. Data down sampling where “(a)” shows actual minute data and “(b)” shows hourly down-sampled data.
Figure 2. Data down sampling where “(a)” shows actual minute data and “(b)” shows hourly down-sampled data.
Sensors 22 06913 g002
Figure 3. Visual results of R-CNN with ML-LSTM for IHEPC (a) hourly and (b) daily.
Figure 3. Visual results of R-CNN with ML-LSTM for IHEPC (a) hourly and (b) daily.
Sensors 22 06913 g003
Figure 4. Visual results of the proposed model over IHEPC dataset (a) AEP, (b) DAYTON, (c) COMED, (d) DOM, (e) DEOK, (f) EKPC, (g) DUQ, (h) PJME, (i) NI, and (j) PJMW dataset.
Figure 4. Visual results of the proposed model over IHEPC dataset (a) AEP, (b) DAYTON, (c) COMED, (d) DOM, (e) DEOK, (f) EKPC, (g) DUQ, (h) PJME, (i) NI, and (j) PJMW dataset.
Sensors 22 06913 g004
Table 1. The internal architecture of R-CNN and ML-LSTM.
Table 1. The internal architecture of R-CNN and ML-LSTM.
LayerFilter-SizeKernel-SizeLayer-Parameter
Input --
Convolutional (conv)_132710,816
conv_232520,544
Add [conv_1, conv_2] --
conv_364312,352
Add [conv_2, conv_3] --
Convolutional_412814160
Add [conv_3, conv_4] --
LSTM100-66,000
LSTM100-80,400
LSTM100-80,400
FC128-12,928
FC60-7740
Total parameters295,340
Table 2. IHEPC dataset attributes and units.
Table 2. IHEPC dataset attributes and units.
AttributesDescriptionUnits
Date and timeComprise the range of datetime valuesdd/mm/yyyy and hh:mm:ss
Global active, reactive power and intensityMinutely averaged Global active1, reactive Power2, and intensity3 valueskilowatt (Kw)1,2
Ampere (A)3
VoltageMinutely averaged voltage valuesVolt(V)
Table 3. Performance comparison of R-CNN and ML-LSTM with state-of-the-art over hourly IHEPC dataset.
Table 3. Performance comparison of R-CNN and ML-LSTM with state-of-the-art over hourly IHEPC dataset.
MethodRMSEMAEMSER2MAPE
Linear regression [35]0.65700.50220.4247-83.74
ANN [56]1.151.08---
CNN [37]0.670.470.37--
CNNLSTM [35]0.5950.33170.3549-32.83
CNN-BDLSTM [36]0.5650.3460.319-29.10
CNNLSTM-autoencoder [37]0.470.310.19--
SE-AE [57]-0.3950.384--
GRU [58]0.410.190.17-34.48
FCRBM [59]0.666--−0.0925
CNNESN [60]0.04720.02660.0022--
Residual CNN Stacked LSTM [61]0.0580.0030.038--
CNN-BiGRU [62]0.420.290.18--
CNN-GRU [63]0.470.330.22--
STLF-Net [64]0.43860.26740.1924-36.24
Residual GRU [65]0.41860.26350.1753--
ESN-CNN [66]0.21530.11370.0463--
Proposed0.03250.01440.00110.98411.024
Table 4. Performance comparison of R-CNN and ML-LSTM with state-of-the-art over daily IHEPC dataset.
Table 4. Performance comparison of R-CNN and ML-LSTM with state-of-the-art over daily IHEPC dataset.
MethodRMSEMAEMSER2MAPE
Linear regression [35]0.50260.39150.2526-52.69
CNN [37]0.070.050.006--
LSTM [35]0.49050.41250.2406-3872
CNN-LSTM [35]0.32210.25690.1037 37.83
FCRBM [59]0.828--0.3304
Proposed0.04470.01320.0020.97952.457
Table 5. Performance comparison of the proposed model with state-of-the-art forecasting methods over the PJM hourly dataset.
Table 5. Performance comparison of the proposed model with state-of-the-art forecasting methods over the PJM hourly dataset.
DatasetMethodRMSEMAEMSER2MAPE
AEPMujeeb et al. [67]0.386---1.08
Gao et al. [68]0.49---1.14
Han et al. [58]0.054----
Khan et al. [61]0.0310.0010.027--
Proposed0.02230.01630.00050.99070.5504
DAYTONKhan et al. [61]0.0460.0330.002--
Proposed0.02060.01440.00040.99110.4982
COMEDKhan et al. [61]0.0440.0300.002--
Proposed0.02160.01310.00050.99060.5475
DOMKhan et al. [61]0.0570.0390.003--
Proposed0.02120.01380.00050.99050.5987
DEOKKhan et al. [61]0.0530.0360.003--
Proposed0.01740.01290.00030.99320.3974
EKPCKhan et al. [61]0.0550.0340.003--
Proposed0.02740.02020.00080.98820.7965
DUQKhan et al. [61]0.0540.0410.003--
Proposed0.04300.02770.00090.99750.8234
PJMEKhan et al. [61]0.0430.0310.002--
Proposed0.01990.01280.00040.99130.4721
NIKhan et al. [61]0.0500.0330.002--
Proposed0.01780.01290.00030.99300.3748
PJMWKhan et al. [61]0.0380.0270.001--
Proposed0.01450.01020.00020.99490.2864
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Alsharekh, M.F.; Habib, S.; Dewi, D.A.; Albattah, W.; Islam, M.; Albahli, S. Improving the Efficiency of Multistep Short-Term Electricity Load Forecasting via R-CNN with ML-LSTM. Sensors 2022, 22, 6913. https://doi.org/10.3390/s22186913

AMA Style

Alsharekh MF, Habib S, Dewi DA, Albattah W, Islam M, Albahli S. Improving the Efficiency of Multistep Short-Term Electricity Load Forecasting via R-CNN with ML-LSTM. Sensors. 2022; 22(18):6913. https://doi.org/10.3390/s22186913

Chicago/Turabian Style

Alsharekh, Mohammed F., Shabana Habib, Deshinta Arrova Dewi, Waleed Albattah, Muhammad Islam, and Saleh Albahli. 2022. "Improving the Efficiency of Multistep Short-Term Electricity Load Forecasting via R-CNN with ML-LSTM" Sensors 22, no. 18: 6913. https://doi.org/10.3390/s22186913

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop