Abstract
Peer-to-Peer (P2P) energy trading has gained much attention recently due to the advanced development of distributed energy resources. P2P enables prosumers to trade their surplus electricity and allows consumers to purchase affordable and locally produced renewable energy. Therefore, it is significant to develop solutions that are able to forecast energy consumption and generation toward better power management, thereby making renewable energy more accessible and empowering prosumers to make an informed decision on their energy management. In this paper, several models for forecasting short-term renewable energy consumption and generating are developed and discussed. Real-time energy datasets were collected from smart meters that were installed in residential premises in Western Australia. These datasets are collected from August 2018 to Apr 2019 at fine time resolution down to 5 s and comprise energy import from the grid, energy export to the grid, energy generation from installed rooftop PV, energy consumption in households, and outdoor temperature. Several models for forecasting short-term renewable energy consumption and generating are developed and discussed. The empirical results demonstrate the superiority of the optimised deep learning-based Long Term Short Memory (LSTM) model in forecasting both energy consumption and generation and outperforms the baseline model as well as the alternative classical and machine learning methods by a substantial margin.
Keywords: Energy consumption, Energy generation, Renewable energy, Time series forecasting, Peer-to-peer energy trading
Energy consumption; Energy generation; Renewable energy; Time series forecasting; Peer-to-peer energy trading.
1. Introduction
The ongoing growth of the world's population has led to a dramatic increase in energy demand which is expected to approach a 25% increase, in comparison to the actual consumption, by 2040 [1]. To meet such demand, efforts have been consolidated toward developing advanced renewable energy systems that have led to establishing new economic alternatives [2, 3]. In this context, sharing economy has exploded in popularity over recent years and is expecting to continue its trend and expand to many market sectors e.g., Peer-to-Peer (P2P) transportation (Uber), P2P accommodation (Airbnb), P2P energy trading, etc. Supply and demand are matched using high-speed ICT in sharing economy which is disrupting traditional business models. In the energy sector, the development and the widespread deployment of smart meters offer the essential infrastructure for sharing economy through P2P energy trading [4]. This leads to affordable clean energy and ultimately to the sustainable energy transition. Further, this commences venues for power system markets as well as potentially allowing a transformation to customers sharing electricity. In addition, P2P electricity markets may allow investing in locally produced renewable energy which provides more flexibility for consumers to choose their source of electric energy. A sharing economy in the energy sector can be seen through a P2P local energy trading market allowing energy buyers and sellers to trade with each other bypassing the central system.
The Australian Government has funded a research project (RENeW Nexus project [5]) to investigate the effect of using blockchain technology as well as data analytics to enable P2P trading of energy and water. In the first phase of the project, electric smart meters were deployed in fifty houses across the city of Fremantle, Western Australia. The energy data was collected in real-time and included: energy data imported from the grid; energy data exported to the grid; energy data generated from the installed rooftop PV; and energy consumed at the households. To enable effective P2P energy trading, real-time forecasting is essential to estimate the likely energy usage and production, forming crucial insights that can substantially increase operation efficiency and make the most efficient energy trading between prosumers and consumers. In a traditional centralised power supply, consumers purchase electricity from utilities and eligible counterparts in fixed tariffs. On contrary, consumers become prosumers (i.e., they can consume as well as produce power) in P2P energy trading and their role can be smoothly switched, thereby allowing to either purchasing power or selling it to achieve a win-win in an open market model [6]. However, decision-making, including electricity price bidding, energy consumption and generation scheduling, etc., in such a decentralised environment is not an easy task; this is due to the complexity, independency, uncertainty, and unpredictability of the market in reality [7, 8]. Further, reliability analysis, security assessment, voltage excursion, thermal overloading, and other network constraints [9, 10, 11, 12, 13] augment the trading complexity. Therefore, real-time energy forecasting is an essential guide for buyers and sellers to assist the bidding process on the spot market, to a better understanding of real-time prosumers behaviour, and to enable peak-shaving and smooth power dispatching [14, 15, 16]. It also allows to attain security of operations conducted under tight reserve margins [17].
In addition, a traditional highly centralised energy management and market is controlled by a few key players. A transition to a more democratic decentralised market is needed. The real-time forecasting for P2P energy consumption and generation is not only enabling the transition but also providing a solution to the phenomenon known as the utility death spiral [18]. This global phenomenon is caused by the increased adoption of rooftop photovoltaic (PV) panels which reduces overall grid electricity demand without affecting peak demand. With fewer paying customers to maintain the necessary infrastructure, utilities are required to increase electricity prices; this in turn encourages greater adoption of PV panels. This is exacerbated by the decreasing costs of PV technology. Battery adoption by PV users can help to reduce peak demand however excess electricity is still exported to the grid in an inefficient manner, at times when demand is low. A study of actual P2P trading system has been reported and published in [4].
In this paper, short-term forecasting of energy consumption and production using actual data is explored. Real-time data were collected from smart meters that were installed in residential premises and used to evaluate the efficacy and effectiveness of statistical and machine learning techniques. Well-known classical (statistical) and AI-based models are used to analyse energy consumption and production. The comprehensive comparison is investigated and illustrated. The following are the key contributions of this paper:
-
•
A two-fold analytical model to forecast energy consumption and generation in P2P settings is designed and implemented.
-
•
An auto-ML model and various classical and advanced AI-based techniques are incorporated, and their utility is evaluated and compared comprehensively.
-
•
To the best of our knowledge, this is the first paper that furnishes a forecasting model for both energy consumption and generation based on real data captured from a P2P grid system in the state of Western Australia.
The rest of this paper is organized as follows: Section 2 offers a review of various intelligent techniques used in energy consumption and generation. Section 3 discusses different forecasting approaches used for time series analysis. Section 4 provides a comprehensive discussion on the models’ development. The carried out experiments to validate the utility of the model are presented and discussed in Section 5. Section 6 discusses the empirical results and point to the limitations of this study. Section 7 concludes the paper.
2. Related works
The application of Artificial Intelligence (AI) has extended to numerous industrial domains. This is due to the sophisticated architecture of AI algorithms that empowers them to address a wide variety of real-life problems [19, 20, 21, 22, 23]. Forecasting of energy consumption and generation is an important venue of research that is used to tackle related issues such as demand response, fault detection and troubleshooting, model predictive control, and energy management and optimization [24]. Hence, renewable energy forecasting has attracted a great deal of interest recently [25, 26, 27, 28, 29]. Energy forecasting techniques are commonly categorized into two key approaches, namely conventional models and AI-based models. In a recent survey [30], it was indicated that the proportion of AI-based models to conventional/classical models, based on a study of 128 models, are 48% and 43%, respectively. This implies the significance of both models in the designated task even though AI-based models are much more sophisticated. In fact, some classical models, under certain assumptions, overshadow advanced AI approaches [31]. In this section, we visit studies that have been carried out in the area of both energy forecasting and consumption using the aforementioned approaches.
In terms of energy consumption forecasting, various studies were proposed to tackle this issue incorporating both classical models [32, 33, 34, 35, 36, 37, 38] and AI-based models [36, 37, 38, 39, 40, 41]. For example [32], proposed a novel time-series smoothing model for medium-term electricity consumption forecasting. The model is based on a semi-parametric model that offers an extension of the semi-parametric and nonparametric vector autoregression model. Long-term energy consumption forecasting was addressed in [33] where authors combined a mixed data sampling model with an autoregressive distributed lag to predict energy demand in China. For energy consumption in Turkey and Pakistan, Akpinar et al. [34] and Hussain et al. [35] applied the ARIMA model for this designated task respectively. Various other conventional models have been incorporated for energy consumption prediction, including logistic regression [36], linear regression [37], and Nonlinear regression (NLR) [38].
Popular AI-based approaches such as Artificial Neural Networks (ANN), Support Vector Machine (SVM), and Random Forest (RF) were also utilized in energy forecasting models. ANN includes various machine learning algorithms that were used in the literature. For example, three layers of Feedforward Neural Network (FFNN) are used in [42]. An optimized hybrid algorithm that uses FFNN was presented by [39] to predicted short-term building energy forecasting. Another optimization algorithm, for short-term energy consumption, was proposed which incorporates FFNN and Bayesian regularization algorithm [40]. Another thread of efforts adopts Deep Learning (DL) techniques leveraged by advanced specifications of modern computers. DL models are mainly represented by the recurrent neural network (RNN) to forecast energy consumption [43]. He [41] incorporated RNN to model the implicit dynamics and obtain the predicted load along with CNN that was used in their work to extract significant features from the historical load sequence. Long Short-Term Memory (LSTM) is a special type of RNN that has been proven to outperform modern models [41, 44]. Authors of [45] employed LSTM architecture toward electricity consumption forecasting. They presented the sequence-to-sequence architecture to obtain a random number of formerly attainable load measurements and use them as input to approximate the load for future time steps.
Statistical models using both classical models as well as AI-based models were utilised to predict energy generation, including ARMAX models [46], ARIMA/SARIMA models [47, 48, 49], multiple regression [50], regression with neural networks [51, 52, 53], support vector machine (SVM) [54], etc. For example, the authors of [55] transformed time series data into stationary data using the ARIMA model. Their model for predicting energy is validated using the Akaike Information Criterion (AIC) and Residual Sum of Squares (SSE). Ayub et al. [56] applied SVM with three parameters (kernel parameter, cost penalty, and incentive loss function parameter) on the electricity load data set. Using various machine learning methods was also reported on a real data set collected in France [57]. Li [58] introduced short-term solar irradiance forecasting algorithms based on Hidden Markov Model and SVM regression. The paper presents an approach to predict Photovoltaic (PV) generation under different weather conditions. Statistical and intelligent models based on machine learning used for PV generation forecasting were also reported in [3, 59, 60, 61].
Table 1 demonstrates a summary of several approaches for forecasting energy generation and consumption. As illustrated in the table, most of the current studies were carried out using limited variations of statistical techniques to predict either energy generation or energy consumption. Also, few attempts were undertaken to conduct very short-term energy forecasting. Implementing short-term energy forecasting approaches benefits real-time or near real-time energy dispatching systems, and it has proven utility in providing better coordination of resources [14]. Our approach, on the other hand, discusses the implementation of short-term (i.e., hourly) power generation as well as power consumption models using a variety of classical and sophisticated AI-based techniques. Further, to the best of our knowledge, this is the first paper that offers a two-fold energy consumption and generation model based on real data captured from a P2P grid system in the state of Western Australia.
Table 1.
Ref. | Forecasting Model | Consumption/Generation | Data Source | Data Temporality | Forecasting Horizon |
Country | Evaluation Metric(s) |
---|---|---|---|---|---|---|---|
[32] | Vector Autoregression | Consumption | Suzhou Municipal Bureau of Statistics | Jan 2004 to Jan 2014 | Monthly | China | MAPE |
[33] | ADL-MIDAS | Consumption | National Bureau of Statistics of China | 2016 | Quarterly | China | RMSFE |
[34] | ARIMA | Consumption | Adapazari Natural Gas Distribution | 2009 to 2012 | Monthly | Turkey | MAPE |
[35] | Holt-Winter and ARIMA | Consumption | Pakistan Economic Survey | 1980 to 2011 | Annually | Pakistan | RMSE, MAPE |
[62] | ARMA + Kalman filter | Consumption | Hellenic Public Power Corporation S.A. | Jan 2004 to Dec 2006 | Daily | Greece | MAPE |
[63] | MA + SARIMA + PSO | Consumption | Power Grids of China | Dec 2003 to Dec 2009 | Monthly | China | MAE, RMSE, MAPE |
[64] | Exponential smoothing model + Bayesian inference | Consumption | IEA website | 1990 to 2014 | Yearly | Japan | AAEP |
[65] | FARX | Consumption | Residential Energy Consumption Survey |
Apr 2018 to Jul 2015 | Hourly | USA | MAPE, RMSE |
[36] | Multi-cycle logistic model | Consumption and Generation | US Energy Information Agency | 1949 to 2015 | Yearly | USA | R-square |
[66] | Holt-Winters exponential smoothing method | Consumption | International Energy Agency | 1993 to 2007 | Yearly | Romania | MAPE, MAE, MSE |
[67] | Adaptive Residual Compensation | Generation | global energy forecasting competition | 2004 to 2014 | Hourly | USA | RMSE, R-square, CRPS |
[39] | FFNN | Consumption | ASHRAE, library building located in Hangzhou, East China | Sep 1989 to Feb 1990 | Hourly | China | MAPE |
[40] | FFNN and Bayesian regularization algorithm | Consumption | Building management system | Jul 2012 | 15-minute | N/A | MBE, RMSE |
[41] | RNN and CNN | Consumption | A City in North China | Feb 2010 to Dec 2012 | Hourly | China | MAPE, MAE |
[45] | LSTM | Consumption | Individual household electric power consumption | Dec 2006 to Nov 2010 | Hourly, and sub-hourly | USA | RMSE |
[55] | ARIMA | Generation | Building in Reese Research Center | Nov 2017 to Nov 2018 | Monthly | USA | MAPE |
[59] | WT, LSTM, SAE | Generation | Energy Information Administration | Jan 1997 to Dec 2022 | Monthly | USA | MAE, RMSE, U1, U2 |
[60] | Ensembled ANN | Generation | Federal Institute of Southern Minas Gerais State | May 2017 to Apr 2019 | Weekly | Brazil | MAPE |
[61] | ARIMA and ANN | Generation | National Climate Data Center | Jan 2014 to Oct 2019 | Daily | Korea | RMSE, MAPE |
[3] | LSTM | Generation | Turkish Electricity Transmission Corporation | Jan 2016 to Dec 2019 | Daily | Turkey | RMSE, MAE MAPE |
This study |
LSTM, ARIMA, VAR LR, Lasso, Ridge, ElasticNet HuberRegressor, Lars, LassoLars, PassiveAggressiveRegressor, RANSACRegressor, SGDRegressor, and RapidMiner |
Consumption and Generation | RENeW Nexus project | Aug 2018 to Apr 2019 | Hourly | Australia | RMSE, MAE |
3. Time series forecasting models and evaluation metrics
In this section, a discussion is provided on the classical (statistical) and AI-based models which are incorporated in the development of the prediction models in this study. Those models are typically used to forecast the energy consumption and generation of buildings, and implemented through four main stages, namely; data acquisition, data pre-processing, model training, and model testing [68]. Further, certain machine learning tools and software platforms will be presented followed by a discussion on key metrics to evaluate the performance of the time series forecasting model.
3.1. Classical time series forecasting models
3.1.1. Univariate and multivariate classical models
Classical time series forecasting models are designed to focus on linear relationships where many of these techniques rely on decomposing time series to three main components; (i) Trend: indicates the increase/decrease of data; (ii) Seasonality: inferring an iterating pattern in a particular time interval; and (iii) Noise: refers to irregular components of data samples [69]. There are various classical time series forecasting models to handle either or both univariate (a single variable) and multivariate (Multiple variables) time series problems.
Autoregressive Integrated Moving Average (ARIMA): ARIMA refers to a particular form of regression-based models which presume that time series values are continuous measures [70]. ARIMA is commonly used to be fitted to the time series dataset to gain a better fathom of the dataset and to forecast future points in the examined time series problem. ARIMA models are widely used due to their extensive harness of sophisticated statistics that permit effective determination of the embodied parameters, as well as inclusive valuation of their suitability [71]. ARIMA is used to model stationary and non-stationary time series datasets, and it is mainly composed of two models; Autoregressive Models (AR): uses observations inferred from preceding time steps in the series to feed a regression linear function, thereby forecasting the value at the next time step; Moving Average (MA): uses the past forecasting errors instead in the regression linear function. ARIMA also includes a pre-processing step (Integration (I)) to make the series stationary by conducting a differencing step. Hence, any typical ARIMA model should identify three main parameters; , thus the notation of ARIMA can be defined as . These parameters formulate the forecasting equation of an ARIMA model as follows:
(1) |
where is the observed output at time and is the error term at time .
Seasonal Autoregressive Integrated Moving-Average (SARIMA): Seasonal ARIMA or SARIMA is a special form of ARIMA model that is used mainly to model a wide range of seasonal data. This is attained by incorporating an additional seasonal parameter to handle a period of seasonality [72]. SARIMA, therefore, has its own configuration and new different hyperparameters are introduced as follows; (i) : Seasonal autoregressive order. : Seasonal difference order; : Seasonal moving average order; and : The number of time steps for a single seasonal period (for example, m = 4 of data is quarterly). The notation of SARIMA can be described as and can be formulated as:
(2) |
where is the original non-stationary output at time is the stationary output at time ., and is the error term at time .
Vector Autoregression (VAR): VAR is used commonly with multiple parallel time series (eg. multivariate time series). It is a generic form of AR for predicting a vector of time series. Hence, it includes one equation for each variable in the model, then VAR generates a forecast for each variable in a repetitive way [72]. VAR model has proven ability to fit several time series problems, this includes analysing a dissimilar number of variables to infer the dynamic relationships among them [73], or investigating whether a certain variable might affect forecasting different variables. In other words, each designated variable is a linear function of previous lags of itself as well as previous lags of other variables. The typical equation for a VAR model can be written as:
(3) |
where is a constant, to are the coefficients of the lags of until order which indicates the number of -lags of used, and indicates the white noise.
3.1.2. Multi-step forecast with linear algorithms
Time series problem is commonly addressed through forecasting one value (observation) in the future (i.e. one-step prediction). However, conducting time series analysis can also include predicting multiple values in the future (i.e., multiple-steps prediction). For example, instead of predicting the temperature of the next day (one-step), forecasting the temperature for the next week (seven-steps) is also valid with multi-step time series. This study will focus on one-step time series forecasting and will leave multi-step forecasting to future work. Python's scikit-learn library1 offers a set of generalized linear regression-based models that can be incorporated with one-step and multi-step time series forecasting problems. These include LinearRegression, Lasso, Ridge, HuberRegressor, to name a few.
3.2. Deep learning - Long Short-Term Memory (LSTM)
Long Short-Term Memory networks - commonly called "LSTMs" - are a special kind of Recurrent Neural Network (RNN), capable of learning long-term dependencies. It was first introduced by Hochreiter & Schmidhuber [74], revised, and circulated by many research fellows at the following work [75, 76, 77]. LSTM layer uses the concept of numerous hidden state types to alter the quantity of information kept across so called states. This can be used for instance when working on sequential data (i.e., time-series or text), since the hidden states can store a given amount of information from previous states beyond the just handled one. This means that for example, a connection between the first word of a long text and the last word can be created even though the complete paragraph is quite long. To account for the importance of in-between states, LSTMs use mechanisms to adjust the importance and amount of influence a hidden state has for the current calculation [78].
Figure 1 portrays the inner cell diagram of an LSTM Network. LSTM preserves a hidden vector, , and a memory vector, , where the state updates and outputs are controlled at each step, respectively. The computation at each time step is formulated as the following:
(4) |
(5) |
(6) |
(7) |
(8) |
(9) |
Where is the activation vector of the input gate, is the activation vector for the forget gate, is the activation vector of the output gate, and is the activation vector of the cell state gate. is the hidden state vector of the LSTM unit, the logistic sigmoid function is indicated by , the elementwise multiplication is embodied by . represent the recurrent weight matrices. Finally, notations indicate the projection matrices.
The ongoing incorporation of deep learning approaches is due to their technical capabilities to learn long-term dependencies as well as nonlinear characteristics that are captured in electric data which commonly leads to obtain accurate forecasting results; thus they overshadow other classical or statistical machine learning models [80, 81]. Further, data in an energy forecasting problem might embody a strong periodicity (i.e., oscillates or fluctuates over time) [82]. This can be observed in electricity consumption that commonly peaks at a certain time of the day. This periodicity can be also perceived seasonally (i.e., monthly, yearly, etc.). The literature also verifies the utility of deep learning models in other metrics including evaluation errors, prediction accuracy, and robust generalisation ability [83, 84, 85].
3.3. Data science tools
With the growing interest in data science and machine learning, the intense rivalry amongst software companies has led those companies to develop cross-platform tools and software systems that are used for conducting machine learning and data mining tasks with minimum efforts. Gartner in their 2021 Magic Quadrant for Data Science and Machine Learning Platforms [86] has placed seventeen distinguished software companies into four different forms of technology providers. As depicted in Figure 2, RapidMiner is amongst four competing companies which are positioned on the Leader quadrant for being drivers for transformation.
RapidMiner™ has been incorporated in this study as an alternative rapid approach for conducting time series forecasting. The aim is to examine the results of the conducted experiments using this software platform, thereby providing another benchmark comparison with the obtained results of classical and deep learning-based forecasting techniques.
3.4. Metrics for models performance evaluation
This study uses a set of evaluation metrics to measure the performance of the implemented prediction module. The following metrics are used to compare the performance of the model for forecasting energy consumption and generation: (i) Root-Mean-Square Error (RMSE): is used to measure the accuracy of model prediction performance. It is commonly used to calculate standard deviation of the prediction errors (residuals); (ii) Mean Square Error (MSE): provides an indication on how well the regression model is by computing the distance between the data point and the regression line; and (iii) Mean Absolute Error (MAE): it is essentially the mean of the absolute differences between forecasted and actual values.
These metrics can be defined through the following formulas:
(10) |
(11) |
(12) |
where = forecasted value, = the actual value, and = number of data samples.
4. Models development
In this section, we discuss various experiments carried out to develop models for forecasting energy consumption and generation. The experiments that incorporate classical and ML techniques were developed using Python. Keras2 is also used as an efficient deep learning-based python library. Another set of experiments was also implemented incorporating the RapidMiner™ software platform.
4.1. Dataset description and Features Selection
The energy data are collected from August 2018 to Apr 2019 (9 months) at fine time resolution down to 5 s and transmitted through the Internet of Thing network. Amazon Web Services (AWS), a Virtual Private Cloud (VPC), is used to store energy data. The data comprises energy import from the grid, energy export to the grid, energy generation from installed rooftop PV, and energy consumption in households. This paper aims to build forecasting models to predict both energy generation and consumption. The dataset also comprises daily temperature observations collected by the Bureau of Meteorology for Perth, Western Australia3.
Datasets Pre-processing: a set of data pre-processing and feature engineering steps were followed before conducting experiments: (1) the energy consumption and generation data were aggregated from different sites and transformed from data with 5 s intervals into data with one-hour intervals. The aggregation of all power consumption and generation for all sites was carried out due to the fact that forecasting accuracy commonly declines as the level of aggregation decreases [87]. (2) Energy consumption and generation dataset were integrated with the temperature data obtained from the weather dataset. (3) The null/missing values in the dataset were identified and imputed. (4) A new feature to represent the day of the week is created.
Features Selection: The resultant dataset comprises three different variables which are selected due to their direct effect on energy generation and consumption. Those include day of the week (Monday to Sunday), outdoor temperature (⁰C), and the hour of the day (0–23). Table 2 shows some statistics of the processed datasets and Figure 3 illustrates a diagram of three subplots showing temperature data, energy consumption and energy generation during the whole 6120 h period (≈9 months).
Table 2.
Weekday | Temp | Energy Generated | Energy Consumed | |
---|---|---|---|---|
mean | 3.00 | 17.46 | 31.38 | 29.25 |
std | 2.00 | 6.35 | 41.62 | 12.13 |
min | 0.00 | 2.80 | 0.00 | 3.27 |
25% | 1.00 | 13.30 | 0.04 | 20.21 |
50% | 3.00 | 17.30 | 2.55 | 27.75 |
75% | 5.00 | 21.20 | 62.64 | 35.07 |
max | 6.00 | 38.00 | 136.29 | 88.24 |
Figure 4 depicts the correlation between temperature and energy (generated and consumed). It is evident that energy generation and consumption are highly dependent on temperature. People tend to consume much energy on hot days (for cooling) and cold days (for heating). Also, as depicted in Figure 4, there is moderate energy consumption in mild weather conditions. On the other hand, renewable energy production surges as temperature rises, thereby augmenting supplied energy (mix energy) to the smart grid.
The dataset from the beginning of Aug 2018 to the end of Feb 2019 was used in the model training process, and the remaining dataset of Mar and Apr 2019 were used to test and evaluate the model performance as depicted in Figure 5.
4.2. Multivariate energy consumption and generation forecasting using deep learning
4.2.1. Dataset preparation
Normalization and reshaping: incorporating LSTM for time series problem necessitates a need to normalize the dataset and to transform the dataset to a supervised learning problem (i.e., dataset is split into a set of variables/features and an output label). This involves learning from the previous data points in the series to forecast the next value in the future, thereby, the sequence of observations inferred from the series should be transformed as feature variables from which LSTM can learn. Based on suggestions provided by [88], we frame our time series problem as forecasting the energy consumption/generation at the current time (t) given the temperature, week of the day, and energy consumption/generation at the former time step.
Data Stationary: the next step is to ensure that the dataset is stationary which facilitates model development. This can be obtained by ensuring that there is no regular increase or decrease in the data over time (i.e., data is time-independent). Figure 6 portrays the relationship between both energy consumption and generation over time. It is evident that both energy generation and consumption do not depict any explicit trend over time, this data is stationary.
Time Series Scaling: LSTM, like any other neural network, requires the dataset to be scaled to fit its activation function. The default activation function of LSTM is function, its output values are in the range of [−1,1] which is a favorable range for time series data. Hence, scaling was applied to the dataset, and min/max coefficients were calculated on both the training and testing dataset.
4.2.2. LSTM network model design
Grid Search Hyperparameters Settings: constructing a neural network is not a conventional task; the selection amongst dissimilar parameters for LSTM setting consumes time and effort, but it is crucial to attain high-quality prediction model. This is because the forecasting performance of a developed model might vary based on dissimilar hyperparameter settings provided, this is due to the stochastic nature of LSTM algorithm where different batch sizes for example lead LSTM to learn differently each time it maps inputs with outputs. Therefore, hyperparameter optimization technique such as grid search was incorporated in this study to examine various LSTM structures of dissimilar hyperparameters settings. In particular, grid search comprises building a distinct LSTM for a combination of different parameters to tuning a neural network, thereby finding the optimal settings for building a competent forecasting model. To obtain this, we build a dictionary depicted in Table 3 containing a list of selected parameters with the incorporated selected settings available in the Keras library.
Table 3.
Hyperparameter | Examined Settings |
---|---|
Batch Sizes | 2,3 |
Number of Neurons | 2,3 |
No of Epochs | 1000,1500,2000 |
Optimization algorithms | SGD, RMSprop, Adagrad, Adadelta, Adam, Adamax, Nadam |
Activation Functions | tanh, softmax, elu, selu, softplus, softsign, relu, sigmoid, |
Losses | mse, mae, mape, logcosh |
Dropout Rate | 0.0, 0.1, 0.2, 0.3, 0.4, 0.5, 0.6, 0.7, 0.8, 0.9 |
4.3. Models development using RapidMiner™
RapidMiner™ was used to forecast both energy consumption and generation using diverse embodied machine learning-based statistical techniques, namely; Generalised Linear Model (GLM), Deep Learning (DL), Random Forest Tree (RFT), Gradient Boosted Tree (GBT), Decision Tree (DT), and Support Vector Machine (SVM). Figure 7 depicts the model development process incorporating RapidMiner studio.
As depicted in Figure 7, the developed framework using RapidMiner™ comprises the following main stages with their embodied operators:
-
(i)
Data Engineering and Pre-processing: in this phase, the dataset of energy consumption and generation passed through the set of data acquisition and pre-processing operators where the dataset is initially loaded with its metadata to the RapidMiner Object. Dataset then is regularised, unified, and mapped. In particular, pre-processing operator alters the type of the numerical attributes to real (float) data type. It also maps all values of these attributes to real values. Dataset Engineering in RapidMiner also includes replacing the missing value with the minimum, maximum or average value of the designated attribute - average value is chosen in our case. Finally, the dataset was split into three different subsets (40% Training, 20% Testing, and 20% Validation)
-
(ii)
Model Training and Testing: after the data engineering and pre-processing phase, the dataset passed through the “Reordering” operator that allows reordering the regular features of the dataset. RapidMiner allows to customize the reordering operator where ordering can be done alphabetically (including Regular Expressions) or with a reference – the default alphabetical ordering was used. Then, the dataset passed through the “Sliding Window Validation” operator. This operator comprises a set of sub-operators encapsulating the sliding windows of both training and test datasets and also is responsible to measure the performance of a prediction operator. The average of the performance measurements is computed after the window is moved over the dataset. Further, as depicted in Figure 7, the Sliding Window Validation operator embodies the set of all Machine Learning modules incorporated for training in this experiment alongside delivering a list of performance values based on the performance settings indicated for each Machine Learning technique.
-
(iii)
Model Evaluation: by completing the previous phase, the final performance of the various machine learning algorithm is tested through several experiments to infer the optimal hyperparameter settings for each module. In Model Evaluation an assessment was carried out to see how well the model is able to generalize. This is through using data samples that were not used to build the model, the aim is to provide an unbiased evaluation of final model effectiveness [89], by means of evaluating the final hypothesis by an independent evaluation [90].
As mentioned previously, the aim of this study is to provide an efficient forecasting platform to predict both renewable energy generation and energy consumption in households. The next section presents several experiments conducted to find the optimal model for energy generation and consumption.
5. Experimental results and discussion
5.1. Forecasting results based on LSTM model
5.1.1. Baseline model of performance
Building a baseline model for a time series problem (or generally any machine learning problem) is crucial as it establishes a benchmark comparison with the developed model. One of the most frequently used baseline models for time series problems is the persistence algorithm which essentially forecasts the value of the next time step (t+1) using the inferred value from the previous time step (t-1). Evaluating this baseline model was carried out using the rolling forecast or walk-forward validation approach as an alternative approach to cross-validation. Cross-validation could not be used in a time series problem as it neglects the fact that temporal factor links all observations of the time series dataset, thus samples are time-dependent. This differs from a typical machine learning problem where samples are independent with commonly no explicit relationships. Figure 8 illustrates the actual and prediction values from both energy consumption and generation tests datasets, also RMSEs values for both baseline models are indicated which will be used in the experiments discussed next section.
5.1.2. Experimental results on energy generation and consumption
The experiments were initiated by developing a model to forecast energy generation and consumption. Grid search optimization technique was used, thereby various LSTM network designs were constructed and tested to infer the optimal structure with the best hyperparameter settings. Table 4 presents the optimal settings for each task obtained by conducting a grid search for the LSTMs’ hyperparameters. Table 4 also illustrates the values of key metrics inferred from the conducted designated experiment for both tasks (i.e., Energy Generation and Energy Consumption).
Table 4.
Task | Hyperparameters |
Evaluation Metric |
|||||||
---|---|---|---|---|---|---|---|---|---|
No. epochs | No. neurons | Batch size | optimizer | Dropout rate | activation | loss | RMSE | MAE | |
Energy Generation |
2000 | 3 | 2 | Adam | 0.2 | tanh | mae | 0.5654 | 0.329 |
Energy Consumption |
2000 | 3 | 3 | SGD | 0.2 | relu | mae | 0.3273 | 0.2410 |
As depicted in Table 4, the developed model has proven the ability to outperform the baseline model as the resultant RMSE is lower than the inferred value from the persistence baseline model for both tasks. Figure 9 and Figure 10 portray the reported loss and validation loss per epoch obtained from training the LSTM models (experiments of the lowest RMSEs) designed for energy generation and consumption forecasting respectively. The loss/validation_loss values were computed on training and validation datasets (summary of the errors inferred by each example in training and validation sets at successive epochs) and the plots show that both models designed for the two tasks are skill on these two datasets.
5.1.3. A comparison with classical time series forecasting algorithms
Another thread of experiments was implemented to compare the deep learning LSTM technique with several classical and linear algorithms. In fact, those models had been implemented initially to scrutiny their performance before conducting any further experiments. This is due to the claim that the classical time series-based approach might outweigh other sophisticated statistical deep learning approaches [91]. However, as depicted in Table 5, the experiment on more than ten classical and linear-based algorithms to forecast both energy generation and consumption did not report high performance in terms of RMSE metric and did not outweigh the LSTM deep learning model for both tasks.
Table 5.
Linear/Classical Algorithm | RMSE |
MAE |
||
---|---|---|---|---|
EG | EC | EG | EC | |
ARIMA | 4.256 | 4.539 | 3.325 | 3.482 |
VAR | 6.452 | 5.983 | 4.254 | 4.198 |
LinearRegression | 2.110 | 2.082 | 1.568 | 1.568 |
Lasso | 2.045 | 2.023 | 1.770 | 1.770 |
Ridge | 2.109 | 2.082 | 1.569 | 1.568 |
ElasticNet | 2.042 | 2.019 | 1.765 | 1.765 |
HuberRegressor | 2.500 | 2.455 | 1.743 | 1.742 |
Lars | 2.110 | 2.082 | 1.568 | 1.568 |
LassoLars | 2.045 | 2.023 | 1.771 | 1.770 |
PassiveAggressiveRegressor | 6.133 | 6.088 | 2.41 | 2.409 |
RANSACRegressor | 2.464 | 2.421 | 1.391 | 1.389 |
SGDRegressor | 2.103 | 2.137 | 1.612 | 1.592 |
5.2. Forecasting results using RapidMiner™
The experiment on RapidMiner™ was carried out based on the methodology discussed in section 4.3 and the approach provided. Table 6 presents a summary of the key parameters and their settings for the experiment conducted using RapidMiner platform. These settings comprise values inferred by RapidMiner using their auto model to fit the designated regression problem.
Table 6.
Parameter | Description | Value |
---|---|---|
Windowing | ||
Window size | number of values per window | 20 |
Step size | Size between the first values of two successive windows | 1 |
Horizon size | The number of values taken as the horizon (i.e. time points). | 24 |
Machine Learning Models | ||
Gineralised Linear Model (GLM) | ||
Solver | For optimisation | IRLSM |
maximum number of threads | Controls parallelism level of building model | 1 |
number of lambdas | Controls the amount of applied regularization | 30 |
Artificial Neoral Network (ANN) | ||
Activation function | Function used by neurons in the hidden layers | Rectifier |
No. of hidden layer | Number of hidden layers in the model | 2 |
No. of neurons per layer | Size of each hidden layer | 50 |
Epochs | Iteration times over dataset | 10 |
L1 | Regularization (absolute value of the weights) | 1.0E-5 |
L2 | Regularization (sum of the squared weights) | 0.0 |
Loss function | loss (error) function | Auto |
Random Forest Tree (RFT) | ||
No. Trees | Number of random generated trees | 20 |
Criterion | On which attribute will be split | least_square |
Max_depth | Depth of the tree | 7 |
Gradient Boosted Tree (GBT) | ||
No. Trees | Number of generated trees | 150 |
Decision Tree (DT) | ||
Criterion | On which attribute will be split | least_square |
Max_depth | Depth of the tree | 15 |
Support Victor Machine (SVM) | ||
Kernel Type | Kernel Function used in the model | Radial |
Kernel gamma | SVM kernel parameter gamma | 1.0000000000000007 |
kernel cache | size of the cache for kernel evaluation (MB) | 200 |
C | SVM complexity constant | 1000 |
RapidMiner computes the performance based on 20% of the unseen dataset. This dataset is embedded for a multi-hold-out-set validation where average performance is computed for seven dissimilar subsets. Table 7 shows the performance metrics of six different machine learning algorithms. As depicted in the table, Gradient Boosted Tree obtains the best performance in both evaluation metrics amongst all techniques to forecast both power consumption and power generation. This is commonly due to the strong scalability and regularization quality of the GLM amongst other techniques [92, 93].
Table 7.
GLM |
ANN |
RF |
GBT |
DT |
SVM |
|||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|
EG | EC | EG | EC | EG | EC | EG | EC | EG | EC | EG | EC | |
RMSE | 10.253 | 9.76 | 12.742 | 7.353 | 10.25 | 8.254 | 8.275 | 4.342 | 12.291 | 7.976 | 14.383 | 6.164 |
MAE | 8.796 | 7.211 | 8.687 | 5.331 | 8.69 | 6.011 | 4.214 | 2.918 | 5.824 | 5.544 | 8.298 | 4.258 |
Figure 11 and Figure 12 portray the predictions plots of the six different regression algorithms tested using RapidMiner on energy generation and energy consumption respectively. These charts show the predictions vs. the actual values for the validation samples. It is evident that Random Forest Tree performance was poor on both tasks, this is due to the fact that RF commonly forecasts an average of unseen training samples since it is not able to extrapolate the normal increasing/decreasing trend in the dataset, thereby does not scale well for time series data.
5.3. Aggregated experimental results
Previous sections present and discuss the experimental results that are implemented using the baseline model, optimised LSTM, classical models, and RapidMiner embedded machine learning models. To obtain a holistic view, we aggregate the resultant values of the evaluation metrics obtained by all models in two figures, thereby providing better insights. Figure 13 demonstrates the aggregated RMSE values obtained by all models in forecasting both energy consumption and generation experiments, while Figure 14 shows the MAE values obtained in the same designated experiments. These figures verify again the superiority of the optimised LSTM and its utility to provide energy consumption and generation predictions with minimum errors. The next section furnishes a further discussion on these empirical results and cast more light on the significance of the study.
6. Discussion, limitations, and future work
In the energy sector, embracing distributed renewable resources is recognised as a potential solution to the climate change problem [94]. Therefore, the energy sector is presently shifting promptly to accommodate sustainable energy transitions driven by forces of technological innovation and advances of digitalisation platforms. Opportunities for distributed energy markets are tremendous and being trialled around the world including pilot projects are being trialled across the world to explore the potential of P2P energy trading, including the RENeW Nexus project in the city of Fremantle, Western Australia [5]. However, to increase operation efficiency (including energy consumption and generation scheduling) as well as make the most efficient energy trading between prosumers and consumers it is imperative to build intelligent systems that are able to estimate energy usage and production in such a decentralised environment. This study attempts to address this issue by developing a correlated array of various classical and advanced AI-based techniques to forecast both energy generation and consumption. The empirical results have concluded with important outcomes that can be summarised as follows.
The first observation is the superiority of the optimised LSTM model in the designated experiments. The abstract structure of the LSTM model enables capturing strong periodicity due to its capacity to maintain the temporal correlation utilising the memory block that is embedded in the recurrent layer [95]. Hence, contextual information represented by the temporal or spatial dependency of data can be attained and well-presented [96]. It is also important to highlight the significance of the hyperparameter optimization process that has been carried out on LSTM to infer the best set of settings so as to fit well the incorporated dataset. The second observation is the poor performance of the autoML tool, namely RapidMiner, and its embedded ML on the designated tasks in comparison to the optimised LSTM as well as the classical models. Figure 11 and Figure 12 provide an evident viewpoint on this inadequacy of RapidMiner's built-in settings to tackle the forecasting problem. Also, this study points to the importance to carry out data preprocessing and preparation to enhance data quality prior to data analytics. This is through data normalisation, reshaping, and scaling as well as ensuring stationarity of data.
This paper attempts to provide an efficient forecasting platform to predict both renewable energy generation and energy consumption in households. However, this study exhibits certain limitations that we hope to address in the future. For example, despite the applicability of LSTM to tackle the designated problem due to its capacity to handle sequence pattern information, LSTM utilises only the attributes given in the training dataset with inadequacy to capture patterns that appear in local and global trends of the time series. Extracting patterns of the data can be attained by implementing a Convolutional Neural Network (CNN) model [81]. Therefore, a promising avenue for future research is to develop a hybrid model incorporating both LSTM and CNN that can enhance forecasting accuracy. Also, bagging and other ensemble machine learning strategies will be investigated. In the same context, other advanced hyperparameter and feature selection strategies including bayesian optimization, random optimization, and evolutionary optimization will be examined. This study utilises one-time step as a forecasting approach. Strategies for multi-step time series forecasting will be scrutinized in future research using large-scale datasets to provide further flexibility, thus forecasting multiple time steps instead of one single time step. Finally, more features/attributes will be incorporated, thereby providing a comprehensive multivariate forecasting approach, for example, by examining behavioral characteristics of households’ residents in terms of their electricity consumption and generation.
7. Conclusion
This study presents several conducted experiments to forecast hourly-based energy consumption and generation. Real-time data were collected from smart meters that were installed in residential premises and used to evaluate the efficacy and effectiveness of various well-known classical (statistical) and AI-based models that are used to predict energy consumption and production. This paper furnishes the following contributions (i) A two-fold analytical model to forecast energy consumption and generation in P2P settings is designed and implemented; (ii) an auto-ML model and various classical and advanced AI-based techniques are incorporated, and their utility is evaluated and compared comprehensively; and (iii) to the best of our knowledge, this is the first paper that furnishes a forecasting model for both energy consumption and generation based on real data captured from a P2P grid system in the state of Western Australia.
Declarations
Author contribution statement
Bilal Abu-Salih: Conceived and designed the experiments; Performed the experiments; Analyzed and interpreted the data; Contributed reagents, materials, analysis tools or data; Wrote the paper.
Pornpit Wongthongtham: Analyzed and interpreted the data; Contributed reagents, materials, analysis tools or data; Wrote the paper.
Greg Morrison and Kevin Coutinho: Contributed reagents, materials, analysis tools or data; Wrote the paper.
Manaf Al-Okaily and Ammar Huneiti: Analyzed and interpreted the data; Wrote the paper.
Funding statement
This research did not receive any specific grant from funding agencies in the public, commercial, or not-for-profit sectors.
Data availability statement
Data will be made available on request.
Declaration of interests statement
The authors declare no conflict of interest.
Additional information
No additional information is available for this paper.
Acknowledgements
N/A.
Footnotes
References
- 1.Santamouris M. Minimizing Energy Consumption, Energy Poverty and Global and Local Change in the Built Environment. Innovating to Zero; 2019. Energy consumption and environmental quality of the building sector. [Google Scholar]
- 2.Council G.W.E. Global Wind Energy Council; Brussels, Belgium: 2021. GWEC| Global Wind Report 2021. [Google Scholar]
- 3.Bilgili M., Yildirim A., Ozbek A., Celebi K., Ekinci F. Long short-term memory (LSTM) neural network and adaptive neuro-fuzzy inference system (ANFIS) approach in modeling renewable electricity generation forecasting. Int. J. Green Energy. 2021;18(6):578–594. [Google Scholar]
- 4.Wongthongtham P., Marrable D., Abu-Salih B., Liu X., Morrison G. Blockchain-enabled Peer-to-Peer energy trading. Comput. Electr. Eng. 2021;94:107299. [Google Scholar]
- 5.Ledger P. Australian government awards grant to $8 million project in the city of Fremantle using the power ledger platform. https://medium.com/power-ledger/australian-government-awards-grant-to-8-million-project-in-the-city-of-fremantle-using-the-power-2dbadfad50ae
- 6.IRENA . 2020. PEER-TO-PEER ELECTRICITY TRADING.https://irena.org/-/media/Files/IRENA/Agency/Publication/2020/Jul/IRENA_Peer-to-peer_trading_2020.pd Accessed: 13/02/2021. [Online]. Available: [Google Scholar]
- 7.Bogensperger A., Ferstl J., Yu Y. 2021. Comparison of Pricing Mechanisms in Peer-To-Peer Energy Communities. [Google Scholar]
- 8.Ahmad T., Zhang H., Yan B. A review on renewable energy and electricity requirement forecasting models for smart grid and buildings. Sustain. Cities Soc. 2020;55:102052. [Google Scholar]
- 9.Zhang C., Wu J., Zhou Y., Cheng M., Long C. Peer-to-Peer energy trading in a Microgrid. Appl. Energy. 2018;220:1–12. [Google Scholar]
- 10.Ertugrul Ö.F. Forecasting electricity load by a novel recurrent extreme learning machines approach. Int. J. Electr. Power Energy Syst. 2016;78:429–435. [Google Scholar]
- 11.Ahmadi S., Bevrani H., Jannaty H. 2012 Second Iranian Conference on Renewable Energy and Distributed Generation. IEEE; 2012. A fuzzy inference model for short-term load forecasting; pp. 39–44. [Google Scholar]
- 12.Amjady N. Short-term hourly load forecasting using time-series modeling with peak load estimation capability. IEEE Trans. Power Syst. 2001;16(3):498–505. [Google Scholar]
- 13.Ceperic E., Ceperic V., Baric A. A strategy for short-term load forecasting by support vector regression machines. IEEE Trans. Power Syst. 2013;28(4):4356–4364. [Google Scholar]
- 14.Said D. Intelligent photovoltaic power forecasting methods for a sustainable electricity market of smart micro-grid. IEEE Commun. Mag. 2021;59(7):122–128. [Google Scholar]
- 15.Barman M., Choudhury N.D., Sutradhar S. A regional hybrid Goa-SVM model based on similar day approach for short-term load forecasting in Assam, India. Energy. 2018;145:710–720. [Google Scholar]
- 16.Elsinga B., van Sark W.G. Short-term peer-to-peer solar forecasting in a network of photovoltaic systems. Appl. Energy. 2017;206:1464–1483. [Google Scholar]
- 17.Høverstad B.A., Tidemann A., Langseth H., Öztürk P. Short-term load forecasting with seasonal decomposition using evolution for parameter tuning. IEEE Trans. Smart Grid. 2015;6(4):1904–1913. [Google Scholar]
- 18.Costello K.W., Hemphill R.C. Electric utilities’‘death spiral’: hyperbole or reality? Electr. J. 2014;27(10):7–26. [Google Scholar]
- 19.Shawabkeh A., Faris H., Aljarah I., Abu-Salih B., Alboaneen D., Alhindawi N. An evolutionary-based random weight networks with taguchi method for Arabic web pages classification. Arabian J. Sci. Eng. 2021;46(4):3955–3980. [Google Scholar]
- 20.Abu-Salih B., et al. Social Big Data Analytics: Practices, Techniques, and Applications. 2021. Predictive analytics using Social Big Data and machine learning; pp. 113–143. [Google Scholar]
- 21.Abu-Salih B., Wongthongtham P., Zhu D., Chan K.Y., Rudra A. Springer Nature; 2021. Social Big Data Analytics: Practices, Techniques, and Applications. [Google Scholar]
- 22.Abu-Salih B., et al. Time-aware domain-based social influence prediction. J. Big Data. 2020;7(1):10. [Google Scholar]
- 23.Abu-Salih B., et al. arXiv preprint arXiv:2201.05203; 2022. An Intelligent System for Multi-Topic Social Spam Detection in Microblogging. [Google Scholar]
- 24.Runge J., Zmeureanu R. Forecasting energy use in buildings using artificial neural networks: a review. Energies. 2019;12(17):3254. [Google Scholar]
- 25.Divina F., García Torres M., Goméz Vela F.A., Vázquez Noguera J.L. A comparative study of time series forecasting methods for short term electric energy consumption prediction in smart buildings. Energies. 2019;12(10):1934. [Google Scholar]
- 26.Ribeiro G.T., Mariani V.C., dos Santos Coelho L. Enhanced ensemble structures using wavelet neural networks applied to short-term load forecasting. Eng. Appl. Artif. Intell. 2019;82:272–281. [Google Scholar]
- 27.Pérez-Chacón R., Asencio-Cortés G., Martínez-Álvarez F., Troncoso A. Information Sciences; 2020. Big Data Time Series Forecasting Based on Pattern Sequence Similarity and its Application to the Electricity Demand. [Google Scholar]
- 28.Hu H., Wang L., Peng L., Zeng Y.-R. Effective energy consumption forecasting using enhanced bagged echo state network. Energy. 2020;193:116778. [Google Scholar]
- 29.Hu H., Wang L., Lv S.-X. Renewable Energy; 2020. Forecasting Energy Consumption and Wind Power Generation Using Deep echo State Network. [Google Scholar]
- 30.Wei N., Li C., Peng X., Zeng F., Lu X. Conventional models and artificial intelligence-based models for energy consumption forecasting: a review. J. Petrol. Sci. Eng. 2019;181:106187. [Google Scholar]
- 31.Wei N., Li C., Peng X., Li Y., Zeng F. Daily natural gas consumption forecasting via the application of a novel hybrid model. Appl. Energy. 2019;250:358–368. [Google Scholar]
- 32.Shao Z., Gao F., Yang S.-L., Yu B.-g. A new semiparametric and EEMD based framework for mid-term electricity demand forecasting in China: hidden characteristic extraction and probability density prediction. Renew. Sustain. Energy Rev. 2015;52:876–889. [Google Scholar]
- 33.He Y., Lin B. Forecasting China's total energy demand and its structure using ADL-MIDAS model. Energy. 2018;151:420–429. [Google Scholar]
- 34.Akpinar M., Yumusak N. 2013 7th International Conference on Application of Information and Communication Technologies. IEEE; 2013. Forecasting household natural gas consumption with ARIMA model: a case study of removing cycle; pp. 1–6. [Google Scholar]
- 35.Hussain A., Rahman M., Memon J.A. Forecasting electricity consumption in Pakistan: the way forward. Energy Pol. 2016;90:73–80. [Google Scholar]
- 36.Harris T.M., Devkota J.P., Khanna V., Eranki P.L., Landis A.E. Logistic growth curve modeling of US energy production and consumption. Renew. Sustain. Energy Rev. 2018;96:46–57. [Google Scholar]
- 37.Kovačič M., Šarler B. Genetic programming prediction of the natural gas consumption in a steel plant. Energy. 2014;66:273–284. [Google Scholar]
- 38.Özmen A., Yılmaz Y., Weber G.-W. Natural gas consumption forecast with MARS and CMARS models for residential users. Energy Econ. 2018;70:357–381. [Google Scholar]
- 39.Li K., Hu C., Liu G., Xue W. Building's electricity consumption prediction using optimized artificial neural networks and principal component analysis. Energy Build. 2015;108:106–113. [Google Scholar]
- 40.Chae Y.T., Horesh R., Hwang Y., Lee Y.M. Artificial neural network model for forecasting sub-hourly electricity usage in commercial buildings. Energy Build. 2016;111:184–194. [Google Scholar]
- 41.He W. Load forecasting via deep neural networks. Procedia Comput. Sci. 2017;122:308–314. [Google Scholar]
- 42.Szoplik J. Forecasting of natural gas consumption with artificial neural networks. Energy. 2015;85:208–220. [Google Scholar]
- 43.Kermanshahi B. Recurrent neural network for forecasting next 10 years loads of nine Japanese utilities. Neurocomputing. 1998;23(1-3):125–133. [Google Scholar]
- 44.Wei N., Li C., Duan J., Liu J., Zeng F. Daily Natural gas load forecasting based on a hybrid deep learning model. Energies. 2019;12(2):218. [Google Scholar]
- 45.Marino D.L., Amarasinghe K., Manic M. IECON 2016-42nd Annual Conference of the IEEE Industrial Electronics Society. IEEE; 2016. Building energy load forecasting using deep neural networks; pp. 7046–7051. [Google Scholar]
- 46.Yang H.-T., Huang C.-M., Huang C.-L. Proceedings of Power Industry Computer Applications Conference. IEEE; 1995. Identification of ARMAX model for short term load forecasting: an evolutionary programming approach; pp. 325–330. [Google Scholar]
- 47.Mohamed Z., Bodger P. 2004. Forecasting Electricity Consumption: A Comparison of Models for New Zealand. [Google Scholar]
- 48.Chujai P., Kerdprasop N., Kerdprasop K. Time series analysis of household electric consumption with ARIMA and ARMA models. Proc. Int. Multi Conf. Eng. Comput. Scient. 2013;1:295–300. [Google Scholar]
- 49.Chakhchoukh Y., Panciatici P., Mili L. Electric load forecasting based on statistical robust methods. IEEE Trans. Power Syst. 2010;26(3):982–991. [Google Scholar]
- 50.Charytoniuk W., Chen M., Van Olinda P. Nonparametric regression based short-term load forecasting. IEEE Trans. Power Syst. 1998;13(3):725–730. [Google Scholar]
- 51.Ko C.-N., Lee C.-M. Short-term load forecasting using SVR (support vector regression)-based radial basis function neural network with dual extended Kalman filter. Energy. 2013;49:413–422. [Google Scholar]
- 52.Sevlian R., Rajagopal R. arXiv preprint arXiv:1404.0058; 2014. Short Term Electricity Load Forecasting on Varying Levels of Aggregation. [Google Scholar]
- 53.Aung Z., Toukhy M., Williams J., Sanchez A., Herrero S. Towards accurate electricity load forecasting in smart grids. Proc. DBKDA. 2012:51–57. [Google Scholar]
- 54.Chen B.-J., Chang M.-W. Load forecasting using support vector machines: a study on EUNITE competition 2001. IEEE Trans. Power Syst. 2004;19(4):1821–1830. [Google Scholar]
- 55.Atique S., Noureen S., Roy V., Subburaj V., Bayne S., Macfie J. 2019 IEEE 9th Annual Computing and Communication Workshop and Conference (CCWC) IEEE; 2019. Forecasting of total daily solar energy generation using ARIMA: a case study; pp. 114–119. [Google Scholar]
- 56.Ayub N., Javaid N., Mujeeb S., Zahid M., Khan W.Z., Khattak M.U. International Conference on Advanced Information Networking and Applications. Springer; 2019. Electricity load forecasting in smart grids using support vector machine; pp. 1–13. [Google Scholar]
- 57.Melzi F.N., Touati T., Same A., Oukhellou L. 2016 15th IEEE International Conference on Machine Learning and Applications (ICMLA) IEEE; 2016. Hourly solar irradiance forecasting based on machine learning models; pp. 441–446. [Google Scholar]
- 58.Li J., Ward J.K., Tong J., Collins L., Platt G. Machine learning for solar irradiance forecasting of photovoltaic system. Renew. Energy. 2016;90:542–553. [Google Scholar]
- 59.Qiao W., Li Z., Liu W., Liu E. Fastest-growing source prediction of US electricity production based on a novel hybrid model using wavelet transform. Int. J. Energy Res. 2021 [Google Scholar]
- 60.Moreira M., Balestrassi P., Paiva A., Ribeiro P., Bonatto B. Design of experiments using artificial neural network ensemble for photovoltaic generation forecasting. Renew. Sustain. Energy Rev. 2021;135:110450. [Google Scholar]
- 61.Lee J., Cho Y. National-scale electricity peak load forecasting: traditional, machine learning, or hybrid model? Energy. 2022;239:122366. [Google Scholar]
- 62.Pappas S.S., et al. Electricity demand load forecasting of the Hellenic power system using an ARMA model. Elec. Power Syst. Res. 2010;80(3):256–264. [Google Scholar]
- 63.Zhu S., Wang J., Zhao W., Wang J. A seasonal hybrid procedure for electricity demand forecasting in China. Appl. Energy. 2011;88(11):3807–3815. [Google Scholar]
- 64.Nadimi R., Tokimatsu K. Analyzing of renewable and non-renewable energy consumption via bayesian inference. Energy Proc. 2017;142:2773–2778. [Google Scholar]
- 65.Chen K., Jiang J., Zheng F., Chen K. A novel data-driven approach for residential electricity consumption prediction based on ensemble learning. Energy. 2018;150:49–60. [Google Scholar]
- 66.Bianco V., Manca O., Nardini S., Minea A.A. Analysis and forecasting of nonresidential electricity consumption in Romania. Appl. Energy. 2010;87(11):3584–3590. [Google Scholar]
- 67.Su H.-Y., Liu T.-Y., Hong H.-H. Adaptive residual compensation ensemble models for improving solar energy generation forecasting. IEEE Trans. Sustain. Energy. 2019;11(2):1103–1105. [Google Scholar]
- 68.Zhao H.-x., Magoulès F. A review on the prediction of building energy consumption. Renew. Sustain. Energy Rev. 2012;16(6):3586–3592. [Google Scholar]
- 69.Barker J., Gajewar A., Golyaev K., Bansal G., Conners M. Thirty-Second AAAI Conference on Artificial Intelligence. 2018. Secure and automated enterprise revenue forecasting. [Google Scholar]
- 70.Beard E., et al. Understanding and using time series analyses in addiction research. Addiction. 2019 doi: 10.1111/add.14643. [DOI] [PubMed] [Google Scholar]
- 71.Tylkowski J., Hojan M. Time decomposition and short-term forecasting of hydrometeorological conditions in the south baltic coastal zone of Poland. Geosciences. 2019;9(2):68. [Google Scholar]
- 72.Hyndman R.J., Athanasopoulos G. OTexts; 2018. Forecasting: Principles and Practice. [Google Scholar]
- 73.Cui H., Wu R., Zhao T. Sustainable development study on an energy-economic-environment system based on a vector autoregression model in shanxi, China. Pol. J. Environ. Stud. 2019;28(3):1623–1635. [Google Scholar]
- 74.Hochreiter S., Schmidhuber J. "Long short-term memory". Neural Comput. Nov 15 1997;9(8):1735–1780. doi: 10.1162/neco.1997.9.8.1735. [DOI] [PubMed] [Google Scholar]
- 75.Greff K., Srivastava R.K., Koutnik J., Steunebrink B.R., Schmidhuber J. LSTM: a search space Odyssey. IEEE Transact. Neural Networks Learn. Syst. Oct 2017;28(10):2222–2232. doi: 10.1109/TNNLS.2016.2582924. [DOI] [PubMed] [Google Scholar]
- 76.Gers F.A., Schmidhuber J., Cummins F. Continual prediction with LSTM; 1999. Learning to Forget. [DOI] [PubMed] [Google Scholar]
- 77.Cho K., et al. arXiv preprint arXiv:1406.1078; 2014. Learning Phrase Representations Using RNN Encoder-Decoder for Statistical Machine Translation. [Google Scholar]
- 78.Ming Y., et al. 2017 IEEE Conference on Visual Analytics Science and Technology (VAST) IEEE; 2017. Understanding hidden memories of recurrent neural networks; pp. 13–24. [Google Scholar]
- 79.Thomas A. Keras lstm tutorial - how to easily build a powerful deep learning language model. http://adventuresinmachinelearning.com/
- 80.Bouktif S., Fiaz A., Ouni A., Serhani M.A. Multi-sequence LSTM-RNN deep learning and metaheuristics for electric load forecasting. Energies. 2020;13(2):391. [Google Scholar]
- 81.Alhussein M., Aurangzeb K., Haider S.I. Hybrid CNN-LSTM model for short-term individual household load forecasting. IEEE Access. 2020;8:180544–180557. [Google Scholar]
- 82.Wang J.Q., Du Y., Wang J. LSTM based long-term energy consumption prediction with periodicity. Energy. 2020;197:117197. [Google Scholar]
- 83.Zang H., Liu L., Sun L., Cheng L., Wei Z., Sun G. Short-term global horizontal irradiance forecasting based on a hybrid CNN-LSTM model with spatiotemporal correlations. Renew. Energy. 2020;160:26–41. [Google Scholar]
- 84.Zheng T., Liu S., Chen Z., Qiao Y., Law R. Forecasting daily room rates on the basis of an LSTM model in difficult times of Hong Kong: evidence from online distribution channels on the hotel industry. Sustainability. 2020;12(18):7334. [Google Scholar]
- 85.Banerjee T., Sinha S., Choudhury P. Applied Intelligence; 2022. Long Term and Short Term Forecasting of Horticultural Produce Based on the LSTM Network Model; pp. 1–31. [Google Scholar]
- 86.Idoine C., Krensky P., Brethenoux E., Hare J., Sicular S., Vashisth S. 2018. Energy consumption and environmental quality of the building sector. Magic Quadrant for data science and machine-learning platforms. (Accessed 23 June 2021) [Google Scholar]
- 87.Marinescu A., Harris C., Dusparic I., Clarke S., Cahill V. 2013 2nd International Workshop on Software Engineering Challenges for the Smart Grid (SE4SG) IEEE; 2013. Residential electrical demand forecasting in very small scale: an evaluation of forecasting methods; pp. 25–32. [Google Scholar]
- 88.Brownlee J. 2018. How to Convert a Time Series to a Supervised Learning Problem in Python. [Google Scholar]
- 89.Kuhn M., Johnson K. Springer; 2013. Applied Predictive Modeling. [Google Scholar]
- 90.Russell S.J., Norvig P. Pearson Education Limited; Malaysia: 2016. Artificial Intelligence: a Modern Approach. [Google Scholar]
- 91.Makridakis S., Spiliotis E., Assimakopoulos V. Statistical and Machine Learning forecasting methods: concerns and ways forward. PLoS One. 2018;13(3):e0194889. doi: 10.1371/journal.pone.0194889. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 92.Chen T., Guestrin C. Proceedings of the 22nd Acm Sigkdd International Conference on Knowledge Discovery and Data Mining. ACM; 2016. Xgboost: a scalable tree boosting system; pp. 785–794. [Google Scholar]
- 93.Rokach L. Decision forest: twenty years of research. Inf. Fusion. 2016;27:111–125. [Google Scholar]
- 94.Halkos G.E., Gkampoura E.-C. Reviewing usage, potentials, and limitations of renewable energy sources. Energies. 2020;13(11):2906. [Google Scholar]
- 95.Fu R., Zhang Z., Li L. 2016 31st Youth Academic Annual Conference of Chinese Association of Automation (YAC) IEEE; 2016. Using LSTM and GRU neural network methods for traffic flow prediction; pp. 324–328. [Google Scholar]
- 96.Srivastava S., Lessmann S. A comparative study of LSTM neural networks in forecasting day-ahead global horizontal irradiance with satellite data. Sol. Energy. 2018;162:232–247. [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Data Availability Statement
Data will be made available on request.