Skip to main content
PLOS ONE logoLink to PLOS ONE
. 2024 Dec 20;19(12):e0314391. doi: 10.1371/journal.pone.0314391

Improving prediction of solar radiation using Cheetah Optimizer and Random Forest

Ibrahim Al-Shourbaji 1,2, Pramod H Kachare 3, Abdoh Jabbari 1, Raimund Kirner 2,*, Digambar Puri 3, Mostafa Mehanawi 1, Abdalla Alameen 4
Editor: Mohamad Abou Houran5
PMCID: PMC11661600  PMID: 39705221

Abstract

In the contemporary context of a burgeoning energy crisis, the accurate and dependable prediction of Solar Radiation (SR) has emerged as an indispensable component within thermal systems to facilitate renewable energy generation. Machine Learning (ML) models have gained widespread recognition for their precision and computational efficiency in addressing SR prediction challenges. Consequently, this paper introduces an innovative SR prediction model, denoted as the Cheetah Optimizer-Random Forest (CO-RF) model. The CO component plays a pivotal role in selecting the most informative features for hourly SR forecasting, subsequently serving as inputs to the RF model. The efficacy of the developed CO-RF model is rigorously assessed using two publicly available SR datasets. Evaluation metrics encompassing Mean Absolute Error (MAE), Mean Squared Error (MSE), and coefficient of determination (R2) are employed to validate its performance. Quantitative analysis demonstrates that the CO-RF model surpasses other techniques, Logistic Regression (LR), Support Vector Machine (SVM), Artificial Neural Network, and standalone Random Forest (RF), both in the training and testing phases of SR prediction. The proposed CO-RF model outperforms others, achieving a low MAE of 0.0365, MSE of 0.0074, and an R2 of 0.9251 on the first dataset, and an MAE of 0.0469, MSE of 0.0032, and R2 of 0.9868 on the second dataset, demonstrating significant error reduction.

Introduction

Over the decade, industrial modernization has constantly increased the energy demand of our economy, while the energy supply amount is limited. The continually diminishing conventional resources like fossil fuels, petroleum, coal, and natural gas have further increased the demand-supply gap, commonly known as the energy crisis. In search of green and sustainable energy, efforts are encouraged to use renewable resources like solar, wind, water, biomass, and geothermal. These renewable resources reduce environmental pollution and stress on conventional resources [14].

Solar Radiation (SR) is a readily and abundantly available renewable resource in most parts of the world. In recent years, SR prediction has received significant attention in academia [58]. Several ML models have been reported to predict SR [9]. Compared support vector machine (SVM) and Extreme Gradient Boosting (XGBoost) for SR prediction. Comparative evaluation using prediction accuracy, stability, and computational time showed that XGBoost outperformed SVM for estimating SR. In [10] work, the authors assessed the ability of kernel-based nonlinear extension of Arps decline and Gradient Boosting (GB) models with categorical features to support daily global SR prediction. Both models were reported to be suitable for SR. In another work [11], bagging and boosting ensemble models improve the performance of SVM, decision trees, and artificial neaural network (ANN) models for hourly SR prediction. The results demonstrated that decision trees improved SR prediction than the other used models.

RF technique is also widely applied for energy systems and SR prediction [1215], used Firefly (FF) algorithm to choose the best number of leaves per tree for the RF. The outcomes showed that the combined model outperformed the conventional Random Forest (RF), and ANN. Also, optimizing the ANN regarding the accuracy and execution speed also enhances the overall performance [16, 17]. In another work [18], Particle Swarm Optimization (PSO) was utilized to optimize RF parameters for SR prediction. The results showed that PSO-RF achieved better SR prediction than the decision trees, RF, and ANN. In [19] work, RF, random tree, reduced error pruning trees, and hybrid random tree reduce model are compared for daily SR prediction using four locations located in Burkina Faso. The results confirmed that RF is a promising approach for predicting SR accurately. In another comparison study [20], RF and three other tree-based machine learning (ML) models for SR prediction in India are compared. The results indicated that RF outperformed the other models used in their study. Benali et al. [21], smart persistence, ANN, and RF are compared to predict global horizontal, diffuse horizontal, and beam normal components and SR in France. The results showed that RF predicted all components more effectively than other methods. In another recent work, [22], the authors compared seven ML models for SR prediction. The obtained results showed that RF performed better than the other compared methods.

A hybrid ML technique is proposed for hour global SR prediction [23]. The developed approach relied on a Convolutional Neural Network, Non-parametric Gaussian Process Regression (GPR), Least square SVM (LS-SVM), and Extreme Learning Machine (ELM). Experimental results showed that the proposed approach attained better accuracy than other popular ML models. In another work, Kuhe et al. [24], the authors attempt to improve the prediction accuracy of SR using ANN’s ensemble. They used Makurdi city in Nigeria data as a case study to evaluate the model. The results showed that the proposed model achieves great accuracy. Pang et al. [25], showed that the Recurrent Neural Network (RNN) can achieve higher accuracy for SR than the ANN model. In [26] work, a system to predict SR using Neural Networks is proposed. The authors used meteorological data from five different cities in Bangladesh. They used RNN, long short term memory (LSTM), and Gated Recurrent Unit (GRU) models to train the meteorological data. Among the used models, the GRU performed the best. Geshnigani et al. [27], improved conventional Multiple Llinear Regression, Multilayer Perceptron Neural Network, and Adaptive Neuro-Fuzzy Inference System (ANFIS) using optimization algorithms. They confirmed that the ANFIS model is effectively enhanced when optimization methods are used. Although deep learning-based models have shown success in a variety of areas over the past few years, these models require large datasets or need optimization of model parameters and latency to achieve generalization [2830].

The metaheuristic models have shown good performance in variation applications [31, 32]. Wu et al. [33], used Ant Colony Optimization (ACO), Cuckoo Search (CS) and Grey Wolf Optimization (GWO) algorithms to optimize the SVM model for SR prediction. Among the hybrid models, GWO-SVM had the best results. In the work of Goliatt and Yaseen [34], a computation intelligent model based on the hybridization of Covariance Matrix Adaptive Evolution Strategies (CMAES) with XGBoost and Multi-Adaptive Regression Splines (MARS) models for building robust predictive models for daily SR prediction is developed. Results showed that the developed approach improves prediction accuracy. Gupta et al. [35], merged Variance Inflation Factor with Mutual Information (VIFMI) as a feature selection method and then the most important features are used as inputs to a Stack Ensemble Extra Tress (SEET) model for estimating SR. Results showed that the developed ensemble model effectively reduced prediction error.

Due to the rapid deployment of solar energy techniques, SR has become common in recent years, and several ML models are widely used to predict SR. However, ML techniques can encounter challenges regarding input data, which can affect the efficacy of these models Banadkooki et al. [36], choosing relevant features is critical for building a learning model, as the input feature set influences it. This plays a pivotal role in maintaining high classification accuracy and reducing the dimensionality of datasets. According to the author’s best knowledge, well-known methods like PSO, ACO, GWO and FF perform well either in local or global searches, and they have poor performance in balancing between local and global searches. CO is a new population-based algorithm introduced in 2022 by Akbari [37] as a powerful method for mimicking specific cheetahs’ hunting strategies. CO shows a better balance between local and global search ability. Hence, in the present work. A new hybridization approach has been developed for more reliable and accurate SR predictions. For this purpose, the hybrid of CO and RF (CO-RF) is employed for predicting SR. The key contributions of this paper can be summarised as follows:

  • Proposing an efficient prediction model, CO-RF, for SR by integrating optimization algorithm rather than the use of existing ML models alone.

  • Evaluating the developed CO-RF model on two publicly available datasets and three evaluation metrics: mean absolute error (MAE), mean square error (MSE), and R2.

  • Investigating CO-RF’s effectiveness in performing SR prediction during training and testing phases and comparing its efficiency to other models before and after using optimization algorithms.

The rest of this paper is structured as follows: Section 2 describes the datasets and methods used and the developed CO-RF model for SR prediction. The statistical metrics to evaluate the ML models are presented in section 3. Comparative analysis of the models is presented in section 4, and section 5 concludes the paper.

Materials and methods

Worldwide, solar energy is an increasingly important renewable energy source. Accurate prediction of SR is crucial for the efficient design and operation of solar power systems. A publicly available SR dataset for measurements is necessary to address this need. In this work, two such datasets are used for experimental evaluations.

Dataset-1

HI-SEAS weather station collected the SR dataset and made it accessible to the scientific community as part of the NASA hackathon task NASA (2023) [38]. The dataset encompasses various meteorological attributes, enabling researchers to examine the intricate relationships between various environmental factors and SR. It comprises eleven distinct features that provide a comprehensive insight into the environmental conditions that influence SR patterns. In preparation for analysis, the UNIX time, a timestamp format, is transformed to facilitate the sorting and organization of recorded measurements. Two critical columns are derived from UNIX time: ‘Month’ and ‘Year,’ enabling temporal analysis of SR trends. A ‘Sun availability’ feature is generated as a difference between sunrise and sunset times, shedding light on the temporal dynamics of sunlight exposure. SR values are scaled to kilo-watts per square meter (kW/m²) to mitigate the impact of large variations in SR values, ultimately enhancing the performance of ML models. The processed dataset comprises 32,686 records and nine distinct features that encapsulate vital meteorological and temporal parameters. A comprehensive statistical analysis of the dataset is given in Table 1.

Table 1. Statistical summary of dataset1.

Feature Unit Mean SD Min 25% 50% 75% Max
Recording time UNIX time 1.48E+9 3.00E+6 1.47E+9 1.47E+9 1.48E+9 1.48E+9 1.48E+9
Temperature Fahrenheit 51.103 6.201 34.000 46.000 50.000 55.000 71.000
Pressure Hg 30.423 0.055 30.190 30.400 30.430 30.460 30.560
Humidity % 75.016 25.990 8.000 56.000 85.000 97.000 103.000
Wind direction Degrees 143.490 83.168 0.090 82.228 147.700 179.310 359.950
Wind speed m/hr 6.244 3.490 0.000 3.370 5.620 7.870 40.500
Hour 24 hr 11.557 6.912 0.000 6.000 12.000 18.000 23.000
Month - 10.489 1.235 1.000 10.000 11.000 11.000 12.000
Sun availability 0-12 hr 11.348 0.476 11.000 11.000 11.000 12.000 12.000

Dataset-2

This dataset is collected and maintained by the King Abdullah City for Atomic and Renewable Energy in Saudi Arabia, provided by the OpenData platform, a government-based repository for open data, for all experimental evaluations [39]. The dataset comprises 1265 records, each characterized by 26 distinct features. These features encapsulate various aspects of solar power generation and meteorological conditions. The data is collected from 41 solar power facilities in Saudi Arabia, as described in Table 2.

Table 2. Distribution of solar radiation records across Saudi Arabia.

Location No. of records Location No. of records
Al-Aflaaj Technical Institute 37 Najran University 31
Al-Baha University 20 Prince Sattam Bin Abdulaziz University 36
Al-Dawadmi College of Technology 36 Princess Norah University 6
Al-Hanakiyah Technical Institute 35 Qassim University 38
Al-Jouf College of Technology 20 Rania Technical Institute 20
Al-Qunfudhah Technical Institute 35 Royal-Commission of Jubail & Yanbu 20
Al-Uyaynah Research Station 42 Saline Water Conversion Corp., Al-Khafji 20
Al-Wajh Technical Institute 34 Saline Water Conversion Corp., Farasan 23
Arar Technical Institute 20 Saline Water Conversion Corp., Hagl 36
Duba Technical Institute 36 Saline Water Conversion Corp., Jubail 35
Hafar Al-Batin Technical College 34 Saline Water Conversion Corp., Umluj 35
Hail College of Technology 19 Shaqra University 36
Jazan University 21 Sharurah Technical Institute 35
K.A.CARE, Olaya 42 Tabuk University 34
K.A.CARE, City Site 42 Taibah University 17
King Abdulaziz University, Osfan 34 Taif University 38
KAU of Science & Technology 38 Timaa Technical Institute 36
King Fahd University of Petroleum Minerals 38 Umm Al-Qura University 17
King Faisal University 38 University of Dammam 38
King Saud University 21 Wadi Addawasir College of Technology 36
Majmaah University 36 Total 1265

Preliminary data analysis revealed that the feature ‘Wind Speed at 3m (SD) Uncertainty (m/s)’ has missing values for all site locations. Hence, this feature is dropped. Also, four records are dropped due to missing solar radiation values. To enhance the utility of the dataset, several preprocessing steps are undertaken. The average values of the corresponding feature impute the missing values in all remaining records. After processing the missing values, each feature is normalized independently to have zero mean and unity standard deviation. The solar radiation values represented by the Global Horizontal Irradiance (GHI), is normalized within the range [-1, +1]. This normalization accounts for variations in power production among different solar stations, ensuring that the dataset maintains consistency and is suitable for machine-learning applications. The processed dataset has 1262 records and 22 features. The GHI features have a mean value of 0.117 and a Standard Deviation (SD) of 0.492. A comprehensive statistics of the dataset is described in Table 3.

Table 3. Statistical summary of dataset2.

Feature Unit Min 25% 50% 75% Max
Air Temperature Celsius -2.546 -0.798 0.133 0.838 1.810
Air Temperature Uncertainty Celsius -0.040 -0.040 -0.040 -0.040 25.100
Wind Direction at 3m Degree -1.542 -0.948 0.000 1.023 1.447
Wind Direction at 3m Uncertainty Degree -13.497 0.164 0.164 0.164 2.896
Wind Speed at 3m. m/s -3.271 -0.676 -0.028 0.405 5.054
Wind Speed at 3m Uncertainty. m/s -1.549 -1.549 0.678 0.678 0.678
Wind Speed at 3m. m/s -3.469 -0.731 -0.099 0.532 4.955
Diffuse Horizontal Irradiance (DHI) Wh/m2 -1.864 -0.867 -0.055 0.735 2.692
DHI Uncertainty Wh/m2 -1.629 -0.804 -0.118 0.688 8.167
DHI Standard Deviation Wh/m2 -2.396 -0.626 0.000 0.507 4.263
Direct Normal Irradiation (DNI) Wh/m2 -2.900 -0.715 -0.015 0.588 3.089
DNI Uncertainty Wh/m2 -1.767 -0.871 0.139 0.525 6.931
DNI Standard Deviation Wh/m2 -2.659 -0.613 0.000 0.566 3.003
Global Horizontal Irradiance (GHI) Wh/m2 -1.000 -0.272 0.154 0.540 1.000
GHI Uncertainty Wh/m2 -0.729 -0.368 -0.032 0.186 19.933
Standard Deviation GHI Wh/m2 -1.790 -0.725 0.000 0.432 4.114
Peak Wind Speed at 3m. m/s -3.978 -0.731 -0.123 0.486 4.746
Peak Wind Speed at 3m Uncertainty m/s -5.274 0.117 0.117 0.117 5.507
Relative Humidity % -1.494 -0.964 -0.072 0.773 2.269
Relative Humidity Uncertainty % -8.375 -0.049 -0.049 -0.049 20.765
Barometric Pressure hPa -2.657 -0.675 -0.258 1.043 1.431
Barometric Pressure Uncertainty hPa -2.459 -0.783 -0.364 0.892 4.662

Cheetah Optimizer (CO)

The CO is a recent metaheuristic method proposed by [37]. This method mimics the behaviours of a cheetah. In this algorithm. a cheetah starts the hunting process by scanning its environment (defined as search space) to detect probable prey. Depending on the distance between the cheetah and the prey, the cheetah may hold the position and wait for the prey to get closer. The rushing and capturing are two stages of the cheetah’s attack. The hunt may be terminated due to reduced energy limits, fast prey feeling, etc. After termination, the cheetah returns home to rest before the next hunt. The pseudo-code of the CO is provided in [37]. The CO chooses one of the following strategies depending on the prey, their condition, search area and distance of the prey:

  • Search strategy:

    Cheetahs find the probable prey in the environment by scanning the area while sitting/standing or patrolling the area actively. The scanning mode is suitable while walking on the plains for dense and grazing prey. On the other hand, scattered and active prey demand more energy-consuming active mode. The cheetah selects a sequence of these modes during the hunt depending on its own condition, the condition of prey, and area coverage. For ith cheetah in jth arrangement, the new position X(i,j)(t+1) is calculated based on their current position Xit as follows:
    X(i,j)(t+1)=Xit+r(i,j)(Δ-1)·a(i,j)t (1)
    Where, t is the hunting time with maximum limit T, r(i,j)(Δ-1) is the random number, and a(i,j)t is step size.
  • Sit and wait strategy:

    When the prey is in the cheetah’s vision field, any movement may reveal the cheetah’s presence to the prey and allow the prey to escape. To avoid the prey’s escape, the cheetah gets closer to the prey by hiding among the bushes or lying on the ground before the ambush. Therefore, in this mode, the cheetah remains at their position and waits for the prey to come nearer, and this behaviour can be represented as:
    X(i,j)(t+1)=X(i,j)t (2)
  • Attack strategy

    Speed and flexibility are two crucial factors in a cheetah’s attack. The cheetah starts the attack by rushing towards the prey at full speed. When the prey starts to flee as soon as it realizes the cheetah’s attack, the cheetah adjusts the path to intervene in the prey. For the current prey position X(P,j)t in jth arrangement, the cheetah’s position according to this attack strategy is mathematically defined as:
    X(i,j)(t+1)=X(P,j)t+γ(i,j)(Δ-1).b(i,j)t (3)
    where, γ(i,j)(Δ-1) and b(i,j)t are turning and interaction factors of ith cheetah in jth arrangement.

The Cheetah Optimizer operates through the following steps:

  • Initialization: The algorithm starts by initializing parameters and setting hyper-parameters such as learning rates and momentum terms.

  • Gradient Computation: At each iteration, the gradient of the objective function with respect to the parameters is computed.

  • Update Step: The optimizer uses a specific update rule, which is influenced by the computed gradient and hyper-parameters, to adjust the parameters.

  • Convergence Check: The optimizer checks whether the changes in the objective function or parameters are below a certain threshold to determine if convergence has been reached.

  • Parameter Update: If the convergence criteria are not met, the algorithm updates the parameters and repeats the process.

Random Forest (RF)

Random Forest [40] is a machine learning model that builds an ensemble of decision trees, with each tree trained using a subset of randomly selected examples and features. This diversity in data for each tree reduces overfitting and improves robustness [41]. The final prediction is typically the pooling of predictions from trees in the forest. The ensemble approach generalizes well to new data, making RF popular for various ML problems [42]. RF builds trees in parallel and, hence, is less prone to overfitting, leading to faster training times. RF does not require extensive tuning of hyperparameters due to built-in overfitting controlling mechanisms, like random subsampling of records and features. The primary hyperparameters for tuning in RF include the number of trees in the ensemble and the maximum depth of each tree. In practice, it uses deep trees, and the number of trees is selected based on model performance [43].

The proposed system for SR prediction uses CO for selecting optimum features and RF for evaluating performance using these optimized features, as shown in system flow in Fig 1. The hyper-parameters model with the number of trees and a maximum depth determined through cross-validation.

Fig 1. System flow of CO-RF model for SR prediction.

Fig 1

Experiments and results

Experimental setup

All the ML models are implemented using the Python-based Scikit-Learn library, capitalizing on its robust model development and evaluation capabilities. The experimentation environment employed was a Windows 10 operating system on an Intel i7 processor clocked at 3.13 GHz and 64 GB of RAM. Lower fitness values indicate better feature selection, enhancing the SR prediction models. These methods are: Particle Swarm Optimization (PSO) [44], ACO [45], Reptile Search Algorithm (RSA) [46], Snake Optimizer (SO) [47], and CO, [37]. The standard parameters of these methods are selected empirically and are set: Population size = 20, number of iterations = 100, and each of them is executed indecently run ten times mitigate biases towards local minima. The settings of these methods are defined based on their implementations in original works, and they are listed in Table 4.

Table 4. Parameters settings.

Algorithm Parameters
PSO c1 = c2 = 2; wmin = 0.1 and Wmax = 0.9
ACO τ0 = 1; p = 0.95; α = 1.2; β = 0.5.
RSA γ = 0.9; θ = 0.5.
SO c1 = 0.5; c2 = 0.05; c3 = 2
CO n = 6; m = 2.
LR Regression type = Lasso (L1); Regression coefficient = 1.
SVM Regularization = 10; kernel = radial basic function; gamma = 0.01.
ANN Hidden layers = (20, 2); activation = ReLU; solver = Adam, batch_size = 200.
RF No. of Trees = 200; impurity = gini, max_features = 14; min_samples_split = 5.
CO-RF Uses the parameters of the CO and RF

Evaluation metrics

To assess prediction results, three different statistical measures are used for the evaluation of the developed CO-RF and the other ML models. These metrics are selected due to their suitability for SR prediction in previous works [22, 33, 48, 49], and they are defined as follows:

MAE=1Ni=1N|mi-pi| (4)
MSE=1Ni=1N(mi-pi)2 (5)
R2=1-i(mi-pi)2i(mi-m¯)2 (6)

where, N is the number of observations, mi is the ith measured SR value, pi is the ith SR value predicted by the model and μim is the mean of the measured SR values.

Results and discussion

The performance of the models in predicting SR is rigorously evaluated and provided in this section. The average convergence behavior of these algorithms is presented in Fig 2. In Fig 2a, the CO algorithm achieves a minimum convergence fitness of 0.0029. PSO and ACO also demonstrated strong feature selection capabilities with fitness values of 0.0052 and 0.0041, respectively. CO stood out as the top-performing algorithm for dataset-1. These results provide valuable insights for researchers and practitioners looking to optimize feature selection for SR prediction tasks in similar datasets. As shown in Fig 2b, CO converges at a fitness value of 0.0027 for dataset-2. The fitness values represent the quality of feature selection for enhancing SR prediction. SO also demonstrated strong feature selection abilities, yielding a fitness value of 0.0029. ACO and RSA reported the same fitness value of 0.0035. PSO achieved the worst performance with a 0.0040 fitness value. The marginally higher convergence value in dataset-1 may be attributed to the larger number of records than in dataset-2.

Fig 2. Convergence behavior of different optimization algorithms as feature selectors for SR prediction for (a) dataset-1 and (b) dataset-2.

Fig 2

CO reported the smallest fitness values for both datasets, indicating the best performance compared to the other optimization algorithms. Hence, further analysis uses only CO as a feature selection method. CO chooses five salient features from the original nine features in dataset-1 and fourteen out of the original twenty-one features in dataset-2. The selected salient features from both datasets are subsequently employed as inputs for various ML algorithms to assess their capabilities for SR prediction.

The computation time required for either dataset is important to justify the model’s scalability. The overall computation time required by the hybrid model can be divided into three parts: time for feature selection by CO, time for training the ML model using the selected features, and time for performance evaluation. Most of the time required was for feature selection due to the iterative nature of the process. The ML model training time varied over different folds depending upon the number of selected features. The testing performance time had almost insignificant variations across different folds. The overall time for dataset-1 was 30.125 ± 2.536 and for dataset-2 was 10.983 ± 0.843. It was observed that almost 90-93% of the overall time was for feature selection, and approximately 1% of the overall time was for performance evaluation. It must be noted that the feature selection process is required only during the model training. In the testing phase, the most important feature already indicated by the training process can only be calculated further to reduce the solar radiation prediction time than the models using the complete feature set.

Predictive performance

The selected features are assessed using MAE, MSE, and R2 measures. Table 5 compares the predictive performance of the ML models using a complete feature set and with a CO-based selected feature set across both datasets. The models depicted with the prefix ‘CO-’ use features optimized with CO. Each model is validated using 10-fold cross-validation and resulting metric in terms of both mean and SD.

Table 5. Comparative performance of ML models with and without CO for SR prediction.
Dataset Model MAE MSE R 2
Dataset-1 LR 0.1523 ± 0.0037 0.0389 ± 0.0036 0.6162 ± 0.0167
SVM 0.0826 ± 0.0115 0.0226 ± 0.0050 0.8163 ± 0.0565
ANN 0.0593 ± 0.0023 0.0128 ± 0.0015 0.9023 ± 0.0129
RF 0.0412 ± 0.0016 0.0095 ± 0.0007 0.9165 ± 0.0071
CO-LR 0.1475 ± 0.0035 0.0382 ± 0.0018 0.6208 ± 0.0176
CO-SVM 0.0729 ± 0.0118 0.0176 ± 0.0051 0.8337 ± 0.0576
CO-ANN 0.0439 ± 0.0023 0.0106 ± 0.0015 0.9175 ± 0.0126
CO-RF 0.0365 ± 0.0015 0.0074 ± 0.0007 0.9251 ± 0.0073
Dataset-2 LR 0.1105 ± 0.0072 0.0102 ± 0.0056 0.8936 ± 0.0165
SVM 0.0762 ± 0.0069 0.0119 ± 0.0045 0.9510 ± 0.0163
ANN 0.0745 ± 0.0059 0.0103 ± 0.0017 0.9571 ±0.0080
RF 0.0469 ± 0.0031 0.0032 ± 0.0004 0.9868 ± 0.0018
CO-LR 0.0927 ± 0.0069 0.0089 ± 0.0045 0.9093 ± 0.0163
CO-SVM 0.0737 ± 0.0069 0.0119 ± 0.0045 0.9510 ± 0.0163
CO-ANN 0.0745 ± 0.0059 0.0103 ± 0.0017 0.9571 ±0.0080
CO-RF 0.0469 ± 0.0031 0.0032 ± 0.0004 0.9868 ± 0.0018

In the evaluation of dataset-1, a distinct performance variation is observed among SR prediction models. LR exhibited MAE of 0.1523, indicating a reasonable prediction accuracy. However, introducing CO hybridization (CO-LR) slightly improved prediction, and the R2 value increased to 0.6208. On the other hand, SVM demonstrated superior performance than LR, with MAE of 0.0826, a lower MSE of 0.0226, and a higher R2 of 0.8163. Applying CO hybridization to SVM (CO-SVM) enhanced its predictive power, resulting in an R2 of 0.8337. ANN and RF models also exhibited commendable performance. The hybridization of CO (CO-ANN and CO-RF) leads to significant improvements.

In dataset-2, LR achieved an MAE of 0.1105, showing a competitive predictive capability. The CO hybridization (CO-LR) slightly decreased the MAE, and the R2 value elevated to 0.9093. SVM continued to excel on this dataset, with a mean MAE of 0.0762 and an impressive R2 of 0.9510. The application of CO hybridization (CO-SVM) maintained high-performance levels, reaffirming its effectiveness. Additionally, ANN and RF models displayed high performance, and the inclusion of CO (CO-ANN and CO-RF) significantly improved their performance.

Dataset 1 showcased model variability, with RF emerging as a standout performer, notably bolstered by CO hybridization. Similarly, dataset-2 demonstrated consistent accuracy across models, with RF again excelling and CO hybridization reaffirming its efficacy. While LR and ANN models showed competitive results, the introduction of CO had a significant impact. To summarize, the dataset-dependent nature of SR prediction models, with CO hybridization potentially enhancing ML-based approaches. Hence, if performed using CO-optimized features, all further analysis is hybridized with four ML models.

Comparative performance of CO-optimized ML models

Visualizations are employed to provide intuitive insights into the CO-based models’ predictive behavior. Fig 3 provides eight scatter plots for essential insights into the performance of four ML models across training data of both SR datasets. These scatter plots employ measured SR values, sourced from the datasets, on the horizontal axis, while the SR estimated by the ML models is plotted on the vertical axis. The ideal scenario in these scatter plots is a perfectly diagonal line where the estimated SR values align precisely with the measured SR values. Such alignment represents accurate predictions by the models. Consequently, the further the data points deviate from the diagonal line, the less accurate the model’s predictions become.

Fig 3.

Fig 3

Convergence behavior of CO-optimized ML models using training data of dataset-1 (first row, (a)-(d)) and dataset-2 (second row, (e)-(h)): CO-LR ((a) & (e)), CO-SVM ((b) & (f)), CO-ANN ((c) & (g)), and CO-RF ((d) & (h)).

In dataset-1, the SR values are measured in kilowatt-hours per square meter (kWh/m2). CO-LR has a maximum deviation from the ideal diagonal arrangement, indicating the worst performance among all models. CO-SVM exhibits a relatively higher spread of data points, indicating a wider dispersion of predictions. Conversely, CO-ANN shows a more clustered distribution of points, suggesting improved predictive consistency. However, the star performer in this dataset is CO-RF, which demonstrates the tightest clustering of points, signifying highly accurate SR predictions. In dataset-2, the SR values are in watt-hours per square meter (Wh/m2), and all three ML models exhibit commendable performance. The scatter plots for CO-SVM, CO-ANN, and CO-RF showcase strong positive correlations between predicted and actual SR values. In this case, CO-RF marginally outperforms the others, demonstrating a slightly tighter clustering of points, reaffirming its effectiveness in SR prediction even in datasets with different units. Fig 4 provides a set of scatter plots using testing data for different ML models applied to both SR datasets. In dataset-1’s, CO-SVM displays a relatively high spread of data points, suggesting some variability in its predictions. CO-ANN exhibits a more clustered distribution, indicating better predictive consistency. However, CO-RF outperforms both, showcasing the tightest clustering of points, signifying exceptionally accurate SR predictions on the testing data. In dataset-2’s, all three models (CO-SVM, CO-ANN, and CO-RF) continue to demonstrate strong positive correlations between their predictions and the actual SR values. While all models perform well, CO-RF exhibits a slightly tighter clustering of data points, reaffirming its effectiveness in SR prediction on the testing data.

Fig 4.

Fig 4

Convergence behavior of CO-optimized ML models using testing data of dataset-1 (first row, (a)-(d)) and dataset-2 (second row, (e)-(h)): CO-LR ((a) & (e)), CO-SVM ((b) & (f)), CO-ANN ((c) & (g)), and CO-RF ((d) & (h)).

In summary, Figs 3 and 4 serve as a visual confirmation of the models’ performance. The closer the data points align with the diagonal line, the more accurate the model’s predictions on unseen data. While CO-SVM, CO-ANN, and CO-RF exhibit their strengths, CO-RF consistently shines as a reliable choice for accurate SR prediction, particularly evident in the tight clustering of data in both datasets.

CO-optimized ML model for temporal SR prediction

The SR prediction capability of different ML models is compared using training data and testing data. In dataset-1, UNIX time is used to arrange the records chronologically. Records corresponding to the first 1000 minutes of the data are used for visualization. In dataset-2, each record’s timestamp contains only the date; hence, no intraday temporal variation can be analysed. Instead, the first 100 records corresponding to different days in chronological order (not always consecutive days) are visualized.

The time series visualizations provide insights into how these models perform over time, aiding in assessing their suitability for SR prediction applications. Fig 5 uses a training subset of the datasets. Each column corresponds to a different dataset, and each row represents a different CO-optimized ML model. In dataset-1, it can be seen that CO-LR and CO-SVM missed several of the measured SR values. These models captured the basic periodic nature of the measured SR pattern but could not match the amplitude variation. CO-ANN, on the hand, performs better than CO-SVM in tracking the periodicity and amplitude variation. Finally, CO-RF almost exactly matches the measured SR values significantly better than the others.

Fig 5.

Fig 5

Measured and predicted solar radiation by CO-optimized ML models using training data of dataset-1 (first column) and dataset-2 (second column): CO-LR ((a) & (b)), CO-SVM ((c) & (d)), CO-ANN ((e) & (f)), and CO-RF ((g) & (h)).

Fig 6 depicts a similar temporal relationship in the testing data. A similar observation can be noted in Fig 6. CO-LR and CO-SVM missed several SR values with higher amplitude correctly estimated by CO-ANN and CO-RF. Although CO-ANN and CO-RF visually miss almost identical SR values, the estimated SR values using CO-RF are much closer to measured SR values than the SR values estimated using CO-ANN. The CO-RF yielded better prediction results, and the temporal and amplitude variations in SR values performed best among the other examined models.

Fig 6.

Fig 6

Measured and predicted solar radiation by CO-optimized ML models using testing data of dataset-1 (first column) and dataset-2 (second column): CO-LR ((a) & (b)), CO-SVM ((c) & (d)), CO-ANN ((e) & (f)), and CO-RF ((g) & (h)).

Comparative analysis with existing work

Building on the advancements in machine learning models, the comparative analysis illustrates a clear trend toward increasing model complexity and predictive power. Table 6 compares existing literature on solar radiation prediction in different regions arranged chronologically. For instance, Marzo et al. [50] improved upon earlier methods with an artificial neural network (R2 = 0.8396), while Sharafati et al. [19] and Hassan et al. [51] adopted random forest and empirical predictive models, achieving impressive R2 values of 0.9312 and 0.9533, respectively. More sophisticated approaches emerged post-2020, such as deep learning techniques applied by Üstün et al. [52] and Ağbulut et al. [53], both of which demonstrated enhanced predictive accuracy (R2 = 0.9647 and 0.9378). The introduction of LSTM models, as seen in Ngoc-Lan Huynh et al. [54] (R2 = 0.9612), and Narvaez et al. [55] with random forest (R2 = 0.9654), showcased the potential of recurrent networks for time-series data.

Table 6. Comparative analysis with existing literature.

Reference Year Case study Model R 2
Olatomiwa et al. [56] 2015 Nigeria Adaptive neuro-fuzzy model 0.6678
Marzo et al. [50] 2017 South Africa ANN 0.8396
Hassan et al. [51] 2018 Egypt Empirical predictive model 0.9533
Sharafati et al. [19] 2019 Burkina Faso Random forest 0.9312
Guijo-Rubio et al. [57] 2020 Spain Hybrid neural network 0.9728
Üstün et al. [52] 2020 Turkey Deep learning 0.9647
Ağbulut et al. [53] 2021 Turkey Deep learning 0.9378
Narvaez et al. [55] 2021 Colombia Random forest 0.9654
Ngoc-Lan Huynh et al. [54] 2021 Vietnam LTSM 0.9612
Rai et al. [58] 2021 Mauritania Bi-LTSM 0.9805
Goliatt et al. [59] 2022 Burkina Faso Multi-adaptive regression splines 0.9800
present work 2024 Soudi Arebia CO with RF 0.9868

Finally, the present work leverages a CO-based random forest model, achieving the highest R2 (0.9868) reported to date, demonstrating the superior predictive capability of combining RF with CO techniques in the context of Saudi Arabia. This underscores the evolution of models towards increasing precision and adaptability across diverse geographical case studies.

Conclusion and future work

SR prediction is well-studied for ML techniques to be utilized due to the capabilities of these models in providing reliable and robust predictions for SR values. This work introduced an efficient hybridization of the optimization algorithm and ML model for SR prediction. Two open-source datasets and three statistical metrics, MAE, MSE, and R2, are employed as evaluation measurements. Comparative analysis of five state-of-the-art optimization algorithms showed that CO is the most efficient for SR prediction. The LR, SVM, ANN, and RF standalone models are compared with hybrid models using CO-based feature selection. The experimental findings showed that CO-RF provides the best prediction performance compared to the other competitive methods with MAE of 0.0365, MSE of 0.0074, and R2 of 0.9251 on dataset-1 and MAE of 0.0469 of MSE of 0.0032, and R2 of 0.9868 on dataset-2. In future research, CO-RF can be applied to a large-scale IoT-based online monitoring system to predict SR based on real-time data from IoT sensors. Another possible research method is to use deep learning and recurrent architecture for SR prediction while powered by optimization algorithms.

Data Availability

All relevant data are within the manuscript and its Supporting Information files.

Funding Statement

The article processing charge for this paper was funded by the University of Hertfordshire.

References

  • 1. Bamisile O, Oluwasanmi A, Ejiyi C, Yimen N, Obiora S, Huang Q. Comparison of machine learning and deep learning algorithms for hourly global/diffuse solar radiation predictions. International Journal of Energy Research. 2022;46(8):10052–10073. doi: 10.1002/er.6529 [DOI] [Google Scholar]
  • 2. Farzin S, Valikhan Anaraki M. Modeling and predicting suspended sediment load under climate change conditions: a new hybridization strategy. Journal of Water and Climate Change. 2021;12(6):2422–2443. doi: 10.2166/wcc.2021.317 [DOI] [Google Scholar]
  • 3. Farrokhi A, Farzin S, Mousavi SF. Meteorological drought analysis in response to climate change conditions, based on combined four-dimensional vine copulas and data mining (VC-DM). Journal of Hydrology. 2021;603:127135. doi: 10.1016/j.jhydrol.2021.127135 [DOI] [Google Scholar]
  • 4. Anaraki MV, Farzin S. Humboldt Squid Optimization Algorithm (HSOA): A novel nature-inspired technique for solving optimization problems. IEEE Access. 2023;11:122069–122115. doi: 10.1109/ACCESS.2023.3328248 [DOI] [Google Scholar]
  • 5.Yang X, Jiang F, Liu H. Short-term solar radiation prediction based on SVM with similar data. In: 2nd IET Renewable Power Generation Conference (RPG 2013). IET; 2013. p. 1–4.
  • 6. Fan J, Wang X, Wu L, Zhou H, Zhang F, Yu X, et al. Comparison of Support Vector Machine and Extreme Gradient Boosting for predicting daily global solar radiation using temperature and precipitation in humid subtropical climates: A case study in China. Energy conversion and management. 2018;164:102–111. doi: 10.1016/j.enconman.2018.02.087 [DOI] [Google Scholar]
  • 7. Kim H, Park S, Park HJ, Son HG, Kim S. Solar radiation forecasting based on the hybrid CNN-CatBoost model. IEEE Access. 2023;11:13492–13500. doi: 10.1109/ACCESS.2023.3243252 [DOI] [Google Scholar]
  • 8. Ghimire S, Deo RC, Casillas-Pérez D, Salcedo-Sanz S. Boosting solar radiation predictions with global climate models, observational predictors and hybrid deep-machine learning algorithms. Applied Energy. 2022;316:119063. doi: 10.1016/j.apenergy.2022.119063 [DOI] [Google Scholar]
  • 9. Fan J, Wu L, Zhang F, Cai H, Wang X, Lu X, et al. Evaluating the effect of air pollution on global and diffuse solar radiation prediction using support vector machine modeling based on sunshine duration and air temperature. Renewable and Sustainable Energy Reviews. 2018;94:732–747. doi: 10.1016/j.rser.2018.06.029 [DOI] [Google Scholar]
  • 10. Wu L, Huang G, Fan J, Zhang F, Wang X, Zeng W. Potential of kernel-based nonlinear extension of Arps decline model and gradient boosting with categorical features support for predicting daily global solar radiation in humid regions. Energy Conversion and Management. 2019;183:280–295. doi: 10.1016/j.enconman.2018.12.103 [DOI] [Google Scholar]
  • 11. Basaran K, Özçift A, Kılınç D. A new approach for prediction of solar radiation with using ensemble learning algorithm. Arabian Journal for Science and Engineering. 2019;44:7159–7171. doi: 10.1007/s13369-019-03841-7 [DOI] [Google Scholar]
  • 12. Tooke TR, Coops NC, Webster J. Predicting building ages from LiDAR data with random forests for building energy modeling. Energy and Buildings. 2014;68:603–610. doi: 10.1016/j.enbuild.2013.10.004 [DOI] [Google Scholar]
  • 13. Sun H, Gui D, Yan B, Liu Y, Liao W, Zhu Y, et al. Assessing the potential of random forest method for estimating solar radiation using air pollution index. Energy Conversion and Management. 2016;119:121–129. doi: 10.1016/j.enconman.2016.04.051 [DOI] [Google Scholar]
  • 14. Kratzenberg M, Zürn HH, Revheim PP, Beyer HG. Identification and handling of critical irradiance forecast errors using a random forest scheme–A case study for southern Brazil. Energy Procedia. 2015;76:207–215. doi: 10.1016/j.egypro.2015.07.900 [DOI] [Google Scholar]
  • 15. Ibrahim IA, Khatib T. A novel hybrid model for hourly global solar radiation prediction using random forests technique and firefly algorithm. Energy Conversion and Management. 2017;138:413–425. doi: 10.1016/j.enconman.2017.02.006 [DOI] [Google Scholar]
  • 16. Al-Shourbaji I, Kachare PH, Abualigah L, Abdelhag ME, Elnaim B, Anter AM, et al. A deep batch normalized convolution approach for improving COVID-19 detection from chest X-ray images. Pathogens. 2022;12(1):17. doi: 10.3390/pathogens12010017 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 17. AlShourbaji I, Kachare P, Zogaan W, Muhammad L, Abualigah L. Learning features using an optimized artificial neural network for breast cancer diagnosis. SN Computer Science. 2022;3(3):229. doi: 10.1007/s42979-022-01129-6 [DOI] [Google Scholar]
  • 18.Gupta S, Katta AR, Baldaniya Y, Kumar R. Hybrid random forest and particle swarm optimization algorithm for solar radiation prediction. In: 2020 IEEE 5th international conference on computing communication and automation (ICCCA). IEEE; 2020. p. 302–307.
  • 19. Sharafati A, Khosravi K, Khosravinia P, Ahmed K, Salman SA, Yaseen ZM, et al. The potential of novel data mining models for global solar radiation prediction. International Journal of Environmental Science and Technology. 2019;16:7147–7164. doi: 10.1007/s13762-019-02344-0 [DOI] [Google Scholar]
  • 20. Srivastava R, Tiwari A, Giri V. Solar radiation forecasting using MARS, CART, M5, and random forest model: A case study for India. Heliyon. 2019;5(10). doi: 10.1016/j.heliyon.2019.e02692 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 21. Benali L, Notton G, Fouilloy A, Voyant C, Dizene R. Solar radiation forecasting using artificial neural network and random forest methods: Application to normal beam, horizontal diffuse and global components. Renewable energy. 2019;132:871–884. doi: 10.1016/j.renene.2018.08.044 [DOI] [Google Scholar]
  • 22. Belmahdi B, BOUARDI AE. Short-term solar radiation forecasting using machine learning models under different sky conditions: evaluations and comparisons. Environmental Science and Pollution Research. 2024;31(1):966–981. doi: 10.1007/s11356-023-31246-5 [DOI] [PubMed] [Google Scholar]
  • 23. Guermoui M, Benkaciali S, Gairaa K, Bouchouicha K, Boulmaiz T, Boland JW. A novel ensemble learning approach for hourly global solar radiation forecasting. Neural Computing and Applications. 2022; p. 1–23. [Google Scholar]
  • 24. Kuhe A, Achirgbenda VT, Agada M. Global solar radiation prediction for Makurdi, Nigeria, using neural networks ensemble. Energy Sources, Part A: Recovery, Utilization, and Environmental Effects. 2021;43(11):1373–1385. doi: 10.1080/15567036.2019.1637481 [DOI] [Google Scholar]
  • 25. Pang Z, Niu F, O’Neill Z. Solar radiation prediction using recurrent neural network and artificial neural network: A case study with comparisons. Renewable Energy. 2020;156:279–289. doi: 10.1016/j.renene.2020.04.042 [DOI] [Google Scholar]
  • 26. Faisal AF, Rahman A, Habib MTM, Siddique AH, Hasan M, Khan MM. Neural networks based multivariate time series forecasting of solar radiation using meteorological data of different cities of Bangladesh. Results in Engineering. 2022;13:100365. doi: 10.1016/j.rineng.2022.100365 [DOI] [Google Scholar]
  • 27. Geshnigani FS, Golabi MR, Mirabbasi R, Tahroudi MN. Daily solar radiation estimation in Belleville station, Illinois, using ensemble artificial intelligence approaches. Engineering Applications of Artificial Intelligence. 2023;120:105839. doi: 10.1016/j.engappai.2023.105839 [DOI] [Google Scholar]
  • 28. Kachare PH, Puri DV, Sangle SB, Al-Shourbaji I, Jabbari A, Kirner R, et al. LCADNet: a novel light CNN architecture for EEG-based Alzheimer disease detection. Physical and Engineering Sciences in Medicine. 2024; p. 1–14. [DOI] [PubMed] [Google Scholar]
  • 29. Kachare PH, Sangle SB, Puri DV, Khubrani MM, Al-Shourbaji I. STEADYNet: Spatiotemporal EEG analysis for dementia detection using convolutional neural network. Cognitive Neurodynamics. 2024; p. 1–14. doi: 10.1007/s11571-024-10153-6 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 30. Puri DV, Kachare PH, Sangle SB, Kirner R, Jabbari A, Al-Shourbaji I, et al. LEADNet: Detection of Alzheimer’s Disease using Spatiotemporal EEG Analysis and Low-Complexity CNN. IEEE Access. 2024;. doi: 10.1109/ACCESS.2024.3435768 [DOI] [Google Scholar]
  • 31. AlShourbaji I, Kachare P, Alshathri S, Duraibi S, Elamin B, Elsayed Abd Elaziz M. An Efficient Parallel Reptile Search Algorithm and Snake Optimizer Approach for Feature Selection. Mathematics. 2022;10:2351. doi: 10.3390/math10132351 [DOI] [Google Scholar]
  • 32. Puri DV, Kachare PH, Nalbalwar SL. Metaheuristic optimized time–frequency features for enhancing Alzheimer’s disease identification. Biomedical Signal Processing and Control. 2024;94:106244. doi: 10.1016/j.bspc.2024.106244 [DOI] [Google Scholar]
  • 33. Wu Z, Cui N, Gong D, Zhu F, Li Y, Xing L, et al. Predicting daily global solar radiation in various climatic regions of China based on hybrid support vector machines with meta-heuristic algorithms. Journal of Cleaner Production. 2023;385:135589. doi: 10.1016/j.jclepro.2022.135589 [DOI] [Google Scholar]
  • 34. Goliatt L, Yaseen ZM. Development of a hybrid computational intelligent model for daily global solar radiation prediction. Expert Systems with Applications. 2023;212:118295. doi: 10.1016/j.eswa.2022.118295 [DOI] [Google Scholar]
  • 35. Gupta R, Yadav AK, Jha S, Pathak PK. A robust regressor model for estimating solar radiation using an ensemble stacking approach based on machine learning. International Journal of Green Energy. 2024;21(8):1853–1873. doi: 10.1080/15435075.2023.2276152 [DOI] [Google Scholar]
  • 36. Banadkooki FB, Ehteram M, Panahi F, Sammen SS, Othman FB, Ahmed ES. Estimation of total dissolved solids (TDS) using new hybrid machine learning models. Journal of Hydrology. 2020;587:124989. doi: 10.1016/j.jhydrol.2020.124989 [DOI] [Google Scholar]
  • 37. Akbari MA, Zare M, Azizipanah-Abarghooee R, Mirjalili S, Deriche M. The cheetah optimizer: A nature-inspired metaheuristic algorithm for large-scale optimization problems. Scientific reports. 2022;12(1):10953. doi: 10.1038/s41598-022-14338-z [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 38.Andrey. Solar Radiation Prediction: Task from NASA Hackathon; 2017. Available from: https://www.kaggle.com/datasets/dronio/SolarEnergy.
  • 39. Al-Shourbaji I, Kachare P, Fadlelseed S, Jabbari A, Hussien AG, Al-Saqqar F, et al. Artificial ecosystem-based optimization with dwarf mongoose optimization for feature selection and global optimization problems. International Journal of Computational Intelligence Systems. 2023;16(1):102. doi: 10.1007/s44196-023-00279-6 [DOI] [Google Scholar]
  • 40. Breiman L. Random forests. Machine learning. 2001;45:5–32. doi: 10.1023/A:1017934522171 [DOI] [Google Scholar]
  • 41. Schonlau M, Zou RY. The random forest algorithm for statistical learning. The Stata Journal. 2020;20(1):3–29. doi: 10.1177/1536867X20909688 [DOI] [Google Scholar]
  • 42. Puri DV, Nalbalwar SL, Nandgaonkar AB, Gawande JP, Wagh A. Automatic detection of Alzheimer’s disease from EEG signals using low-complexity orthogonal wavelet filter banks. Biomedical Signal Processing and Control. 2023;81:104439. doi: 10.1016/j.bspc.2022.104439 [DOI] [Google Scholar]
  • 43. Zang C, Ma Y. Ensemble machine learning: Methods and application. Springer publication, DOI. 2012;10:978–1. [Google Scholar]
  • 44.Kennedy J, Eberhart R. Particle swarm optimization. In: Proceedings of ICNN’95-international conference on neural networks. vol. 4. ieee; 1995. p. 1942–1948.
  • 45. Dorigo M, Maniezzo V, Colorni A. Ant system: optimization by a colony of cooperating agents. IEEE transactions on systems, man, and cybernetics, part b (cybernetics). 1996;26(1):29–41. doi: 10.1109/3477.484436 [DOI] [PubMed] [Google Scholar]
  • 46. Abualigah L, Abd Elaziz M, Sumari P, Geem ZW, Gandomi AH. Reptile Search Algorithm (RSA): A nature-inspired meta-heuristic optimizer. Expert Systems with Applications. 2022;191:116158. doi: 10.1016/j.eswa.2021.116158 [DOI] [Google Scholar]
  • 47. Hashim FA, Hussien AG. Snake Optimizer: A novel meta-heuristic optimization algorithm. Knowledge-Based Systems. 2022;242:108320. doi: 10.1016/j.knosys.2022.108320 [DOI] [Google Scholar]
  • 48. Tao H, Ewees AA, Al-Sulttani AO, Beyaztas U, Hameed MM, Salih SQ, et al. Global solar radiation prediction over North Dakota using air temperature: development of novel hybrid intelligence model. Energy Reports. 2021;7:136–157. doi: 10.1016/j.egyr.2020.11.033 [DOI] [Google Scholar]
  • 49. Solano ES, Affonso CM. Solar irradiation forecasting using ensemble voting based on machine learning algorithms. Sustainability. 2023;15(10):7943. doi: 10.3390/su15107943 [DOI] [Google Scholar]
  • 50. Marzo A, Trigo-Gonzalez M, Alonso-Montesinos J, Martínez-Durbán M, López G, Ferrada P, et al. Daily global solar radiation estimation in desert areas using daily extreme temperatures and extraterrestrial radiation. Renewable Energy. 2017;113:303–311. doi: 10.1016/j.renene.2017.01.061 [DOI] [Google Scholar]
  • 51. Hassan MA, Khalil A, Kaseb S, Kassem MA. Independent models for estimation of daily global solar radiation: A review and a case study. Renewable and Sustainable Energy Reviews. 2018;82:1565–1575. doi: 10.1016/j.rser.2017.07.002 [DOI] [Google Scholar]
  • 52.Üstün İ, Üneş F, Mert İ, Karakuş C. A comparative study of estimating solar radiation using machine learning approaches: DL, SMGRT, and ANFIS. Energy Sources, Part A: Recovery, Utilization, and Environmental Effects. 2022;44(4):10322–10345. doi: 10.1080/15567036.2020.1781301 [DOI] [Google Scholar]
  • 53. Ümit Ağbulut, Gürel AE, Biçen Y. Prediction of daily global solar radiation using different machine learning algorithms: Evaluation and comparison. Renewable and Sustainable Energy Reviews. 2021;135:110114. doi: 10.1016/j.rser.2020.110114 [DOI] [Google Scholar]
  • 54. Ngoc-Lan Huynh A, Deo RC, Ali M, Abdulla S, Raj N. Novel short-term solar radiation hybrid model: Long short-term memory network integrated with robust local mean decomposition. Applied Energy. 2021;298:117193. doi: 10.1016/j.apenergy.2021.117193 [DOI] [Google Scholar]
  • 55. Narvaez G, Giraldo LF, Bressan M, Pantoja A. Machine learning for site-adaptation and solar radiation forecasting. Renewable Energy. 2021;167:333–342. doi: 10.1016/j.renene.2020.11.089 [DOI] [Google Scholar]
  • 56. Olatomiwa L, Mekhilef S, Shamshirband S, Petković D. Adaptive neuro-fuzzy approach for solar radiation prediction in Nigeria. Renewable and Sustainable Energy Reviews. 2015;51:1784–1791. doi: 10.1016/j.rser.2015.05.068 [DOI] [Google Scholar]
  • 57. Guijo-Rubio D, Durán-Rosal AM, Gutiérrez PA, Gómez-Orellana AM, Casanova-Mateo C, Sanz-Justo J, et al. Evolutionary artificial neural networks for accurate solar radiation prediction. Energy. 2020;210:118374. doi: 10.1016/j.energy.2020.118374 [DOI] [Google Scholar]
  • 58. Ibrahim RA, Abd Elaziz M, Ewees AA, El-Abd M, Lu S. New feature selection paradigm based on hyper-heuristic technique. Applied Mathematical Modelling. 2021;98:14–37. doi: 10.1016/j.apm.2021.04.018 [DOI] [Google Scholar]
  • 59. Goliatt L, Sulaiman SO, Khedher KM, Farooque AA, Yaseen ZM. Estimation of natural streams longitudinal dispersion coefficient using hybrid evolutionary machine learning model. Engineering Applications of Computational Fluid Mechanics. 2021;15(1):1298–1320. doi: 10.1080/19942060.2021.1972043 [DOI] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Data Availability Statement

All relevant data are within the manuscript and its Supporting Information files.


Articles from PLOS ONE are provided here courtesy of PLOS

RESOURCES