TY - JOUR AU - Alameen, Abdalla AB - Introduction Over the decade, industrial modernization has constantly increased the energy demand of our economy, while the energy supply amount is limited. The continually diminishing conventional resources like fossil fuels, petroleum, coal, and natural gas have further increased the demand-supply gap, commonly known as the energy crisis. In search of green and sustainable energy, efforts are encouraged to use renewable resources like solar, wind, water, biomass, and geothermal. These renewable resources reduce environmental pollution and stress on conventional resources [1–4]. Solar Radiation (SR) is a readily and abundantly available renewable resource in most parts of the world. In recent years, SR prediction has received significant attention in academia [5–8]. Several ML models have been reported to predict SR [9]. Compared support vector machine (SVM) and Extreme Gradient Boosting (XGBoost) for SR prediction. Comparative evaluation using prediction accuracy, stability, and computational time showed that XGBoost outperformed SVM for estimating SR. In [10] work, the authors assessed the ability of kernel-based nonlinear extension of Arps decline and Gradient Boosting (GB) models with categorical features to support daily global SR prediction. Both models were reported to be suitable for SR. In another work [11], bagging and boosting ensemble models improve the performance of SVM, decision trees, and artificial neaural network (ANN) models for hourly SR prediction. The results demonstrated that decision trees improved SR prediction than the other used models. RF technique is also widely applied for energy systems and SR prediction [12–15], used Firefly (FF) algorithm to choose the best number of leaves per tree for the RF. The outcomes showed that the combined model outperformed the conventional Random Forest (RF), and ANN. Also, optimizing the ANN regarding the accuracy and execution speed also enhances the overall performance [16, 17]. In another work [18], Particle Swarm Optimization (PSO) was utilized to optimize RF parameters for SR prediction. The results showed that PSO-RF achieved better SR prediction than the decision trees, RF, and ANN. In [19] work, RF, random tree, reduced error pruning trees, and hybrid random tree reduce model are compared for daily SR prediction using four locations located in Burkina Faso. The results confirmed that RF is a promising approach for predicting SR accurately. In another comparison study [20], RF and three other tree-based machine learning (ML) models for SR prediction in India are compared. The results indicated that RF outperformed the other models used in their study. Benali et al. [21], smart persistence, ANN, and RF are compared to predict global horizontal, diffuse horizontal, and beam normal components and SR in France. The results showed that RF predicted all components more effectively than other methods. In another recent work, [22], the authors compared seven ML models for SR prediction. The obtained results showed that RF performed better than the other compared methods. A hybrid ML technique is proposed for hour global SR prediction [23]. The developed approach relied on a Convolutional Neural Network, Non-parametric Gaussian Process Regression (GPR), Least square SVM (LS-SVM), and Extreme Learning Machine (ELM). Experimental results showed that the proposed approach attained better accuracy than other popular ML models. In another work, Kuhe et al. [24], the authors attempt to improve the prediction accuracy of SR using ANN’s ensemble. They used Makurdi city in Nigeria data as a case study to evaluate the model. The results showed that the proposed model achieves great accuracy. Pang et al. [25], showed that the Recurrent Neural Network (RNN) can achieve higher accuracy for SR than the ANN model. In [26] work, a system to predict SR using Neural Networks is proposed. The authors used meteorological data from five different cities in Bangladesh. They used RNN, long short term memory (LSTM), and Gated Recurrent Unit (GRU) models to train the meteorological data. Among the used models, the GRU performed the best. Geshnigani et al. [27], improved conventional Multiple Llinear Regression, Multilayer Perceptron Neural Network, and Adaptive Neuro-Fuzzy Inference System (ANFIS) using optimization algorithms. They confirmed that the ANFIS model is effectively enhanced when optimization methods are used. Although deep learning-based models have shown success in a variety of areas over the past few years, these models require large datasets or need optimization of model parameters and latency to achieve generalization [28–30]. The metaheuristic models have shown good performance in variation applications [31, 32]. Wu et al. [33], used Ant Colony Optimization (ACO), Cuckoo Search (CS) and Grey Wolf Optimization (GWO) algorithms to optimize the SVM model for SR prediction. Among the hybrid models, GWO-SVM had the best results. In the work of Goliatt and Yaseen [34], a computation intelligent model based on the hybridization of Covariance Matrix Adaptive Evolution Strategies (CMAES) with XGBoost and Multi-Adaptive Regression Splines (MARS) models for building robust predictive models for daily SR prediction is developed. Results showed that the developed approach improves prediction accuracy. Gupta et al. [35], merged Variance Inflation Factor with Mutual Information (VIFMI) as a feature selection method and then the most important features are used as inputs to a Stack Ensemble Extra Tress (SEET) model for estimating SR. Results showed that the developed ensemble model effectively reduced prediction error. Due to the rapid deployment of solar energy techniques, SR has become common in recent years, and several ML models are widely used to predict SR. However, ML techniques can encounter challenges regarding input data, which can affect the efficacy of these models Banadkooki et al. [36], choosing relevant features is critical for building a learning model, as the input feature set influences it. This plays a pivotal role in maintaining high classification accuracy and reducing the dimensionality of datasets. According to the author’s best knowledge, well-known methods like PSO, ACO, GWO and FF perform well either in local or global searches, and they have poor performance in balancing between local and global searches. CO is a new population-based algorithm introduced in 2022 by Akbari [37] as a powerful method for mimicking specific cheetahs’ hunting strategies. CO shows a better balance between local and global search ability. Hence, in the present work. A new hybridization approach has been developed for more reliable and accurate SR predictions. For this purpose, the hybrid of CO and RF (CO-RF) is employed for predicting SR. The key contributions of this paper can be summarised as follows: Proposing an efficient prediction model, CO-RF, for SR by integrating optimization algorithm rather than the use of existing ML models alone. Evaluating the developed CO-RF model on two publicly available datasets and three evaluation metrics: mean absolute error (MAE), mean square error (MSE), and R2. Investigating CO-RF’s effectiveness in performing SR prediction during training and testing phases and comparing its efficiency to other models before and after using optimization algorithms. The rest of this paper is structured as follows: Section 2 describes the datasets and methods used and the developed CO-RF model for SR prediction. The statistical metrics to evaluate the ML models are presented in section 3. Comparative analysis of the models is presented in section 4, and section 5 concludes the paper. Materials and methods Worldwide, solar energy is an increasingly important renewable energy source. Accurate prediction of SR is crucial for the efficient design and operation of solar power systems. A publicly available SR dataset for measurements is necessary to address this need. In this work, two such datasets are used for experimental evaluations. Dataset-1 HI-SEAS weather station collected the SR dataset and made it accessible to the scientific community as part of the NASA hackathon task NASA (2023) [38]. The dataset encompasses various meteorological attributes, enabling researchers to examine the intricate relationships between various environmental factors and SR. It comprises eleven distinct features that provide a comprehensive insight into the environmental conditions that influence SR patterns. In preparation for analysis, the UNIX time, a timestamp format, is transformed to facilitate the sorting and organization of recorded measurements. Two critical columns are derived from UNIX time: ‘Month’ and ‘Year,’ enabling temporal analysis of SR trends. A ‘Sun availability’ feature is generated as a difference between sunrise and sunset times, shedding light on the temporal dynamics of sunlight exposure. SR values are scaled to kilo-watts per square meter (kW/m²) to mitigate the impact of large variations in SR values, ultimately enhancing the performance of ML models. The processed dataset comprises 32,686 records and nine distinct features that encapsulate vital meteorological and temporal parameters. A comprehensive statistical analysis of the dataset is given in Table 1. Download: PPT PowerPoint slide PNG larger image TIFF original image Table 1. Statistical summary of dataset1. https://doi.org/10.1371/journal.pone.0314391.t001 Dataset-2 This dataset is collected and maintained by the King Abdullah City for Atomic and Renewable Energy in Saudi Arabia, provided by the OpenData platform, a government-based repository for open data, for all experimental evaluations [39]. The dataset comprises 1265 records, each characterized by 26 distinct features. These features encapsulate various aspects of solar power generation and meteorological conditions. The data is collected from 41 solar power facilities in Saudi Arabia, as described in Table 2. Download: PPT PowerPoint slide PNG larger image TIFF original image Table 2. Distribution of solar radiation records across Saudi Arabia. https://doi.org/10.1371/journal.pone.0314391.t002 Preliminary data analysis revealed that the feature ‘Wind Speed at 3m (SD) Uncertainty (m/s)’ has missing values for all site locations. Hence, this feature is dropped. Also, four records are dropped due to missing solar radiation values. To enhance the utility of the dataset, several preprocessing steps are undertaken. The average values of the corresponding feature impute the missing values in all remaining records. After processing the missing values, each feature is normalized independently to have zero mean and unity standard deviation. The solar radiation values represented by the Global Horizontal Irradiance (GHI), is normalized within the range [-1, +1]. This normalization accounts for variations in power production among different solar stations, ensuring that the dataset maintains consistency and is suitable for machine-learning applications. The processed dataset has 1262 records and 22 features. The GHI features have a mean value of 0.117 and a Standard Deviation (SD) of 0.492. A comprehensive statistics of the dataset is described in Table 3. Download: PPT PowerPoint slide PNG larger image TIFF original image Table 3. Statistical summary of dataset2. https://doi.org/10.1371/journal.pone.0314391.t003 Cheetah Optimizer (CO) The CO is a recent metaheuristic method proposed by [37]. This method mimics the behaviours of a cheetah. In this algorithm. a cheetah starts the hunting process by scanning its environment (defined as search space) to detect probable prey. Depending on the distance between the cheetah and the prey, the cheetah may hold the position and wait for the prey to get closer. The rushing and capturing are two stages of the cheetah’s attack. The hunt may be terminated due to reduced energy limits, fast prey feeling, etc. After termination, the cheetah returns home to rest before the next hunt. The pseudo-code of the CO is provided in [37]. The CO chooses one of the following strategies depending on the prey, their condition, search area and distance of the prey: Search strategy: Cheetahs find the probable prey in the environment by scanning the area while sitting/standing or patrolling the area actively. The scanning mode is suitable while walking on the plains for dense and grazing prey. On the other hand, scattered and active prey demand more energy-consuming active mode. The cheetah selects a sequence of these modes during the hunt depending on its own condition, the condition of prey, and area coverage. For ith cheetah in jth arrangement, the new position is calculated based on their current position as follows: (1) Where, t is the hunting time with maximum limit T, is the random number, and is step size. Sit and wait strategy: When the prey is in the cheetah’s vision field, any movement may reveal the cheetah’s presence to the prey and allow the prey to escape. To avoid the prey’s escape, the cheetah gets closer to the prey by hiding among the bushes or lying on the ground before the ambush. Therefore, in this mode, the cheetah remains at their position and waits for the prey to come nearer, and this behaviour can be represented as: (2) Attack strategy Speed and flexibility are two crucial factors in a cheetah’s attack. The cheetah starts the attack by rushing towards the prey at full speed. When the prey starts to flee as soon as it realizes the cheetah’s attack, the cheetah adjusts the path to intervene in the prey. For the current prey position in jth arrangement, the cheetah’s position according to this attack strategy is mathematically defined as: (3) where, and are turning and interaction factors of ith cheetah in jth arrangement. The Cheetah Optimizer operates through the following steps: Initialization: The algorithm starts by initializing parameters and setting hyper-parameters such as learning rates and momentum terms. Gradient Computation: At each iteration, the gradient of the objective function with respect to the parameters is computed. Update Step: The optimizer uses a specific update rule, which is influenced by the computed gradient and hyper-parameters, to adjust the parameters. Convergence Check: The optimizer checks whether the changes in the objective function or parameters are below a certain threshold to determine if convergence has been reached. Parameter Update: If the convergence criteria are not met, the algorithm updates the parameters and repeats the process. Random Forest (RF) Random Forest [40] is a machine learning model that builds an ensemble of decision trees, with each tree trained using a subset of randomly selected examples and features. This diversity in data for each tree reduces overfitting and improves robustness [41]. The final prediction is typically the pooling of predictions from trees in the forest. The ensemble approach generalizes well to new data, making RF popular for various ML problems [42]. RF builds trees in parallel and, hence, is less prone to overfitting, leading to faster training times. RF does not require extensive tuning of hyperparameters due to built-in overfitting controlling mechanisms, like random subsampling of records and features. The primary hyperparameters for tuning in RF include the number of trees in the ensemble and the maximum depth of each tree. In practice, it uses deep trees, and the number of trees is selected based on model performance [43]. The proposed system for SR prediction uses CO for selecting optimum features and RF for evaluating performance using these optimized features, as shown in system flow in Fig 1. The hyper-parameters model with the number of trees and a maximum depth determined through cross-validation. Download: PPT PowerPoint slide PNG larger image TIFF original image Fig 1. System flow of CO-RF model for SR prediction. https://doi.org/10.1371/journal.pone.0314391.g001 Dataset-1 HI-SEAS weather station collected the SR dataset and made it accessible to the scientific community as part of the NASA hackathon task NASA (2023) [38]. The dataset encompasses various meteorological attributes, enabling researchers to examine the intricate relationships between various environmental factors and SR. It comprises eleven distinct features that provide a comprehensive insight into the environmental conditions that influence SR patterns. In preparation for analysis, the UNIX time, a timestamp format, is transformed to facilitate the sorting and organization of recorded measurements. Two critical columns are derived from UNIX time: ‘Month’ and ‘Year,’ enabling temporal analysis of SR trends. A ‘Sun availability’ feature is generated as a difference between sunrise and sunset times, shedding light on the temporal dynamics of sunlight exposure. SR values are scaled to kilo-watts per square meter (kW/m²) to mitigate the impact of large variations in SR values, ultimately enhancing the performance of ML models. The processed dataset comprises 32,686 records and nine distinct features that encapsulate vital meteorological and temporal parameters. A comprehensive statistical analysis of the dataset is given in Table 1. Download: PPT PowerPoint slide PNG larger image TIFF original image Table 1. Statistical summary of dataset1. https://doi.org/10.1371/journal.pone.0314391.t001 Dataset-2 This dataset is collected and maintained by the King Abdullah City for Atomic and Renewable Energy in Saudi Arabia, provided by the OpenData platform, a government-based repository for open data, for all experimental evaluations [39]. The dataset comprises 1265 records, each characterized by 26 distinct features. These features encapsulate various aspects of solar power generation and meteorological conditions. The data is collected from 41 solar power facilities in Saudi Arabia, as described in Table 2. Download: PPT PowerPoint slide PNG larger image TIFF original image Table 2. Distribution of solar radiation records across Saudi Arabia. https://doi.org/10.1371/journal.pone.0314391.t002 Preliminary data analysis revealed that the feature ‘Wind Speed at 3m (SD) Uncertainty (m/s)’ has missing values for all site locations. Hence, this feature is dropped. Also, four records are dropped due to missing solar radiation values. To enhance the utility of the dataset, several preprocessing steps are undertaken. The average values of the corresponding feature impute the missing values in all remaining records. After processing the missing values, each feature is normalized independently to have zero mean and unity standard deviation. The solar radiation values represented by the Global Horizontal Irradiance (GHI), is normalized within the range [-1, +1]. This normalization accounts for variations in power production among different solar stations, ensuring that the dataset maintains consistency and is suitable for machine-learning applications. The processed dataset has 1262 records and 22 features. The GHI features have a mean value of 0.117 and a Standard Deviation (SD) of 0.492. A comprehensive statistics of the dataset is described in Table 3. Download: PPT PowerPoint slide PNG larger image TIFF original image Table 3. Statistical summary of dataset2. https://doi.org/10.1371/journal.pone.0314391.t003 Cheetah Optimizer (CO) The CO is a recent metaheuristic method proposed by [37]. This method mimics the behaviours of a cheetah. In this algorithm. a cheetah starts the hunting process by scanning its environment (defined as search space) to detect probable prey. Depending on the distance between the cheetah and the prey, the cheetah may hold the position and wait for the prey to get closer. The rushing and capturing are two stages of the cheetah’s attack. The hunt may be terminated due to reduced energy limits, fast prey feeling, etc. After termination, the cheetah returns home to rest before the next hunt. The pseudo-code of the CO is provided in [37]. The CO chooses one of the following strategies depending on the prey, their condition, search area and distance of the prey: Search strategy: Cheetahs find the probable prey in the environment by scanning the area while sitting/standing or patrolling the area actively. The scanning mode is suitable while walking on the plains for dense and grazing prey. On the other hand, scattered and active prey demand more energy-consuming active mode. The cheetah selects a sequence of these modes during the hunt depending on its own condition, the condition of prey, and area coverage. For ith cheetah in jth arrangement, the new position is calculated based on their current position as follows: (1) Where, t is the hunting time with maximum limit T, is the random number, and is step size. Sit and wait strategy: When the prey is in the cheetah’s vision field, any movement may reveal the cheetah’s presence to the prey and allow the prey to escape. To avoid the prey’s escape, the cheetah gets closer to the prey by hiding among the bushes or lying on the ground before the ambush. Therefore, in this mode, the cheetah remains at their position and waits for the prey to come nearer, and this behaviour can be represented as: (2) Attack strategy Speed and flexibility are two crucial factors in a cheetah’s attack. The cheetah starts the attack by rushing towards the prey at full speed. When the prey starts to flee as soon as it realizes the cheetah’s attack, the cheetah adjusts the path to intervene in the prey. For the current prey position in jth arrangement, the cheetah’s position according to this attack strategy is mathematically defined as: (3) where, and are turning and interaction factors of ith cheetah in jth arrangement. The Cheetah Optimizer operates through the following steps: Initialization: The algorithm starts by initializing parameters and setting hyper-parameters such as learning rates and momentum terms. Gradient Computation: At each iteration, the gradient of the objective function with respect to the parameters is computed. Update Step: The optimizer uses a specific update rule, which is influenced by the computed gradient and hyper-parameters, to adjust the parameters. Convergence Check: The optimizer checks whether the changes in the objective function or parameters are below a certain threshold to determine if convergence has been reached. Parameter Update: If the convergence criteria are not met, the algorithm updates the parameters and repeats the process. Random Forest (RF) Random Forest [40] is a machine learning model that builds an ensemble of decision trees, with each tree trained using a subset of randomly selected examples and features. This diversity in data for each tree reduces overfitting and improves robustness [41]. The final prediction is typically the pooling of predictions from trees in the forest. The ensemble approach generalizes well to new data, making RF popular for various ML problems [42]. RF builds trees in parallel and, hence, is less prone to overfitting, leading to faster training times. RF does not require extensive tuning of hyperparameters due to built-in overfitting controlling mechanisms, like random subsampling of records and features. The primary hyperparameters for tuning in RF include the number of trees in the ensemble and the maximum depth of each tree. In practice, it uses deep trees, and the number of trees is selected based on model performance [43]. The proposed system for SR prediction uses CO for selecting optimum features and RF for evaluating performance using these optimized features, as shown in system flow in Fig 1. The hyper-parameters model with the number of trees and a maximum depth determined through cross-validation. Download: PPT PowerPoint slide PNG larger image TIFF original image Fig 1. System flow of CO-RF model for SR prediction. https://doi.org/10.1371/journal.pone.0314391.g001 Experiments and results Experimental setup All the ML models are implemented using the Python-based Scikit-Learn library, capitalizing on its robust model development and evaluation capabilities. The experimentation environment employed was a Windows 10 operating system on an Intel i7 processor clocked at 3.13 GHz and 64 GB of RAM. Lower fitness values indicate better feature selection, enhancing the SR prediction models. These methods are: Particle Swarm Optimization (PSO) [44], ACO [45], Reptile Search Algorithm (RSA) [46], Snake Optimizer (SO) [47], and CO, [37]. The standard parameters of these methods are selected empirically and are set: Population size = 20, number of iterations = 100, and each of them is executed indecently run ten times mitigate biases towards local minima. The settings of these methods are defined based on their implementations in original works, and they are listed in Table 4. Download: PPT PowerPoint slide PNG larger image TIFF original image Table 4. Parameters settings. https://doi.org/10.1371/journal.pone.0314391.t004 Evaluation metrics To assess prediction results, three different statistical measures are used for the evaluation of the developed CO-RF and the other ML models. These metrics are selected due to their suitability for SR prediction in previous works [22, 33, 48, 49], and they are defined as follows: (4) (5) (6) where, N is the number of observations, mi is the ith measured SR value, pi is the ith SR value predicted by the model and μim is the mean of the measured SR values. Results and discussion The performance of the models in predicting SR is rigorously evaluated and provided in this section. The average convergence behavior of these algorithms is presented in Fig 2. In Fig 2a, the CO algorithm achieves a minimum convergence fitness of 0.0029. PSO and ACO also demonstrated strong feature selection capabilities with fitness values of 0.0052 and 0.0041, respectively. CO stood out as the top-performing algorithm for dataset-1. These results provide valuable insights for researchers and practitioners looking to optimize feature selection for SR prediction tasks in similar datasets. As shown in Fig 2b, CO converges at a fitness value of 0.0027 for dataset-2. The fitness values represent the quality of feature selection for enhancing SR prediction. SO also demonstrated strong feature selection abilities, yielding a fitness value of 0.0029. ACO and RSA reported the same fitness value of 0.0035. PSO achieved the worst performance with a 0.0040 fitness value. The marginally higher convergence value in dataset-1 may be attributed to the larger number of records than in dataset-2. Download: PPT PowerPoint slide PNG larger image TIFF original image Fig 2. Convergence behavior of different optimization algorithms as feature selectors for SR prediction for (a) dataset-1 and (b) dataset-2. https://doi.org/10.1371/journal.pone.0314391.g002 CO reported the smallest fitness values for both datasets, indicating the best performance compared to the other optimization algorithms. Hence, further analysis uses only CO as a feature selection method. CO chooses five salient features from the original nine features in dataset-1 and fourteen out of the original twenty-one features in dataset-2. The selected salient features from both datasets are subsequently employed as inputs for various ML algorithms to assess their capabilities for SR prediction. The computation time required for either dataset is important to justify the model’s scalability. The overall computation time required by the hybrid model can be divided into three parts: time for feature selection by CO, time for training the ML model using the selected features, and time for performance evaluation. Most of the time required was for feature selection due to the iterative nature of the process. The ML model training time varied over different folds depending upon the number of selected features. The testing performance time had almost insignificant variations across different folds. The overall time for dataset-1 was 30.125 ± 2.536 and for dataset-2 was 10.983 ± 0.843. It was observed that almost 90-93% of the overall time was for feature selection, and approximately 1% of the overall time was for performance evaluation. It must be noted that the feature selection process is required only during the model training. In the testing phase, the most important feature already indicated by the training process can only be calculated further to reduce the solar radiation prediction time than the models using the complete feature set. Predictive performance. The selected features are assessed using MAE, MSE, and R2 measures. Table 5 compares the predictive performance of the ML models using a complete feature set and with a CO-based selected feature set across both datasets. The models depicted with the prefix ‘CO-’ use features optimized with CO. Each model is validated using 10-fold cross-validation and resulting metric in terms of both mean and SD. Download: PPT PowerPoint slide PNG larger image TIFF original image Table 5. Comparative performance of ML models with and without CO for SR prediction. https://doi.org/10.1371/journal.pone.0314391.t005 In the evaluation of dataset-1, a distinct performance variation is observed among SR prediction models. LR exhibited MAE of 0.1523, indicating a reasonable prediction accuracy. However, introducing CO hybridization (CO-LR) slightly improved prediction, and the R2 value increased to 0.6208. On the other hand, SVM demonstrated superior performance than LR, with MAE of 0.0826, a lower MSE of 0.0226, and a higher R2 of 0.8163. Applying CO hybridization to SVM (CO-SVM) enhanced its predictive power, resulting in an R2 of 0.8337. ANN and RF models also exhibited commendable performance. The hybridization of CO (CO-ANN and CO-RF) leads to significant improvements. In dataset-2, LR achieved an MAE of 0.1105, showing a competitive predictive capability. The CO hybridization (CO-LR) slightly decreased the MAE, and the R2 value elevated to 0.9093. SVM continued to excel on this dataset, with a mean MAE of 0.0762 and an impressive R2 of 0.9510. The application of CO hybridization (CO-SVM) maintained high-performance levels, reaffirming its effectiveness. Additionally, ANN and RF models displayed high performance, and the inclusion of CO (CO-ANN and CO-RF) significantly improved their performance. Dataset 1 showcased model variability, with RF emerging as a standout performer, notably bolstered by CO hybridization. Similarly, dataset-2 demonstrated consistent accuracy across models, with RF again excelling and CO hybridization reaffirming its efficacy. While LR and ANN models showed competitive results, the introduction of CO had a significant impact. To summarize, the dataset-dependent nature of SR prediction models, with CO hybridization potentially enhancing ML-based approaches. Hence, if performed using CO-optimized features, all further analysis is hybridized with four ML models. Comparative performance of CO-optimized ML models. Visualizations are employed to provide intuitive insights into the CO-based models’ predictive behavior. Fig 3 provides eight scatter plots for essential insights into the performance of four ML models across training data of both SR datasets. These scatter plots employ measured SR values, sourced from the datasets, on the horizontal axis, while the SR estimated by the ML models is plotted on the vertical axis. The ideal scenario in these scatter plots is a perfectly diagonal line where the estimated SR values align precisely with the measured SR values. Such alignment represents accurate predictions by the models. Consequently, the further the data points deviate from the diagonal line, the less accurate the model’s predictions become. Download: PPT PowerPoint slide PNG larger image TIFF original image Fig 3. Convergence behavior of CO-optimized ML models using training data of dataset-1 (first row, (a)-(d)) and dataset-2 (second row, (e)-(h)): CO-LR ((a) & (e)), CO-SVM ((b) & (f)), CO-ANN ((c) & (g)), and CO-RF ((d) & (h)). https://doi.org/10.1371/journal.pone.0314391.g003 In dataset-1, the SR values are measured in kilowatt-hours per square meter (kWh/m2). CO-LR has a maximum deviation from the ideal diagonal arrangement, indicating the worst performance among all models. CO-SVM exhibits a relatively higher spread of data points, indicating a wider dispersion of predictions. Conversely, CO-ANN shows a more clustered distribution of points, suggesting improved predictive consistency. However, the star performer in this dataset is CO-RF, which demonstrates the tightest clustering of points, signifying highly accurate SR predictions. In dataset-2, the SR values are in watt-hours per square meter (Wh/m2), and all three ML models exhibit commendable performance. The scatter plots for CO-SVM, CO-ANN, and CO-RF showcase strong positive correlations between predicted and actual SR values. In this case, CO-RF marginally outperforms the others, demonstrating a slightly tighter clustering of points, reaffirming its effectiveness in SR prediction even in datasets with different units. Fig 4 provides a set of scatter plots using testing data for different ML models applied to both SR datasets. In dataset-1’s, CO-SVM displays a relatively high spread of data points, suggesting some variability in its predictions. CO-ANN exhibits a more clustered distribution, indicating better predictive consistency. However, CO-RF outperforms both, showcasing the tightest clustering of points, signifying exceptionally accurate SR predictions on the testing data. In dataset-2’s, all three models (CO-SVM, CO-ANN, and CO-RF) continue to demonstrate strong positive correlations between their predictions and the actual SR values. While all models perform well, CO-RF exhibits a slightly tighter clustering of data points, reaffirming its effectiveness in SR prediction on the testing data. Download: PPT PowerPoint slide PNG larger image TIFF original image Fig 4. Convergence behavior of CO-optimized ML models using testing data of dataset-1 (first row, (a)-(d)) and dataset-2 (second row, (e)-(h)): CO-LR ((a) & (e)), CO-SVM ((b) & (f)), CO-ANN ((c) & (g)), and CO-RF ((d) & (h)). https://doi.org/10.1371/journal.pone.0314391.g004 In summary, Figs 3 and 4 serve as a visual confirmation of the models’ performance. The closer the data points align with the diagonal line, the more accurate the model’s predictions on unseen data. While CO-SVM, CO-ANN, and CO-RF exhibit their strengths, CO-RF consistently shines as a reliable choice for accurate SR prediction, particularly evident in the tight clustering of data in both datasets. CO-optimized ML model for temporal SR prediction The SR prediction capability of different ML models is compared using training data and testing data. In dataset-1, UNIX time is used to arrange the records chronologically. Records corresponding to the first 1000 minutes of the data are used for visualization. In dataset-2, each record’s timestamp contains only the date; hence, no intraday temporal variation can be analysed. Instead, the first 100 records corresponding to different days in chronological order (not always consecutive days) are visualized. The time series visualizations provide insights into how these models perform over time, aiding in assessing their suitability for SR prediction applications. Fig 5 uses a training subset of the datasets. Each column corresponds to a different dataset, and each row represents a different CO-optimized ML model. In dataset-1, it can be seen that CO-LR and CO-SVM missed several of the measured SR values. These models captured the basic periodic nature of the measured SR pattern but could not match the amplitude variation. CO-ANN, on the hand, performs better than CO-SVM in tracking the periodicity and amplitude variation. Finally, CO-RF almost exactly matches the measured SR values significantly better than the others. Download: PPT PowerPoint slide PNG larger image TIFF original image Fig 5. Measured and predicted solar radiation by CO-optimized ML models using training data of dataset-1 (first column) and dataset-2 (second column): CO-LR ((a) & (b)), CO-SVM ((c) & (d)), CO-ANN ((e) & (f)), and CO-RF ((g) & (h)). https://doi.org/10.1371/journal.pone.0314391.g005 Fig 6 depicts a similar temporal relationship in the testing data. A similar observation can be noted in Fig 6. CO-LR and CO-SVM missed several SR values with higher amplitude correctly estimated by CO-ANN and CO-RF. Although CO-ANN and CO-RF visually miss almost identical SR values, the estimated SR values using CO-RF are much closer to measured SR values than the SR values estimated using CO-ANN. The CO-RF yielded better prediction results, and the temporal and amplitude variations in SR values performed best among the other examined models. Download: PPT PowerPoint slide PNG larger image TIFF original image Fig 6. Measured and predicted solar radiation by CO-optimized ML models using testing data of dataset-1 (first column) and dataset-2 (second column): CO-LR ((a) & (b)), CO-SVM ((c) & (d)), CO-ANN ((e) & (f)), and CO-RF ((g) & (h)). https://doi.org/10.1371/journal.pone.0314391.g006 Comparative analysis with existing work Building on the advancements in machine learning models, the comparative analysis illustrates a clear trend toward increasing model complexity and predictive power. Table 6 compares existing literature on solar radiation prediction in different regions arranged chronologically. For instance, Marzo et al. [50] improved upon earlier methods with an artificial neural network (R2 = 0.8396), while Sharafati et al. [19] and Hassan et al. [51] adopted random forest and empirical predictive models, achieving impressive R2 values of 0.9312 and 0.9533, respectively. More sophisticated approaches emerged post-2020, such as deep learning techniques applied by Üstün et al. [52] and Ağbulut et al. [53], both of which demonstrated enhanced predictive accuracy (R2 = 0.9647 and 0.9378). The introduction of LSTM models, as seen in Ngoc-Lan Huynh et al. [54] (R2 = 0.9612), and Narvaez et al. [55] with random forest (R2 = 0.9654), showcased the potential of recurrent networks for time-series data. Download: PPT PowerPoint slide PNG larger image TIFF original image Table 6. Comparative analysis with existing literature. https://doi.org/10.1371/journal.pone.0314391.t006 Finally, the present work leverages a CO-based random forest model, achieving the highest R2 (0.9868) reported to date, demonstrating the superior predictive capability of combining RF with CO techniques in the context of Saudi Arabia. This underscores the evolution of models towards increasing precision and adaptability across diverse geographical case studies. Experimental setup All the ML models are implemented using the Python-based Scikit-Learn library, capitalizing on its robust model development and evaluation capabilities. The experimentation environment employed was a Windows 10 operating system on an Intel i7 processor clocked at 3.13 GHz and 64 GB of RAM. Lower fitness values indicate better feature selection, enhancing the SR prediction models. These methods are: Particle Swarm Optimization (PSO) [44], ACO [45], Reptile Search Algorithm (RSA) [46], Snake Optimizer (SO) [47], and CO, [37]. The standard parameters of these methods are selected empirically and are set: Population size = 20, number of iterations = 100, and each of them is executed indecently run ten times mitigate biases towards local minima. The settings of these methods are defined based on their implementations in original works, and they are listed in Table 4. Download: PPT PowerPoint slide PNG larger image TIFF original image Table 4. Parameters settings. https://doi.org/10.1371/journal.pone.0314391.t004 Evaluation metrics To assess prediction results, three different statistical measures are used for the evaluation of the developed CO-RF and the other ML models. These metrics are selected due to their suitability for SR prediction in previous works [22, 33, 48, 49], and they are defined as follows: (4) (5) (6) where, N is the number of observations, mi is the ith measured SR value, pi is the ith SR value predicted by the model and μim is the mean of the measured SR values. Results and discussion The performance of the models in predicting SR is rigorously evaluated and provided in this section. The average convergence behavior of these algorithms is presented in Fig 2. In Fig 2a, the CO algorithm achieves a minimum convergence fitness of 0.0029. PSO and ACO also demonstrated strong feature selection capabilities with fitness values of 0.0052 and 0.0041, respectively. CO stood out as the top-performing algorithm for dataset-1. These results provide valuable insights for researchers and practitioners looking to optimize feature selection for SR prediction tasks in similar datasets. As shown in Fig 2b, CO converges at a fitness value of 0.0027 for dataset-2. The fitness values represent the quality of feature selection for enhancing SR prediction. SO also demonstrated strong feature selection abilities, yielding a fitness value of 0.0029. ACO and RSA reported the same fitness value of 0.0035. PSO achieved the worst performance with a 0.0040 fitness value. The marginally higher convergence value in dataset-1 may be attributed to the larger number of records than in dataset-2. Download: PPT PowerPoint slide PNG larger image TIFF original image Fig 2. Convergence behavior of different optimization algorithms as feature selectors for SR prediction for (a) dataset-1 and (b) dataset-2. https://doi.org/10.1371/journal.pone.0314391.g002 CO reported the smallest fitness values for both datasets, indicating the best performance compared to the other optimization algorithms. Hence, further analysis uses only CO as a feature selection method. CO chooses five salient features from the original nine features in dataset-1 and fourteen out of the original twenty-one features in dataset-2. The selected salient features from both datasets are subsequently employed as inputs for various ML algorithms to assess their capabilities for SR prediction. The computation time required for either dataset is important to justify the model’s scalability. The overall computation time required by the hybrid model can be divided into three parts: time for feature selection by CO, time for training the ML model using the selected features, and time for performance evaluation. Most of the time required was for feature selection due to the iterative nature of the process. The ML model training time varied over different folds depending upon the number of selected features. The testing performance time had almost insignificant variations across different folds. The overall time for dataset-1 was 30.125 ± 2.536 and for dataset-2 was 10.983 ± 0.843. It was observed that almost 90-93% of the overall time was for feature selection, and approximately 1% of the overall time was for performance evaluation. It must be noted that the feature selection process is required only during the model training. In the testing phase, the most important feature already indicated by the training process can only be calculated further to reduce the solar radiation prediction time than the models using the complete feature set. Predictive performance. The selected features are assessed using MAE, MSE, and R2 measures. Table 5 compares the predictive performance of the ML models using a complete feature set and with a CO-based selected feature set across both datasets. The models depicted with the prefix ‘CO-’ use features optimized with CO. Each model is validated using 10-fold cross-validation and resulting metric in terms of both mean and SD. Download: PPT PowerPoint slide PNG larger image TIFF original image Table 5. Comparative performance of ML models with and without CO for SR prediction. https://doi.org/10.1371/journal.pone.0314391.t005 In the evaluation of dataset-1, a distinct performance variation is observed among SR prediction models. LR exhibited MAE of 0.1523, indicating a reasonable prediction accuracy. However, introducing CO hybridization (CO-LR) slightly improved prediction, and the R2 value increased to 0.6208. On the other hand, SVM demonstrated superior performance than LR, with MAE of 0.0826, a lower MSE of 0.0226, and a higher R2 of 0.8163. Applying CO hybridization to SVM (CO-SVM) enhanced its predictive power, resulting in an R2 of 0.8337. ANN and RF models also exhibited commendable performance. The hybridization of CO (CO-ANN and CO-RF) leads to significant improvements. In dataset-2, LR achieved an MAE of 0.1105, showing a competitive predictive capability. The CO hybridization (CO-LR) slightly decreased the MAE, and the R2 value elevated to 0.9093. SVM continued to excel on this dataset, with a mean MAE of 0.0762 and an impressive R2 of 0.9510. The application of CO hybridization (CO-SVM) maintained high-performance levels, reaffirming its effectiveness. Additionally, ANN and RF models displayed high performance, and the inclusion of CO (CO-ANN and CO-RF) significantly improved their performance. Dataset 1 showcased model variability, with RF emerging as a standout performer, notably bolstered by CO hybridization. Similarly, dataset-2 demonstrated consistent accuracy across models, with RF again excelling and CO hybridization reaffirming its efficacy. While LR and ANN models showed competitive results, the introduction of CO had a significant impact. To summarize, the dataset-dependent nature of SR prediction models, with CO hybridization potentially enhancing ML-based approaches. Hence, if performed using CO-optimized features, all further analysis is hybridized with four ML models. Comparative performance of CO-optimized ML models. Visualizations are employed to provide intuitive insights into the CO-based models’ predictive behavior. Fig 3 provides eight scatter plots for essential insights into the performance of four ML models across training data of both SR datasets. These scatter plots employ measured SR values, sourced from the datasets, on the horizontal axis, while the SR estimated by the ML models is plotted on the vertical axis. The ideal scenario in these scatter plots is a perfectly diagonal line where the estimated SR values align precisely with the measured SR values. Such alignment represents accurate predictions by the models. Consequently, the further the data points deviate from the diagonal line, the less accurate the model’s predictions become. Download: PPT PowerPoint slide PNG larger image TIFF original image Fig 3. Convergence behavior of CO-optimized ML models using training data of dataset-1 (first row, (a)-(d)) and dataset-2 (second row, (e)-(h)): CO-LR ((a) & (e)), CO-SVM ((b) & (f)), CO-ANN ((c) & (g)), and CO-RF ((d) & (h)). https://doi.org/10.1371/journal.pone.0314391.g003 In dataset-1, the SR values are measured in kilowatt-hours per square meter (kWh/m2). CO-LR has a maximum deviation from the ideal diagonal arrangement, indicating the worst performance among all models. CO-SVM exhibits a relatively higher spread of data points, indicating a wider dispersion of predictions. Conversely, CO-ANN shows a more clustered distribution of points, suggesting improved predictive consistency. However, the star performer in this dataset is CO-RF, which demonstrates the tightest clustering of points, signifying highly accurate SR predictions. In dataset-2, the SR values are in watt-hours per square meter (Wh/m2), and all three ML models exhibit commendable performance. The scatter plots for CO-SVM, CO-ANN, and CO-RF showcase strong positive correlations between predicted and actual SR values. In this case, CO-RF marginally outperforms the others, demonstrating a slightly tighter clustering of points, reaffirming its effectiveness in SR prediction even in datasets with different units. Fig 4 provides a set of scatter plots using testing data for different ML models applied to both SR datasets. In dataset-1’s, CO-SVM displays a relatively high spread of data points, suggesting some variability in its predictions. CO-ANN exhibits a more clustered distribution, indicating better predictive consistency. However, CO-RF outperforms both, showcasing the tightest clustering of points, signifying exceptionally accurate SR predictions on the testing data. In dataset-2’s, all three models (CO-SVM, CO-ANN, and CO-RF) continue to demonstrate strong positive correlations between their predictions and the actual SR values. While all models perform well, CO-RF exhibits a slightly tighter clustering of data points, reaffirming its effectiveness in SR prediction on the testing data. Download: PPT PowerPoint slide PNG larger image TIFF original image Fig 4. Convergence behavior of CO-optimized ML models using testing data of dataset-1 (first row, (a)-(d)) and dataset-2 (second row, (e)-(h)): CO-LR ((a) & (e)), CO-SVM ((b) & (f)), CO-ANN ((c) & (g)), and CO-RF ((d) & (h)). https://doi.org/10.1371/journal.pone.0314391.g004 In summary, Figs 3 and 4 serve as a visual confirmation of the models’ performance. The closer the data points align with the diagonal line, the more accurate the model’s predictions on unseen data. While CO-SVM, CO-ANN, and CO-RF exhibit their strengths, CO-RF consistently shines as a reliable choice for accurate SR prediction, particularly evident in the tight clustering of data in both datasets. Predictive performance. The selected features are assessed using MAE, MSE, and R2 measures. Table 5 compares the predictive performance of the ML models using a complete feature set and with a CO-based selected feature set across both datasets. The models depicted with the prefix ‘CO-’ use features optimized with CO. Each model is validated using 10-fold cross-validation and resulting metric in terms of both mean and SD. Download: PPT PowerPoint slide PNG larger image TIFF original image Table 5. Comparative performance of ML models with and without CO for SR prediction. https://doi.org/10.1371/journal.pone.0314391.t005 In the evaluation of dataset-1, a distinct performance variation is observed among SR prediction models. LR exhibited MAE of 0.1523, indicating a reasonable prediction accuracy. However, introducing CO hybridization (CO-LR) slightly improved prediction, and the R2 value increased to 0.6208. On the other hand, SVM demonstrated superior performance than LR, with MAE of 0.0826, a lower MSE of 0.0226, and a higher R2 of 0.8163. Applying CO hybridization to SVM (CO-SVM) enhanced its predictive power, resulting in an R2 of 0.8337. ANN and RF models also exhibited commendable performance. The hybridization of CO (CO-ANN and CO-RF) leads to significant improvements. In dataset-2, LR achieved an MAE of 0.1105, showing a competitive predictive capability. The CO hybridization (CO-LR) slightly decreased the MAE, and the R2 value elevated to 0.9093. SVM continued to excel on this dataset, with a mean MAE of 0.0762 and an impressive R2 of 0.9510. The application of CO hybridization (CO-SVM) maintained high-performance levels, reaffirming its effectiveness. Additionally, ANN and RF models displayed high performance, and the inclusion of CO (CO-ANN and CO-RF) significantly improved their performance. Dataset 1 showcased model variability, with RF emerging as a standout performer, notably bolstered by CO hybridization. Similarly, dataset-2 demonstrated consistent accuracy across models, with RF again excelling and CO hybridization reaffirming its efficacy. While LR and ANN models showed competitive results, the introduction of CO had a significant impact. To summarize, the dataset-dependent nature of SR prediction models, with CO hybridization potentially enhancing ML-based approaches. Hence, if performed using CO-optimized features, all further analysis is hybridized with four ML models. Comparative performance of CO-optimized ML models. Visualizations are employed to provide intuitive insights into the CO-based models’ predictive behavior. Fig 3 provides eight scatter plots for essential insights into the performance of four ML models across training data of both SR datasets. These scatter plots employ measured SR values, sourced from the datasets, on the horizontal axis, while the SR estimated by the ML models is plotted on the vertical axis. The ideal scenario in these scatter plots is a perfectly diagonal line where the estimated SR values align precisely with the measured SR values. Such alignment represents accurate predictions by the models. Consequently, the further the data points deviate from the diagonal line, the less accurate the model’s predictions become. Download: PPT PowerPoint slide PNG larger image TIFF original image Fig 3. Convergence behavior of CO-optimized ML models using training data of dataset-1 (first row, (a)-(d)) and dataset-2 (second row, (e)-(h)): CO-LR ((a) & (e)), CO-SVM ((b) & (f)), CO-ANN ((c) & (g)), and CO-RF ((d) & (h)). https://doi.org/10.1371/journal.pone.0314391.g003 In dataset-1, the SR values are measured in kilowatt-hours per square meter (kWh/m2). CO-LR has a maximum deviation from the ideal diagonal arrangement, indicating the worst performance among all models. CO-SVM exhibits a relatively higher spread of data points, indicating a wider dispersion of predictions. Conversely, CO-ANN shows a more clustered distribution of points, suggesting improved predictive consistency. However, the star performer in this dataset is CO-RF, which demonstrates the tightest clustering of points, signifying highly accurate SR predictions. In dataset-2, the SR values are in watt-hours per square meter (Wh/m2), and all three ML models exhibit commendable performance. The scatter plots for CO-SVM, CO-ANN, and CO-RF showcase strong positive correlations between predicted and actual SR values. In this case, CO-RF marginally outperforms the others, demonstrating a slightly tighter clustering of points, reaffirming its effectiveness in SR prediction even in datasets with different units. Fig 4 provides a set of scatter plots using testing data for different ML models applied to both SR datasets. In dataset-1’s, CO-SVM displays a relatively high spread of data points, suggesting some variability in its predictions. CO-ANN exhibits a more clustered distribution, indicating better predictive consistency. However, CO-RF outperforms both, showcasing the tightest clustering of points, signifying exceptionally accurate SR predictions on the testing data. In dataset-2’s, all three models (CO-SVM, CO-ANN, and CO-RF) continue to demonstrate strong positive correlations between their predictions and the actual SR values. While all models perform well, CO-RF exhibits a slightly tighter clustering of data points, reaffirming its effectiveness in SR prediction on the testing data. Download: PPT PowerPoint slide PNG larger image TIFF original image Fig 4. Convergence behavior of CO-optimized ML models using testing data of dataset-1 (first row, (a)-(d)) and dataset-2 (second row, (e)-(h)): CO-LR ((a) & (e)), CO-SVM ((b) & (f)), CO-ANN ((c) & (g)), and CO-RF ((d) & (h)). https://doi.org/10.1371/journal.pone.0314391.g004 In summary, Figs 3 and 4 serve as a visual confirmation of the models’ performance. The closer the data points align with the diagonal line, the more accurate the model’s predictions on unseen data. While CO-SVM, CO-ANN, and CO-RF exhibit their strengths, CO-RF consistently shines as a reliable choice for accurate SR prediction, particularly evident in the tight clustering of data in both datasets. CO-optimized ML model for temporal SR prediction The SR prediction capability of different ML models is compared using training data and testing data. In dataset-1, UNIX time is used to arrange the records chronologically. Records corresponding to the first 1000 minutes of the data are used for visualization. In dataset-2, each record’s timestamp contains only the date; hence, no intraday temporal variation can be analysed. Instead, the first 100 records corresponding to different days in chronological order (not always consecutive days) are visualized. The time series visualizations provide insights into how these models perform over time, aiding in assessing their suitability for SR prediction applications. Fig 5 uses a training subset of the datasets. Each column corresponds to a different dataset, and each row represents a different CO-optimized ML model. In dataset-1, it can be seen that CO-LR and CO-SVM missed several of the measured SR values. These models captured the basic periodic nature of the measured SR pattern but could not match the amplitude variation. CO-ANN, on the hand, performs better than CO-SVM in tracking the periodicity and amplitude variation. Finally, CO-RF almost exactly matches the measured SR values significantly better than the others. Download: PPT PowerPoint slide PNG larger image TIFF original image Fig 5. Measured and predicted solar radiation by CO-optimized ML models using training data of dataset-1 (first column) and dataset-2 (second column): CO-LR ((a) & (b)), CO-SVM ((c) & (d)), CO-ANN ((e) & (f)), and CO-RF ((g) & (h)). https://doi.org/10.1371/journal.pone.0314391.g005 Fig 6 depicts a similar temporal relationship in the testing data. A similar observation can be noted in Fig 6. CO-LR and CO-SVM missed several SR values with higher amplitude correctly estimated by CO-ANN and CO-RF. Although CO-ANN and CO-RF visually miss almost identical SR values, the estimated SR values using CO-RF are much closer to measured SR values than the SR values estimated using CO-ANN. The CO-RF yielded better prediction results, and the temporal and amplitude variations in SR values performed best among the other examined models. Download: PPT PowerPoint slide PNG larger image TIFF original image Fig 6. Measured and predicted solar radiation by CO-optimized ML models using testing data of dataset-1 (first column) and dataset-2 (second column): CO-LR ((a) & (b)), CO-SVM ((c) & (d)), CO-ANN ((e) & (f)), and CO-RF ((g) & (h)). https://doi.org/10.1371/journal.pone.0314391.g006 Comparative analysis with existing work Building on the advancements in machine learning models, the comparative analysis illustrates a clear trend toward increasing model complexity and predictive power. Table 6 compares existing literature on solar radiation prediction in different regions arranged chronologically. For instance, Marzo et al. [50] improved upon earlier methods with an artificial neural network (R2 = 0.8396), while Sharafati et al. [19] and Hassan et al. [51] adopted random forest and empirical predictive models, achieving impressive R2 values of 0.9312 and 0.9533, respectively. More sophisticated approaches emerged post-2020, such as deep learning techniques applied by Üstün et al. [52] and Ağbulut et al. [53], both of which demonstrated enhanced predictive accuracy (R2 = 0.9647 and 0.9378). The introduction of LSTM models, as seen in Ngoc-Lan Huynh et al. [54] (R2 = 0.9612), and Narvaez et al. [55] with random forest (R2 = 0.9654), showcased the potential of recurrent networks for time-series data. Download: PPT PowerPoint slide PNG larger image TIFF original image Table 6. Comparative analysis with existing literature. https://doi.org/10.1371/journal.pone.0314391.t006 Finally, the present work leverages a CO-based random forest model, achieving the highest R2 (0.9868) reported to date, demonstrating the superior predictive capability of combining RF with CO techniques in the context of Saudi Arabia. This underscores the evolution of models towards increasing precision and adaptability across diverse geographical case studies. Conclusion and future work SR prediction is well-studied for ML techniques to be utilized due to the capabilities of these models in providing reliable and robust predictions for SR values. This work introduced an efficient hybridization of the optimization algorithm and ML model for SR prediction. Two open-source datasets and three statistical metrics, MAE, MSE, and R2, are employed as evaluation measurements. Comparative analysis of five state-of-the-art optimization algorithms showed that CO is the most efficient for SR prediction. The LR, SVM, ANN, and RF standalone models are compared with hybrid models using CO-based feature selection. The experimental findings showed that CO-RF provides the best prediction performance compared to the other competitive methods with MAE of 0.0365, MSE of 0.0074, and R2 of 0.9251 on dataset-1 and MAE of 0.0469 of MSE of 0.0032, and R2 of 0.9868 on dataset-2. In future research, CO-RF can be applied to a large-scale IoT-based online monitoring system to predict SR based on real-time data from IoT sensors. Another possible research method is to use deep learning and recurrent architecture for SR prediction while powered by optimization algorithms. TI - Improving prediction of solar radiation using Cheetah Optimizer and Random Forest JF - PLoS ONE DO - 10.1371/journal.pone.0314391 DA - 2024-12-20 UR - https://www.deepdyve.com/lp/public-library-of-science-plos-journal/improving-prediction-of-solar-radiation-using-cheetah-optimizer-and-0udPGRoBAk SP - e0314391 VL - 19 IS - 12 DP - DeepDyve ER -