4.1. Effect of Using Last Hour’s Weather Variables on Forecasting
To study this effect on Saudi datasets, ten forecasting models were trained and tested twice with the same records. First, training was conducted using 18 features as shown in
Table 8, which include temporal features, the last hour’s weather variables, and WS values of the previous 5 hours besides the WS value of the same hour last day. In the second trial, only WS values of the previous 5 hours were used, as highlighted in
Table 8. For the Caracas dataset, 20 features were used in the first trial (with WS_lag6, and WS_lag7 added to
Table 8 features) and 7 features in the second trial (WS_lag1 to WS_lag7). For the Toronto dataset, 24 features were used in the first trial (with WS_lag8 to WS_lag12 added to
Table 8 features and WS_1D removed). In the second trial for the Toronto dataset, only 12 features were used (WS_lag1 to WS_lag12).
Figure 28 shows the average MAE results of 20 runs of the six DL-based forecasting models and four ML-based models when weather features were used besides WS lagged features, whereas
Figure 29 shows the same when only WS lagged features were used.
For Alghat dataset, we noted that using weather features has improved the MAE results for all six DL-based forecasting models by 33% at most as with the GRU model and 20% at least as with the CNN-LSTM model. Using weather features in ML-based models improved the MAE results for all four models by 30% at most as with the XGB model and 5% at least as with MLR model. The best MAE value is 0.14 achieved by LSTM, GRU, BiLSTM, BiGRU, and XGB models, while the worst MAE value is 0.20 and associated with the MLR model.
For Dumat Aljandal dataset, using weather features has improved MAE results for all six DL-based forecasting models by 25% at most as with LSTM, BiLSTM, BiGRU, and LSTM-AE models and 15% at least as with CNN-LSTM model. Using weather features in ML-based models improved the MAE results for all four models by 24% at most as with the XGB model and 5% at least as with MLR model. The best MAE value is 0.15 achieved by LSTM, BiLSTM, BiGRU, and LSTM-AE models, while the worst MAE value is 0.20 and associated with the MLR model.
For the Waad Alshamal dataset, using weather features has improved MAE results for all six DL-based forecasting models by 32% at most as with BiLSTM model and 16% at least as with CNN-LSTM model. Using weather features in ML-based models improved the MAE results for all four models by 27% at most, as with the RFR model and 5% at least as with the MLR model. The best MAE value is 0.13, achieved by BiLSTM, while the worst MAE value is 0.20 and associated with the MLR model.
For the Yanbu dataset, using weather features has improved MAE results for all DL-based forecasting models, except CNN-LSTM model, by 18% at most as with LSTM, GRU, BiLSTM, and BiGRU models. Using weather features in ML-based models improved the MAE results for all four models by 18% at most, as with SVR and 6% at least as with the MLR model. The best MAE value is 0.14, achieved by LSTM, GRU, BiLSTM, BiGRU, and SVR models, while the worst MAE value is 0.20 and associated with the MLR model.
For the Caracas dataset, using weather features has not improved MAE results, except for GRU and RFR models, which were improved by 14%. The best MAE value is 0.06, achieved by all models with weather features.
For the Toronto dataset, using weather features has not improved MAE results, except for CNN-LSTM and MLR models, which were improved by 5%. In fact, LSTM, GRU, BiLSTM, and LSTM-AE models achieved better results using only lagged features. Weather features worsened the results. The best MAE value is 0.18, achieved by the SVR model with only lagged features.
For Saudi datasets, we can summarize MAE results that using weather features has improved all DL- and ML-based models for all four locations, but the improvement percentage is the highest with the Alghat dataset and the lowest with the Yanbu dataset. We might relate low improvement with the Yanbu dataset to the lower correlation between WS and T_lag1 and between WS and DHI_lag1 (see
Figure 14) compared to other locations. Also, Yanbu is a coastal city, unlike the other three locations, and there are no significant changes in Yanbu weather from season to season. For example, the average temperature is 32° C in August and 21° C in January. With weather features, all models have similar MAE results, except the MLR model, which achieved the worst MAE value for all four datasets. The BiLSTM model is the best, which attained the best MAE value for all Saudi locations.
To summarize MAE results for the Caracas dataset, weather features have not improved MAE results because WS has strong correlations with its lagged seven features (see
Figure 15), which makes it easy to predict the next value with no extra features. Also, the MAE value is 0.06 for all models. This is a low value compared to other locations because the maximum WS in Caracas is 2.9 (see
Table 6). Using weather features in the Toronto dataset has made MAE results worse in most of the cases because WS has strong correlations with its lagged twelve features (see
Figure 16). Also, MAE values for Toronto are the largest because the maximum WS is 15.6—the highest among all locations (see
Table 6).
Figure 30 shows the average RMSE results of 20 runs of the six DL-based forecasting models and four ML-based models when weather features were used besides WS lagged features, whereas
Figure 31 shows the same when only WS lagged features were used.
For the Alghat dataset, we noted that using weather features has improved the RMSE results for all six DL-based forecasting models by 32% at most as with GRU model and 21% at least as with CNN-LSTM model. Using weather features in ML-based models improved the RMSE results for all four models by 27% at most, as with the RFR model and 4% at least as with the MLR model. The best RMSE value is 0.19 achieved by LSTM, GRU, BiLSTM, and BiGRU models, while the worst RMSE value is 0.27 and associated with the MLR model.
For Dumat Aljandal dataset, using weather features has improved RMSE results for all six DL-based forecasting models by 25% at most as with the LSTM model and 18% at least as with the CNN-LSTM model. Using weather features in ML-based models improved the RMSE results for all four models by 24% at most as with the XGB model and 3% at least as with MLR model. The best RMSE value is 0.21 achieved by LSTM, BiLSTM, and LSTM-AE models, while the worst RMSE value is 0.28 and associated with the MLR model.
For Waad Alshamal dataset, using weather features has improved RMSE results for all six DL-based forecasting models by 33% at most as with LSTM, GRU, and BiLSTM models and 19% at least as with the CNN-LSTM model. Using weather features in ML-based models improved the RMSE results for all four models by 30% at most as with XGB model and 3% at least as with MLR model. The best RMSE value is 0.18 achieved by LSTM, GRU, BiLSTM, and BiGRU models, while the worst RMSE value is 0.28 and associated with the MLR model.
For the Yanbu dataset, using weather features has improved RMSE results for all DL-based forecasting models, except CNN-LSTM model, by 22% at most, as with the LSTM model. Using weather features in ML-based models improved the RMSE results for all four models by 18% at most, as with SVR and 8% at least as with the MLR model. The best RMSE value is 0.18 achieved by LSTM, BiLSTM, and SVR models, while the worst RMSE value is 0.23 and is associated with the RFR model.
For the Caracas dataset, using weather features has improved RMSE results only for LSTM, GRU, SVR, MLR, and XGB models by 13% and RFR model by 22% The best RMSE value is 0.07 achieved by six models with weather features.
For the Toronto dataset, using weather features has not improved the RMSE results, except for CNN-LSTM, SVR, and XGB models, which were improved by 3% at least. In fact, GRU, BiLSTM, and LSTM-AE models achieved better results with lagged features only and using weather features worsened the results. The best RMSE value is 0.30 achieved by the XGB model with weather features and achieved by GRU and BiLSTM with lagged features only.
For Saudi datasets, we can summarize RMSE results that using weather features has improved all DL-based models and ML-based models for all four locations. However, the improvement percentage is the highest with Alghat and Waad Alshamal datasets and the lowest with the Yanbu dataset. We might relate low improvement with the Yanbu dataset to the lower correlation between WS and T_lag1 and between WS and DHI_lag1 (see
Figure 14) compared to other locations. Also, Yanbu is a coastal city, unlike the other three locations, and there are no significant changes in Yanbu weather from season to season. For example, the average temperature is 32° C in August and 21° C in January. With weather features, DL-based models have similar RMSE results, except CNN-LSTM model and ML-based models have similar RMSE results, except MLR model. MLR model achieved the worst RMSE value for three datasets, while LSTM and BiLSTM models attained the best RMSE value for all Saudi locations.
To summarize the RMSE results for the Caracas dataset, weather features have not improved RMSE results because WS has strong correlations with its lagged seven features (see
Figure 15), which makes it easy to predict the next value with no extra features. Also, the RMSE value is 0.07 or 0.08 for all models and is a low value compared to other locations because the maximum WS in Caracas is 2.9 (see
Table 6). Using weather features in the Toronto dataset has not improved RMSE results for most of the models because WS has strong correlations with its lagged twelve features (see
Figure 16). Also, RMSE values for Toronto are the largest because the maximum WS is 15.6 and is the highest among all locations (see
Table 6).
Figure 32 shows the average MAPE results of 20 runs of the six DL-based forecasting models and four ML-based models when weather features were used besides WS lagged features, whereas
Figure 33 shows the same when only WS lagged features were used.
For Alghat dataset, noted that using weather features has improved the MAPE results for all six DL-based forecasting models by 25% at most as with LSTM and GRU models and 15% at least as with the CNN-LSTM model. Using weather features in ML-based models improved the RMSE results for all four models by 25% at most as with XGB and RFR models and 2% at least as with MLR model. The best MAPE value is 5.91 achieved by the LSTM model, while the worst MAPE value is 8.31 and is associated with the MLR model.
For Dumat Aljandal dataset, using weather features has improved MAPE results for all six DL-based forecasting models by 18% at most as with the LSTM model and 11% at least as with the CNN-LSTM model. Using weather features in ML-based models improved the MAPE results for the RFR model by 15% and for XGB and SVR models by 9% at least. The best MAPE value is 7.66 achieved by the LSTM model, while the worst RMSE value is 10.05 and is associated with the MLR model.
For the Waad Alshamal dataset, using weather features has improved MAPE results for all six DL-based forecasting models by 31% at most as with BiLSTM model and 15% at least as with CNN-LSTM model. Using weather features in ML-based models improved the MAPE results for all four models by 28% at most as with the XGB model and 3% at least as with MLR model. The best MAPE value is 5.41 achieved by the BiLSTM model, while the worst MAPE value is 8.22 and is associated with the MLR model.
For the Yanbu dataset, using weather features has improved MAPE results for all DL-based forecasting models by 23% at most as with the LSTM model and by 5% at least as with the CNN-LSTM model. Using weather features in ML-based models improved the MAPE results for all four models by 17% at most as with SVR and 6% at least as with the MLR model. The best MAPE value is 6.67 achieved by the LSTM model, while the worst MAPE value is 8.25 and is associated with the RFR model.
For the Caracas dataset, using weather features has improved MAPE results, except for BiLSTM and CNN-LSTM models. The highest improvement percentage is 12% for the RFR model and the lowest is 2% for the MLR model. The best MAPE value is 4.68, achieved by the SVR model with weather features.
For the Toronto dataset, using weather features has not improved MAPE results, except for ML-based models, which were improved by 2% at most. In fact, DL-based models achieved better results with lagged features only and using weather features worsened the results. The best MAPE value is 5.37, achieved by the SVR model with weather features.
For Saudi datasets, we can summarize MAPE results that using weather features has improved all DL- and ML-based models for all four locations. However, the improvement percentage is the highest with the Alghat and Waad Alshamal datasets and the lowest with the Yanbu dataset. We might relate low improvement with the Yanbu dataset to the lower correlation between WS and T_lag1 and between WS and DHI_lag1 (see
Figure 14) compared to other locations. Also, Yanbu is a coastal city, unlike the other three locations, and there are no significant changes in Yanbu weather from season to season. For example, the average temperature is 32° C in August and 21° C in January. With weather features, all models have similar MAPE results, except MLR and RFR models. MLR model achieved the worst MAPE value for three datasets, while LSTM attained the best MAPE value for three datasets out of four Saudi locations.
To summarize MAPE results for the Caracas dataset, weather features have improved MAPE results because WS has strong correlations with its lagged seven features (see
Figure 15), which makes it easy to predict the next value with no extra features. Using weather features in the Toronto dataset has not improved MAPE results for most of the models because WS has strong correlations with its lagged twelve features (see
Figure 16).
From MAE, RMSE, and MAPE results in this section, we note that using weather features has improved the forecasting results of all models for Saudi locations by around 30% at most. However, DL-based models experienced higher improvement than ML-based models did. This may be related to DL-based models’ ability to handle high dimensionality. Also, the Yanbu dataset has the least improvement percentage because, as explained earlier, Yanbu is a coastal city, unlike the other three locations, and there are no significant changes in Yanbu weather from season to season. This makes weather features less important than WS lagged features in predicting the next value of WS. This is reflected in the lower correlation between WS and T_lag1 and between WS and DHI_lag1 (see
Figure 14) compared to other locations. Weather features with Caracas improved the forecasting results slightly, while it has worsened the results with Toronto for most of the models. The reason behind this is strong WS correlations with its lagged features. We used seven lagged features for Caracas and twelve for Toronto (see
Figure 10). Therefore, the results of ML-based models are better or similar to the results of DL-based models for both locations. We can conclude that when wind speed has strong correlations with its lagged values, ML-based models’ performance would be satisfactory (i.e., SVR and XGB models) while DL-based models are needed with less strong or weak correlations. The same applies to weather features, which can improve the forecasting results more if there are less strong correlations between WS and its lagged features.
4.3. Effect of Using Decomposition Methods on Forecasting
To study this effect, three decomposition methods (described in
Section 3.1.4) are combined with the LSTM model.
Section 3.2.1.6. describes the structure of these hybrid models in detail. The features used to train and test the three hybrid models are the last five hours’ WS values in Saudi locations, the last seven values in Caracas, and the last twelve values in Toronto. The forecasting results of the three hybrid models are compared to six DL-based models and four ML-based models (the same results appeared in
Section 4.1 for WS lagged features in
Figure 29,
Figure 31, and
Figure 33).
Figure 36 shows the MAE results of three hybrid models (EMD-LSTM, CEEMDAN-LSTM, VMD-LSTM), six DL-based models (LSTM, GRU, BiLSTM, BiGRU, LSTM-AE, CNN-LSTM), and four ML-based models (SVR, MLR, XGB, RFR) for all datasets.
From
Figure 36, we note that the best performing model for all Saudi locations is VMD-LSTM model, and the worst is RFR model. The hybrid model of VMD-LSTM achieved MAE value equals to 0.12, which improved the forecasting results over LSTM model by 40% for Alghat, Dumat Aljandal, and Waad Alshamal. It also achieved MAE value equals to 0.09 for Yanbu, which improved the forecasting results over the LSTM model by 47%. Regarding the Caracas dataset, all three hybrid models achieved the same MAE value equals to 0.03, which provided 50% improvement over the LSTM model result. With the Toronto dataset, the hybrid model of CEEMDAN-LSTM achieved better MAE value than the other two hybrid models, which provided 42% improvement over the LSTM model result.
Figure 37 shows the RMSE results of three hybrid models (EMD-LSTM, CEEMDAN-LSTM, VMD-LSTM), six DL-based models (LSTM, GRU, BiLSTM, BiGRU, LSTM-AE, CNN-LSTM), and four ML-based models (SVR, MLR, XGB, RFR) for all datasets.
From
Figure 37, we note that the best performing model for all Saudi locations is the VMD-LSTM model, and the worst is the RFR model. The hybrid model of VMD-LSTM achieved RMSE value equals to 0.15, which improved the forecasting results over LSTM model by 44% for Alghat and Waad Alshamal. It also achieved RMSE value equals to 0.16 for Dumat Aljandal and 0.13 for Yanbu, which improved the forecasting results over LSTM model by 43%. Regarding the Caracas and the Toronto datasets, the hybrid model of CEEMDAN-LSTM achieved slightly better RMSE value than the other two hybrid models, which considered 63% improvement in Caracas and 39% improvement in Toronto forecasting results over LSTM model.
Figure 38 shows the MAPE results of three hybrid models (EMD-LSTM, CEEMDAN-LSTM, VMD-LSTM), six DL-based models (LSTM, GRU, BiLSTM, BiGRU, LSTM-AE, CNN-LSTM), and four ML-based models (SVR, MLR, XGB, RFR) for all datasets.
From
Figure 38, we note that the best performing model for all Saudi locations is the VMD-LSTM model, and the worst is the RFR model. The hybrid model of VMD-LSTM achieved MAPE value equals to 4.81 for Alghat and 5.39 for Dumat Aljandal, which improved the forecasting results over the LSTM model by 39% and 37% for both locations. It also achieved MAPE value equals to 4.7 for Waad Alshamal and 4.66 for Yanbu, which improved the forecasting results over the LSTM model by 41% and 46% for both locations. Regarding the Caracas and the Toronto datasets, the hybrid model of CEEMDAN-LSTM achieved better MAPE value than the other two hybrid models, which considered 58% improvement in Caracas and 41% improvement in Toronto forecasting results over the LSTM model.
From MAE, RMSE, and MAPE results in this section, we conclude that using a hybrid model of LSTM and a decomposition method always achieves better results than using the LSTM model alone. In Saudi locations, the best hybrid model is VMD-LSTM according to all evaluation metrics, with improvement percentage ranges from 39% to 47% over the LSTM model. This observation agrees with the performance comparison done in [
29] between EMD, Ensemble EMD, Wavelet Packet Decomposition, and VMD, in which VMD achieved the most accurate and stable performance. Also, in [
48], VMD compared well to Empirical Wavelet Transform, Complementary Ensemble Empirical Mode Decomposition, and Ensemble Intrinsic Time-scale Decomposition. VMD outperformed EMD in [
49]. In this field, many works show that VMD-based models perform better compared with Wavelet Transform-based and EMD-based models [
29,
50,
51]. The reason behind the VMD superiority is its ability to decompose nonstationary and nonlinear time series and its robustness handling data noise.
Regarding the Caracas and the Toronto datasets, the best hybrid model is CEEMDAN-LSTM according to all evaluation metrics with improvement percentage ranges from 50% to 63% over LSTM model in Caracas and from 39% to 42% over LSTM model in Toronto.
We studied decomposition methods for forecasting by comparing the results of hybrid models to the results of DL- and ML-based models using only lagged WS values. We wonder which method is better: hybrid models with decomposition methods or weather variables with DL-based models (as done in
Section 4.1). To answer this question, we compared the best performing hybrid models VMD-LSTM and CEEMDAN-LSTM to the LSTM model that was trained and tested using weather variables for each dataset in
Figure 39.
Figure 39 (a) shows MAE results for these three models, while RMSE and MAPE results are shown in (b) and (c). From the figure, we can see that the VMD-LSTM model achieved the best forecasting accuracy for Saudi datasets, while the CEEMDAN-LSTM model achieved the same for Caracas and Toronto datasets. Therefore, we can conclude that hybrid models with decomposition methods achieved better results than using weather variables with DL-based models.