Submitted:
20 March 2025
Posted:
21 March 2025
You are already at the latest version
Abstract
Keywords:
I. Introduction
II. Feature Enginnerin in Solar Forecasting
III. Machine Learning Models
A. Multiple Linear Regression
- The strength of the relationship among multiple independent variables and a single dependent variable, exemplified by the interplay of rainfall, temperature, and fertilizer quantity on crop grmvth.
- The projected value of the dependent variable given specific values of the independent variables. For instance, it can predict crop yield based on distinct levels of rainfall, temperature, and fertilizer application.
- -
- y stands for the predicted value of the dependent variable.
- -
- f3o signifies they-intercept, representing the value ofy when aU other factors are zero.
- -
- B, to Bk represent the regression coefficients associated vvith each independent variable (xi to Xk).
- -
- i.:: denotes the model error, porirnying the degree of variation in our estimations.
- Deriving regression coefficients that minimize the overall model error.
- Calculating the t statistic of the overarching model.
- Assessing the corresponding p value, which gauges the likelihood of the t statistic arising due to chance, assuming tbe null hypothesis that no relationship exists between the independent and dependent variables.
B. Ridge Regression
- -
- Y denotes the response variable.
- -
- XJ stands for the jth predictor variable.
- -
- p j represents the average influence on Y when a one-unit increase is observed in XJ, \vhile keeping other predictors constant”
- -
- i.:: represents the eITor term.

- -
- :E signifies the sunm1ation operation.
- -
- Yi represents the actual response value for the ith observation.
- -
- Yi symbolizes the predicted response value generated by the multiple linear regression model.

- -
- j ranges from 1 to p.
- -
- I, (lambda) assumes a non-negative value.
C. Lasso Regression
- -
- Y: Represents the response variable.
- -
- XJ: Denotes the jth predictor variable.
- -
- f3j: Signifies the average impact on Y resulting from a one-unit increase in Xj, keeping aU other predictors constant.
- -
- s: Stands for the error term.
- -
- I: represents the summation symbol.
- -
- Yi signifies the actual response value for the ith observation.
- -
- Yi stands fiff the predicted response value derived from the multiple linear regression model.

- -
- j ranges from 1 to p.
- -
- I, (lambda) assumes a non-negative value.
D. Decision Tree Regression



E. Support Vector Regression
- 3.
-
Data Collection:Gathering a training set composed of samples that serve as the basis for prediction. It’s important that the features in the training set appropriately cover the domain of interest since SVR interpolates only within this domain.
- 4.
-
Kernel Selection:Choosing a suitable kernel, such as Sigmoid, Polynomial, Gaussian, etc., based on the problem’s nature. Kernels have hyperparameters that require tuning. Here, the Gaussian Kernel is taken as an example.
- 5.
-
Creation of Correlation Matrix:Constructing a correlation matrix by evaluating pairs of data points from the training set. Regularization is introduced on the diagonal. This yields a semidefinite matrix representing correlations in a higher-dimensional space than the original training data.
- 6.
-
Solving for Estimator:Solving for the coefficients a in the equation:Wherey=Ka,y represents vector values corresponding to the training set,K is the correlation matrix, anda is the set of coefficients to be determined. Efficient methods like QR/Cholesky can be used to invert the correlation matrix.
- 7.
-
Forming the Estimator:Once a is known, the estimator can be formulated using the coefficients and the chosen kernel. For test points, the estimator calculates y* using a and the kernel function. The result of this estimation isy* =a* k.
IV. Ensemble Learning Models
A. Random Forest
- There should be some actual values in the feature variable of the dataset so that the classifier can predict accurate results rather than a guessed result.
- The predictions from each tree must have very low correlations.
- It takes less training time as compared to other algorithms.
- It predicts output with high accuracy, even for the large dataset
- It runs efficiently.It can also maintain accuracy when a large proportion of data is missing.

B. Bagging Regressor
C. ADA Boost Regressor
- Initialize sample weights as wnl =I for n=l..N.
- For m=l..M, calculate:
- Sample probabilities pn = wnm / In wnm for all n.
- Sample data (Xm ,ym ) by selecting N samples from (X,y) using pn
- Fit weak learner fin to (Xm ,ym ).
- Compute loss in for each sample using predictions y’ from fin and adhering to equation (3).
- Compute average loss i- and terminate if i- 0.5, as indicated in step 4.
- Calculate confidence measure m = i- I (I - i-).
- Update sample weights wnm+ I = wnm m I-in
D. Gradient Boosting Regressor
J: Steepest Descent:For M-stage gradient boosting, steepest descent identifies where is a constant step length, and gim is the gradient of loss function L(f).V. Performance Metrics
A. Root Mean Square Error (RMSE)
B. Mean Squared Error (MSE)
C. Mean Absolute Error (MAE)
VI. Dataset Information
A. Numerical Weather Prediction Data
VII. Results and Discussion
A. Machine Learning Models




- -
- Decision Tree Regression(DTR) has the lowest RMSE, MSE, and MAE values, indicating better predictive accuracy compared to other algorithms.
- -
- MLR, Ridge, and Lasso have fairly similar performance in terms ofRMSE, MSE, and MAE, with Ridge slightly outperforming Lasso.
- -
- Support Vector Regression(SVR) has higher RMSE, MSE, and MAE values, indicating that it might not be capturing the underlying patterns as effectively as the linear-based methods or SVR.
- -
- Decision Tree Regression also has the highest R-squared value, implying a better fit to the data compared to other algorithms.
B. Ensemble Learning Models




- -
- Gradient Boosting Regressor has the lowest RMSE, MSE, and MAE values, indicating better predictive accuracy compared to other algorithms.
- -
- ADA Boost Regressor has higher RMSE, MSE, and MAE values, indicating that it might not be capturing the underlying patterns as effectively as the linear-based methods or SVR.
- -
- Gradient Boostig Regressor also has the highest R-squared value, implying a better fit to the data compared to other algorithms.
| MODEL | MSE | RMSE | MAE | W’2 |
|---|---|---|---|---|
| MULTIPLE LINEAR REGRESSION | 352465.7 | 559.6 | 412.27 | 0.6 |
| RIDGE REGRESSION | 252465.6 | 502.46 | 386.69 | 0.73 |
| LASSO REGRESSION | 252465.6 | 502.46 | 386.69 | 0.73 |
| SUPPORT VECTOR REGRESSION | 557635.43 | 746.75 | 616.352 | 0.39 |
| DECISION TREE REGRESSION | 222190.13 | 471.370 | 286.028 | 0.75 |
| RANDOM FOREST REGRESSION | 164088.8 | 405.07 | 258.06 | 0.82 |
| BAGGING REGRESSOR | 178778.3 | 422.82 | 265.82 | 0.81 |
| ADA BOOST REGRESSOR | 284581.40 | 533.46 | 427.64 | 0.69 |
| GRADIENT BOOSTING REGRESSOR | 158559.33 | 399.44 | 253.62 | 0.83 |
VIII. Conclusion
IX. Future Scope
References
- waone Gaboitaolelwe, Adamu Murtala Zungeru, Abid Yahya,” Machine Learning Based Solar PhotovoltaicPower Forecasting: A Review and Comparison”. [CrossRef]
- Rai, A. Srivastava, and K. C. Jana, “An empirical analysis of machine learning algorithms for solar power forecasting in a high dimensional uncertain environment,” IETE Tech. Rev., pp. 1-16, Nov. 2022. [CrossRef]
- S. Ghazi and K. Ip, “The effect ofweather conditions on the efficiency ofPV panels in the southeast ofU.K.,” Renew. Energy, vol. 69, pp. 50-59, Sep. 2014. [CrossRef]
- Sheth, K., & Patel, D. (2024). Comprehensive examination of solar panel design: A focus on thermal dynamics.
- Smart Grid and Renewable Energy, 15(1). [CrossRef]
- D. Yang, J. Kleissl, C. A. Gueymard, H. T. C. Pedro, and C. F. M. Coimbra, “History and trends in solar irradiance and PV power forecasting: A preliminary assessment and review using text mining,’’ Sol. Energy, vol. 168, pp. 60-101, Jul. 2018. [CrossRef]
- S. Leva, A. Dolara, F. Grimaccia, M. Mussetta, and E. Ogliari, “Analysis and validation of 24 hours ahead neural network forecasting of photovoltaic output power,” Math. Comput. Simul., vol. 131, pp. 88-100, Jan. 2017. [CrossRef]
- D. V. Pombo, H. W. Bindner, S. V. Spataru, P. E. S0rensen, and P. Bacher, “Increasing the accuracy of hourly multi-output solar power forecast with physics-informed machine learning,” Sensors, vol. 22, no. 3, p. 749, Jan. 2022. [CrossRef]
- Nouri, H., Hodge, B.-M., & Karimzadeh “Solar Power Forecasting: A Review of State-of-the-Art Methodologies” (2020). [CrossRef]
- Inman, R. H., & Pedro, H. T. C.”Machine Leaming for Solar Power Forecasting: A Comprehensive Review” (2020). [CrossRef]
- Amjady, N., & Keynia, “An Overview of Solar Power Forecasting and Prediction Intervals” (2020).
- Renewable and Sustainable Energy Reviews, 131, 110029. [CrossRef]
- Aghaei, J., & Fotuhi-Firuzabad,”Short-Term Solar Power Forecasting Using Machine Leaming Techniques” (2020), Energy Reports, 6, 1738-1753. [CrossRef]
- Sheth, K., & Patel, D. (2024). Strategic placement of charging stations for enhanced electric vehicle adoption in San Diego, California. Journal of Transportation Technologies, 14(1), Article 141005. [CrossRef]
- Swami, G., Sheth, K., & Patel, D. (2024). PV capacity evaluation using ASTM E2848: Techniques for accuracy and reliability in bifacial systems. Smart Grid and Renewable Energy, 15(9), Article 159012. [CrossRef]
| Variable Name | NWP | Unit |
|---|---|---|
| temperature_2_m_above_gnd relative_humidity_2_m_above_gn d mean_sea_level_pressure_MSL total_precipitation_sfc snowfall amount sfc - - total cloud cover sfc - - - high_cloud_cover_high_cld_lay medium cloud cover mid cld la - - - - -y low_cloud_cover_low_cld_lay shortwave radiation backwards s - - - fc wind_speed_ IO_m_above_gnd wind_direction_ IO_m_above_gnd wind_speed_80_m_above_gnd wind_direction_80_m_above_gnd wind_speed_900_mb wind direction 900 mb - - - wind_gust_ IO_m_above_gnd angle_of_incidence zenith azimuth generated_power kw |
Temperature At 2m Above Ground Relative Humidity At Above Ground Mean Sea Level Pressure Total Precipitation Snowfall Amount Total Cloud Cover High Cloud Cover Medium Cloud Cover Low Cloud Cover Short Wave Radiation Backwards Wind Spedd At IOm Above Ground Wind Direction At IOm Above Ground Wind Speed 80m At Above Ground Wind Direction At 80m Above Ground Wind Speed Wind Direction Wind Gust At IOm Above Ground Angle Of Incidence Zenith Azimuth Generated power |
C % % % % % % % %W/m2 mis mis mis mis mis mis mis degree degree degree KW |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).