0% found this document useful (0 votes)
21 views16 pages

Final Research Paper

Uploaded by

sirajuddina065
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
21 views16 pages

Final Research Paper

Uploaded by

sirajuddina065
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 16

Page 1 of 16 - Cover Page Submission ID trn:oid:::1:3046018070

Amit Agrawal
Project 2025
Shikhar

Final Year Project

ABES Engineering College

Document Details

Submission ID

trn:oid:::1:3046018070 13 Pages

Submission Date 3,929 Words

Oct 18, 2024, 1:29 PM GMT+5:30


22,007 Characters

Download Date

Oct 18, 2024, 1:32 PM GMT+5:30

File Name

final_research_paper.pdf

File Size

598.5 KB

Page 1 of 16 - Cover Page Submission ID trn:oid:::1:3046018070


Page 2 of 16 - Integrity Overview Submission ID trn:oid:::1:3046018070

48% Overall Similarity


The combined total of all matches, including overlapping sources, for each database.

Filtered from the Report


Bibliography

Quoted Text

Match Groups Top Sources

80 Not Cited or Quoted 46% 9% Internet sources


Matches with neither in-text citation nor quotation marks
47% Publications
6 Missing Quotations 2% 1% Submitted works (Student Papers)
Matches that are still very similar to source material

0 Missing Citation 0%
Matches that have quotation marks, but no in-text citation

0 Cited and Quoted 0%


Matches with in-text citation present, but no quotation marks

Integrity Flags
0 Integrity Flags for Review
Our system's algorithms look deeply at a document for any inconsistencies that
No suspicious text manipulations found. would set it apart from a normal submission. If we notice something strange, we flag
it for you to review.

A Flag is not necessarily an indicator of a problem. However, we'd recommend you


focus your attention there for further review.

Page 2 of 16 - Integrity Overview Submission ID trn:oid:::1:3046018070


Page 3 of 16 - Integrity Overview Submission ID trn:oid:::1:3046018070

Match Groups Top Sources

80 Not Cited or Quoted 46% 9% Internet sources


Matches with neither in-text citation nor quotation marks
47% Publications
6 Missing Quotations 2% 1% Submitted works (Student Papers)
Matches that are still very similar to source material

0 Missing Citation 0%
Matches that have quotation marks, but no in-text citation

0 Cited and Quoted 0%


Matches with in-text citation present, but no quotation marks

Top Sources
The sources with the highest number of matches within the submission. Overlapping sources will not be displayed.

1 Publication

Amit Rai, Ashish Shrivastava, K. C. Jana. "An Empirical Analysis of Machine Learni… 44%

2 Internet

www.researchgate.net 2%

3 Publication

Pawan Singh Mehra, Dhirendra Kumar Shukla. "Artificial Intelligence, Blockchain,… 1%

4 Publication

Enas Raafat Maamoun Shouman. "Chapter 2 Solar Power Prediction with Artificial… 1%

5 Internet

www.streetinsider.com 0%

6 Internet

iea-shc.org 0%

7 Internet

sedok.narod.ru 0%

Page 3 of 16 - Integrity Overview Submission ID trn:oid:::1:3046018070


Page 4 of 16 - Integrity Submission Submission ID trn:oid:::1:3046018070

2
Analysis Of Machine Learning Algorithms For Solar Power
Forecasting

ABSTRACT
6 In the realm of energy generation, solar power stands out as one of the most promising clean
1 energy alternatives to non-renewable sources. However, its reliance on environmental factors
4 introduces uncertainty in energy output. In this context, solar power forecasting plays a crucial
role in reducing this uncertainty and enhancing the overall stability of energy systems. Solar
2 power forecasting helps reduce uncertainty and improve system stability. Recently, machine
learning (ML) models have been widely used for solar power forecasting and design. However,
2 careful consideration of data pre-processing, forecast horizons, and the evaluation of ML model
performance is necessary to identify the most accurate model. This paper offers a comparative
1 analysis of various ML models for solar power forecasting, providing insights into future
research and guiding the selection of appropriate methods based on the strengths and
2 weaknesses of each ML model. Therefore, an effective forecasting approach is determined by
factors such as error performance, convergence time, and computational complexity. This study
evaluates various ML models based on error metrics and convergence time. Additionally, cross-
validation and hyperparameter tuning are explored for the top five performing models to
provide a more thorough evaluation. This offers valuable insights to stakeholders involved in
solar power plant modelling.
KEYWORDS Bagging; linear regression; machine learning; solar power; regularization
method.
ABBREVIATION
LR Linear regression SVM Support vector machine
DT Decision tree SP Solar power
ML Machine learning KNN K Neighbors regression
GBR Gradient boosting regression EN Elastic net
MAE Mean absolute error LAR Least angle regression
SPF Solar power forecasting OMP Orthogonal matching pursuit
MA Moving average LASSO Lasso regression

1. INTRODUCTION
The global demand for electricity is rising due to increasing industrialization and population
growth, making it a key factor in a nation's overall development. Between 2000 and 2018,
energy generation increased by about 3% per year, reflecting worldwide economic growth
1 during that period [1]. However, coal-based power plants produce greenhouse gas emissions
1 and rely on non-renewable resources, which are finite. As a result, the focus is now shifting
toward renewable energy systems [2]. At present, hydropower plays a significant role among
1 renewable energy sources. However, solar power is more widely available and holds great
potential to meet the increasing global energy demand [3]. In recent years, the contribution of
1 solar power plants (SPPs) to the energy grid has been steadily growing. This rise in solar power
brings economic, environmental, and installation advantages, especially in remote areas.

Page 4 of 16 - Integrity Submission Submission ID trn:oid:::1:3046018070


Page 5 of 16 - Integrity Submission Submission ID trn:oid:::1:3046018070

However, the reliance on environmental factors makes solar power generation unpredictable,
which can pose challenges to grid stability [4]. To address the stability issues caused by the
1 unpredictability of solar power, a backup power source or battery storage system is needed [5].
However, these solutions are not cost-effective alternatives. Therefore, solar power forecasting
is a crucial step in managing the unpredictable output of solar power plants (SPPs) and
increasing profitability [6,7]. These methods work well with a single input variable, but the
complexity grows as more independent variables are added. Additionally, the limited
1 availability of data from solar power plants (SPPs) reduces the accuracy of solar power
1 predictions [8]. These traditional models rely on past data and use it to predict future solar
1 power (SP) output for the next time interval [9]. In data-driven approaches, various inputs like
solar radiation, solar power, wind speed, humidity, sun angle, temperature, and weather
1 conditions are provided to the algorithm. The algorithm then establishes a regression
1 relationship between the input and output variables to predict solar power (SP) [10]. The auto-
regressive class of statistical methods analyses components like auto-regressive (AR),
integrated (I), and moving averages (MA) from time-series data. These statistical variables are
then used to create a relationship between the input and target variables. ARIMA and SARIMA
models are commonly used for solar power forecasting in the literature [11-14].
While these methods are effective for short-term predictions, their performance tends to decline
for long-term forecasts, and existing literature often presents a biased comparison of their
effectiveness. Additionally, adjusting the statistical parameters can be quite challenging. Linear
regression offers an advantage over AR models because it doesn't require parameter tuning. It
1 seeks to establish a linear relationship between input and target variables. These methods are
1 straightforward, easy to implement, and require fewer resources compared to advanced
machine learning (ML) techniques, making them widely used in solar power forecasting [15-
4 21]. Non-linear methods like support vector machines, random forests, ensemble learning, and
gradient boosting offer advantages over the previously mentioned techniques by effectively
capturing the non-linear patterns in solar power or radiation data [22].
1 The input parameters for predicting solar power are quite extensive and vary across different
1 studies. Most research typically includes solar radiation, wind speed, and ambient temperature
as key input factors for prediction models. The performance metrics used in these studies
mainly include root mean square error (RMSE), mean absolute error (MAE), mean square error
(MSE), and mean absolute percentage error (MAPE) to assess the proposed models.

1.1 Contribution of Present Work


Solar power forecasting (SPF) is essential for the stable and profitable operation of solar power
plants. Machine learning models are increasingly favoured for their robustness and accuracy,
yet existing research often overlooks the importance of computational time in these techniques.
1 (1) This work also offers a thorough assessment of machine learning models using five
different statistical error metrics. The top five ML models are evaluated further through
residual error curves and validation curves to confirm the best regression model
available.
(2) Additionally, the strengths and weaknesses of various machine learning methods are
discussed to enhance understanding and help in choosing the most suitable approach.

Page 5 of 16 - Integrity Submission Submission ID trn:oid:::1:3046018070


Page 6 of 16 - Integrity Submission Submission ID trn:oid:::1:3046018070

2. METHODS AND MATERIAL


1 This section outlines the data used for analysis, the machine learning methods employed, and
the statistical errors considered for comparing the ML algorithms.
2.1 Data
The accuracy and performance of any prediction method largely rely on the input data and the
choice of suitable machine learning techniques. Therefore, the size of the dataset and its pre-
processing are key factors in the success of any machine learning-based solar power forecasting
1 method. This study uses a publicly available dataset from PV-GIS [23]. Preparing and cleaning
1 input data is essential for enhancing prediction accuracy in forecasting. For this study, the input
data were obtained from PVGIS for the Ghaziabad region in India, located at a latitude of 28.7°
and a longitude of 77.391°. The solar power plant has a capacity of 10,000 kWp and uses
5 crystalline silicon panels. The dataset covers the period from January 1, 2015, to December 31,
1 2016. It includes seven feature vectors: solar power (in watts), reflected radiation (in W/m²),
direct radiation (in W/m²), diffuse radiation (in W/m²), sun height (in degrees), temperature (in
degrees Celsius), and wind speed at a height of 10 meters (in m/s).

1 Figure 2: Correlation matrix of input parameters

Page 6 of 16 - Integrity Submission Submission ID trn:oid:::1:3046018070


Page 7 of 16 - Integrity Submission Submission ID trn:oid:::1:3046018070

1 2.2 Statistical Method of Solar Power Forecasting


Statistical time-series forecasting relies on historical data. It can be divided into traditional
methods and machine learning methods. In this context, the classical approach being discussed
is the ARIMA model.
2.2.1 ARIMA Method
1 ARIMA is a statistical method used for prediction that relies on historical time-series data. Its
accuracy in forecasting depends on selecting the moving average (ap) and autoregressive (bq)
parameters based on the statistical properties of the dataset. The general formula for the
predicted target value (y^) is:
7 y^=c+a1yt-1+….+apyt-p-b1et-1-….-bqet-q
where e represents the error with lagged values.
One advantage of ARIMA is that it requires fewer datasets. However, its prediction accuracy
is heavily influenced by the correlation between present and past time points, making it less
suitable for long-term forecasts. Additionally, estimating the parameters for ARIMA can be
complicated, which can impact forecasting accuracy. For these reasons, many researchers
1 prefer machine learning methods, which automatically adjust the weights and biases of the
model.
2.2.2 Machine Learning Method
1 Machine learning is one of the fastest-growing fields in computer science and is becoming
increasingly popular in various engineering and social decision-making areas. In machine
learning, computers learn from training data. When given enough input data, machine learning
algorithms can make predictions that are often more accurate than those produced by other
methods.
1 First, the imported dataset is checked for any missing values. These missing values can be
replaced with the mean or median of the entire dataset. Using these methods to handle missing
1 values does not affect the overall mean or median. After addressing the missing values, the data
are normalized or standardized. This step ensures that all variables in the dataset are on the
same scale and helps reduce prediction variance. Standardization works best for datasets with
constant means, while normalization is often more accurate. Normalization is especially
important for renewable energy datasets due to their non-linear characteristics.
1 These steps are part of the preprocessing stage. Next, the pre-processed data are divided into
training and testing sets. During the training phase, machine learning models learn from the
training data and develop a relationship to predict the target variable. In the following step,
1 they make predictions on unseen data for validation. Similar training and validation errors
indicate the effectiveness of the machine learning model for predictions.
1 In machine learning analysis and prediction, a preprocessing step is conducted to clean and
prepare the input data. The process begins with importing the raw data into the model.
1 Regression maps input space x(i)=x1,x2,x3,…..xm,to output y(i)=y1,y2,y3,…..ym through function
1 space f(x(i)). Fa….Fn are sample features and y^1,y2^,…..,y^n are predicted vectors.

Page 7 of 16 - Integrity Submission Submission ID trn:oid:::1:3046018070


Page 8 of 16 - Integrity Submission Submission ID trn:oid:::1:3046018070

Figure3: Flow chart of the steps involved in machine learning prediction.

2.3 Statistical Errors


1 With many machine learning algorithms available, choosing the most accurate model for
1 predicting solar power is the hardest part of forecasting. As shown in Table 1, addressing both
bias and variance simultaneously is a challenge for each machine learning method. Therefore,
statistical errors are the key factor in evaluating the performance of any machine learning
algorithm.
Table 1:
Regression Advantages Disadvantages
Techniques

Linear Regression 1-Easy implementation 1-Valid for linear problems.


2-Testing is fast 2-Leck of correlation.

Page 8 of 16 - Integrity Submission Submission ID trn:oid:::1:3046018070


Page 9 of 16 - Integrity Submission Submission ID trn:oid:::1:3046018070

1 k-NN 1- Easy implementation 1-The value of k is crucial.


2-Less hyperparameters are required 2-More time required.

DT 1-No need of pre-processing 1-More tree more complex.


2-Easy to explain 2-Outliers affect the answer.

SVR 1-More accurate 1-Difficult to interpret.


2-Good performance for high data 2-Slower rate of testing.

1 Regularization 1-Efficient 1-Error due to reduction.


2-Computation is easy 2-Bias-mean/variance trade-
off is tough.

1 The performance of machine learning models is assessed using standard statistical indicators
[34].

1 Here, S(t)i is the actual value solar power and S^i(t) is the predicted output.

The performance of machine learning models is measured using various statistical indicators,
1 such as mean absolute percentage error (MAPE), root mean squared error (RMSE), mean
absolute error (MAE), R² error, and root mean squared logarithmic error (RMSLE). MAE
1 represents the average of absolute prediction errors, as shown in Equation (3). R² indicates the
1 proportion of variation between the true and predicted values, shown in Equation (5). RMSE
is the square root of the mean squared error, and for an unbiased estimator, it reflects the
1 standard deviation of the predicted values, as shown in Equation (4). RMSLE measures the
1 logarithmic difference between actual and predicted values and is useful when both values are
large, as shown in Equation (6).

3.RESULT
Solar power is first predicted using the ARIMA model. The base parameters of the ARIMA
1 model are evaluated using the Augmented Dickey-Fuller (ADF) test and Akaike’s Information
Criterion (AIC), as shown in Appendix B. The best model identified is ARIMA (0,1,1) (1,1,0)

Page 9 of 16 - Integrity Submission Submission ID trn:oid:::1:3046018070


Page 10 of 16 - Integrity Submission Submission ID trn:oid:::1:3046018070

1 [12] which indicates a seasonal component in the dataset. This is also evident in the rolling
mean estimation, as shown in Figure 4.

1 Figure 5 compares the actual solar power to the predicted solar power. The prediction errors,
measured by MAPE and R², are 7.986 and 0.696, respectively.

Figure 4: seasonal trends in data set

1
3 In this section, the performance of machine learning models is evaluated on the dataset for 6-
hour predictions. The models are implemented in the Spyder notebook using the Python
programming language, along with the NumPy, pandas, and PyCaret libraries [24]. To assess
the performance of the machine learning algorithms, the models are implemented in PyCaret.

1 The models analysed include Extra Trees (ET), Light Gradient Boosting Machine (LGBM),
Extreme Gradient Boosting (XGB), Random Forest (RF), Gradient Boosting Regression
(GBR), Linear Regression (LR), Bayesian Ridge (BR), Least Absolute Shrinkage and Selection
Operator (LASSO), Ridge Regression (RR), Decision Tree (DT), k-Nearest Neighbors (KNN),
AdaBoost Regression (ABR), Elastic Net (EN), Orthogonal Matching Pursuit (OMP), Least
Angle Regression (LAR), and Support Vector Machine with a linear kernel (SVML). No single
1 machine learning model performs best across all types of time-series datasets. To allow for a
more intuitive comparison, single-step cross-validations with 5 and 10 steps are conducted on
the top five machine learning models.

Page 10 of 16 - Integrity Submission Submission ID trn:oid:::1:3046018070


Page 11 of 16 - Integrity Submission Submission ID trn:oid:::1:3046018070

1 Table 2: Output comparison of different ML models

Table 2 presents the performance metrics of 16 machine learning regression models. The
algorithms are compared using six different performance indicators: MAE, RMSE, R²,
RMSLE, MAPE, and computation time (in seconds). The table is sorted by MAE in ascending
order, from the lowest to the highest values.
As shown in Table 3, the Extra Trees Regression (ETR) model outperforms other machine
1 learning models in terms of MAE, RMSE, R², and MAPE, with values of 1.89×10⁴ W, 4.12×10⁴
W, 0.9997, and 0.0136, respectively. The RMSLE value is 0.1007, which measures the
logarithmic difference between actual and predicted values, and it doesn't penalize large
1 differences, making it useful when both values are large. It is also evident that ensemble
1 machine learning models perform better than other models in terms of MAE, RMSE, RMSLE,
1 and MAPE. The Support Vector Machine with a linear kernel (SVML) shows the poorest
performance in terms of statistical errors.
1 Figure 6 illustrates the performance of different categories of machine learning regression
1 techniques in terms of MAE and RMSE. The figure shows that the average MAE and RMSE
values for ensemble methods (ETR, LightGBM, XGBoost, RF, AdaBoost, and GBR) are
61,237.33 W and 87,019.41 W, respectively, which outperform other types of regression
models.
1 Figure 6: Error performance of different ML

Page 11 of 16 - Integrity Submission Submission ID trn:oid:::1:3046018070


Page 12 of 16 - Integrity Submission Submission ID trn:oid:::1:3046018070

Computational Models

SVM Ensemble ML KNN LR Regularization

1 Figure 7: Time taken by different ML methods

Ensemble regression methods show better error performance, although their computational
time is higher compared to other regression methods due to the use of multiple tree-based
decisions for prediction, as shown in Figure 6. However, in terms of computational time,
regularization and linear regression methods outperform other machine learning models, with
average performance errors illustrated in Figure 7.
1 K-fold cross-validation involves splitting the dataset into K parts for training machine learning
models. Each part is used once for training and once for validating the algorithms. This method
1 shuffles the dataset and divides it into K groups, retaining each group's performance until
training is complete. It helps reduce bias in the prediction method.
1 Residuals in regression indicate the vertical distance between the predicted values and the
regression line. They clearly show the error between the observed and predicted values. Ideally,
1 the residuals of any machine learning model should be normally distributed and independent.
Additionally, the points should be densely packed around the origin and exhibit symmetry or a
1 normal distribution around it. The variation of the R² value for the training data of the ensemble
methods ranges from 1 to 0.979, while for the test data, it ranges from 0.974 to 0.977. These
R² values indicate the effectiveness of the top five ensemble regression techniques.
1 Table 5: Conclusion of hyperparameter tuning of top 5 ML models

Page 12 of 16 - Integrity Submission Submission ID trn:oid:::1:3046018070


Page 13 of 16 - Integrity Submission Submission ID trn:oid:::1:3046018070

1
3 The performance of machine learning models depends on the selection of model parameters
and hyperparameters. The choice of hyperparameter values guides the training process on the
dataset and influences how well the model converges to an optimal solution. Each machine
1 learning model has different hyperparameters, and their selection can vary based on the
characteristics of the datasets. In this study, the hyperparameters were evaluated using the
PyCaret package. The random search cross-validation (CV) method was used to explore the
1 default combinations of hyperparameters provided by the scikit-learn library. The reason for
choosing the default scikit-learn search space is that it offers a wide range of hyperparameter
options.
1 Table 5 presents the results of hyperparameter tuning for most of the ensemble methods. Only
3 ensemble methods were selected for tuning due to their superior performance compared to
other machine learning models. The tuning process improved the results of Gradient Boosting
1 Regression (GBR) and LightGBM by 50.47% and 1.75%, respectively, in terms of the loss
function. In this context, "Ht" refers to the tuned model output, while "D" indicates the default
parameters.

1 Figure 8: Power prediction of 6 hours

4. CONCLUSION

Page 13 of 16 - Integrity Submission Submission ID trn:oid:::1:3046018070


Page 14 of 16 - Integrity Submission Submission ID trn:oid:::1:3046018070

Choosing an accurate machine learning model for solar power forecasting is a challenging task,
and many approaches have been explored in various studies using different ML models. This
1 work offers a detailed review and empirical analysis of solar power forecasting techniques,
including ARIMA-based classical time series methods and 16 different ML models. ARIMA-
based models require relatively fewer data points, but their limitations in long-term predictions
and parameter estimation reduce their effectiveness.
1 While machine learning models are already employed for solar power prediction, many
advanced ML techniques have not been thoroughly investigated or compared in a single study.
1 This work aims to provide a more comprehensive review of these techniques. The performance
evaluation of the ML models indicates that ensemble models offer better forecasting accuracy
than other ML models. The ability to train multiple nodes and combine their predictions gives
ensemble methods an advantage over other approaches.

1 Among the ensemble methods, Extra Trees Regression has outperformed the other models.
However, computational time is a significant limitation for the real-time implementation of
machine learning models. Regularization ML models have the lowest computational time
compared to all ML models. While ensemble models provide accurate forecasting results, they
1 require more computational time. This work offers a practical approach by providing a
systematic review of future research and opportunities for stakeholders to choose an accurate
machine learning model for solar power prediction.

REFERENCES

1. IEA, International Renewable Energy Agency, United Nations Statistics Division, The
World Bank, and World Health Organization, “The energy progress report,” IEA, IRENA,
UNSD, WB, WHO (2019), Track. SDG 7 Energy Progress Report. 2019, Washingt. DC,
2019.
2. Z. Şen, "Solar energy in progress and future research trends," Prog. Energy Combust. Sci.
, vol. 30, no. 4, pp. 367-416, 2004, DOI: 10.1016/j.pecs.2004.02.004.
3. IRENA, Renewable Capacity Statistics 2019.AbuDhabi: International Renewable Energy
Agency (IRENA), 2019. Available: https://www.irena.org/publications/2019/Mar/
Renewable-Capacity-Statistics-2019.
4. S. B. Nam, and J. Hur, “A hybrid spatio-temporal forecasting of solar generating resources
for grid integration,” Energy, Vol. 177, pp. 503–10, 2019. DOI:
10.1016/j.energy.2019.04.127.
5. V.Bagalini,B.Y.Zhao,R.Z.Wang,andU.Desideri,“Solar PV-battery-electric grid-based
energy system for residential applications: system configuration and viability,” Research,
Vol. 2019, pp. 1–17, 2019.DOI:10.34133/2019/ 3838603.
6. N.Dong, J.F.Chang, A.G.Wu,and Z.K.Gao,“A novel convolutional neural network
framework based solar irradiance prediction method,” Int. J. Electr. Power Energy Syst.,
Vol. 114, no. July 2019, pp. 105411, 2020. DOI: 10.1016/j.ijepes.2019.105411.

Page 14 of 16 - Integrity Submission Submission ID trn:oid:::1:3046018070


Page 15 of 16 - Integrity Submission Submission ID trn:oid:::1:3046018070

7. A. E. Gürel, Ü Ağbulut, and Y. Biçen, “Assessment of machine learning, time series,


response surface method ology and empirical models in prediction of global solar
radiation,” J. Clean. Prod., Vol. 277, pp. 122353, 2020.DOI:
10.1016/j.jclepro.2020.122353.
8. Y. Hao, and C. Tian, “A novel two-stage forecasting model based on error factor and
ensemble method for multi-step wind power forecasting,” Appl. Energy, Vol. 238, no. July
2018, pp. 368–83, 2019.DOI:10.1016/j.apenergy.2019.01. 063.
9. U. K.Das, et al., “Forecasting of photovoltaic power gener ation and model optimization:
A review,” Renew. Sustain. Energy Rev., Vol. 81, no. June 2017, pp. 912–28, 2018.DOI:
10.1016/j.rser.2017.08.017.
10. . A.Ahmed, and M.Khalid,“A review on the selected applications of forecasting models in
renewable power systems,” Renew. Sustain. Energy Rev., Vol. 100, no. September 2018,
pp. 9–21, 2019.DOI:10.1016/j.rser.2018.09.046.
11. G. Reikard, “Predicting solar radiation at high resolutions: A comparison of time series
forecasts,” Sol. Energy,Vol. 83, no. 3, pp. 342–9, 2009.DOI: 10.1016/j.solener.2008.
08.007.
12. R. Huang, T. Huang, R. Gadh, and N. Li, “Solar generation prediction using the ARMA
model in a laboratory-level micro-grid,” in 2012 IEEE 3rd International Conference on
Smart Grid Communication SmartGridComm 2012, 2012, pp. 528–33. DOI:
10.1109/SmartGridComm.2012.6486 039.
13. .M.David, F. Ramahatana, P. J. Trombe, andP. Lauret, “Probabilistic forecasting of the
solar irradiance with recursive ARMA and GARCH models,” Sol. Energy, Vol. 133, pp.
55–72, 03.064.
14. M.Bouzerdoum, A. Mellit, and A. Massi Pavan, “A hybrid model (SARIMA-SVM) for
short-term power forecasting of a small-scale grid-connected photovoltaic plant,” Sol.
Energy, Vol. 98, no. PC, pp. 226–35, 2013. DOI: 10.1016/j.solener.2013.10.002.
15. D.Yang,P.Jirutitijaroen,andW.M.Walsh,“Hourly solar irradiance time series forecasting
using cloud cover index,” Sol. Energy, Vol. 86, no. 12, pp. 3531–43, 2012. DOI:
10.1016/j.solener.2012.07.029.
16. . M. Abuella, and B. Chowdhury, “Solar power probabilistic forecasting by using multiple
linear regression analysis,” in Conference proceedings– IEEE SOUTHEAST CON, vol.
2015-June, no. June, 2015, pp. 5–9, DOI: 10.1109/SECON.2015.7132869.
17. H.Sheng, J.Xiao,Y.Cheng,Q.Ni,andS.Wang, “Short term solar power forecasting based on
weighted Gaussian process regression,” IEEE Trans. Ind. Electron., Vol. 65, no. 1, pp. 300–
8, 2018.DOI:10.1109/TIE.2017.2714127.
18. C. Persson, P. Bacher, T. Shiga, and H. Madsen, “Multi site solar power forecasting using
gradient boosted regres sion trees,” Sol. Energy, Vol. 150, pp. 423–36, 2017.DOI:
10.1016/j.solener.2017.04.066.
19. . M.Zamo,O.Mestre, P. Arbogast, and O. Pannekoucke, “A benchmark of statistical
regression methods for short-term forecasting of photovoltaic electricity production. Part
II: Probabilistic forecast of daily production,” Sol. Energy,Vol. 105, pp. 804–
16,2014.DOI:10.1016/j.solener.2014.03.026.
20. G.Wang,Y.Su,andL.Shu,“One-day-ahead daily power forecasting of photovoltaic systems
based on partial functional linear regression models,”Renew.Energy,Vol.96,pp. 469–78,
2016.DOI:10.1016/j.renene.2016.04.089.

Page 15 of 16 - Integrity Submission Submission ID trn:oid:::1:3046018070


Page 16 of 16 - Integrity Submission Submission ID trn:oid:::1:3046018070

21. . R.AmaroeSilva, andM.C.Brito, “Impact of network layout and time resolution on spatio-
temporal solar fore casting,” Sol. Energy, Vol. 163, no. November 2017, pp. 329–37,
2018.DOI:10.1016/j.solener.2018.01.095.
22. M. H. D. M. Ribeiro, and L. dos Santos Coelho, “Ensem ble approach based on bagging,
boosting and stacking for short-term prediction in agribusiness time series,” Appl. Soft
Comput. J., Vol. 86, pp. 105837, 2020.DOI: 10.1016/j.asoc.2019.105837.
23. T. a Huld, M. Súri, E. D. Dunlop, M. Albuisson, and L. Wald, “Integration of
HELIOCLIM-1 database into PV GIS to estimate solar electricity potential in Africa,”
in20th European Photovoltaic Solar Energy Conference Exhibition, 2005, pp. 2989.
Available: c:/pdflib/00018470.pdf.
24. M. Ali, “PyCaret.” 2020. Available: https://pycaret.org/.

Page 16 of 16 - Integrity Submission Submission ID trn:oid:::1:3046018070

You might also like