Prediction of Wind Farm Power Ramp Rates: a Data-Mining Approach

Haiyang Zheng Andrew Kusiak e-mail: andrew-kusiak@uiowa. edu Department of Mechanical and Industrial Engineering, 3131 Seamans Center, University of Iowa, Iowa City, IA 52242-1527 Prediction of Wind Farm Power Ramp Rates: A Data-Mining Approach In this paper, multivariate time series models were built to predict the power ramp rates of a wind farm. The power changes were predicted at 10 min intervals. Multivariate time series models were built with data-mining algorithms. Five different data-mining algorithms were tested using data collected at a wind farm.

We Will Write a Custom Essay Specifically
For You For Only $13.90/page!


order now

The support vector machine regression algorithm performed best out of the ? ve algorithms studied in this research. It provided predictions of the power ramp rate for a time horizon of 10–60 min. The boosting tree algorithm selects parameters for enhancement of the prediction accuracy of the power ramp rate. The data used in this research originated at a wind farm of 100 turbines. The test results of multivariate time series models were presented in this paper. Suggestions for future research were provided. DOI: 10. 1115/1. 142727 Keywords: power ramp rate prediction, wind farm, data-mining algorithms, multivariate time series model, parameter selection 1 Introduction Wind power generation is rapidly expanding and is becoming a noticeable contributor to the electric grid. The fact that most largescale wind farms were developed in recent years has made studies of their performance overdue. Given the changing nature of the wind regime, wind farm power varies across all time scales. The ? uctuating power of wind farms is usually balanced by the power produced by the traditional power plants to meet the grid requirements.

The change of power output in time is referred to as ramping and it is measured with the power ramp rate PRR . The prediction of PRR at 10 min intervals is of interest to the wind industry due to the tightening electric grid requirements 1 . Though the power prediction research has a long tradition in the wind industry, the interest in prediction of power ramps is emerging. There is no industry standard for PRR prediction. Power ramp rate on 10 min intervals is to bene? t the gird management and power scheduling in the wind industry. The literature related to power ramps is discussed next.

Svoboda et al. 2 proposed a Lagrangian relaxation method to solve hydrothermal generation scheduling problems. Three PRR constraints were considered and illustrated with a numerical example. Ummels et al. 3 presented a simulation method to evaluate the integration of large-scale wind farm power with the conventional power generation sources from a cost, reliability, and environmental perspective. Based on the PRR constraints for the reserve activation and generation schedule, the capability of a thermal generation system for balancing a wind power was investigated.

Potter and Negnevitsky 4 applied an adaptive-neuron-fuzzy inference approach to forecast short-term wind speed and direction. Torres et al. 5 used transformed data to build the autoregressive moving average ARMA time series model for prediction of mean hourly wind speed of up to 10 h into the future. Sfetsos 6 presented a novel method for forecasting mean hourly wind speed based on the time series analysis data and showed that the developed model outperformed the conventional forecasting models.

Lange and Focken 7 presented various models for short-term wind power prediction, including physics-based, fuzzy, and neuContributed by the Solar Energy Engineering Division of ASME for publication in the JOURNAL OF SOLAR ENERGY ENGINEERING. Manuscript received August 10, 2008; ? nal manuscript received March 6, 2009; published online July 9, 2009. Review conducted by Spyros Voutsinas. rofuzzy models. Using meteorological data, Barbounis et al. 8 constructed a local recurrent neural network model for long-term wind speed and power forecasting. Hourly wind farm forecasts of up to 72 h were produced.

Developing power and PRR prediction models for wind farms is challenging, as power output is known to undergo rapid variations due to changes in the wind speed, e. g. , due to gusts. The power output strongly depends on the wind conditions and the changing environment of the wind farm. The stochastic nature of a wind farm environment calls for new modeling approaches to accurately predict the power ramp rate. Data mining is a promising approach for modeling wind farm performance. Numerous applications of data mining in manufacturing, marketing, medical informatics, and energy industry proved successful 9–14 .

In this paper, a data-mining approach was applied to build a multivariate time series model to predict power ramp rates of a wind farm over 10 min intervals. Five different data-mining algorithms for the PRR prediction were employed. The boosting tree algorithm was used to reduce the dimensionality of the input and to enhance prediction accuracy. The models were built using historical data collected by the supervisory control and data acquisition SCADA system installed at a wind farm. 2 Basic Methodologies for PRR Prediction 2. 1 Time Series Prediction Modeling.

Time series prediction 15 focuses on determining future events based on known observations, measured typically at successive time intervals often uniform . Time series models are generally applicable to monitoring industrial processes and tracking time-based business metrics. There are two types of time series models: univariate and multivariate models. The univariate time series model consists of observations of a single parameter recorded sequentially over equal time increments. In the multivariate time series model, observations are ? xed-dimension vectors of different parameter values.

The univariate time series prediction model 15,16 is expressed as follows: ? y t + wT = f y t ,y t ? T , . . . ,y t ? mT 1 where T is the sampling time interval , wT is the prediction horizon for example, for w = 2 and T = 10 min, the prediction hori? zon is 20 min , y t + wT is the predicted parameter, y t , y t AUGUST 2009, Vol. 131 / 031011-1 Journal of Solar Energy Engineering Copyright © 2009 by ASME Downloaded 02 Sep 2009 to 128. 255. 53. 136. Redistribution subject to ASME license or copyright; see http://www. asme. org/terms/Terms_Use. cfm ? T , . . . y t ? mT are the current and past observed parameters, and m + 1 is the number of inputs predictors of the model. The multivariate time series model 15 is formulated as follows: ? y t + wT = f y t ,y t ? T , . . . ,y t ? mT ;x1 t ,x1 t ? T , . . . , x1 t ? mT ;x2 t ,x2 t ? T , . . . ,x2 t ? mT ; . . . ; xn t ,xn t ? T , . . . ,xn t ? mT 2 where T is the sampling time interval , wT is the prediction horizon, x1 . . . , xn , y and n + 1 are the observations of the time series ? forming the n + 1 dimensional vector, y t + wT is the predicted parameter, y t , y t ?

T , . . . , y t ? mT are the current and past observed values of y, x1 t , x1 t ? T , . . . , x1 t ? mT are the current and past observed values of parameters x1 , . . . , xn, and m + 1 n + 1 is the number of inputs predictors of the model. To obtain an accurate prediction model with the data-mining approach, appropriate parameters predictors need to be selected. Data mining offers different algorithms to perform this task. For example, the boosting tree algorithm 17,18 and the wrapper approach 19,20 , utilizing the genetic or the ? st best search algorithm 13,21 select the important predictors. The total number of all possible predictors m + 1 n+1 forms a high-dimensional input to the time series model, and therefore the performance of the resultant model is likely to be inferior. To maximize performance of the prediction model, a boosting tree algorithm is employed to select a set of the most important predictors among the m + 1 n + 1 ones in Eq. 2 : y t ,y t ? T , . . . ,y t ? mT ;x1 t ,x1 t ? T , . . . , x1 t ? mT ; . . . ;xn t ,xn t ? T , . . . ,xn t ? mT 2. 2 Prediction Accuracy Metrics.

Two main metrics, the mean absolute error MAE and the standard deviation Std of the absolute error AE , were used to measure prediction accuracy of different data-mining algorithms. The small value of MAE and Std imply the superior prediction performance of the models extracted by data-mining algorithms. In fact, MAE and Std based on absolute error are widely used in the wind industry. Their de? nitions are expressed as ? AE = y t + wT ? y t + wT N 3 Fig. 1 Typical power, power ramp rate, and wind speed plots: „a… wind farm power, „b… power ramp rate, and „c… wind speed AE i MAE =

N i=1 N 4 of each turbine is 1. 5 MW, the capacity of the wind farm is 133. 5 MW. The power ramp rate used in this paper is de? ned as the rate of change of wind farm power during a 10 min interval the standard time interval in wind energy industry and is expressed in kW/ min: PRR = P t + 10 ? P t 10 6 AE i ? MAE Std = i=1 N? 1 5 ? where y t + wT is the predicted PRR, y t + wT is the observed measured PRR, and N is the number of test data points for the prediction model. The data set used by the PRR prediction models is divided into training and test data sets. 2. 3 Data Description.

The data used in this research were generated at a wind farm with 100 turbines. Though the data were sampled at high frequency, e. g. , 2 s, it was averaged and stored at 10 min intervals referred to as the 10 min average data . The data used in this research were collected over a period of 1 month for all turbines of the wind farm. Some data contained many missing values or abnormal values outside of the normal physical range, and thus 89 turbines were selected for the study. For example, the SCADA recorded wind speed should be in the range 0–20 m/s, and the power should be in the range 0–1600 kW.

As the rated power 031011-2 / Vol. 131, AUGUST 2009 where P t + 10 is the wind farm power at time t + 10 time t plus 10 min and P t is the wind farm power at time t. The power ramp rate expresses the rate of change of the wind farm power due to the stochastic nature of the wind. Figure 1 a illustrates the power produced by a wind farm over 10 min intervals. Figure 1 b shows the power ramp rate corresponding to the power presented in Fig. 1 a . Figure 1 c shows the wind speed for the time period considered in Figs. 1 b and 1 c . Ignoring the power consumed by the wind farm, the power produced is always positive Fig. a ; however, the PRR can be positive or negative. The positive PRR indicates increasing power over time, while the negative PRR value means that the wind farm power is decreasTransactions of the ASME Downloaded 02 Sep 2009 to 128. 255. 53. 136. Redistribution subject to ASME license or copyright; see http://www. asme. org/terms/Terms_Use. cfm Table 1 List of parameters Parameter Mean Std Max Min Power PRR Description Mean wind speed of a turbine Standard deviation of the wind speed of a turbine Maximum wind speed of a turbine Minimum wind speed of a turbine Wind farm power Power ramp rate of the wind farm Unit

Table 3 The importance index of predictors generated by the boosting tree algorithm for t + 10 model Predictor m/s m/s m/s m/s kW kW/min PRR-1 PRR-2 PRR-3 PRR-4 PRR-5 Mean-1 Mean-2 Mean-3 Mean-4 Mean-5 Min-1 Min-2 Min-3 Min-4 Min-5 Max-1 Max-2 Max-3 Max-4 Max-5 Std-1 Std-2 Std-3 Std-4 Std-5 Power-1 Power-2 Power-3 Power-4 Power-5 Variable rank 100 100 66 53 71 44 49 38 41 37 67 52 49 44 42 45 48 37 42 40 43 51 45 43 36 40 54 48 41 39 Importance 1. 00 1. 00 0. 66 0. 53 0. 71 0. 44 0. 49 0. 38 0. 41 0. 37 0. 67 0. 52 0. 49 0. 44 0. 42 0. 45 0. 48 0. 37 0. 42 0. 40 0. 3 0. 51 0. 45 0. 43 0. 36 0. 40 0. 54 0. 48 0. 41 0. 39 Table 2 The data set description Data set 1 2 3 Start time stamp 1/1/07 1:40 a. m. End time stamp 1/31/07 11:50 p. m. Description Total data set; 4455 observations Training data set; 3568 1/1/07 1:40 a. m. 1/25/07 8:00 p. m. observations Test data set; 887 1/25/07 8:10 p. m. 1/31/07 11:50 p. m. observations ing. The larger the absolute value of PRR, the faster the power surge or drop . The wind speeds of 89 turbines, the wind speed statistics, and the power collected by the SCADA system were used in data mining.

In this paper, six different parameters were used to build the multivariate time series model. The mean, Std, max, min, and power are the ? rst ? ve parameters x1 , . . . , x5 and the PRR is the sixth parameter y of model 2 . Table 1 lists all the parameters used in this paper. The number of parameters is limited by the data available in this research. The model accuracy could be enhanced if more data were available. The six parameters recorded at 10 min intervals resulted in 4455 instances data set 1 in Table 2 , beginning from “1/1/07 at 1:40 a. m. ” and continuing to “1/31/07 at 11:50 p. . ” During this time period, the overall wind farm performance was considered to be normal. Data set 1 was divided into two subsets: data set 2 and data set 3. Data set 2 contains 3568 data points and were used to develop a prediction model with data-mining algorithms. Data set 3 includes 887 data points and were used to test the prediction performance of the model extracted from data set 2. For the test data set, the MAE Eq. 4 and Std Eq. 5 were the metrics used to evaluate the data-mining algorithms applied to learn multivariate time series model of Sec. 2. 1. 2. 4 Parameter Selection.

Due to the high-dimensionality of the input vector of predictors of the multivariate time series model, the number of inputs was reduced. The quality of the mod- Fig. 2 The importance of predictors generated by the boosting tree algorithm for the t + 10 model Journal of Solar Energy Engineering AUGUST 2009, Vol. 131 / 031011-3 Downloaded 02 Sep 2009 to 128. 255. 53. 136. Redistribution subject to ASME license or copyright; see http://www. asme. org/terms/Terms_Use. cfm Fig. 3 Illustration of the multiperiod multivariate time series prediction model: „a… the t + 10 min PRR prediction and „b… the t + 20 min PRR prediction ls learned from high- and reduced-dimensionality data were compared in Secs. 3. 1 and 3. 2. The most signi? cant predictors were determined by the boosting tree algorithm 17,18 . The same approach was shown to be successful in a previous research 14 . The basic idea of the boosting tree algorithm is to build a number of trees e. g. , binary trees splitting the data set and to approximate the underlying function. The importance of each predictor is measured by its contribution to the prediction accuracy of the training data set.

To build a multivariate t + 10 time series model for 10 min ahead predictions , the value of m = 5 used in the multivariate model is selected, which means that four values observed in the past and one current value of each parameter are considered. In total, six different parameters of the multivariate model were considered and thus it contains 5 6 = 30 predictors. The 30dimensional input is reduced by the boosting tree algorithm. Table 3 shows the importance index of 30 predictors computed by the boosting tree algorithm based on data set 2 of Table 2.

The index “-1” in Table 3 indicates the observation sampled 10 min earlier, “-2” indicates the observation sampled 20 min earlier, and “-3, -4, and -5” indicate the observations sampled 30 min, 40 min, and 50 min earlier, respectively. Note that all the parameter values used in this paper were all average values over the 10 min interval. Figure 2 shows the importance of all 30 predictors for the t + 10 min models ranked from the largest to the smallest one. To maximize prediction accuracy it is important to select important predictors among the ones on the list y t ,y t ? T , . . . ,y t ? mT ;x1 t ,x1 t ?

T , . . . , x1 t ? mT ; . . . ;xn t ,xn t ? T , . . . ,xn t ? mT A threshold value of 0. 50 was established heuristically to select the predictors for the time series models. The predictors selected by the boosting tree algorithm for the t + 10 min PRR are PPR-1, PPR-2, PPR-5, Min-1, PPR-3, Power-2, PRR-4, Min-2, and Std-2. The number of predictors was reduced from 30 to 9. The threshold value of 0. 50 used in the computation produced good quality results. A lower threshold value would lead to more Table 4 Prediction error of the t + 10 models without parameter selection generated by the ? e different algorithms Absolute error kW/min MLP SVM C&R Fig. 4 Prediction results produced by the t + 10 model without parameter selection: „a… prediction performance of the ? ve different algorithms for the test data set of Table 2 and „b… the observed and predicted PPRs by the SVM algorithm predictors that could degrade performance of the models due to the “curse of dimensionality” principle 19,22 , which means that high-dimension input could negatively impact performance of the model built by the data-mining algorithm. 2. 5 Multiperiod Predictions With a Multivariate Time Series Model.

The t + 10 min prediction model is not suf? cient for integration of the wind farm with the power grid. Six different multivariate time series models are needed to predict the PRR at t + 10– t + 60 min intervals. For t + 10 interval prediction, data set 2 in Table 2 is used for parameter selection and building time series models with data-mining algorithms, and the test data data set 3 in Table 2 were used to validate performance of the models. For t + 20– t + 60 predictions, the training data set remains the same; however, the test data set containing 887 points is reduced by one for each of the next 10 min period predictions.

Figure 3 illustrates the concept of a multiperiod prediction for PRR over 10 min intervals. In this model, the sampling time period T is 10 min. Using the 10 min average measured values including mean, Std, max, min, power, and PRR in Table 1 at the intervals t = ? 50, t = ? 40 , . . . , t = ? 10, t = 0? , the average PRR value at the subsequent interval t + 10 is predicted Fig. 3 a . In Table 5 Prediction error of the t + 10 model with selected parameters generated by ? ve different algorithms Absolute error kW/min MLP SVM C&R MAE 340. 66 298. 94 360. 19 396. 62 312. 44 Std 448. 9 323. 32 407. 56 396. 62 342. 33 Maximum 5119. 73 2512. 34 2657. 89 4236. 02 3516. 80 Minimum 0. 03 0. 15 0. 15 0. 38 0. 03 MAE 280. 13 243. 14 307. 97 356. 79 290. 57 Std 309. 38 276. 39 335. 56 323. 92 318. 37 Maximum 3248. 12 2817. 77 3860. 94 3516. 65 3270. 62 Minimum 0. 16 0. 03 0. 61 0. 15 0. 03 Random forest tree Pace regression Random forest tree Pace regression 031011-4 / Vol. 131, AUGUST 2009 Transactions of the ASME Downloaded 02 Sep 2009 to 128. 255. 53. 136. Redistribution subject to ASME license or copyright; see http://www. asme. org/terms/Terms_Use. cfm

Table 6 The importance index of predictors generated by the boosting tree algorithm for t + 20 model Predictor Mean-1 Mean-2 Mean-3 Mean-4 Mean-5 Std-1 Std-2 Std-3 Std-4 Std-5 Max-1 Max-2 Max-3 Max-4 Max-5 Min-1 Min-2 Min-3 Min-4 Min-5 PRR-1 PRR-2 PRR-3 PRR-4 PRR-5 Power-1 Power-2 Power-3 Power-4 Power-5 Variable rank 54 50 41 39 31 40 46 48 46 32 68 61 42 47 36 33 46 31 32 28 100 72 26 49 38 68 57 46 47 40 Importance 0. 54 0. 50 0. 41 0. 39 0. 31 0. 40 0. 46 0. 48 0. 46 0. 32 0. 68 0. 61 0. 42 0. 47 0. 36 0. 33 0. 38 0. 31 0. 32 0. 28 1. 00 0. 72 0. 26 0. 52 0. 38 0. 68 0. 57 0. 50 0. 51 0. 40 Fig. The prediction results of the t + 10 model with parameter selection: „a… prediction performance of the ? ve algorithms for the test data set of Table 2 and „b… observed and predicted PRRs by the SVM algorithm Fig. 3 b , based on the measured values including mean, Std, max, min, power, and PRR in Table 1 at the intervals t = ? 50, t = ? 40 , . . . , t = ? 10, t = 0 , the average PRR value at the subsequent interval t + 20 is predicted. Similarly, with the same input and different models, the 10 min average PRR values at intervals t + 30, t + 40, and t + 50 are predicted. 3 Industrial Case Study 3. The t + 10 min PRR Prediction Without Parameter Selection. To compare the accuracy of models built before and after parameters selection, the original 30 predictors were used as inputs to construct a multivariate time series model. Five different data-mining algorithms were applied to build PRR prediction models for a wind farm based on data set 2 of Table 2. These algorithms include the multilayer perceptron algorithm MLP 23,24 , the support vector machine SVM regression 25,26 , the random forest 27,28 , the classi? cation and regression C&R tree 13,29 , and the pace regression algorithm 13,30 .

The ? ve algorithms used in this research are representative of different classes of data-mining algorithms. The MLP algorithm is usually used in nonlinear regression and classi? cation modeling. The SVM is a supervised learning algorithm used in classi? cation and regression. It constructs a linear discriminant function that separates instances as widely as possible. The C&R tree builds a decision tree to predict either classes classi? cation or Gaussians regression . The random forest algorithm grows many classi? cation trees to classify a new object from an input vector. Each tree Fig. The importance of predictors computed by the boosting tree algorithm Journal of Solar Energy Engineering AUGUST 2009, Vol. 131 / 031011-5 Downloaded 02 Sep 2009 to 128. 255. 53. 136. Redistribution subject to ASME license or copyright; see http://www. asme. org/terms/Terms_Use. cfm Table 7 Prediction error for the t + 20 models generated by the ? ve different algorithms Absolute error kW/min MLP SVM C&R MAE 362. 52 301. 31 364. 28 336. 25 336. 79 Std 360. 21 319. 48 366. 12 340. 41 347. 08 Maximum 3960. 36 3635. 03 4067. 49 4473. 17 4023. 24 Minimum 1. 27 0. 10 0. 88 1. 34 0. 65 Random forest tree Pace regression otes for every class, and ? nally the forest chooses the classi? cation having the most votes over all the trees in the forest. The pace regression algorithm consists of a group of estimators that are either optimal overall or optimal under certain conditions. It is a new approach to ? tting linear models in high-dimensional spaces. To test the accuracy of these algorithms, models trained from data set 2 of Table 2 were tested on data set 3 from Table 2. Table 4 shows the prediction accuracy of the models generated by the ? ve algorithms. Figure 4 a illustrates the absolute error of different algorithms. The ? st 100 observed PPRs and those predicted by the SVM algorithm for data set 3 were shown in Fig. 4 b . It can be seen from Table 4 and Fig. 4 that the SVM algorithm outperforms the other four algorithms. The C&R tree algorithm produces the worst predictions, and the pace regression algorithm performs quite well. The model can be updated to re? ect the process change over time. The update frequency could be, e. g. , 3 weeks. Alternatively, a separate routine could monitor the model performance and refresh the model once its performance would degrade. 3. 2 The t + 10 min Prediction With Parameter Selection.

In this section, the predictors as input for the multivariate time series model are selected by the boosting tree algorithm. As described in Sec. 2. 3, 9 out of 30 predictors were selected to build the time series model. The nine selected predictors are PPR-1, PPR-2, PPR-5, Min-1, PPR-3, Power-2, PRR-4, Min-2, and Std-2. To test the difference between t + 10 min prediction models built with and without parameter selection, the ? ve data-mining algorithms in Sec. 3. 1 were used. Multivariate models were retrained from data set 2 of Table 2 and were tested on data set 3 from Table 2.

Table 5 shows the prediction accuracy of the models generated by the ? ve algorithms. Figure 5 a illustrates the absolute error of the ? ve algorithms, while Fig. 5 b shows the ? rst 100 observed PPRs and those predicted by the SVM algorithm for data set 3. The results in Tables 4 and 5, and Figs. 4 and 5 demonstrate that the prediction accuracy of all ? ve algorithms was improved after parameter selection by the boosting tree algorithm. The SVM algorithm outperformed the other four algorithms in both scenarios, i. e. , with and without parameter selection. 3. 3 The t + 20 min Prediction With Parameter Selection.

To build a multivariate time series model for t + 20 min PRR prediction, parameter selection is performed by the boosting tree algorithm. Table 6 shows the importance of 30 predictors computed by the boosting tree algorithm based on data set 2 in Table 2 and t + 20 prediction horizons. In Table 6, -1 denotes the observation sampled 10 min earlier, 2 denotes the observation sampled 20 min earlier, and -3, -4, and -5 denote the observations sampled 30 min, 40 min, and 50 min in the past, respectively. Figure 6 shows the importance index of the 30 predictors for t + 20 PRR predictions ranked from the largest to the smallest one.

When comparing the results in Figs. 6 and 2, and Tables 6 and 3, the importance of predictors varies for the t + 10 and t + 20 models. Similar to Sec. 2. 4, 0. 5 was established as a threshold to select signi? cant predictors for t + 20 model. The boosting tree algorithm selected seven predictors and provided the following ranking: PPR-1, PPR-2, Max-1, Power-1, Max-2, Power-2, and Mean-1. 031011-6 / Vol. 131, AUGUST 2009 Fig. 7 Observed and predicted PRRs from the t + 20 models with selected parameters: „a… MLP algorithm, „b… SVM algorithm, „c… random forest algorithm, „d… C&R tree algorithm, and „e… pace regression algorithm

Transactions of the ASME Downloaded 02 Sep 2009 to 128. 255. 53. 136. Redistribution subject to ASME license or copyright; see http://www. asme. org/terms/Terms_Use. cfm Table 8 Absolute error statistics for multiperiod models Absolute error kW/min t + 30 t + 40 t + 50 t + 60 min min min min prediction prediction prediction prediction MAE 329. 83 347. 92 387. 45 458. 70 Std 347. 03 418. 41 404. 92 469. 24 Maximum 4109. 27 4600. 32 4566. 47 4972. 20 Minimum 0. 59 1. 94 0. 02 0. 62 Table 7 shows the prediction error of the models generated by the ? e algorithms the same as in Sec. 3. 2 . Figure 7 shows the ? rst 100 observed and predicted PRR values for data set 3 in Table 2. The SVM algorithm outperformed the other four; however, the accuracy decreased compared with the t + 10 results reported in Sec. 3. 2. 3. 4 Multiperiod Prediction With Parameter Selection. As the SVM algorithm performed better for both t + 10 and t + 20 predictions. Therefore, it was selected to build multivariate time series PRR models for t + 30– t + 60 min intervals. After parameter selection with the same parameter importance threshold of 0. , the 30 predictors were reduced to a seven-dimensional input with the boosting tree algorithm. For the t + 30 min model, the seven predictors were ranked as follows: Min-3, Min-1, Min-2, PRR-2, PRR-3, Max-3, and PRR-1. For the t + 40 min model, the ranking is PRR-2, PRR-4, PRR-1, Max-1, Power-1, PRR-3, and Mean-1. For the t + 50 min model, the ranking is PRR-1, Max-1, Mean-1, PRR-3, Std-1, PRR-4, and Power-5. And for the t + 60 min model, the ranking is Std-2, PRR-2, Mean-2, Max-2, Power-4, Power-5, and Max-3. The boosting tree algorithm selects different parameters over different periods of the PRR prediction, i. . , the results depend on the data set properties. Using the selected parameters, multiperiod prediction models were built by the SVM algorithm. The test data set used for the t + 10 min model of Sec. 3. 2 containing 887 points was reduced by 1 for each of the next 10 min period predictions. Table 8 shows the absolute error statistics for the multivariate time series prediction over four different 10 min intervals. Figures 8 a –8 d show the ? rst 100 observed and predicted PRRs over t + 30 min, t + 40 min, t + 50 min, and t + 60 min intervals, respectively.

The mean, the standard deviation, and the maximum error all increase as the prediction horizon lengthens. However, the minimum error remains relatively stable. The multivariate model provides accurate PRR prediction at the t + 10 to t + 40 intervals; however, the accuracy at the t + 50 and t + 60 intervals deteriorates. It appears that for longer horizon predictions, weather forecasting data may be useful. 4 Conclusion In this paper, multivariate time series models for power ramp rate prediction at different time horizons, from 10 min to 60 min, were constructed.

Five different data-mining algorithms were used to build the PRR prediction models. The boosting tree algorithm selected important predictors. After parameter selection, the original 30-dimensional input was signi? cantly reduced, and thus the accuracy of the multivariate time series model was improved. The SVM algorithm outperformed the other four algorithms studied in this paper. The multivariate time series model for PRR prediction built by the SVM algorithm turned out to be accurate and robust. The models constructed in the paper predicted the power ramp at t + 10– t + 60 min intervals.

A comprehensive comparative analysis of the multivariate models built with different data-mining algorithms was reported in this paper. The time series models accurately predicted the power ramp rate of the wind farm at t + 10– t + 40 horizons; however, the accuracy at t + 50 min and t + 60 min horizons degrades. The extracted Journal of Solar Energy Engineering Fig. 8 Observed and predicted PRRs for different periods for the ? rst 100 test data points: „a… the t + 30 min PRR model, „b… the t + 40 min PRR model, „c… the t + 50 min PRR model, and „d… the t + 60 min PRR model models are essential in power grid integration and management.

The multivariate time series prediction model may become a basis for predictive control aimed at optimizing the power ramp rate. The current wind farm power prediction models usually estimate the power at 1 h or 3 h intervals based on weather forecastAUGUST 2009, Vol. 131 / 031011-7 Downloaded 02 Sep 2009 to 128. 255. 53. 136. Redistribution subject to ASME license or copyright; see http://www. asme. org/terms/Terms_Use. cfm ing data. These predictions reveal power ramps over long time horizons. Prediction of power ramp rates at shorter intervals, e. g. , 10 min, is of importance to the electric grid.

The model built in this research does not use weather forecasting data, and it provides valuable ramp rate prediction on 10 min intervals. One avenue to be pursued in future research is the transformation of the time series data, e. g. , using wavelets or Kalman ? lters. One disadvantage of the proposed approach is that the multivariate time series model used different parameters, and therefore updating the model with most current data is important. As the number of prediction steps increases, the error increases. The models investigated in this research were intended for predicting the power ramp rate at relatively short horizons.

One possible mitigation strategy is to incorporate weather forecasting and additional off-site observation data, all at additional computational cost. Other research questions, including the seasonal performance of the proposed approach, could be addressed, provided that the appropriate data would be available. Acknowledgment The research reported in the paper has been partially supported by funding from the Iowa Energy Center Grant No. 07-01. References 1 David, A. S. , 1994, Wind Turbine Technology: Fundamental Concepts of Wind Turbine Engineering, ASME, New York, p. 638. 2 Svoboda, A. J. , Tseng, C. , Li, C. , and Johnson, R. B. 1997, “Short-Term Resource Scheduling With Ramp Constraints,” IEEE Trans. Power Syst. , 12 1 , pp. 77–83. 3 Ummels, B. C. , Gibescu, M. , Pelgrum, E. , Kling, W. L. , and Brand, A. J. , 2007, “Impacts of Wind Power on Thermal Generation Unit Commitment and Dispatch,” IEEE Trans. Energy Convers. , 22 1 , pp. 44–51. 4 Potter, C. W. , and Negnevitsky, M. , 2006, “Very Short-Term Wind Forecasting for Tasmanian Power Generation,” IEEE Trans. Power Syst. , 21 2 , pp. 965– 972. 5 Torres, J. L. , Garcia, A. , De Blas, M. , and De Francisco, A. , 2005, “Forecast of Hourly Average Wind Speed With ARMA Models in Spain,” Sol. Energy, 79 1 , pp. 5–77. 6 Sfetsos, A. , 2002, “A Novel Approach for the Forecasting of the Mean Hourly Wind Speed Time Series,” Renewable Energy, 27 2 , pp. 163–174. 7 Lange, M. , and Focken, U. , 2006, Physical Approach to Short-Term Wind Power Prediction, Springer-Verlag, Berlin, p. 208. 8 Barbounis, T. G. , Theocharis, J. B. , Alexiadis, M. C. , and Dokopoulos, P. S. , 2006, “Long-Term Wind Speed and Power Forecasting Using Local Recurrent Neural Network Models,” IEEE Trans. Energ. Convers. , 21 1 , pp. 273–284. 9 Kusiak, A. , and Song, Z. , 2006, “Combustion Ef? ciency Optimization and Virtual Testing: A Data-Mining Approach,” IEEE Trans. Ind.

Informat. , 2 3 , pp. 176–184. 10 Kusiak, A. , 2006, “Data Mining: Manufacturing and Service Applications,” Int. J. Prod. Res. , 44 18–19 , pp. 4175–4191. 11 Berry, M. J. A. , and Linoff, G. S. , 2004, Data Mining Techniques: For Marketing, Sales, and Customer Relationship Management, 2nd ed. , Wiley, New York. 12 Backus, P. , Janakiram, M. , Mowzoon, S. , Runger, G. C. , and Bhargava, A. , 2006, “Factory Cycle-Time Prediction With Data-Mining Approach,” IEEE Trans. Semicond. Manuf. , 19 2 , pp. 252–258. 13 Tan, P. N. , Steinbach, M. , and Kumar, V. , 2006, Introduction to Data Mining, Pearson/Addison Wesley, Boston, MA. 14 Kusiak, A. Zheng, H. , and Song, Z. , 2009, “Wind Farm Power Prediction: A Data Mining Approach,” Wind Energy, 12 3 , pp. 275–293. 15 Box, J. E. P. , and Jenkins, G. M. , 1976, Time Series Analysis Forecasting and Control, Holden-Day, San Francisco, CA. 16 Brown, B. G. , Katz, R. W. , and Murphy, A. H. , 1984, “Time Series Prediction Model to Simulate and Forecast Wind Speed and Wind Power,” J. Clim. Appl. Meteorol. , 23 8 , pp. 1184–1195. 17 Friedman, J. H. , 2002, “Stochastic Gradient Boosting,” Comput. Stat. Data Anal. , 38 4 , pp. 367–378. 18 Friedman, J. H. , 2001, “Greedy Function Approximation: A Gradient Boosting Machine,” Ann.

Stat. , 29 5 , pp. 1189–1232. 19 Witten, I. H. , and Frank, E. , 2005, Data Mining: Practical Machine Learning Tools and Techniques, 2nd ed. , Morgan Kaufmann, San Francisco, CA. 20 Kohavi, R. , and John, G. H. , 1997, “Wrappers for Feature Subset Selection,” Artif. Intell. , 97 1–2 , pp. 273–324. 21 Espinosa, J. , Vandewalle, J. , and Wertz, V. , 2005, Fuzzy Logic, Identi? cation and Predictive Control, Springer-Verlag, London, UK. 22 http://en. wikipedia. org/wiki/Curse_of_dimensionality. 23 Bishop, C. M. , 1995, Neural Networks for Pattern Recognition, Oxford University, New York. 24 Seidel, P. , Seidel, A. , and Herbarth, O. 2007, “Multilayer Perceptron Tumor Diagnosis Based on Chromatography Analysis of Urinary Nucleoside,” Neural Networks, 20 5 , pp. 646–651. 25 Smola, A. J. , and Schoelkopf, B. , 2004, “A Tutorial on Support Vector Regression,” Stat. Comput. , 14 3 , pp. 199–222. 26 Cristianini, N. , and Shawe-Taylor, J. , 2000, An Introduction to Support Vector Machines and Other Kernel-Based Learning Methods, Cambridge University, New York, p. 189. 27 Prasad, A. M. , Iverson, L. R. , and Liaw, A. , 2006, “Newer Classi? cation and Regression Tree Techniques: Bagging and Random Forests for Ecological Prediction,” Ecosystems, 9 2 , pp. 81–189. 28 Breiman, L. , 2001, “Random Forest,” Mach. Learn. , 45 1 , pp. 5–32. 29 Breiman, L. , Friedman, J. , Olshen, R. A. , and Stone, C. J. , 1984, Classi? cation and Regression Trees, Wadsworth International, Monterey, CA. 30 Wang, Y. , and Witten, I. H. , 2002, “Modeling for Optimal Probability Prediction,” Proceedings of the 19th International Conference in Machine Learning, Sydney, Australia, pp. 650–657. 031011-8 / Vol. 131, AUGUST 2009 Transactions of the ASME Downloaded 02 Sep 2009 to 128. 255. 53. 136. Redistribution subject to ASME license or copyright; see http://www. asme. org/terms/Terms_Use. cfm

x

Hi!
I'm Kira

Would you like to get such a paper? How about receiving a customized one?

Check it out