Öz An area's wind speed forecasting is very important to investigate whether the area is available for wind power. The wind speed estimation has been carried out by means of other machine learning methods, mostly artificial neural networks. Because, in such methods, it is aimed to estimate the wind speed with the highest accuracy along with the decimal. However, if a wind farm is to be installed, the wind speed, which is a variable in the range of 0-20 m / s, can easily be estimated with round values. If the wind speed values obtained with round values are forecasted with a high accuracy rate, the wind speed that is necessary for the establishment of a wind power plant in a region is obtained by a shorter and easier way. In this study, the decision tree method was used in order to reach wind speed information with an easier method and with a very short training period. Decision tree methods were examined in three different structures and three different decision tree models were designed. Additionally the estimation results of all three methods were very high, the most accurate estimation was obtained by the “Coarse Decision Tree” method which is much simpler and faster than the others.
Gupta, B.,, Rawat, A., Jain, J., Arora, A., Dhami, N. (2017). Analysis of Various Decision Tree Algorithms for Classification in Data Mining. International Journal of Computer Applications, 163(8), 15-19.
Ray, P.K., Kishor, N. (2014). Optimal Feature and Decision Tree-Based Classiﬁcation of Power Quality Disturbances in Distributed Generation Systems. IEEE Transactions on Sustainable on Energy, 5(1), 200-208.
Sangita B.P., Deshmukh, S.R. (2011). Use of Support Vector Machine, decision tree and Naive Bayesian techniques for wind speed classification. International Conference on Power and Energy Systems Power and Energy Systems (ICPS).
Retscreen Engineering & Cases Textbook, Clean Energy Project Analysis, Clean Energy Decision Support Centre ISBN: 0-662-35670-5 Catalogue no.: M39-97/2003E-PDF, © Minister of Natural Resources Canada 2001-2004. http://unfccc.int/resource/cd_roms/na1/mitigation/Module_5/Module_5_1/b_tools/RETScreen/Manuals/Wind.pdf, last acces date: Feb 27th, 2019.
Chien, C. F., Chen, L. F. 2008. Data Mining to Improve Personnel Selection and Enhance Human Capital: A Case Study in High-Technology Industry. Expert Systems with Applications, 34, 280-290.
Tsang, S., Kao, B., Yip, K.Y., Ho, W.S., Lee, S.D. (2011). Decision Trees for Uncertain Data, IEEE Transaction on Knowledge and Data Engineering, 23(1), 67-78.
Rokach, L., Maimon, O. (2014). Data Mining with Decision Trees Theory and Applications. 2nd edition, 81, World Scientific Publishing Co. Pte. Ltd.
Quinlan J.R, (1993). C4.5: Programs for Machine Learning. Morgan Kaufmann, San Mateo, CA, 302.
Dougherty, G., Pattern Recognition and Classiﬁcation. Springer New York Heidelberg Dordrecht London.
Loh, W.Y., Shih, Y.S. (1997). Split Selection Methods for Classification Trees, Statistica Sinica, 7(4), 815-840.
Friedl, M.A., Brodley C.E., (1997). Decision tree classification of land cover from remotely sensed data. Remote Sensing of Environment, 61, 399–409.
Safavian S.R., Landgrebe D. (1991). A survey of decision tree classifier methodology. IEEE Transactions on Systems Man and Cybernetics, 21, 660-674.
Tan, P.N., Steinbach, M., Kumar, V. (2006). Introduction to Data Mining. Pearson Addison-Wesley.
Maimon, O., Rokach, L. (2010). Data Mining and Knowledge Discovery Handbook, Springer 2010.
Suneetha, N., Hari, Ch.V.M.V., Kumar, S. (2010). Modified Gini Index Classification: A Case Study of Heart Disease Dataset. International Journal on Computer Science and Engineering, 2(6),1959-1965
Breiman L., Friedman J.H., Olshen R.A. and Stone C.J. (1984). Classification and Regression Trees. Monterey, CA: Wadsworth, 358 s.
Raileanu L.E., Stoffel, K., Theoretical Comparison between the Gini Index and Information Gain Criteria. Annals of Mathematics and Artificial Intelligence, 41(1):77-93.
Teknomo, K. (2012). Decision Tree Tutorial. www. revoledu.com Online edition.
Mingers J., (1989). An empirical comparison of pruning methods for decision tree induction. Machine Learning, 4, 227–243.