EDUCATIONAL DATA MINING METHODS FOR TIMSS 2015 MATHEMATICS SUCCESS: TURKEY CASE

Educational data mining (EDM) is an important research area which has an ability of analyzing and modeling educational data. Obtained outputs from EDM help researchers and education planners understand and revise the systematic problems of current educational strategies. This study deals with an important international study, namely Trends International Mathematics and Science Study (TIMSS). EDM methods are applied to last released TIMSS 2015 8th grade Turkish students' data. The study has mainly twofold: to find best performer algorithm(s) for classifying students’ mathematic success and to extract important features on success. The most appropriate algorithm is found as logistic regression and also support vector machines - polynomial kernel and support vector machines - Pearson VII function-based universal kernel give similar performances with logistic regression. Different feature selection methods are used in order to extract the most effective features in classification among all features in the original dataset. “Home Educational Resources”, “Student Confident in Mathematics” and “Mathematics Achievement Too Low for Estimation” are found the most important features in all feature selection methods.

___

  • [1] Mullis, I.V., Martin, M.O., Foy, P., Arora, A., (2012) TIMSS 2011 international results in mathematics. International Association for the Evaluation of Educational Achievement. Herengracht 487, Amsterdam, 1017 BT, The Netherlands.
  • [2] Han, J., Kamber, M., Pei, J., (2012) Data mining: Concept and techniques, (3rd ed.). MA: Morgan Kaufmann Publishers, Burlington.
  • [3] Sinharay, S., (2016) An NCME instructional module on data mining methods for classification and regression. Educational Measurement: Issues and Practice 35, 38-54.
  • [4] Ramaswami, M., Bhaskaran, R., (2012) A CHAID based performance prediction model in educational data mining. arXiv preprint arXiv:1002.1144.
  • [5] Romero, C., Ventura, S., (2010) Educational data mining: a review of the state of the art. IEEE Transactions on Systems, Man, and Cybernetics, Part C (Applications and Reviews) 40, 601-618.
  • [6] Romero, C., Ventura, S., (2007) Educational data mining: A survey from 1995 to 2005. Expert systems with applications 33, 135-146.
  • [7] Baker, R.S., Yacef, K., (2009) The state of educational data mining in 2009: A review and future visions. JEDM Journal of Educational Data Mining 1, 3-17.
  • [8] Siemens, G., Baker, R.S., (2012) Learning analytics and educational data mining: towards communication and collaboration. In Proceedings of the 2nd international conference on learning analytics and knowledge, 2012, April, ACM.
  • [9] Mohamad, S.K., Tasir, Z., (2013) Educational data mining: A review. Procedia-Social and Behavioral Sciences 97, 320-324.
  • [10] Peña-Ayala, A., (2014) Educational data mining: A survey and a data mining-based analysis of recent works. Expert systems with applications 41, 1432-1462.
  • [11] Shahiri, A.M., Husain, W., (2015) A review on predicting student's performance using data mining techniques. Procedia Computer Science 72, 414-422.
  • [12] Kotsiantis, S., Pierrakeas, C., Pintelas, P., (2004) Predicting Students' performance in Distance Learning Using Machine Learning Techniques. Applied Artificial Intelligence 18, 411-426.
  • [13] Baradwaj, B.K., Pal, S., (2011) Mining educational data to analyze students' performance. International Journal of Advanced Computer Science and Applications 2, 63-69.
  • [14] Rajni, J., Malaya, D.B., (2015) Predictive analytics in a higher education context. IT Professional 17, 24-33.
  • [15] Martínez Abad, F., Chaparro Caso López, A.A., (2017) Data-mining techniques in detecting factors linked to academic achievement. School Effectiveness and School Improvement 28, 39-55.
  • [16] Cortez, P., Silva, A.M.G., (2008) Using data mining to predict secondary school student performance. In A. Brito & J. Teixeira (Eds.), Proceedings of 5th Annual Future Business Technology. Conference. Porto, Portugal: EUROSIS, 5-12.
  • [17] Osmanbegović, E., Suljić, M., (2012) Data mining approach for predicting student performance. Economic Review 10, 3-12.
  • [18] Hammouri, H., (2010) Attitudinal and motivational variables related to mathematics achievement in Jordan: Findings from the Third International Mathematics and Science Study (TIMSS). Educational Research 46, 241-257.
  • [19] Liu, S., Meng, L., (2010) Re‐examining factor structure of the attitudinal items from TIMSS 2003 in cross‐cultural study of mathematics self‐concept. Educational Psychology 30, 699-712.
  • [20] Askin, O.E., Gokalp, F., (2013) Comparing the predictive and classification performances of logistic regression and neural networks: a case study on timss 2011. Procedia-Social and Behavioral Sciences 106, 667-676.
  • [21] Topçu, M.S., Erbilgin, E., Arıkan, S., (2016) Factors Predicting Turkish and Korean Students' Science and Mathematics Achievement in TIMSS 2011. Eurasia Journal of Mathematics, Science & Technology Education 12, 1711-1737.
  • [22] Kılıç-Depren, S., Askin, Ö.E., Öz, E., (2017) Identifying the Classification Performances of Educational Data Mining Methods: A Case Study for TIMSS. Educational Sciences: Theory & Practice 17, 1605-1623.
  • [23] Filiz, E., Oz, E., (2019) Finding The Best Algorithms and Effective Factors in Classifıcation of Turkish Science Student Success. Journal of Baltic Science Education 18, 239-253.
  • [24] LaRoche, S., Joncas, M., Foy, P., (2016) Sample Design in TIMSS 2015. Martin, M. O., Mullis, I.V.S., and Hooper, M. (Eds.), Methods and Procedures in TIMSS 2015. Retrieved from Boston College, TIMSS & PIRLS International Study Center.
  • [25] Masters, G.N., Wright, B.D., (1997) The partial credit model. In M.J. van de Linden & R.K. Hambleton (Eds.), Handbook of modern item response theory. Berlin: Springer.
  • [26] Filiz, E., Öz, E., (2017) Classification of BIST-100 Index' Changes via Machine Learning Methods. Marmara University Journal of Economic & Administrative Sciences 39, 117-129.
  • [27] Jiang, S., Pang, G., Wu, M., Kuang, L., (2012) An improved K-nearest-neighbor algorithm for text categorization. Expert Systems with Applications 39, 1503-1509.
  • [28] Li, B., Yu, S., Lu, Q., (2003) An improved k-nearest neighbor algorithm for text categorization, Proceedings of the 20th International Conference on Computer Processing of Oriental Languages, 3-6 August 2003, Shenyang, China.
  • [29] John, G.H., Langley, P., (1995) Estimating continuous distributions in Bayesian classifiers. In Proceedings of the Eleventh conference on Uncertainty in artificial intelligence, 1995, August, Morgan Kaufmann Publishers Inc.
  • [30] Zhang, H., (2004) The optimality of naive Bayes. AA 1, 3.
  • [31] Haykin, S., (1999) Neural Networks: A comprehensive Foundation, Prentice Hall International. Inc., Englewood Cliffs.
  • [32] Alpaydın, E., (2004) Introduction to machine learning. MIT press, Cambridge.
  • [33] Shawe-Taylor, J., Bartlett, P.L., Williamson, R.C., Anthony, M., (1998) Structural risk minimization over data-dependent hierarchies. IEEE transactions on Information Theory 44, 1926-1940.
  • [34] Varshney, P.K., Arora, M.K., (2004) Advanced image processing techniques for remotely sensed hyperspectral data. Springer Science & Business Media.
  • [35] Rokach, L., Maimon, O., (2005) Decision trees. In Data mining and knowledge discovery handbook (pp. 165-192) Springer, Boston, MA.
  • [36] Quinlan, J.R., (2014) C4.5: programs for machine learning. Elsevier.
  • [37] Kalmegh, S., (2015) Analysis of WEKA data mining algorithm REPTree, Simple CART and RandomTree for classification of Indian news. Int. J. Innov. Sci. Eng. Technol. 2, 438-446.
  • [38] Srinivasan, D.B., Mekala, P., (2014) Mining Social Networking Data for Classification Using REPTree. International Journal of Advance Research in Computer Science and Management Studies 2, 155-160.
  • [39] Chen, X.W., Liu, M., (2005) Prediction of protein–protein interactions using random decision forest framework. Bioinformatics 21, 4394-4400.
  • [40] Breiman, L., (2001) Random forests. Machine learning 45, 5-32.
  • [41] Hosmer, D.W., Lemeshow, S., (2000) Applied Logistic Regression, 2nd ed.; Hoboken, NJ: John Wiley & Sons, Inc.
  • [42] Balaban, M.E., Kartal, E., (2015) Basic Algorithms of Data Mining and Machine Learning and Applications with R Language. Çağlayan Kitabevi, İstanbul.
  • [43] Donner, A., Klar, N., (1996) The statistical analysis of kappa statistics in multiple samples. Journal of clinical epidemiology 49, 1053-1058.
  • [44] Turanoğlu-Bekar, E., Ulutagay, G., Kantarcı-Savas, S., (2016) Classification of thyroid disease by using data mining models: A comparison of decision tree algorithms. Oxford Journal of Intelligent Decision and Data Sciences 2, 13-28.
  • [45] Willmott, C.J., Matsuura, K., (2005) Advantages of the mean absolute error (MAE) over the root mean square error (RMSE) in assessing average model performance. Climate research 30, 79-82.
  • [46] Gennari, J.H., Langley, P., Fisher, D., (1989) Models of incremental concept formation. Artificial intelligence 40, 11-61.
  • [47] Gümüşçü, A., Aydilek, İ.B., Taşaltın, R., (2016) Comparison of Feature Selection Algorithms on Microarray Data Classification. Harran University Journal of Engineering 1, 1-7.
  • [48] Kononenko, I., (1994) Estimating attributes: analysis and extensions of RELIEF, In European conference on machine learning, Springer: Berlin, Heidelberg.
  • [49] Cover, T.M., Thomas, J.A., (2012) Elements of information theory. John Wiley & Sons.
  • [50] Aktas, M.S., Kalıpsız, O., (2015) Research and Comparative Practice on the Application of Feature Selection Techniques to Banking Data in Data Mining. In Proceedings of the 9th Turkish National Software Engineering Symposium, 2015, September; Yasar University: Izmir, Turkey.
  • [51] Muda, Z., Yassin, W., Sulaiman, M.N., Udzir, N.I., (2011) Intrusion detection based on k-means clustering and OneR classification, 7th International Conference on Information Assurance and Security IAS2011, 5-8 December 2011, Malacca, Malaysia.
  • [52] Kabakchieva, D., (2013) Predicting student performance by using data mining methods for classification. Cybernetics and information technologies 13, 61-72.
  • [53] Saltelli, A., Chan, K., Scott, E.M., (2000) Sensitivity analysis (Vol. 1). New York: Wiley.
  • [54] Gondra, I., (2008) Applying machine learning to software fault-proneness prediction. Journal of Systems and Software 81, 186-195.
  • [55] Yao, J.T., (2003) Sensitivity analysis for data mining, NAFIPS 2003 22nd International Conference of the North American Fuzzy Information Processing Society, 24-26 July 2003, Chicago, IL, USA.
  • [56] Gazi, V.E., (2007) Data Mining Sensitivity. Master’s thesis, Institute of Science and Technology Istanbul Technical University, Turkey.
  • [57] Hall, M.A., (2000) Correlation-based Feature Selection for Discrete and Numeric Class Machine Learning, The University of Waikato, Working Paper 00/8. Hamilton, New Zealand.
  • [58] Frank, E., Mark, A.H., Ian, H.W., (2016) The WEKA Workbench. Online Appendix for “Data Mining: Practical Machine Learning Tools and Techniques”, Fourth Edition; Morgan Kaufmann.
  • [59] Schreiber, J.B., (2002) Scoring Above the International Average: A Logistic Regression Model of the TIMSS Advanced Mathematics Exam. Multiple Linear Regression Viewpoints 28, 22-30.
  • [60] Yoo J.E., (2018) TIMSS 2011 Student and Teacher Predictors for Mathematics Achievement Explored and Identified via Elastic Net. Frontiers in psychology 9, 317.
  • [61] Delen, D. A., (2010) comparative analysis of machine learning techniques for student retention management. Decision Support Systems 49, 498-506.