A Comprehensive Evolution for Applicability of Machine Learning Algorithms on Various Domains

Machine learning algorithms are able to learn from data, make decision and improve what they learn by having experience without human intervention. Machine learning techniques are becoming increasingly important nowadays that everything is going to be fully automated. They have been used in various fields such as recommendation engines, self-driving cars, offering personal suggestions from retailers, cyber fraud detection, face recognition, and etc. This study presents some of the most commonly used machine learning techniques from supervised and unsupervised learning classes such as linear regression, logistic regression, neural networks, and self-organizing map. In linear regression technique, it is tackled to fit a linear function to user data in order to model the relationship between variables. It can be a useful technique to make weather estimation, to understand marketing effectiveness and to model consumer behavior. Logistic regression is a statistical model that uses a logistic function and is appropriate when dependent variable is binary. Neural networks mimic the operation of human brain to recognize patterns from the underlying data. They have wide range of application such as cancer diagnosis, e-mail spam filtering, and signal classification. Self-organizing map, a special type of neural networks, is utilized to achieve dimensionality reduction that generally used for seismic analysis, project prioritization, and image processing such as color reduction. Each implementation in this study shows that the success of the results obtained by applying machine learning techniques depends on using the right technique in the appropriate area.

___

  • Shalev-Shwartz, S., & Ben-David, S., Understanding Machine Learning: From Theory to Algorithms, Cambridge: Cambridge University Press, 2014.
  • Smola, A. and Vishwanathan, S.V.N., Introduction to Machine Learning, Cambridge University Press, 2008.
  • Alpaydin, Ethem, Introduction to Machine Learning, Third Edition, MIT Press, 2014.
  • Ng, Andrew, Machine Learning Yearning, 2018.
  • Mohan, V. and Gupta, M., Data Regression with Normal Equation on GPU using CUDA, International Journal of Computer Science, Information Technology and Security (IJCSITS), Vol. 2, No.2, April 2012.
  • Monahan, J. F., A Primer on Linear Models, CRC Press, 2008.
  • Rao, C.R., Toutenburg, H., Shalabh, and Heumann, C., Linear Models and Generalizations - Least Squares and Alternatives, Springer, 2008.
  • Khuri, Andre I., Linear Model Methodology, CRC Press, 2010.
  • Seltman, H. J., Experimental Design and Analysis, 2018 Available from: http://www.stat.cmu.edu/∼hseltman/309/Book/Book.pdf
  • Montgomery, Douglas C., Peck, Elizabeth A., Vining, G. Geoffrey, Introduction to Linear Regression Analysis, Fifth Edition, John Wiley & Sons, Inc., 2012.
  • https://www.worldometers.info/coronavirus/country/us/
  • Wilson, R.A. and Keil, F.C., The MIT Encyclopedia of the Cognitive Sciences, 1999.
  • Toussaint, M., Lecture Notes: Some notes on gradient descent, Berlin, May 3, 2012.
  • Wang, X., Method of Steepest Descent and its Applications, 2008.
  • The Free Dictionary [Internet]. "Normal Equations". McGraw-Hill Dictionary of Scientific & Technical Terms, 6E, The McGraw-Hill Companies, Inc., 2003 [cited 27 Apr. 2020]. Available from: https://encyclopedia2.thefreedictionary.com/Normal+Equations
  • Rawlings, J. O., Sastry, G. P., Dickey, D. A., Applied Regression Analysis: A Research Tool, Second Edition, 8 Springer-Verlag New York, Inc., 1998.
  • Russell, S. and Norvig, P., Artificial Intelligence: A Modern Approach, Third Edition, Prentice Hall, 2010.
  • Kaplan, D. T., Statistical Modelling A Fresh Approach, E-book version of Second Edition, Project MOSAIC Books, 2017.
  • Sperandei, S., Understanding logistic regression analysis, Biochemia medica, 24. 12-8. 10.11613/BM.2014.003, 2014.
  • Dua, D. and Graff, C., UCI Machine Learning Repository. Irvine, CA: University of California, School of Information and Computer Science, 2019. Available from: http://archive.ics.uci.edu/ml
  • Gonzalez, R.C. and Woods, R.E., Digital image processing, Third Edition, Upper Saddle River, N.J.: Prentice Hall, 2008.
  • Rojas, R., Neural Networks: A Systematic Introduction, Springer-Verlag, Berlin, 1996.
  • Kohonen, T., Self-organized formation of topologically correct feature maps, Biological Cybernetics, vol.66, pp.59–69, 1982.
  • Kohonen, T., Self-Organizing Maps, Springer-Verlag, Berlin, 2001.
  • Chetia, S. and Sarma, K. K., PCA and SOM based Dimension Reduction Techniques for Quaternary Protein Structure Prediction, International Journal of Computer Applications (0975 – 8887), Volume 46– No.18, May 2012.
  • Gan, G., Ma, C. and Wu, J., Data Clustering: Theory, Algorithms, and Applications, SIAM, 2007.
  • Celebi, M.E., Improving the performance of k-means for color quantization, Image and Vision Computing, vol.29, pp. 260–271, 2011.
  • Atsalakis, A. and Papamarkos N., Color reduction by using a new self-growing and self-organized neural network, Vision, Video and Graphics, 2005.
  • Zagoris, K., Papamarkos1, N. and Koustoudis, I., Color Reduction Using the Combination of the Kohonen Self-Organized Feature Map and the Gustafson-Kessel Fuzzy Algorithm, P. Perner (Ed.): MLDM 2007, LNAI 4571, pp. 703–715, Springer-Verlag Berlin Heidelberg, 2007.
  • Chang, C.H., Xu, P., Xiao, R. and Srikanthan, T., New adaptive color quantization method based on self-organizing maps, IEEE Transactions on Neural Networks, vol. 16, no. 1, pp. 237–249, 2005.
  • Hu, Y.-C., Lee, M.-G., K-means based color palette design scheme with the use of stable flags, Journal of Electronic Imaging, vol. 16 (3), 2007.
  • Hu, Y.-C. and Su, B.-H., Accelerated k-means clustering algorithm for colour image quantization, Imaging Science Journal, 56 (1) pp. 29–40, 2008.
  • Park, H.J., Kim, K.B., and Cha, E.-Y., An Effective Color Quantization Method Using Color Importance-Based Self-Organizing Maps, Neural Network World, 121-137, 2015.