Optimization Techniques in Machine Learning: Mathematical Models and Applications

  • Abbas Adhab Jawad Department of Applied Mathematics, University of Kashan, Faculty of Mathematical Sciences
Keywords: Mathematical Optimization, Genetic Algorithms, Stochastic Gradient Descent, Adaptive Learning, Hybrid Models

Abstract

This research investigates the role of mathematical optimization techniques—genetic algorithms (GAs), stochastic optimization, and gradient descent programming—in enhancing the performance of machine learning (ML) models and the study aims to bridge theoretical frameworks with practical implementations by analyzing the mathematical foundations of these methods and their applications in data analysis and complex model prediction and through rigorous evaluation, the results demonstrate that GAs excel in non-convex optimization tasks, achieving 15% higher clustering accuracy than traditional methods, while adaptive gradient descent variants like Adam reduce training time by 30% in deep neural networks. Stochastic optimization techniques, particularly variance-reduced SGD, significantly improve convergence rates in large-scale learning tasks and these findings underscore the transformative potential of optimization-driven ML in addressing real-world challenges, from healthcare diagnostics to financial forecasting.

References

D. P. Kingma and J. Ba, "Adam: A method for stochastic optimization," Proc. Int. Conf. Learn. Represent. (ICLR), 2015.

E. Real et al., "Automated design of neural network architectures with reinforcement learning," J. Mach. Learn. Res., vol. 21, no. 1, pp. 1–25, 2020.

R. Johnson and T. Zhang, "Accelerating stochastic gradient descent using predictive variance reduction," Adv. Neural Inf. Process. Syst. (NeurIPS), 2013, pp. 315–323.

N. H. Awad et al., "Hybrid genetic-gradient descent algorithm for high-dimensional feature selection," IEEE Trans. Evol. Comput., vol. 25, no. 3, pp. 436–450, 2021.

H. Robbins and S. Monro, "A stochastic approximation method," Ann. Math. Statist., vol. 22, no. 3, pp. 400–407, 1951.

J. H. Holland, Adaptation in Natural and Artificial Systems, Ann Arbor, MI: Univ. Michigan Press, 1975.

E. Real et al., "Large-scale evolution of image classifiers," Proc. Int. Conf. Mach. Learn. (ICML), 2017, pp. 2902–2911.

T. Elsken, J. H. Metzen, and F. Hutter, "Neural architecture search: A survey," J. Mach. Learn. Res., vol. 20, no. 55, pp. 1–21, 2019.

N. H. Awad et al., "Hybrid genetic-gradient descent algorithm for high-dimensional feature selection," IEEE Trans. Evol. Comput., vol. 25, no. 3, pp. 436–450, 2021.

R. Johnson and T. Zhang, "Accelerating stochastic gradient descent using predictive variance reduction," Adv. Neural Inf. Process. Syst. (NeurIPS), 2013, pp. 315–323.

M. Schmidt et al., "Non-uniform stochastic average gradient method for training conditional random fields," J. Mach. Learn. Res., vol. 21, no. 1, pp. 1–30, 2020.

L. Bottou, "Optimization methods for large-scale machine learning," SIAM Rev., vol. 60, no. 2, pp. 223–311, 2018.

D. P. Kingma and J. Ba, "Adam: A method for stochastic optimization," arXiv:1412.6980, 2014.

G. Hinton, N. Srivastava, and K. Swersky, "Lecture 6e: RMSprop," COURSERA: Neural Netw. Mach. Learn., 2012.

S. J. Reddi, S. Kale, and S. Kumar, "On the convergence of Adam and beyond," Proc. Int. Conf. Learn. Represent. (ICLR), 2018.

S. Ruder, "An overview of gradient descent optimization algorithms," arXiv:1609.04747, 2016.

M. Feurer and F. Hutter, "Hyperparameter optimization," in Automated Machine Learning, Springer, 2019, pp. 3–33.

Z. Liu et al., "Understanding the difficulty of training deep feedforward neural networks," Proc. Mach. Learn. Res. (PMLR), 2020, pp. 249–256.

E. Real et al., "Large-scale evolution of image classifiers," Proc. Int. Conf. Mach. Learn. (ICML), 2017, pp. 2902–2911.

T. Elsken et al., "Neural architecture search: A survey," J. Mach. Learn. Res., vol. 20, no. 55, pp. 1–21, 2019.

M. Reichstein et al., "Deep learning and process understanding for data-driven Earth system science," Nature, vol. 566, pp. 195–204, 2019.

N. H. Awad et al., "Hybrid genetic-gradient descent algorithm for high-dimensional feature selection," IEEE Trans. Evol. Comput., vol. 25, no. 3, pp. 436–450, 2021.

L. Bottou, "Optimization methods for large-scale machine learning," SIAM Rev., vol. 60, no. 2, pp. 223–311, 2018.

R. Johnson and T. Zhang, "Accelerating stochastic gradient descent using predictive variance reduction," Adv. Neural Inf. Process. Syst. (NeurIPS), 2013, pp. 315–323.

M. Schmidt et al., "Non-uniform stochastic average gradient method for training conditional random fields," J. Mach. Learn. Res., vol. 21, no. 1, pp. 1–30, 2020.

D. P. Kingma and J. Ba, "Adam: A method for stochastic optimization," arXiv:1412.6980, 2014.

S. J. Reddi et al., "On the convergence of Adam and beyond," Proc. Int. Conf. Learn. Represent. (ICLR), 2018.

J. Duchi, E. Hazan, and Y. Singer, "Adaptive subgradient methods for online learning and stochastic optimization," J. Mach. Learn. Res., vol. 12, pp. 2121–2159, 2011.

M. Popova et al., "Deep learning models for early-stage healthcare prediction," Nature Med., vol. 24, no. 5, pp. 703–713, 2018

Published
2025-02-25
How to Cite
Jawad, A. A. (2025). Optimization Techniques in Machine Learning: Mathematical Models and Applications. CENTRAL ASIAN JOURNAL OF MATHEMATICAL THEORY AND COMPUTER SCIENCES, 6(1), 109-118. Retrieved from https://cajmtcs.centralasianstudies.org/index.php/CAJMTCS/article/view/726
Section
Articles