ANALYSIS OF NEURAL NETWORK HYPERPARAMETERS FOR PREDICTING USER GAZE DIRECTION IN ADAPTIVE LEARNING SYSTEMS
PDF

Keywords

gaze direction
adaptive learning
CNN
personalization
machine learning
LMS

How to Cite

[1]
O. Barkovska, Y. Liapin, I. Ruban, D. Rosinskiy, and V. Tkachov, “ANALYSIS OF NEURAL NETWORK HYPERPARAMETERS FOR PREDICTING USER GAZE DIRECTION IN ADAPTIVE LEARNING SYSTEMS”, ITLT, vol. 108, no. 4, pp. 263–278, Sep. 2025, doi: 10.33407/itlt.v108i4.6145.

Abstract

This study explores the application of convolutional neural networks (CNNs) for real-time gaze direction prediction within adaptive learning environments. In response to the increasing demand for personalized education, the ability to monitor learners' attention presents novel opportunities for dynamic content adaptation in learning management systems (LMS) and training management systems (TMS). The research evaluates the performance of four CNN architectures – LeNet, AlexNet, VGGNet, and ResNet – under varying hyperparameter configurations, including batch size, optimizer type, and activation function. Experiments employ the synthetic UnityEyes dataset alongside a custom test set comprising video recordings of actual user behavior. The results indicate that ResNet and VGGNet achieve the highest accuracy (up to 85%) and the lowest loss values when combined with Swish or GELU activation functions and AdamW or SGD optimizers. These findings underscore the importance of selecting appropriate architectures and hyperparameters for effective gaze-tracking applications. Furthermore, the study proposes a novel method for determining the user's area of visual attention through gaze vector analysis, implemented using TensorFlow, Keras, OpenCV, and Mediapipe. This approach enables real-time adaptation of learning content – such as prompts, interface modifications, or supplementary explanations – based on observed visual activity. In contrast to traditional static methods, the proposed solution enables dynamic personalization of the learning experience. A comparative analysis of model accuracy and training efficiency demonstrates the potential of gaze-based systems to enhance inclusive educational technologies. Future research will aim to improve system robustness against variations in lighting, head orientation, and partial facial occlusion, as well as to integrate gaze-controlled content navigation modules tailored for learners with special educational needs.

PDF

References

M. Faisal, A.W. AlAmeeri, A. Alsumait, “An adaptive e-learning framework: crowdsourcing approach,” in Proc. 17th Int. Conf. on Inf. Integration and Web-based Appl. & Services, New York, NY, USA, 2015, Article 13, pp. 1–5, doi: 10.1145/2837185.2837249. (in English)

O. Barkovska, H. Ivashchenko, D. Rosinskiy, D. Zakharov, “Educational Training Simulator for Monitoring Reading Technique and Speed Based on Speech-To-Text (STT) Methods,” Inf. Technologies and Learn. Tools, vol. 103, no. 5, pp. 21–38, Oct. 2024, doi: 10.33407/itlt.v103i5.5647. (in English)

O. Barkovska, Y. Liapin, T. Muzyka, I. Ryndyk, P. Botnar, “Gaze Direction Monitoring Model in Computer System for Academic Performance Assessment,” Inf. Technologies and Learn. Tools, vol. 99, no. 1, pp. 63–75, Feb. 2024, doi: 10.33407/itlt.v99i1.5503. (in English)

K. Vanlehn, “The Relative Effectiveness of Human Tutoring, Intelligent Tutoring Systems, and Other Tutoring Systems,” Educational Psychologist, vol. 46, no. 4, pp. 197–221, 2011, doi: 10.1080/00461520.2011.611369. (in English)

V. Aleven, R. Baraniuk, E. Brunskill, S. Crossley, D. Demszky, S. Fancsali, S. Gupta, K. Koedinger, C. Piech, S. Ritter, D. Thomas, S. Woodhead, W. Xing, “Towards the Future of AI-Augmented Human Tutoring in Math Learning,” in Artif. Intell. in Educ. Posters and Late Breaking Results, Workshops and Tut., Industry and Innov. Tracks, Practitioners, Doctoral Consortium and Blue Sky, Springer, Cham, 2023, pp. 26–31, doi: 10.1007/978-3-031-36336-8_3. (in English)

B. Handini, N. Nurhasanah, F. Panly, “The Effect of Artificial Intelligent Technology Used (Duolingo Application) To Enhance English Learning,” ELLITE: J. of English Lang., Literature, and Teaching, vol. 7, no. 2, pp. 86–94, 2022, doi: 10.32528/ellite.v7i2.8354. (in English)

S. Dutta, S. Ranjan, S. Mishra, V. Sharma, P. Hewage, C. Iwendi, “Enhancing Educational Adaptability: A Review and Analysis of AI-Driven Adaptive Learning Platforms,” in 2024 4th Int. Conf. on Innovative Practices in Technol. and Manage., Noida, India, 2024, pp. 1–5, doi: 10.1109/ICIPTM59628.2024.10563448. (in English)

V. Kirov, B. Malamin, “Are translators afraid of artificial intelligence?” Societies, vol. 12, no. 2, p. 70, Apr. 2022, doi: 10.3390/soc12020070. (in English)

R.R. Saqr; S.A. Al-Somali; M.Y. Sarhan, “Exploring the Acceptance and User Satisfaction of AI-Driven e-Learning Platforms (Blackboard, Moodle, Edmodo, Coursera and edX): An Integrated Technology Model,” Sustainability, vol. 16, no. 1, p. 204, 2024, doi: 10.3390/su16010204. (in English)

H. M. Šola, F. H. Qureshi, S. Khawaja, “AI-Powered Eye Tracking for Bias Detection in Online Course Reviews: A Udemy Case Study,” Big Data and Cogn. Comput., vol. 8, no. 11, p. 144, 2024, doi: 10.3390/bdcc8110144. (in English)

N. Kaur, P. Singh, “Conventional and contemporary approaches used in text to speech synthesis: a review,” Artif. Intell. Rev., vol. 56, pp. 5837–5880, 2023, doi: 10.1007/s10462-022-10315-0. (in English)

J. Jeuring, H. Keuning, S. Marwan, D. Bouvier, C. Izu, N. Kiesler, T. Lehtinen, D. Lohr, A. Peterson, S. Sarsa. “Towards Giving Timely Formative Feedback and Hints to Novice Programmers,” in Proc. of the 2022 Working Group Reports on Innov. and Technol. in Comput. Sci. Educ., New York, NY, USA, 2022, pp. 95–115 doi: 10.1145/3571785.3574124. (in English)

P. Lameras, S. Arnab, “Power to the Teachers: An Exploratory Review on Artificial Intelligence in Education,” Information, vol. 13, no. 1, p. 14, 2022, doi: 10.3390/info13010014. (in English)

E. Wood, T. Baltrušaitis, L.-P. Morency, P. Robinson, A. Bulling. “Learning an appearance-based gaze estimator from one million synthesised images,” in Proc. of the Ninth Biennial ACM Symp. on Eye Tracking Res. & Appl., New York, NY, USA, 2016, pp. 131–138, doi: 10.1145/2857491.2857492. (in English)

Creative Commons License

This work is licensed under a Creative Commons Attribution-NonCommercial-ShareAlike 4.0 International License.

Copyright (c) 2025 Olesia Barkovska, Yaroslav Liapin, Igor Ruban, Dmytro Rosinskiy, Vitalii Tkachov

Downloads

Download data is not yet available.