A Comparative Study of YOLO, SSD, Faster R-CNN, and More for Optimized Eye-Gaze Writing

Eye-gaze writing technology holds significant promise but faces several limitations. Existing eye-gaze-based systems often suffer from slow performance, particularly under challenging conditions such as low-light environments, user fatigue, or excessive head movement and blinking. These factors nega...

Full description

Saved in:
Bibliographic Details
Main Authors: Walid Abdallah Shobaki, Mariofanna Milanova
Format: Article
Language:English
Published: MDPI AG 2025-04-01
Series:Sci
Subjects:
Online Access:https://www.mdpi.com/2413-4155/7/2/47
Tags: Add Tag
No Tags, Be the first to tag this record!
Description
Summary:Eye-gaze writing technology holds significant promise but faces several limitations. Existing eye-gaze-based systems often suffer from slow performance, particularly under challenging conditions such as low-light environments, user fatigue, or excessive head movement and blinking. These factors negatively impact the accuracy and reliability of eye-tracking technology, limiting the user’s ability to control the cursor or make selections. To address these challenges and enhance accessibility, we created a comprehensive dataset by integrating multiple publicly available datasets, including the Eyes Dataset, Dataset-Pupil, Pupil Detection Computer Vision Project, Pupils Computer Vision Project, and MPIIGaze dataset. This combined dataset provides diverse training data for eye images under various conditions, including open and closed eyes and diverse lighting environments. Using this dataset, we evaluated the performance of several computer vision algorithms across three key areas. For object detection, we implemented YOLOv8, SSD, and Faster R-CNN. For image segmentation, we employed DeepLab and U-Net. Finally, for self-supervised learning, we utilized the SimCLR algorithm. Our results indicate that the Haar classifier achieves the highest accuracy (0.85) with a model size of 97.358 KB, while YOLOv8 demonstrates competitive accuracy (0.83) alongside an exceptional processing speed and the smallest model size (6.083 KB), making it particularly suitable for cost-effective real-time eye-gaze applications.
ISSN:2413-4155