I2DNet - Design and real-time evaluation of an appearance-based gaze estimation system
DOI:
https://doi.org/10.16910/jemr.14.4.2Keywords:
Eye tracking, web-cam based eye tracking, Convolutional Neural Networks, Usability evaluationAbstract
Gaze estimation problem can be addressed using either model-based or appearance-based approaches. Model-based approaches rely on features extracted from eye images to fit a 3D eye-ball model to obtain gaze point estimate while appearance-based methods attempt to directly map captured eye images to gaze point without any handcrafted features. Recently, availability of large datasets and novel deep learning techniques made appearance-based methods achieve superior accuracy than model-based approaches. However, many appearance-based gaze estimation systems perform well in within-dataset validation but fail to provide the same degree of accuracy in cross-dataset evaluation. Hence, it is still unclear how well the current state-of-the-art approaches perform in real-time in an interactive setting on unseen users. This paper proposes I2DNet, a novel architecture aimed to improve subject-independent gaze estimation accuracy that achieved a state-of-the-art 4.3 and 8.4 degree mean angle error on the MPIIGaze and RT-Gene datasets respectively. We have evaluated the proposed system as a gaze-controlled interface in real-time for a 9-block pointing and selection task and compared it with Webgazer.js and OpenFace 2.0. We have conducted a user study with 16 participants, and our proposed system reduces selection time and the number of missed selections statistically significantly compared to other two systems.
Downloads
Additional Files
Published
Issue
Section
License
Copyright (c) 2021 L R D Murthy, Siddhi Brahmbhatt, Somnath Arjun, Pradipta Biswas
This work is licensed under a Creative Commons Attribution 4.0 International License.