31575
An Open-Source, Computing Platform-Agnostic, Calibration-Free Preferential Gaze Detection Approach for Social Preference Assessment
Objectives: To develop a portable, fast, open-source preferential gaze tracking pipeline for visual-stimuli guided social preference assessments.
Methods: The proposed gaze-tracking pipeline was deployed on data sourced from the ‘preferential-looking’ task of START (“Screening Tool for Autism Risk using Technology)”, a tablet-based app usable by non-specialists in home settings. The task was administered to a group of 127 children aged 2-7 years at their homes in India. The preferential looking task presents the child with simultaneous social and nonsocial videos on the tablet screen. The tablet camera records the child’s face during the task. This recording is analyzed asynchronously by our gaze tracking algorithm to predict the child’s visual preference in each frame. The algorithm implemented in python is deployed offline on a desktop system, and processes the data for multiple children in a batch mode. The average processing time for the algorithm is 90ms per frame. The algorithm pipeline, consisting of existing open-source modules,
(a) extracts (using OpenCV[1] library calls) regions corresponding to the face and the eyes,
(b) obtains gaze coordinates[2] in pixels based on the device used (no reference needed for the subject or the camera parameters), and
(c) transforms, with temporal smoothing by a median filter, raw pixel coordinates to higher level preference prediction, enabling social preference assessment.
Results: The algorithm achieved 91% accuracy (percentage of frames with correct classification of gaze preference) on a manually annotated dataset with 5000 frames. The frame-drop rate (failure to detect faces due to poor lighting, occlusions, multiple faces, etc.) is observed to be no more than 5% for more than 73.2% of the on-field dataset, and it exhibits a decreasing trend in this range (Figure 1).
Conclusions: Our computer program (available freely[3]) works on video frames, and can, therefore, be used in any system as long as video frames are provided. The video can be captured by non-specialists in casual environments regardless of the hardware employed, making the system cost-effective for a screening measure. The approach is calibration-free, and can be executed asynchronously on video inputs.
References:
[1] Bradski, G. The OpenCV Library, Dr. Dobb's Journal of Software Tools, 2000
[2] Kyle Krafka, Aditya Khosla, Petr Kellnhofer, Harini Kannan, Suchi Bhandarkar, Wojciech Matusik and Antonio Torralba. “Eye Tracking for Everyone”. IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2016.
[3] The Apache Software Foundation License: https://www.apache.org/licenses (Last Accessed: Nov 10, 2018)