Appearance-based Gaze Estimation and Applications in Healthcare

dc.contributor.advisor

Sapiro, Guillermo

dc.contributor.author

Chang, Zhuoqing

dc.date.accessioned

2020-06-09T18:00:21Z

dc.date.available

2022-05-27T08:17:23Z

dc.date.issued

2020

dc.department

Electrical and Computer Engineering

dc.description.abstract

Gaze estimation, the ability to predict where a person is looking, has become an indispensable technology in healthcare research. Current tools for gaze estimation rely on specialized hardware and are typically used in well-controlled laboratory settings. Novel appearance-based methods directly estimate a person's gaze from the appearance of their eyes, making gaze estimation possible with ubiquitous, low-cost devices, such as webcams and smartphones. This dissertation presents new methods on appearance-based gaze estimation as well as applying this technology to solve challenging problems in practical healthcare applications.

One limitation of appearance-based methods is the need to collect a large amount of training data to learn the highly variant eye appearance space. To address this fundamental issue, we develop a method to synthesize novel images of the eye using data from a low-cost RGB-D camera and show that this data augmentation technique can improve gaze estimation accuracy significantly. In addition, we explore the potential of utilizing visual saliency information as a means to transparently collect weakly-labelled gaze data at scale. We show that the collected data can be used to personalize a generic gaze estimation model to achieve better performance on an individual.

In healthcare applications, the possibility of replacing specialized hardware with ubiquitous devices when performing eye-gaze analysis is a major asset that appearance-based methods brings to the table. In the first application, we assess the risk of autism in toddlers by analyzing videos of them watching a set of expert-curated stimuli on a mobile device. We show that appearance-based methods can be used to estimate their gaze position on the device screen and that differences between the autistic and typically-developing populations are significant. In the second application, we attempt to detect oculomotor abnormalities in people with cerebellar ataxia using video recorded from a mobile phone. By tracking the iris movement of participants while they watch a short video stimuli, we show that we are able to achieve high sensitivity and specificity in differentiating people with smooth pursuit oculomotor abnormalities from those without.

dc.identifier.uri

https://hdl.handle.net/10161/21030

dc.subject

Electrical engineering

dc.subject

Computer science

dc.subject

Autism

dc.subject

Computer vision

dc.subject

eye tracking

dc.subject

Gaze estimation

dc.subject

Machine learning

dc.title

Appearance-based Gaze Estimation and Applications in Healthcare

dc.type

Dissertation

duke.embargo.months

23.572602739726026

Files

Original bundle

Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
Chang_duke_0066D_15775.pdf
Size:
17.61 MB
Format:
Adobe Portable Document Format

Collections