Appearance-based Gaze Estimation and Applications in Healthcare
dc.contributor.advisor | Sapiro, Guillermo | |
dc.contributor.author | Chang, Zhuoqing | |
dc.date.accessioned | 2020-06-09T18:00:21Z | |
dc.date.available | 2022-05-27T08:17:23Z | |
dc.date.issued | 2020 | |
dc.department | Electrical and Computer Engineering | |
dc.description.abstract | Gaze estimation, the ability to predict where a person is looking, has become an indispensable technology in healthcare research. Current tools for gaze estimation rely on specialized hardware and are typically used in well-controlled laboratory settings. Novel appearance-based methods directly estimate a person's gaze from the appearance of their eyes, making gaze estimation possible with ubiquitous, low-cost devices, such as webcams and smartphones. This dissertation presents new methods on appearance-based gaze estimation as well as applying this technology to solve challenging problems in practical healthcare applications. One limitation of appearance-based methods is the need to collect a large amount of training data to learn the highly variant eye appearance space. To address this fundamental issue, we develop a method to synthesize novel images of the eye using data from a low-cost RGB-D camera and show that this data augmentation technique can improve gaze estimation accuracy significantly. In addition, we explore the potential of utilizing visual saliency information as a means to transparently collect weakly-labelled gaze data at scale. We show that the collected data can be used to personalize a generic gaze estimation model to achieve better performance on an individual. In healthcare applications, the possibility of replacing specialized hardware with ubiquitous devices when performing eye-gaze analysis is a major asset that appearance-based methods brings to the table. In the first application, we assess the risk of autism in toddlers by analyzing videos of them watching a set of expert-curated stimuli on a mobile device. We show that appearance-based methods can be used to estimate their gaze position on the device screen and that differences between the autistic and typically-developing populations are significant. In the second application, we attempt to detect oculomotor abnormalities in people with cerebellar ataxia using video recorded from a mobile phone. By tracking the iris movement of participants while they watch a short video stimuli, we show that we are able to achieve high sensitivity and specificity in differentiating people with smooth pursuit oculomotor abnormalities from those without. | |
dc.identifier.uri | ||
dc.subject | Electrical engineering | |
dc.subject | Computer science | |
dc.subject | Autism | |
dc.subject | Computer vision | |
dc.subject | eye tracking | |
dc.subject | Gaze estimation | |
dc.subject | Machine learning | |
dc.title | Appearance-based Gaze Estimation and Applications in Healthcare | |
dc.type | Dissertation | |
duke.embargo.months | 23.572602739726026 |
Files
Original bundle
- Name:
- Chang_duke_0066D_15775.pdf
- Size:
- 17.61 MB
- Format:
- Adobe Portable Document Format