Weakly-Supervised Physically Unconstrained Gaze Estimation
About
A major challenge for physically unconstrained gaze estimation is acquiring training data with 3D gaze annotations for in-the-wild and outdoor scenarios. In contrast, videos of human interactions in unconstrained environments are abundantly available and can be much more easily annotated with frame-level activity labels. In this work, we tackle the previously unexplored problem of weakly-supervised gaze estimation from videos of human interactions. We leverage the insight that strong gaze-related geometric constraints exist when people perform the activity of "looking at each other" (LAEO). To acquire viable 3D gaze supervision from LAEO labels, we propose a training algorithm along with several novel loss functions especially designed for the task. With weak supervision from two large scale CMU-Panoptic and AVA-LAEO activity datasets, we show significant improvements in (a) the accuracy of semi-supervised gaze estimation and (b) cross-domain generalization on the state-of-the-art physically unconstrained in-the-wild Gaze360 gaze estimation benchmark. We open source our code at https://github.com/NVlabs/weakly-supervised-gaze.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Gaze Estimation | Gaze360 (test) | MAE (All 360°)13.2 | 40 | |
| Gaze Estimation | Gaze360 frontal face crops 1.0 (test) | Gaze Error (deg)10.1 | 12 | |
| Gaze Estimation | Gaze360 Front Facing | Mean Angular Error10.1 | 11 | |
| Gaze Estimation | Gaze360 (All360°) | Angular Error13.2 | 3 |