Poster + Presentation + Paper
9 October 2021 Multi-feature fusion gaze estimation based on attention mechanism
Zhangfang Hu, Yanling Xia, Yuan Luo, Lan Wang
Author Affiliations +
Conference Poster
Abstract
The variable head pose and low-quality eye images in natural scenes can lead to low accuracy of gaze estimation. In this paper, we propose a multi-feature fusion gaze estimation model based on the attention mechanism. First, face and eye feature extractors based on the group convolution channel and spatial attention mechanism (GCCSAM) are designed to use channel and spatial information to adaptively select and enhance important features in face images and two eye images, and suppress information irrelevant to gaze estimation. Then we design two feature fusion networks to fuse the features of face, two eyes and pupil center position, thus avoiding the effects of two-eye asymmetry and inaccurate head pose estimation on gaze estimation. The average angular error of the proposed method is 4.1° on MPIIGaze and 5.2° on EyeDiap. Compared with the current mainstream methods, our method effectively improves the accuracy and robustness of gaze estimation in natural scenes.
Conference Presentation
© (2021) COPYRIGHT Society of Photo-Optical Instrumentation Engineers (SPIE). Downloading of the abstract is permitted for personal use only.
Zhangfang Hu, Yanling Xia, Yuan Luo, and Lan Wang "Multi-feature fusion gaze estimation based on attention mechanism", Proc. SPIE 11897, Optoelectronic Imaging and Multimedia Technology VIII, 118970Z (9 October 2021); https://doi.org/10.1117/12.2602019
Advertisement
Advertisement
RIGHTS & PERMISSIONS
Get copyright permission  Get copyright permission on Copyright Marketplace
KEYWORDS
Eye

Convolution

Head

Error analysis

Eye models

Convolutional neural networks

Feature extraction

Back to Top