TY - GEN
T1 - Estimate gaze density by incorporating emotion
AU - Liu, Huiying
AU - Xu, Min
AU - He, Xiangjian
AU - Wang, Jinqiao
PY - 2014/11/3
Y1 - 2014/11/3
N2 - Gaze density estimation has attracted many research efforts in the past years. The factors considered in the existing methods include low level feature saliency, spatial position, and objects. Emotion, as an important factor driving attention, has not been taken into account. In this paper, we are the first to estimate gaze density through incorporating emotion. To estimate the emotion intensity of each position in an image, we consider three aspects, generic emotional content, facial expression intensity, and emotional objects. Generic emotional content is estimated by using Multiple instance learning, which is employed to train an emotion detector from weakly labeled images. Facial expression intensity is estimated by using a ranking method. Emotional objects are detected, by taking blood/injury and worm/snake as examples. Finally, emotion intensity, low level feature saliency, and spatial position, are fused, through a linear support vector machine, to estimate gaze density. The performance is tested on public eye tracking dataset. Experimental results indicate that incorporating emotion does improve the performance of gaze density estimation.
AB - Gaze density estimation has attracted many research efforts in the past years. The factors considered in the existing methods include low level feature saliency, spatial position, and objects. Emotion, as an important factor driving attention, has not been taken into account. In this paper, we are the first to estimate gaze density through incorporating emotion. To estimate the emotion intensity of each position in an image, we consider three aspects, generic emotional content, facial expression intensity, and emotional objects. Generic emotional content is estimated by using Multiple instance learning, which is employed to train an emotion detector from weakly labeled images. Facial expression intensity is estimated by using a ranking method. Emotional objects are detected, by taking blood/injury and worm/snake as examples. Finally, emotion intensity, low level feature saliency, and spatial position, are fused, through a linear support vector machine, to estimate gaze density. The performance is tested on public eye tracking dataset. Experimental results indicate that incorporating emotion does improve the performance of gaze density estimation.
KW - Emotion
KW - Visual attention
KW - Visual saliency
UR - http://www.scopus.com/inward/record.url?scp=84913598136&partnerID=8YFLogxK
U2 - 10.1145/2647868.2654983
DO - 10.1145/2647868.2654983
M3 - Conference contribution
AN - SCOPUS:84913598136
T3 - MM 2014 - Proceedings of the 2014 ACM Conference on Multimedia
SP - 1113
EP - 1116
BT - MM 2014 - Proceedings of the 2014 ACM Conference on Multimedia
PB - Association for Computing Machinery
T2 - 2014 ACM Conference on Multimedia, MM 2014
Y2 - 3 November 2014 through 7 November 2014
ER -