A copy of this work was available on the public web and has been preserved in the Wayback Machine. The capture dates from 2017; you can also visit the original URL.
The file type is application/pdf
.
An efficient audiovisual saliency model to predict eye positions when looking at conversations
2015
2015 23rd European Signal Processing Conference (EUSIPCO)
Classic models of visual attention dramatically fail at predicting eye positions on visual scenes involving faces. While some recent models combine faces with low-level features, none of them consider sound as an input. Yet it is crucial in conversation or meeting scenes. In this paper, we describe and refine an audiovisual saliency model for conversation scenes. This model includes a speaker diarization algorithm which automatically modulates the saliency of conversation partners' faces and
doi:10.1109/eusipco.2015.7362640
dblp:conf/eusipco/CoutrotG15
fatcat:k4pma5eskrcrrknzwidi4llpzu