A copy of this work was available on the public web and has been preserved in the Wayback Machine. The capture dates from 2018; you can also visit the original URL.
The file type is application/pdf
.
Prediction-Based Audiovisual Fusion for Classification of Non-Linguistic Vocalisations
2016
IEEE Transactions on Affective Computing
Prediction plays a key role in recent computational models of the brain and it has been suggested that the brain constantly makes multisensory spatiotemporal predictions. Inspired by these findings we tackle the problem of audiovisual fusion from a new perspective based on prediction. We train predictive models which model the spatiotemporal relationship between audio and visual features by learning the audio-to-visual and visual-to-audio feature mapping for each class. Similarly, we train
doi:10.1109/taffc.2015.2446462
fatcat:2xtesk55lfewnn66jzo6fu45q4