A copy of this work was available on the public web and has been preserved in the Wayback Machine. The capture dates from 2019; you can also visit the original URL.
The file type is application/pdf
.
Towards expressive musical robots: a cross-modal framework for emotional gesture, voice and music
2012
EURASIP Journal on Audio, Speech, and Music Processing
It has been long speculated that expression of emotions from different modalities have the same underlying 'code', whether it be a dance step, musical phrase, or tone of voice. This is the first attempt to implement this theory across three modalities, inspired by the polyvalence and repeatability of robotics. We propose a unifying framework to generate emotions across voice, gesture, and music, by representing emotional states as a 4-parameter tuple of speed, intensity, regularity, and extent
doi:10.1186/1687-4722-2012-3
fatcat:gz6vyoj42rh7ze2fno524aivgq