Filters








2,451 Hits in 13.2 sec

Towards Facial Gestures Generation by Speech Signal Analysis Using HUGE Architecture [chapter]

Goranka Zoric, Karlo Smid, Igor S. Pandzic
2009 Lecture Notes in Computer Science  
Information needed for generation of facial gestures is extracted from speech prosody by analyzing natural speech in real-time.  ...  In our current work we concentrate on finding correlation between speech signal and occurrence of facial gestures. Motivation behind this work is computer-generated human correspondent, ECA.  ...  The work was partly carried out within the research project "Embodied Conversational Agents as interface for networked and mobile services" supported by the Ministry of Science, Education and Sports of  ... 
doi:10.1007/978-3-642-00525-1_11 fatcat:tizzuzvgvfettm6622e53jxvpa

Automatic Technologies for Processing Spoken Sign Languages

Alexey Karpov, Irina Kipyatkova, Milos Zelezny
2016 Procedia Computer Science  
Sign languages use visual-kinetic clues for human-to-human communication combining hand gestures with lips articulation and facial mimics.  ...  They also possess a special grammar that is quite different from that of speech-based spoken languages.  ...  codes by OpenGL 15 ; multi-modal user interface, which integrates all the components for automatic generation of SL gestures, auditory and visual speech (articulation and facial expressions) in the signing  ... 
doi:10.1016/j.procs.2016.04.050 fatcat:bn7zmnnlvzg6zkpj7itl3km2x4

Towards a Multimedia Knowledge-Based Agent with Social Competence and Human Interaction Capabilities

Leo Wanner, Ioannis Kompatsiaris, Elisabeth André, Florian Lingenfelser, Gregor Mehlmann, Andries Stam, Ludo Stellingwerff, Bianca Vieru, Lori Lamel, Wolfgang Minker, Louisa Pragst, Josep Blat (+8 others)
2016 Proceedings of the 1st International Workshop on Multimedia Analysis and Retrieval for Multimodal Interaction - MARMI '16  
It is composed of an ontology-based and reasoning-driven dialogue manager, multimodal communication analysis and generation modules and a search engine for the retrieval of multimedia background content  ...  Analysis of facial and gestural communication Facial expressions and gestures communicate semantic and/ or affective information.  ...  a combination of facial, gestural and multilingual verbal signals, embedded into a specific cultural, social and emotional context; (iii) plan the dialogue using ontology-based reasoning techniques in  ... 
doi:10.1145/2927006.2927011 dblp:conf/mir/WannerBDDLMSKVK16 fatcat:h5znrszjszcujenxbw7zl2w7ui

Multilingual and Multimodal Corpus-Based Text-to-Speech System - PLATTOS - [chapter]

Matej Rojc, Izidor Mlakar
2011 Speech and Language Technologies  
Heterogeneous relation graphs used in the TTS architecture All TTS processing steps contribute to the linguistic information used for generating the speech signal.  ...  At the end, only the correct order from the input must be preserved, before playing-out generated speech signals.  ...  By using emphasis markers word/phrase-break markers, ECA EVA can generate different speech-driven pointing gestures that can visually emphasize a certain word/phrase.  ... 
doi:10.5772/16245 fatcat:ga3vmxvvevcx3a4352oehbwcky

Towards More Realistic Human-Robot Conversation: A Seq2Seq-based Body Gesture Interaction System [article]

Minjie Hua, Fuyuan Shi, Yibing Nan, Kai Wang, Hao Chen, Shiguo Lian
2019 arXiv   pre-print
Both models are adapted from the sequence-to-sequence (seq2seq) architecture to synthesize body gestures represented by the movements of twelve upper-body keypoints.  ...  The proposed system consists of a listening model and a speaking model used in corresponding conversational phases.  ...  They adopted RNN encoder-decoder architecture to generate both verbal responses and facial expressions for a chatting avatar.  ... 
arXiv:1905.01641v3 fatcat:yeldr5hszvh2xov7cgpmw7g6je

The 'All-at-Onceness' of Embodied, Faceto- Face Interaction

Liesbet Quaeghebeur
2012 Cognitive Semiotics  
I survey existing theories about this matter - mainly coming from gesture studies - and extend McNeill's (1992, 2005) intrapersonal language-thought-hand nexus to the interpersonal realm: the speech/gesture  ...  talk involving co-expressive and synchronized facial expressions, gesticulation, gaze, posture, and other bodily means of conveying information.  ...  These accounts of how speech and gesture (or multimodal expression in general) are produced simultaneously differ in the cognitive architectures they propose; yet they share a common concept of communication  ... 
doi:10.1515/cogsem.2012.4.1.167 fatcat:vkvybwnnabf3tpryuzqbtus32u

The 'All-at-Onceness' of Embodied, Faceto- Face Interaction

Liesbet Quaeghebeur
2009 Cognitive Semiotics  
the speech/gesture unit does not transport thought from one mind into another; rather, embodied interactions constitute shared thoughts.  ...  talk involving co-expressive and synchronized facial expressions, gesticulation, gaze, posture, and other bodily means of conveying information.  ...  These accounts of how speech and gesture (or multimodal expression in general) are produced simultaneously differ in the cognitive architectures they propose; yet they share a common concept of communication  ... 
doi:10.1515/cogsem.2009.4.1.167 fatcat:cemyq3t6vrejzptrj5bkrk7fxe

Modeling Social Signals and Contexts in Robotic Socially Believable Behaving Systems [chapter]

Anna Esposito, Lakhmi C. Jain
2016 Intelligent Systems Reference Library  
For example, in speech, the influence of visual on auditory signals perception is proved by the McGurk effect [28] .  ...  Gestures act in partnership with speech, building up shared knowledge and meanings when the interactional exchange is successful [10, 15, 23] .  ... 
doi:10.1007/978-3-319-31053-4_2 fatcat:mcghgv4vfjbahi7c2higpdbrni

Multimodal Language Processing in Human Communication

Judith Holler, Stephen C. Levinson
2019 Trends in Cognitive Sciences  
posed by a turn-taking system.  ...  The natural ecology of human language is face-to-face interaction comprising the exchange of a plethora of multimodal signals.  ...  of visual signals alone is predictive of questionhood at the speech-act level (indicated by the black broken arrow pointing up towards the social-action level).  ... 
doi:10.1016/j.tics.2019.05.006 pmid:31235320 fatcat:ilu3sewylrc5dmzcvs7fivba44

A Review on Deep Learning Algorithms for Speech and Facial Emotion Recognition

Charlyn Pushpa Latha, Mohana Priya
2016 APTIKOM Journal on Computer Science and Information Technologies  
Facial Electromyogram (FEMG) signals are used to detect the different emotions of humans.  ...  This paper focuses on the review of some of the deep learning techniques used by various researchers which paved the way to improve the classification accuracy of the FEMG signals as well as the speech  ...  This paper attempts to present various techniques that can be used to recognize the emotions using FEMG signals and also using speech signals.  ... 
doi:10.11591/aptikom.j.csit.118 fatcat:gerpzx54qrgrtf3pqnnzgjywim

A Review on Deep Learning Algorithms for Speech and Facial Emotion Recognition

Charlyn Pushpa Latha, Mohana Priya
2020 APTIKOM Journal on Computer Science and Information Technologies  
This paperfocuses on the review of some of the deep learning techniques used by various researchers which paved the way toimprove the classification accuracy of the FEMG signals as well as the speech signals  ...  Facial Electromyogram (FEMG) signals are used to detect the different emotionsof humans.  ...  This paper attempts to present various techniques that can be used to recognize the emotions using FEMG signals and also using speech signals.  ... 
doi:10.34306/csit.v1i3.55 fatcat:l2tska7j5ferna4wupt3f2jcp4

Building Expression into Virtual Characters [article]

V. Vinayagamoorthy, M. Gillies, A. Steed, E. Tanguy, X. Pan, C. Loscos, M. Slater
2006 Eurographics State of the Art Reports  
We cover models of individual characters' emotion and personality, models of interpersonal behaviour and methods for generating expression.  ...  In particular, we assume that a virtual character representation is already available, and we describe a variety of models and methods that are used to give the characters more "depth" so that they are  ...  simultaneously generating speech and gestures.  ... 
doi:10.2312/egst.20061052 fatcat:hps7ewshhrdi3jbi5ros57lf2q

A Systematic Review on Affective Computing: Emotion Models, Databases, and Recent Advances [article]

Yan Wang, Wei Song, Wei Tao, Antonio Liotta, Dawei Yang, Xinlei Li, Shuyong Gao, Yixuan Sun, Weifeng Ge, Wei Zhang, Wenqiang Zhang
2022 arXiv   pre-print
However, it is hard to reveal one's inner emotion hidden purposely from facial expressions, audio tones, body gestures, etc.  ...  Firstly, we introduce two typical emotion models followed by commonly used databases for affective computing.  ...  Spectral features are often obtained by transforming the time-domain speech signal into the frequency-domain speech signal using the Fourier transform [173] . Bitouk et al.  ... 
arXiv:2203.06935v3 fatcat:h4t3omkzjvcejn2kpvxns7n2qe

Automatic Sign Language Gesture Recognition using Prewitt & Morphological Dilation

2020 International Journal of Engineering and Advanced Technology  
They are visual languages that rely on hand gestures as well as on bodily and facial expressions.  ...  Automatic sign language gesture recognition is an approach for recognizing gestures and converts it to its actual meaning and convey either through speech or text as per requirements.  ...  Automatic Sign Language Gesture Recognition using Prewitt & Morphological Dilation Although signaling is primarily used by the deaf and hard of hearing, it is also used by hearing people, such as those  ... 
doi:10.35940/ijeat.a1884.1010120 fatcat:63xveelibbft5ghi5oxlxardsq

Towards an Articulation-Based Developmental Robotics Approach for Word Processing in Face-to-Face Communication

Bernd J. Kröger, Peter Birkholz, Christiane Neuschaefer-Rube
2011 Paladyn: Journal of Behavioral Robotics  
and human-like behavior like e.g. speaking and co-speech gesturing.  ...  This is a fruitful basic scenario not only for learning to speak, but also for learning to communicate in general, including to produce co-verbal manual gestures and to produce co-verbal facial expressions  ...  Acknowledgments This work was supported in part by German Research Council (DFG) grant Kr 1439/13-1 and grant Kr 1439/15-1 and in part by COST-action 2102.  ... 
doi:10.2478/s13230-011-0016-6 fatcat:vpg24n4lcvfy7lhlwcjad5u2qe
« Previous Showing results 1 — 15 out of 2,451 results