64,953 Hits in 4.7 sec

Speech based Emotion Recognition using various Features and SVM Classifier

Supriya Jagtap
2019 International Journal for Research in Applied Science and Engineering Technology  
Basically, aim of this system to recognize the emotions and estimate the various features namely formant frequency, energy, pitch and MFCC from speech signal. accuracy of emotion detection system using  ...  In this paper methodology for human emotion recognizes by extracting the speech signal.  ...  CONCLUSIONS In this work, emotion detection based on speech analysis using SVM classifier is bestowed. The performance evaluated shows better accuracy.  ... 
doi:10.22214/ijraset.2019.3018 fatcat:3fjxhihrhvbv7dgax4fmm2nwf4

Instantaneous Emotion Detection System using Vocalizations

M. JayaLakshmi
2012 IOSR Journal of Engineering  
This paper describes an experimental study on the detection of emotion from speech.  ...  With this work both the negative and non-negative emotions can be easily detected.  ...  The phrases were all collected in English language and are throughout acted emotions.  ... 
doi:10.9790/3021-02721114 fatcat:72dcc23jhjd4rprtb33l2cezgm

Regional language Speech Emotion Detection using Deep Neural Network

Sweta Padman, Dhiraj Magare, M.D. Patil, V.A. Vyawahare
2022 ITM Web of Conferences  
Emotions assist people in communicating and understanding others' viewpoints by transmitting sentiments and providing feedback.The basic objective of speech emotion recognition is to enable computers to  ...  Voice emotion recognition is used in a variety of applications such as voice monitoring, online learning, clinical investigations, deception detection, entertainment, computer games, and call centres.  ...  Automatic Speech Emotion Recognition Speech emotion recognition systems use a person's speech to automatically detect his or her emotional state.  ... 
doi:10.1051/itmconf/20224403071 fatcat:tccammmy65eutkfabdfewaaiby

SPARTA: Speaker Profiling for ARabic TAlk [article]

Wael Farhan, Muhy Eddin Za'ter, Qusai Abu Obaidah, Hisham al Bataineh, Zyad Sober, Hussein T. Al-Natsheh
2020 arXiv   pre-print
This paper proposes a novel approach to an automatic estimation of three speaker traits from Arabic speech: gender, emotion, and dialect.  ...  First, The datasets were edited and thoroughly divided into train, development, and test sets (open to the public), and a benchmark was set for each task and dataset throughout the paper.  ...  Moreover, despite the significant shortage of publicly available corpora for Arabic speech, speech corpora that can be used for gender, emotion, and dialect detection tasks are assembled with benchmarks  ... 
arXiv:2012.07073v1 fatcat:dkbtd25njjbpfk4czisxglpxca

Emotional Techy Basyang: An Automated Filipino Narrative Storyteller

John Christopher P. Gonzaga, Jemimah A. Seguerra, Jhonnel A. Turingan, Mel Patrick A. Ulit, Ria A. Sagum
2014 International Journal of Future Computer and Communication  
Concatenative synthesis was the algorithm used in the TTS process wherein every speech audio that represents each syllables will be concatenated to each other with some pauses for speech turns and delimiters  ...  This study is specifically concerned in developing a storyteller application which uses sentiment analysis and includes Text-to-Speech (TTS) that converts the input text story into its audio output.  ...  to our families for their love and support during our thesis development, to Ernesto Rondon High School for allowing us to implement in their school and to our Almighty Lord Jesus Christ who gave us the  ... 
doi:10.7763/ijfcc.2014.v3.310 fatcat:r6wna7rz6fad7ddl46npsiyire

Punjabi Emotional Speech Database:Design, Recording and Verification

Kamaldeep Kaur
2021 International Journal of Intelligent Systems and Applications in Engineering  
This paper introduces Punjabi Emotional Speech Database that has been created to evaluate the recognition of emotions in speech, by the humans and the computer system.  ...  The database has been designed, recorded and verified using various standards. The results set a standard for identifying emotions from Punjabi speech.  ...  We are highly obliged to all the speakers and listeners for their active participation throughout the whole process.  ... 
doi:10.18201/ijisae.2021473641 fatcat:x4kmwfiwrbh4bpvni7ewajgfay

Speech Emotion Recognition using GFCC and BPNN

Shaveta Sharma, Parminder Singh
2014 International Journal of Engineering Trends and Technoloy  
From the past years, researchers have showed a very interest in the speech recognition systems based on the emotions.  ...  When the verbal content is well recognized on the speaker's emotion, a promising enhancement of such systems would come. So recognition in speech is a crucial step.  ...  Besides, the detection of files, video games and psychiatric aid are often claimed as further scenarios for emotion recognition.  ... 
doi:10.14445/22315381/ijett-v18p265 fatcat:qdzirswtibggveudurxoucfz24

Discriminating Neutral and Emotional Speech using Neural Networks

Sudarsana Reddy Kadiri, P. Gangamohan, B. Yegnanarayana
2014 International Conference on Natural Language Processing  
In this paper, we address the issue of speaker-specific emotion detection (neutral vs emotion) from speech signals with models for neutral speech as reference.  ...  As emotional speech is produced by the human speech production mechanism, the emotion information is expected to lie in the features of both excitation source and the vocal tract system.  ...  The changes are not sustainable for longer periods, and hence are not likely to be present throughout. This is due to an extra effort needed to produce the emotional speech.  ... 
dblp:conf/icon-nlp/KadiriGY14 fatcat:wm4tpxu7fbhflkw2p7zzvmogx4

Audio-Visual Recognition of Emotional Engagement of People with Dementia

Lars Steinert, Felix Putze, Dennis Küster, Tanja Schultz
2021 Conference of the International Speech Communication Association  
We further discuss the opportunities and challenges of detecting emotional engagement from speech in PwD.  ...  We demonstrate that they retained their ability to verbally express emotional engagement even at severe stages of the disease.  ...  We also gratefully acknowledge the support of the Leibniz ScienceCampus Bremen Digital Public Health (, which is jointly funded by the Leibniz Association (W4/2018), the Federal State of Bremen  ... 
doi:10.21437/interspeech.2021-567 dblp:conf/interspeech/SteinertPKS21 fatcat:ubvm36kcafeszgu2plonhm47yy

Emotion Recognition from Natural Phone Conversations in Individuals with and without Recent Suicidal Ideation

John Gideon, Heather T. Schatten, Melvin G. McInnis, Emily Mower Provost
2019 Interspeech 2019  
Participants selfreport their emotion periodically throughout the study. However, the dataset is relatively small and has uncertain labels.  ...  Suicidal ideation is also associated with emotion dysregulation. Therefore, in this work, we focus on the detection of emotion from speech and its relation to suicide.  ...  The data collection effort was reviewed and approved by the IRBs of Butler Hospital and the University of Michigan (HUM00052163).  ... 
doi:10.21437/interspeech.2019-1830 dblp:conf/interspeech/GideonSMP19 fatcat:y4begxgxmzemdcbyo3npfnjb5m

Exploration of Affect Sensing from Speech and Metaphorical Text [chapter]

Li Zhang
2009 Lecture Notes in Computer Science  
The detected affective states from text also play an important role in producing emotional animation for users' avatars. Evaluation of the affect detection from speech and text is provided.  ...  We report new developments on affect detection from textual metaphorical affective expression and affect sensing from speech.  ...  enjoyment with the emphasis of users' notice of the AI character's contribution throughout.  ... 
doi:10.1007/978-3-642-03364-3_31 fatcat:iojivos54vcx5bu4zoehpur2fi

Artificial Sensory Head

Sayani Manna
2015 International Journal of Signal Processing, Image Processing and Pattern Recognition  
All throughout the process, we have an face model that can be activated by speech and image.  ...  On receiving input from user of the system the type of emotion is detected. A bit pattern corresponding to each emotion is set. This generated bit is sent to the target application through usb port.  ...  The inputs are:  Speech  Still digital image When speech is given as input, it is processed and accordingly the type of emotion is detected.  ... 
doi:10.14257/ijsip.2015.8.11.34 fatcat:6qkf6xjsaza6dctr5ue4guz22m

Emotive Driver Advisor System (EDAS) [chapter]

Oleg Gusikhin, Erica Klampfl, Dimitar Filev, Yifan Chen
2011 Lecture Notes in Electrical Engineering  
, e.g. send Words Interpretation Spoken Dialogue System (SDS) Speech Signal Emotion recognition software Driver's emotional state EDAS system queries the web sends key words in text  ...  and/or button based interaction) • Context aware & emotive dialogue Emotion Recognition of Driver Ways to detect emotions: -Facial expressions: captured by camera -Voice: input through microphone  ... 
doi:10.1007/978-3-642-19539-6_2 fatcat:p4cxvf3qpfdwpl5r4ysevp6cja

Vocal Emotion Recognition Based on HMM and GMM for Mandarin Speech

Menghan Sun, Baochen Jiang, Jing Yuan
2012 International Journal of Education and Management Engineering  
The recognition of emotions from speech is a challenging issue.  ...  In this paper, two Hidden Markov Modelbased vocal emotion classifiers are trained and evaluated by an emotional mandarin speech corpus based on Mel-Frequency Cepstral Coefficient features.  ...  To evaluate the recognition performance, the speech utterances should be labeled to compare the detected emotion with the true emotion.  ... 
doi:10.5815/ijeme.2012.03.04 fatcat:uk47plenovabpncan6y5sf6t3y

A Storyteller's Tale: Literature Audiobooks Genre Classification Using CNN and RNN Architectures

Nehory Carmi, Azaria Cohen, Mireille Avigal, Anat Lerner
2019 Interspeech 2019  
Note that, throughout the reading, there are sections that are more typical to the book's genre than others.  ...  As the samples were taken sequentially throughout the reading of the books and were short in duration, we did not expect high classification rates.  ...  For audio analysis, it is used in emotion detection.  ... 
doi:10.21437/interspeech.2019-1154 dblp:conf/interspeech/CarmiCAL19 fatcat:63haf4c5i5bctjd5psl7mfdore
« Previous Showing results 1 — 15 out of 64,953 results