Filters








972 Hits in 5.2 sec

Intelligent Audio, Speech, and Music Processing Applications

Woon S. Gan, Sen M. Kuo, John H. L. Hansen
2008 EURASIP Journal on Audio, Speech, and Music Processing  
This special issue on intelligent audio, speech, and music processing (IASMP) consists of 13 papers that reflect a diverse range of disciplines in speech, audio, and music processing.  ...  These intelligent audio, speech, and music processing applications create an environment that is sensitive, adaptive, and responsive to the presence of users.  ...  This paper introduces a new 2 EURASIP Journal on Audio, Speech, and Music Processing frequency-domain adaptive algorithm for network echo cancellation.  ... 
doi:10.1155/2008/854716 fatcat:jtkbcmp4uzgdfgua5qzsowecdi

Perceptual Models for Speech, Audio, and Music Processing

Jont B. Allen, Wai-Yip Geoffrey Chan, Stephen Voran
2007 EURASIP Journal on Audio, Speech, and Music Processing  
This special issue on Perceptual Models for Speech, Audio, and Music Processing contains seven papers that exemplify the breadth and depth of current work in perceptual modeling and its applications.  ...  New understandings of human auditory perception have recently contributed to advances in numerous areas related to audio, speech, and music processing.  ...  or apply such models to the important open problems in speech, audio, and music processing.  ... 
doi:10.1155/2007/12687 fatcat:y4dtql6w45grhi35p3twgwjdry

Atypical Speech

Georg Stemmer, Elmar Nöth, Vijay Parsa
2010 EURASIP Journal on Audio, Speech, and Music Processing  
The article On the impact of children's emotional speech on acoustic and language models by S. Steidl, A. Batliner, D. Seppi, and B.  ...  The majority of the articles analyses the influence of atypical speech on automatic speech recognition performance in great detail, and different methods to reduce the error rates for atypical speech are  ...  When the speech recognizer is trained on neutral speech, the somewhat surprising result for a collection of spontaneous utterances from children is 2 EURASIP Journal on Audio, Speech, and Music Processing  ... 
doi:10.1155/2010/835974 fatcat:trnubdf6k5ditchy3grxg7obom

Data-based spatial audio processing

Maximo Cobos, Jens Ahrens, Konrad Kowalczyk, Archontis Politis
2022 EURASIP Journal on Audio, Speech, and Music Processing  
and enhanced speech quality.  ...  The article by Zieliński et al. addresses the automatic disambiguation between front and back audio sources in binaural music recordings.  ...  Publisher's Note Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.  ... 
doi:10.1186/s13636-022-00248-5 dblp:journals/ejasmp/CobosAKP22a fatcat:izhfdqyjbrdmvikt45fcmojwsu

An imperceptible and robust audio watermarking algorithm

Ali Al-Haj
2014 EURASIP Journal on Audio, Speech, and Music Processing  
In this paper, we propose a semi-blind, imperceptible, and robust digital audio watermarking algorithm.  ...  Imperceptibility, robustness, and high data payload of the proposed algorithm are demonstrated using different musical clips.  ...  Accordingly, we evaluated the performance of the proposed algorithm using three mono audio signals representing pop music, instrumental music, and speech.  ... 
doi:10.1186/s13636-014-0037-2 fatcat:eli3icb5wfe3ln6pvquqczp74u

Multimicrophone Speech Dereverberation: Experimental Validation

Koen Eneman, Marc Moonen
2007 EURASIP Journal on Audio, Speech, and Music Processing  
and a speech recognition system.  ...  Dereverberation is required in various speech processing applications such as handsfree telephony and voice-controlled systems, especially when signals are applied that are recorded in a moderately or  ...  Jacques Duchateau for making available the automatic speech recognition system of the ESAT-PSI research group.  ... 
doi:10.1155/2007/51831 fatcat:t2u5qhuwsfbifm5mo3jxd6kkma

Wideband Speech Recovery Using Psychoacoustic Criteria

Visar Berisha, Andreas Spanias
2007 EURASIP Journal on Audio, Speech, and Music Processing  
In this paper, we propose a novel speech compression method that relies on bandwidth extension.  ...  Many modern speech bandwidth extension techniques predict the high-frequency band based on features extracted from the lower band.  ...  An elaborate loudness estimation model is used to predict how much a particular frame of audio will 2 EURASIP Journal on Audio, Speech, and Music Processing benefit from a more precise representation of  ... 
doi:10.1155/2007/16816 fatcat:pa5egdsi7rdejk2ksop3ihgig4

Environmental Sound Synthesis, Processing, and Retrieval

Andrea Valle
2010 EURASIP Journal on Audio, Speech, and Music Processing  
Apart by the specific solutions proposed by the authors, the reader's perspective is enriched also by the different focus EURASIP Journal on Audio, Speech, and Music Processing 3 of the two contributions  ...  In order to deal with this complexity, it is possible to cite at least some fields and authors that have provided a general frame for the understanding of (environmental) 2 EURASIP Journal on Audio, Speech  ... 
doi:10.1155/2010/178164 fatcat:hdsbdzl2uvam5ppq6kecixu3ke

An audio watermark-based speech bandwidth extension method

Zhe Chen, Chengyong Zhao, Guosheng Geng, Fuliang Yin
2013 EURASIP Journal on Audio, Speech, and Music Processing  
A novel speech bandwidth extension method based on audio watermark is presented in this paper.  ...  The time-domain and frequency-domain envelope parameters are extracted from the high-frequency components of speech signal, and then these parameters are embedded in the corresponding narrowband speech  ...  Conclusions A speech bandwidth extension method based on the modified audio watermark is proposed in this paper.  ... 
doi:10.1186/1687-4722-2013-10 fatcat:db6etypoujbgtgdt3tetns3d6e

The Latin Music Mood Database

Carolina L. dos Santos, Carlos N. Silla
2015 EURASIP Journal on Audio, Speech, and Music Processing  
The method for assigning mood labels to the musical recordings is based on the knowledge of a professionally trained Brazilian musician and the identification of the predominant emotion perceived in each  ...  In this paper we present the Latin Music Mood Database, an extension of the Latin Music Database but for the task of music mood/emotion classification.  ...  Acknowledgements We thank the anonymous reviewers for their insightful and very valuable feedback on the earlier versions of this manuscript. Received: 14 August 2014 Accepted: 24 July 2015  ... 
doi:10.1186/s13636-015-0065-6 fatcat:gxdmv6jdsjgjdjqlcprcudlij4

Semantic Labeling of Nonspeech Audio Clips

Xiaojuan Ma, Christiane Fellbaum, Perry Cook
2010 EURASIP Journal on Audio, Speech, and Music Processing  
Finally, we examined which kinds of entities and events, when captured by nonlinguistic acoustic clips, appear to be well-suited to elicit information for communication, and which ones are less discriminable  ...  We analyzed the human labels in terms of their lexical and semantic properties to ascertain that the audio clips do evoke the information suggested by their pre-defined captions.  ...  They are grateful to the Kimberly and Frank H.  ... 
doi:10.1155/2010/404860 fatcat:b6zuse7gnff47dt4jxkbteihey

Semantic Labeling of Nonspeech Audio Clips

Xiaojuan Ma, Christiane Fellbaum, Perry Cook
2010 EURASIP Journal on Audio, Speech, and Music Processing  
Finally, we examined which kinds of entities and events, when captured by nonlinguistic acoustic clips, appear to be well-suited to elicit information for communication, and which ones are less discriminable  ...  We analyzed the human labels in terms of their lexical and semantic properties to ascertain that the audio clips do evoke the information suggested by their pre-defined captions.  ...  They are grateful to the Kimberly and Frank H.  ... 
doi:10.1186/1687-4722-2010-404860 fatcat:nkqffoh2kvcvxgbkiwsoyrhr4i

Speech enhancement methods based on binaural cue coding

Xianyun Wang, Changchun Bao
2019 EURASIP Journal on Audio, Speech, and Music Processing  
After that, they are combined one by one to form a codebook.  ...  features of noisy speech and the T-F masks.  ...  ., grant no. 61831019 and grant no.61471014). Authors' contributions All authors read and approved the final manuscript.  ... 
doi:10.1186/s13636-019-0164-x fatcat:3ysievp6ibfr5oy5c25cec4m3y

Audio bandwidth extension based on temporal smoothing cepstral coefficients

Xin Liu, Chang-Chun Bao
2014 EURASIP Journal on Audio, Speech, and Music Processing  
In this paper, we propose a wideband (WB) to super-wideband audio bandwidth extension (BWE) method based on temporal smoothing cepstral coefficients (TSCC).  ...  The resulting 'steady-state' spectrum is processed by frequency weighting, and the temporal smoothing cepstral coefficients are obtained by means of the power-law loudness function and cepstral normalization  ...  Six test signals including pop music, guitar, sax, drums, and speech were selected at random from the MPEG database, and the level of the original test signals and the processed signals was normalized  ... 
doi:10.1186/s13636-014-0041-6 fatcat:p57hkkwg3jfj3dpgpxz7lhdw54

Multiple-Description Multistage Vector Quantization

Pradeepa Yahampath
2007 EURASIP Journal on Audio, Speech, and Music Processing  
Multistage vector quantization (MSVQ) is a technique for low complexity implementation of high-dimensional quantizers, which has found applications within speech, audio, and image coding.  ...  An algorithm for optimizing the codebooks of an MD-MSVQ for a given packet-loss probability is suggested, and a practical example involving quantization of speech line spectral frequency (LSF) vectors  ...  The side distortion in System D is due to the error in predicting the current LSF, based on the 6 EURASIP Journal on Audio, Speech, and Music Processing previously reconstructed LSF (which depends on the  ... 
doi:10.1155/2007/67146 fatcat:oofebovm75cchhfp4lw2uw3sce
« Previous Showing results 1 — 15 out of 972 results