Filters








379 Hits in 5.8 sec

Ultra2Speech – A Deep Learning Framework for Formant Frequency Estimation and Tracking from Ultrasound Tongue Images [article]

Pramit Saha, Yadong Liu, Bryan Gick, Sidney Fels
2020 arXiv   pre-print
This work addresses the articulatory-to-acoustic mapping problem based on ultrasound (US) tongue images for the development of a silent-speech interface (SSI) that can provide them with an assistance in  ...  We use a novel deep learning architecture to map US tongue images from the US probe placed beneath a subject's chin to formants that we call, Ultrasound2Formant (U2F) Net.  ...  This work was funded by the Natural Sciences and Engineering Research Council (NSERC) of Canada and Canadian Institutes for Health Research (CIHR).  ... 
arXiv:2006.16367v1 fatcat:7g57iknkh5ah3dvc2bpv2sgnnq

Automated detection of the tongue surface in sequences of ultrasound images

Michael Unser, Maureen Stone
1992 Journal of the Acoustical Society of America  
It is designed to process sequences of sagittal tongue sections that are digitized in real time and stored in standard tagged image file format (TIFF).  ...  Results obtained using ultrasound data are presented.  ...  Earlier manual tracking of the tongue surface profile in a sequence of scans was prohibitively slow, as even one second of data provided 60 tongue surface contours (video fields).  ... 
doi:10.1121/1.402934 pmid:1629491 fatcat:cfmeloeirzg5za4wcnoalu436y

Recording and analyzing kinematic data in children and adults with SOLLAR: Sonographic & Optical Linguo-Labial Articulation Recording system

Aude Noiray, Jan Ries, Mark Tiede, Elina Rubertus, Catherine Laporte, Lucie Ménard
2020 Laboratory Phonology  
Our gratitude to Anthony de Simone for constructing the probe holder and pedestal used in our research.  ...  We also thank all students at LOLA who have contributed to improve the SOLLAR platform over the past couple of years and to a broader extent, researchers in the ultrasound imaging research community who  ...  automatically tracing tongue contours in ultrasound video data.  ... 
doi:10.5334/labphon.241 fatcat:l2uuj3gznzazpgu75pwswhnsi4

Tongue `N' Groove: An Ultrasound Based Music Controller

Florian Vogt, Graeme Mccaig, Mir A. Ali, Sidney S. Fels
2002 Zenodo  
Here we propose a novel musical controller which acquires imaging data of the tongue with a two-dimensional medical ultrasound scanner.  ...  We evaluate the mapping space between tongue shape and controller parameters and its expressive characteristics.  ...  ACKNOWLEDGMENTS We thank Perry Cook for his contribution of SPASM and Bryan Gick for help with the ultrasound scanner. We also thank Paula Wirth for creating the illustration in Figure 2 .  ... 
doi:10.5281/zenodo.1176468 fatcat:52i6tfx4uvalbczw7ugbbu6wiu

Fully-automated tongue detection in ultrasound images

Elham Karimi, Lucie Ménard, Catherine Laporte
2019 Computers in Biology and Medicine  
Two novel quality measures were also developed that predict the reliability of the segmentation result so that an image with a reliable contour can be chosen to confidently initialize fully automated tongue  ...  This work introduces a new method for extracting tongue contours in ultrasound images that requires no training nor manual intervention.  ...  semi-automatic approach to accurately track the tongue contours in US video sequences.  ... 
doi:10.1016/j.compbiomed.2019.103335 pmid:31279163 fatcat:w447rlh3ebaqbnrjoycoetmvue

Graph-based tracking of the tongue contour in ultrasound sequences with adaptive temporal regularization

Lisa Tang, Ghassan Hamarneh
2010 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition - Workshops  
We propose a graph-based approach for semi-automatic tracking of the human tongue in 2D+time ultrasound image sequences.  ...  Tongue contour tracking is formulated as a graph-labeling problem, where each vertex is labeled with a displacement vector describing its motion.  ...  Tim Bressmann from the Department of Speech-Language Pathology, University of Toronto, for provision of the US data. We also thank Bianca Herold and Matt Toom for assistance in data preparation.  ... 
doi:10.1109/cvprw.2010.5543597 dblp:conf/cvpr/TangH10 fatcat:t4hrellao5hb3amnouhp6g73zy

Development of a silent speech interface driven by ultrasound and optical images of the tongue and lips

Thomas Hueber, Elie-Laurent Benaroya, Gérard Chollet, Bruce Denby, Gérard Dreyfus, Maureen Stone
2010 Speech Communication  
Visual features are extracted from ultrasound images of the tongue and from video images of the lips using a PCA-based image coding technique.  ...  In the recognition stage, the visual HMMs are used to identify phonetic targets in a sequence of visual features.  ...  Acknowledgements This work was supported by the French Department of Defense (DGA), the "Centre de Microélectronique de Paris Ile-de-France" (CEMIP) and the French National  ... 
doi:10.1016/j.specom.2009.11.004 fatcat:rhbuz6kr4vhgbiffxkorticmyu

Increased midsagittal tongue velocity as indication of articulatory compensation in patients with lateral partial glossectomies

Orchid Rastadmehr, Tim Bressmann, Ron Smyth, Jonathan C. Irish
2008 Head and Neck  
Using B-mode ultrasound, the midsagittal tongue movement of 10 patients with lateral partial glossectomy during a standardized reading passage was analyzed before and after surgery.  ...  The main outcome measure was the tongue velocity during speech. The technique of defect reconstruction (local vs flap) was included as a covariate in the analysis. Results.  ...  Pascal Van Lieshout, Department of Speech-Language Pathology, University of Toronto, and 2 anonymous Head & Neck reviewers for helpful comments and feedback.  ... 
doi:10.1002/hed.20772 pmid:18213728 fatcat:275vz2yjdvgtpowwxlnzxcf5eq

2003: Designing, Playing, and Performing with a Vision-Based Mouth Interface [chapter]

Michael J. Lyons, Michael Hähnel, Nobuji Tetsutani
2017 Current Research in Systematic Musicology  
The role of the face and mouth in speech production as well as non-verbal communication suggests the use of facial action t o control musical sound.  ...  We report our experience with various gesture-to-sound mappings and musical applications, and describe a live performance which used the Mouthesizer interface.  ...  With the Tongue 'n' Groove, an ultrasound device is held below the jaw and an image of the tongue contour reconstructed, or alternatively, optical flow due to tongue motion is calculated.  ... 
doi:10.1007/978-3-319-47214-0_8 fatcat:jyzkc3lzi5av3lk4hckew4ybje

Dynamic 3-D Visualization of Vocal Tract Shaping During Speech

Yinghua Zhu, Yoon-Chul Kim, M. I. Proctor, S. S. Narayanan, K. S. Nayak
2013 IEEE Transactions on Medical Imaging  
after manual segmentation of targeted articulators and smoothing.  ...  We present a novel method for the creation of 3-D dynamic movies of vocal tract shaping based on the acquisition of 2-D dynamic data from parallel slices and temporal alignment of the image sequences using  ...  ACKNOWLEDGMENT The authors acknowledge the support and collaboration of the SPAN (Speech Production and Articulation kNowledge) group at the University of Southern California.  ... 
doi:10.1109/tmi.2012.2230017 pmid:23204279 pmcid:PMC3896513 fatcat:d455sbyk5rc5lobaii6wo7mkui

Native Language Influence on Brass Instrument Performance: An Application of Generalized Additive Mixed Models (GAMMs) to Midsagittal Ultrasound Images of the Tongue

Matthias Heyne, Donald Derrick, Jalal Al-Tamimi
2019 Frontiers in Psychology  
After normalizing to account for differences in vocal tract shape and ultrasound transducer orientation, we used generalized additive mixed models (GAMMs) to estimate average tongue surface shapes used  ...  While the New Zealand English-speaking participants employed a playing tongue shape approximating schwa and the vowel used in the word 'lot,' the Tongan participants used a tongue shape loosely patterning  ...  Tongue Contour Tracing and Outlier Removal It is important to understand that ultrasound measurements are usually exported as sequences of individual images (or videos) with almost all information contained  ... 
doi:10.3389/fpsyg.2019.02597 pmid:31827453 pmcid:PMC6890863 fatcat:t7ygcotfx5dmpmnccc4wcd3glu

International Conference on Image Processing

1996 Proceedings of 3rd IEEE International Conference on Image Processing  
for perceptual processing of video sequences Error visualisation and presentation in video compression systems Image quality prediction for bitrate allocation On the partition of binary edge maps as a  ...  structure for video data: a formal specification Volume data coding based on region segmentation using finite mixture model Buffer-constrained coding of video sequences with quasi-constant quality Automatic  ... 
doi:10.1109/icip.1996.560353 fatcat:le3ysy6wxrfr7nq56ueropy7tu

Acoustic and articulatory analysis of French vowels produced by congenitally blind adults and sighted adults

Lucie Ménard, Corinne Toupin, Shari R. Baum, Serge Drouin, Jérôme Aubin, Mark Tiede
2013 Journal of the Acoustical Society of America  
Synchronous ultrasound, acoustic, and video recordings of the participants articulating the ten French oral vowels were made.  ...  Furthermore, blind speakers use smaller differences in lip protrusion but larger differences in tongue position and shape than their sighted peers to produce rounding and place of articulation contrasts  ...  ACKNOWLEDGMENTS This work was supported by the Social Sciences and Humanities Research Council of Canada and the Natural Sciences and Engineering Research Council of Canada.  ... 
doi:10.1121/1.4818740 pmid:24116433 fatcat:6ozlx6aek5cf3kbjthhirmyh3e

Proceedings of 3rd IEEE International Conference on Image Processing

1996 Proceedings of 3rd IEEE International Conference on Image Processing ICIP-96  
for perceptual processing of video sequences Error visualisation and presentation in video compression systems Image quality prediction for bitrate allocation On the partition of binary edge maps as a  ...  and estimation in object-oriented video coding Segmentation of an image sequence using multi-dimensional image attributes Scalable video with background segmentation Sven Siggelkow, Rolf-Rainer Griga  ... 
doi:10.1109/icip.1996.559416 fatcat:jb4cdydgf5edtdljfuzj423ozu

Back From the Future: Nonlinear Anticipation in Adults' and Children's Speech

Aude Noiray, Martijn Wieling, Dzhuma Abakarova, Elina Rubertus, Mark Tiede
2019 Journal of Speech, Language and Hearing Research  
Method The technique of ultrasound imaging was employed to record tongue movement at 5 time points throughout short utterances of the form V1#CV2.  ...  In children, this differentiation is not yet mature: Vowels show greater prominence over time and seem activated more in phase with those of previous segments relative to adults.  ...  In the same way that isolines are used in topographic maps to represent locations sharing the same altitude, the red contour lines connect points that have a similar (predicted, based on all trials) tongue  ... 
doi:10.1044/2019_jslhr-s-csmc7-18-0208 pmid:31465705 fatcat:4onzqcn5hfhcfkfgh7ehypbhe4
« Previous Showing results 1 — 15 out of 379 results