Share Email Print

Proceedings Paper

Computer-aided segmentation and 3D analysis of in vivo MRI examinations of the human vocal tract during phonation
Author(s): Axel Wismüller; Johannes Behrends; Phil Hoole; Gerda L. Leinsinger; Anke Meyer-Baese; Maximilian F. Reiser
Format Member Price Non-Member Price
PDF $17.00 $21.00

Paper Abstract

We developed, tested, and evaluated a 3D segmentation and analysis system for in vivo MRI examinations of the human vocal tract during phonation. For this purpose, six professionally trained speakers, age 22-34y, were examined using a standardized MRI protocol (1.5 T, T1w FLASH, ST 4mm, 23 slices, acq. time 21s). The volunteers performed a prolonged (≥21s) emission of sounds of the German phonemic inventory. Simultaneous audio tape recording was obtained to control correct utterance. Scans were made in axial, coronal, and sagittal planes each. Computer-aided quantitative 3D evaluation included (i) automated registration of the phoneme-specific data acquired in different slice orientations, (ii) semi-automated segmentation of oropharyngeal structures, (iii) computation of a curvilinear vocal tract midline in 3D by nonlinear PCA, (iv) computation of cross-sectional areas of the vocal tract perpendicular to this midline. For the vowels /a/,/e/,/i/,/o/,/ø/,/u/,/y/, the extracted area functions were used to synthesize phoneme sounds based on an articulatory-acoustic model. For quantitative analysis, recorded and synthesized phonemes were compared, where area functions extracted from 2D midsagittal slices were used as a reference. All vowels could be identified correctly based on the synthesized phoneme sounds. The comparison between synthesized and recorded vowel phonemes revealed that the quality of phoneme sound synthesis was improved for phonemes /a/ and /y/, if 3D instead of 2D data were used, as measured by the average relative frequency shift between recorded and synthesized vowel formants (p<0.05, one-sided Wilcoxon rank sum test). In summary, the combination of fast MRI followed by subsequent 3D segmentation and analysis is a novel approach to examine human phonation in vivo. It unveils functional anatomical findings that may be essential for realistic modelling of the human vocal tract during speech production.

Paper Details

Date Published: 12 March 2008
PDF: 5 pages
Proc. SPIE 6916, Medical Imaging 2008: Physiology, Function, and Structure from Medical Images, 69160T (12 March 2008);
Show Author Affiliations
Axel Wismüller, Univ. of Rochester (United States)
Univ. of Munich (Germany)
Johannes Behrends, Univ. of Munich (Germany)
Phil Hoole, Univ. of Munich (Germany)
Gerda L. Leinsinger, Univ. of Munich (Germany)
Anke Meyer-Baese, Florida State Univ. (United States)
Maximilian F. Reiser, Univ. of Munich (Germany)

Published in SPIE Proceedings Vol. 6916:
Medical Imaging 2008: Physiology, Function, and Structure from Medical Images
Xiaoping P. Hu; Anne V. Clough, Editor(s)

© SPIE. Terms of Use
Back to Top
Sign in to read the full article
Create a free SPIE account to get access to
premium articles and original research
Forgot your username?