MODELING THE IDENTIFICATION OF CONCURRENT VOWELS WITH DIFFERENT FUNDAMENTAL FREQUENCIES

被引:155
作者
MEDDIS, R
HEWITT, MJ
机构
[1] Department of Human Sciences, University of Technology, Loughborough
关键词
D O I
10.1121/1.402767
中图分类号
O42 [声学];
学科分类号
070206 ; 082403 ;
摘要
Human listeners are better able to identify two simultaneous vowels if the fundamental frequencies of the vowels are different. A computational model is presented which, for the first time, is able to simulate this phenomenon at least qualitatively. The first stage of the model is based upon a bank of bandpass filters and inner hair-cell simulators that simulate approximately the most relevant characteristics of the human auditory periphery. The output of each filter/hair-cell channel is then autocorrelated to extract pitch and timbre information. The pooled autocorrelation function (ACF) based on all channels is used to derive a pitch estimate for one of the component vowels from a signal composed of two vowels. Individual channel ACFs showing a pitch peak at this value are combined and used to identify the first vowel using a template matching procedure. The ACFs in the remaining channels are then combined and used to identify the second vowel. Model recognition performance shows a rapid improvement in correct vowel identification as the difference between the fundamental frequencies of two simultaneous vowels increases from zero to one semitone in a manner closely resembling human performance. As this difference increases up to four semitones, performance improves further only slowly, if at all.
引用
收藏
页码:233 / 245
页数:13
相关论文
共 46 条