Head-transducer models for speech translation and their automatic acquisition from bilingual data

被引:5
作者
Alshawi, H. [1 ]
Bangalore, S. [1 ]
Douglas, S. [1 ]
机构
[1] AT and T Labs Research, 180 Park Avenue, Florham Park, NJ 07932, United States
关键词
Algorithms - Dynamic programming - Mathematical models - Speech processing - Speech recognition - Statistical methods - Transducers;
D O I
10.1023/A:1011187330969
中图分类号
学科分类号
摘要
This article presents statistical language translation models, called dependency transduction models, based on collections of head transducers. Head transducers are middle-out finite-state transducers which translate a head word in a source string into its corresponding head in the target language, and further translate sequences of dependents of the source head into sequences of dependents of the target head. The models are intended to capture the lexical sensitivity of direct statistical translation models, while at the same time taking account of the hierarchical phrasal structure of language. Head transducers are suitable for direct recursive lexical translation, and are simple enough to be trained fully automatically. We present a method for fully automatic training of dependency transduction models for which the only input is transcribed and translated speech utterances. The method has been applied to create English-Spanish and English-Japanese translation models for speech translation applications. The dependency transduction model gives around 75% accuracy for an English-Spanish translation task (using a simple string edit-distance measure) and 70% for an English-Japanese translation task. Enhanced with target n-grams and a case-based component, English-Spanish accuracy is over 76%; for English-Japanese it is 73% for transcribed speech, and 60% for translation from recognition word lattices.
引用
收藏
页码:105 / 124
相关论文
empty
未找到相关数据