Using Recurrent Neural Networks for Slot Filling in Spoken Language Understanding

被引:359
作者
Mesnil, Gregoire [1 ,2 ]
Dauphin, Yann [3 ]
Yao, Kaisheng [4 ]
Bengio, Yoshua [3 ]
Deng, Li [4 ]
Hakkani-Tur, Dilek [5 ]
He, Xiaodong [4 ]
Heck, Larry [7 ]
Tur, Gokhan [6 ]
Yu, Dong [4 ]
Zweig, Geoffrey [4 ]
机构
[1] Univ Rouen, Dept Comp Sci, F-76821 Mont St Aignan, France
[2] Univ Montreal, Dept Comp Sci, Montreal, PQ H3T 1J4, Canada
[3] Univ Montreal, Dept Comp Sci & Operat Res, Montreal, PQ H3T 1J4, Canada
[4] Microsoft Res, Redmond, WA 98052 USA
[5] Microsoft Res, Mountain View, CA 94043 USA
[6] Apple, Cupertino, CA 95014 USA
[7] Google, Mountain View, CA 94043 USA
关键词
Recurrent neural network (RNN); slot filling; spoken language understanding (SLU); word embedding; DEEP CONVEX NETWORKS;
D O I
10.1109/TASLP.2014.2383614
中图分类号
O42 [声学];
学科分类号
070206 ; 082403 ;
摘要
Semantic slot filling is one of the most challenging problems in spoken language understanding (SLU). In this paper, we propose to use recurrent neural networks (RNNs) for this task, and present several novel architectures designed to efficiently model past and future temporal dependencies. Specifically, we implemented and compared several important RNN architectures, including Elman, Jordan, and hybrid variants. To facilitate reproducibility, we implemented these networks with the publicly available Theano neural network toolkit and completed experiments on the well-known airline travel information system (ATIS) benchmark. In addition, we compared the approaches on two custom SLU data sets from the entertainment and movies domains. Our results show that the RNN-based models outperform the conditional random field (CRF) baseline by 2% in absolute error reduction on the ATIS benchmark. We improve the state-of-the-art by 0.5% in the Entertainment domain, and 6.7% for the movies domain.
引用
收藏
页码:530 / 539
页数:10
相关论文
共 70 条
  • [1] [Anonymous], 2012, Adv. Neural Inf. Process. Syst.
  • [2] [Anonymous], 2012, SEMANTIC COMPOSITION, DOI DOI 10.1162/153244303322533223
  • [3] [Anonymous], 2010, P PYTHON SCI COMPUTI
  • [4] [Anonymous], PROCEEDINGS
  • [5] [Anonymous], 2014, P ACL
  • [6] [Anonymous], 2009, LEARNING DEEP ARCHIT
  • [7] [Anonymous], 2014, P 2014 C EMP METH NA, DOI DOI 10.3115/V1/D14-1002
  • [8] [Anonymous], 2013, P INTERSPEECH
  • [9] [Anonymous], P ACM INT C INF KNOW
  • [10] [Anonymous], P INTERSPEECH