A Bi-Directional LSTM-CNN Model with Attention for Aspect-Level Text Classification

被引:29
作者
Zhu, Yonghua [1 ]
Gao, Xun [2 ]
Zhang, Weilin [2 ]
Liu, Shenkai [1 ]
Zhang, Yuanyuan [3 ]
机构
[1] Shanghai Univ, Shanghai Film Acad, Shanghai 200444, Peoples R China
[2] Shanghai Univ, Sch Comp Engn & Sci, Shanghai 200444, Peoples R China
[3] Zhejiang Chinese Med Univ, Coll Informat Technol, Hangzhou 310053, Zhejiang, Peoples R China
来源
FUTURE INTERNET | 2018年 / 10卷 / 12期
关键词
attention mechanism; NLP; aspect-level sentiment classification;
D O I
10.3390/fi10120116
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
The prevalence that people share their opinions on the products and services in their daily lives on the Internet has generated a large quantity of comment data, which contain great business value. As for comment sentences, they often contain several comment aspects and the sentiment on these aspects are different, which makes it meaningless to give an overall sentiment polarity of the sentence. In this paper, we introduce Attention-based Aspect-level Recurrent Convolutional Neural Network (AARCNN) to analyze the remarks at aspect-level. The model integrates attention mechanism and target information analysis, which enables the model to concentrate on the important parts of the sentence and to make full use of the target information. The model uses bidirectional LSTM (Bi-LSTM) to build the memory of the sentence, and then CNN is applied to extracting attention from memory to get the attentive sentence representation. The model uses aspect embedding to analyze the target information of the representation and finally the model outputs the sentiment polarity through a softmax layer. The model was tested on multi-language datasets, and demonstrated that it has better performance than conventional deep learning methods.
引用
收藏
页数:11
相关论文
共 24 条
  • [1] [Anonymous], 2005, P 4 SIGHANWORKSHOP C
  • [2] Fast Image Reconstruction With L2-Regularization
    Bilgic, Berkin
    Chatnuntawech, Itthi
    Fan, Audrey P.
    Setsompop, Kawin
    Cauley, Stephen F.
    Wald, Lawrence L.
    Adalsteinsson, Elfar
    [J]. JOURNAL OF MAGNETIC RESONANCE IMAGING, 2014, 40 (01) : 181 - 191
  • [3] Dong L, 2014, PROCEEDINGS OF THE 52ND ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, VOL 2, P49
  • [4] Du J, 2017, P NAT CCF C NAT LANG, P183
  • [5] Goldhahn D., 2012, P 8 INT LANG RESS EV
  • [6] Graves A, 2012, STUD COMPUT INTELL, V385, P1, DOI [10.1007/978-3-642-24797-2, 10.1162/neco.1997.9.1.1]
  • [7] Jiang Long, 2011, P 49 ANN M ASS COMP
  • [8] Joachims T, 1999, MACHINE LEARNING, PROCEEDINGS, P200
  • [9] Krizhevsky A., 2017, ADV NEURAL INFORM PR, V60, DOI [10.1145/3065386, DOI 10.1145/3065386]
  • [10] Ma DH, 2017, PROCEEDINGS OF THE TWENTY-SIXTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, P4068