LCF: A Local Context Focus Mechanism for Aspect-Based Sentiment Classification

被引:166
作者
Zeng, Biqing [1 ]
Yang, Heng [2 ]
Xu, Ruyang [2 ]
Zhou, Wu [2 ]
Han, Xuli [2 ]
机构
[1] South China Normal Univ, Sch Software, Foshan 528225, Peoples R China
[2] South China Normal Univ, Sch Comp, Guangzhou 510631, Guangdong, Peoples R China
来源
APPLIED SCIENCES-BASEL | 2019年 / 9卷 / 16期
基金
中国国家自然科学基金;
关键词
aspect-level sentiment classification; local context focus; self-attention; pretrained BERT;
D O I
10.3390/app9163389
中图分类号
O6 [化学];
学科分类号
0703 ;
摘要
Aspect-based sentiment classification (ABSC) aims to predict sentiment polarities of different aspects within sentences or documents. Many previous studies have been conducted to solve this problem, but previous works fail to notice the correlation between the aspect's sentiment polarity and the local context. In this paper, a Local Context Focus (LCF) mechanism is proposed for aspect-based sentiment classification based on Multi-head Self-Attention (MHSA). This mechanism is called LCF design, and utilizes the Context features Dynamic Mask (CDM) and Context Features Dynamic Weighted (CDW) layers to pay more attention to the local context words. Moreover, a BERT-shared layer is adopted to LCF design to capture internal long-term dependencies of local context and global context. Experiments are conducted on three common ABSC datasets: the laptop and restaurant datasets of SemEval-2014 and the ACL twitter dataset. Experimental results demonstrate that the LCF baseline model achieves considerable performance. In addition, we conduct ablation experiments to prove the significance and effectiveness of LCF design. Especially, by incorporating with BERT-shared layer, the LCF-BERT model refreshes state-of-the-art performance on all three benchmark datasets.
引用
收藏
页数:22
相关论文
共 34 条
[1]  
[Anonymous], 2007, ACL 07
[2]  
[Anonymous], 2017, NEURIPS
[3]  
[Anonymous], 2014, P 8 INT WORKSH SEM E
[4]  
[Anonymous], P 24 INT JOINT C ART
[5]  
Bahdanau Dzmitry, 2015, 3 INT C LEARN REPR I
[6]   A neural probabilistic language model [J].
Bengio, Y ;
Ducharme, R ;
Vincent, P ;
Jauvin, C .
JOURNAL OF MACHINE LEARNING RESEARCH, 2003, 3 (06) :1137-1155
[7]  
Bo Pang, 2008, Foundations and Trends in Information Retrieval, V2, P1, DOI 10.1561/1500000001
[8]  
Cho K., 2014, INT C MACH LEARN ICM
[9]  
Devlin J, 2019, 2019 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL HLT 2019), VOL. 1, P4171
[10]  
Dong L, 2014, PROCEEDINGS OF THE 52ND ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, VOL 2, P49