中文预训练模型研究进展

被引:12
作者
侯钰涛
阿布都克力木·阿布力孜
哈里旦木·阿布都克里木
机构
[1] 新疆财经大学信息管理学院
关键词
中文预训练模型; 自然语言处理; 词向量; 预处理; 深度学习;
D O I
暂无
中图分类号
TP391.1 [文字信息处理];
学科分类号
081203 ; 0835 ;
摘要
近年来,预训练模型在自然语言处理领域蓬勃发展,旨在对自然语言隐含的知识进行建模和表示,但主流预训练模型大多针对英文领域。中文领域起步相对较晚,鉴于其在自然语言处理过程中的重要性,学术界和工业界都开展了广泛的研究,提出了众多的中文预训练模型。文中对中文预训练模型的相关研究成果进行了较为全面的回顾,首先介绍预训练模型的基本概况及其发展历史,对中文预训练模型主要使用的两种经典模型Transformer和BERT进行了梳理,然后根据不同模型所属类别提出了中文预训练模型的分类方法,并总结了中文领域的不同评测基准,最后对中文预训练模型未来的发展趋势进行了展望。旨在帮助科研工作者更全面地了解中文预训练模型的发展历程,继而为新模型的提出提供思路。
引用
收藏
页码:148 / 163
页数:16
相关论文
共 22 条
[1]  
ERNIE 2.0: A Continual Pre-Training Framework for Language Understanding[J] . Yu Sun,Shuohuan Wang,Yukun Li,Shikun Feng,Hao Tian,Hua Wu,Haifeng Wang.Proceedings of the AAAI Conference on Artificial Intelligence . 2020 (05)
[2]  
ERNIE: Enhanced Representation through Knowledge Integration[J] . Yu Sun,Shuohuan Wang,Yukun Li,Shikun Feng,Xuyi Chen,Han Zhang,Xin Tian,Danxiang Zhu,Hao Tian,Hua Wu 0003.CoRR . 2019
[3]  
NEZHA: Neural Contextualized Representation for Chinese Language Understanding[J] . Junqiu Wei,Xiaozhe Ren,Xiaoguang Li,Wenyong Huang,Yi Liao,Yasheng Wang,Jiashu Lin,Xin Jiang,Xiao Chen,Qun Liu.CoRR . 2019
[4]  
Transformer-XL: Attentive Language Models Beyond a Fixed-Length Context[J] . Zihang Dai,Zhilin Yang,Yiming Yang,Jaime G. Carbonell,Quoc V. Le,Ruslan Salakhutdinov.CoRR . 2019
[5]  
RoBERTa: A Robustly Optimized BERT Pretraining Approach[J] . Yinhan Liu,Myle Ott,Naman Goyal,Jingfei Du,Mandar Joshi,Danqi Chen,Omer Levy,Mike Lewis,Luke Zettlemoyer,Veselin Stoyanov.CoRR . 2019
[6]  
ALBERT: A Lite BERT for Self-supervised Learning of Language Representations[J] . Zhenzhong Lan,Mingda Chen,Sebastian Goodman,Kevin Gimpel,Piyush Sharma,Radu Soricut.CoRR . 2019
[7]  
Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer[J] . Colin Raffel,Noam Shazeer,Adam Roberts,Katherine Lee,Sharan Narang,Michael Matena,Yanqi Zhou,Wei Li,Peter J. Liu.CoRR . 2019
[8]  
XLNet: Generalized Autoregressive Pretraining for Language Understanding[J] . Zhilin Yang,Zihang Dai,Yiming Yang,Jaime G. Carbonell,Ruslan Salakhutdinov,Quoc V. Le.CoRR . 2019
[9]  
Unified Language Model Pre-training for Natural Language Understanding and Generation[J] . Li Dong 0004,Nan Yang,Wenhui Wang,Furu Wei,Xiaodong Liu,Yu Wang,Jianfeng Gao,Ming Zhou,Hsiao-Wuen Hon.CoRR . 2019
[10]  
StructBERT: Incorporating Language Structures into Pre-training for Deep Language Understanding[J] . Wei Wang,Bin Bi,Ming Yan,Chen Wu,Zuyi Bao,Liwei Peng,Luo Si.CoRR . 2019