共 22 条
[1]
ERNIE 2.0: A Continual Pre-Training Framework for Language Understanding[J] . Yu Sun,Shuohuan Wang,Yukun Li,Shikun Feng,Hao Tian,Hua Wu,Haifeng Wang.Proceedings of the AAAI Conference on Artificial Intelligence . 2020 (05)
[2]
ERNIE: Enhanced Representation through Knowledge Integration[J] . Yu Sun,Shuohuan Wang,Yukun Li,Shikun Feng,Xuyi Chen,Han Zhang,Xin Tian,Danxiang Zhu,Hao Tian,Hua Wu 0003.CoRR . 2019
[3]
NEZHA: Neural Contextualized Representation for Chinese Language Understanding[J] . Junqiu Wei,Xiaozhe Ren,Xiaoguang Li,Wenyong Huang,Yi Liao,Yasheng Wang,Jiashu Lin,Xin Jiang,Xiao Chen,Qun Liu.CoRR . 2019
[4]
Transformer-XL: Attentive Language Models Beyond a Fixed-Length Context[J] . Zihang Dai,Zhilin Yang,Yiming Yang,Jaime G. Carbonell,Quoc V. Le,Ruslan Salakhutdinov.CoRR . 2019
[5]
RoBERTa: A Robustly Optimized BERT Pretraining Approach[J] . Yinhan Liu,Myle Ott,Naman Goyal,Jingfei Du,Mandar Joshi,Danqi Chen,Omer Levy,Mike Lewis,Luke Zettlemoyer,Veselin Stoyanov.CoRR . 2019
[6]
ALBERT: A Lite BERT for Self-supervised Learning of Language Representations[J] . Zhenzhong Lan,Mingda Chen,Sebastian Goodman,Kevin Gimpel,Piyush Sharma,Radu Soricut.CoRR . 2019
[7]
Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer[J] . Colin Raffel,Noam Shazeer,Adam Roberts,Katherine Lee,Sharan Narang,Michael Matena,Yanqi Zhou,Wei Li,Peter J. Liu.CoRR . 2019
[8]
XLNet: Generalized Autoregressive Pretraining for Language Understanding[J] . Zhilin Yang,Zihang Dai,Yiming Yang,Jaime G. Carbonell,Ruslan Salakhutdinov,Quoc V. Le.CoRR . 2019
[9]
Unified Language Model Pre-training for Natural Language Understanding and Generation[J] . Li Dong 0004,Nan Yang,Wenhui Wang,Furu Wei,Xiaodong Liu,Yu Wang,Jianfeng Gao,Ming Zhou,Hsiao-Wuen Hon.CoRR . 2019
[10]
StructBERT: Incorporating Language Structures into Pre-training for Deep Language Understanding[J] . Wei Wang,Bin Bi,Ming Yan,Chen Wu,Zuyi Bao,Liwei Peng,Luo Si.CoRR . 2019