publication . Article . 2021

Review of Pre-training Techniques for Natural Language Processing

CHEN Deguang, MA Jinlin, MA Ziping, ZHOU Jie;
Open Access Chinese
  • Published: 01 Aug 2021 Journal: Jisuanji kexue yu tansuo, volume 15, issue 8, pages 1,359-1,389 (issn: 1673-9418, Copyright policy)
  • Publisher: Journal of Computer Engineering and Applications Beijing Co., Ltd., Science Press
Abstract
In the published reviews of natural language pre-training technology, most literatures only elaborate neural network pre-training technologies or a brief introduction to traditional pre-training technologies, which may result in the development process of natural language pre-training dissected artificially from natural language processing. Therefore, in order to avoid this phenomenon, this paper covers the process of natural language pre-training with four points as follows. Firstly, the traditional natural language pre-training technologies and neural network pre-training technologies are introduced according to the updating route of pre-training technology. With the characteristics of related technologies analyzed, compared, this paper sums up the process of development context and trend of natural language processing technology. Secondly, based on the improved BERT (bidirectional encoder representation from transformers), this paper mainly introduces the latest natural language processing models from two aspects and sums up these models from pre-training mechanism, advantages and disadvantages, performance and so on. The main application fields of natural language processing are presented. Furthermore, this paper explores the challenges and corresponding solutions to natural language processing models. Finally, this paper summarizes the work of this paper and prospects the future development direction, which can help researchers understand the development of pre-training technologies of natural language more comprehensively and provide some ideas to design new models and new pre-training methods.
Subjects
free text keywords: pre-training techniques, natural language processing, neural network, Electronic computers. Computer science, QA75.5-76.95
Communities
  • Digital Humanities and Cultural Heritage
Download from
Any information missing or wrong?Report an Issue