Skip to Main Content
 

Global Search Box

 
 
 
 

ETD Abstract Container

Abstract Header

Pretraining Deep Learning Models for Natural Language Understanding

Abstract Details

2020, BA, Oberlin College, Computer Science.
Since the first bidirectional deep learn- ing model for natural language understanding, BERT, emerged in 2018, researchers have started to study and use pretrained bidirectional autoencoding or autoregressive models to solve language problems. In this project, I conducted research to fully understand BERT and XLNet and applied their pretrained models to two language tasks: reading comprehension (RACE) and part-of-speech tagging (The Penn Treebank). After experimenting with those released models, I implemented my own version of ELECTRA, a pretrained text encoder as a discriminator instead of a generator to improve compute-efficiency, with BERT as its underlying architecture. To reduce the number of parameters, I replaced BERT with ALBERT in ELEC- TRA and named the new model, ALE (A Lite ELECTRA). I compared the performance of BERT, ELECTRA, and ALE on GLUE benchmark dev set after pretraining them with the same datasets for the same amount of training FLOPs.
John L. Donaldson (Advisor)
9 p.

Recommended Citations

Citations

  • Shao, H. (2020). Pretraining Deep Learning Models for Natural Language Understanding [Undergraduate thesis, Oberlin College]. OhioLINK Electronic Theses and Dissertations Center. http://rave.ohiolink.edu/etdc/view?acc_num=oberlin158955297757398

    APA Style (7th edition)

  • Shao, Han. Pretraining Deep Learning Models for Natural Language Understanding. 2020. Oberlin College, Undergraduate thesis. OhioLINK Electronic Theses and Dissertations Center, http://rave.ohiolink.edu/etdc/view?acc_num=oberlin158955297757398.

    MLA Style (8th edition)

  • Shao, Han. "Pretraining Deep Learning Models for Natural Language Understanding." Undergraduate thesis, Oberlin College, 2020. http://rave.ohiolink.edu/etdc/view?acc_num=oberlin158955297757398

    Chicago Manual of Style (17th edition)