https://arxiv.org/abs/2003.10555
ELECTRA: Pre-training Text Encoders as Discriminators Rather Than Generators
Masked language modeling (MLM) pre-training methods such as BERT corrupt the input by replacing some tokens with [MASK] and then train a model to reconstruct the original tokens. While they produce good results when transferred to downstream NLP tasks, the
arxiv.org
https://blog.pingpong.us/electra-review/
꼼꼼하고 이해하기 쉬운 ELECTRA 논문 리뷰
Review of ELECTRA: Pre-training Text Encoders as Discriminators Rather Than Generators
blog.pingpong.us
https://monologg.kr/2020/05/02/koelectra-part1/
11월 15일, 16일 주제 - BertGCN (0) | 2021.11.14 |
---|---|
Electra (0) | 2021.11.09 |
BERT (0) | 2021.11.03 |
11월 3,4,5일 주제 - BERT (0) | 2021.11.03 |
BART (0) | 2021.11.01 |
댓글 영역