상세 컨텐츠

본문 제목

11월 7일, 8일 주제 - ELECTRA

Data Scientist

by 자색 고구마칩 2021. 11. 7. 16:34

본문

https://arxiv.org/abs/2003.10555

 

ELECTRA: Pre-training Text Encoders as Discriminators Rather Than Generators

Masked language modeling (MLM) pre-training methods such as BERT corrupt the input by replacing some tokens with [MASK] and then train a model to reconstruct the original tokens. While they produce good results when transferred to downstream NLP tasks, the

arxiv.org

https://www.youtube.com/watch?v=BGRculoppT8&ab_channel=%EB%94%A5%EB%9F%AC%EB%8B%9D%EB%85%BC%EB%AC%B8%EC%9D%BD%EA%B8%B0%EB%AA%A8%EC%9E%84 

https://blog.pingpong.us/electra-review/

 

꼼꼼하고 이해하기 쉬운 ELECTRA 논문 리뷰

Review of ELECTRA: Pre-training Text Encoders as Discriminators Rather Than Generators

blog.pingpong.us

https://monologg.kr/2020/05/02/koelectra-part1/

 

'Data Scientist' 카테고리의 다른 글

11월 15일, 16일 주제 - BertGCN  (0) 2021.11.14
Electra  (0) 2021.11.09
BERT  (0) 2021.11.03
11월 3,4,5일 주제 - BERT  (0) 2021.11.03
BART  (0) 2021.11.01

관련글 더보기

댓글 영역