Skip to main content
RoBERTa: A Robustly Optimized BERT Pretraining Approach

RoBERTa: A Robustly Optimized BERT Pretraining Approach

Yinhan Liu, et al.

00
2019-07-26
nlptransformers

Abstract

This paper introduces and evaluates the idea described in “RoBERTa: A Robustly Optimized BERT Pretraining Approach”, and reports empirical results that helped shape subsequent work in nlp, transformers.