BioMed-RoBERTa (Gururangan et al., 2020) is a recent model based on RoBERTa-base. BioMed- RoBERTa is initialized from RoBERTa-base, with an additional pretraining of 12.5K steps with a batch size of 2048, using a corpus of 2.7M scien- tific papers from Semantic Scholar (Ammar et al., 2018).
BioMed-RoBERTa-base BioMed-RoBERTa-base is a language model based on the RoBERTa-base (Liu et. al, 2019) architecture. We adapt RoBERTa-base to 2.68 millionĀ
BioMed-RoBERTa-base is a language model based on the RoBERTa-base (Liu et. al, 2019) architecture. We adapt RoBERTa-base to 2.68 million scientific papers from the Semantic Scholar corpus via continued pretraining. This amounts to 7.55B tokens and 47GB of data.
BioMed-RoBERTa-base is continuously pre-trained on scientific biomedical articles based on the RoBERTa-base architecture. Both models have obtained good performance on biomedical domain tasks (Liu et al., 2019; Gururangan et al., 2020) including the relation extraction task we are studying.