metadata
language: en
license: apache-2.0
datasets:
- bookcorpus
- wikipedia
tags:
- fill-mask
library_name: transformers
This model was derived from the bert-base-uncased checkpoint by replacing the GELU with ReLU activation function and further pre-training through several iterations to adapt it to the change of the activation function.