mpiorczynski's picture
Update README.md
1a7fb59 verified
|
raw
history blame
345 Bytes
metadata
language: en
license: apache-2.0
datasets:
  - bookcorpus
  - wikipedia
tags:
  - fill-mask
library_name: transformers

This model was derived from the bert-base-uncased checkpoint by replacing the GELU with ReLU activation function and further pre-training through several iterations to adapt it to the change of the activation function.