Fill-Mask
Transformers
PyTorch
English
bert
pretraining

80% 1x4 Block Sparse BERT-Base (uncased) Prune OFA

This model is was created using Prune OFA method described in Prune Once for All: Sparse Pre-Trained Language Models presented in ENLSP NeurIPS Workshop 2021.

For further details on the model and its result, see our paper and our implementation available here.

Downloads last month
6
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Datasets used to train Intel/bert-base-uncased-sparse-80-1x4-block-pruneofa

Collection including Intel/bert-base-uncased-sparse-80-1x4-block-pruneofa