--- license: mit --- # Distilled-RoBERTa The DistilBERT model is a [RoBERTa](https://huggingface.co/deepset/roberta-base-squad2-distilled) model, which is trained on the SQuAD 2.0 training set, fine-tuned on the [NewsQA](https://huggingface.co/datasets/lucadiliello/newsqa) dataset.