From 935ac13b473164bb9d578640e33d9f21144c365e Mon Sep 17 00:00:00 2001 From: Julien Chaumond Date: Fri, 11 Dec 2020 22:23:53 +0100 Subject: [PATCH] Migrate model card from transformers-repo Read announcement at https://discuss.huggingface.co/t/announcement-all-model-cards-will-be-migrated-to-hf-co-model-repos/2755 Original file history: https://github.com/huggingface/transformers/commits/master/model_cards/distilbert-base-cased-README.md --- README.md | 40 ++++++++++++++++++++++++++++++++++++++++ 1 file changed, 40 insertions(+) create mode 100644 README.md diff --git a/README.md b/README.md new file mode 100644 index 0000000..184ee3a --- /dev/null +++ b/README.md @@ -0,0 +1,40 @@ +--- +language: en +license: apache-2.0 +datasets: +- bookcorpus +- wikipedia +--- + +# DistilBERT base model (cased) + +This model is a distilled version of the [BERT base model](https://huggingface.co/bert-base-cased). +It was introduced in [this paper](https://arxiv.org/abs/1910.01108). +The code for the distillation process can be found +[here](https://github.com/huggingface/transformers/tree/master/examples/distillation). +This model is cased: it does make a difference between english and English. + +All the training details on the pre-training, the uses, limitations and potential biases are the same as for [DistilBERT-base-uncased](https://huggingface.co/distilbert-base-uncased). +We highly encourage to check it if you want to know more. + +## Evaluation results + +When fine-tuned on downstream tasks, this model achieves the following results: + +Glue test results: + +| Task | MNLI | QQP | QNLI | SST-2 | CoLA | STS-B | MRPC | RTE | +|:----:|:----:|:----:|:----:|:-----:|:----:|:-----:|:----:|:----:| +| | 81.5 | 87.8 | 88.2 | 90.4 | 47.2 | 85.5 | 85.6 | 60.6 | + +### BibTeX entry and citation info + +```bibtex +@article{Sanh2019DistilBERTAD, + title={DistilBERT, a distilled version of BERT: smaller, faster, cheaper and lighter}, + author={Victor Sanh and Lysandre Debut and Julien Chaumond and Thomas Wolf}, + journal={ArXiv}, + year={2019}, + volume={abs/1910.01108} +} +```