sshleifer/distilbart-cnn-12-6 is a forked repo from huggingface. License: apache-2-0
Go to file
patil-suraj a4f8f3ea90 add flax model 2021-06-14 07:51:12 +00:00
.gitattributes add flax model 2021-06-14 07:51:12 +00:00
README.md Update README.md 2020-07-14 17:35:15 +00:00
config.json add flax model 2021-06-14 07:51:12 +00:00
flax_model.msgpack add flax model 2021-06-14 07:51:12 +00:00
merges.txt Update merges.txt 2020-06-23 23:23:42 +00:00
pytorch_model.bin Update pytorch_model.bin 2020-06-23 23:24:57 +00:00
rust_model.ot addition of Rust model 2020-11-19 18:06:20 +01:00
tokenizer_config.json Update tokenizer_config.json 2020-06-23 23:23:42 +00:00
vocab.json Update vocab.json 2020-06-23 23:23:42 +00:00

README.md

language tags license datasets thumbnail
en
summarization
apache-2.0
cnn_dailymail
xsum
https://huggingface.co/front/thumbnails/distilbart_medium.png

Usage

This checkpoint should be loaded into BartForConditionalGeneration.from_pretrained. See the BART docs for more information.

Metrics for DistilBART models

Model Name MM Params Inference Time (MS) Speedup Rouge 2 Rouge-L
distilbart-xsum-12-1 222 90 2.54 18.31 33.37
distilbart-xsum-6-6 230 132 1.73 20.92 35.73
distilbart-xsum-12-3 255 106 2.16 21.37 36.39
distilbart-xsum-9-6 268 136 1.68 21.72 36.61
bart-large-xsum (baseline) 406 229 1 21.85 36.50
distilbart-xsum-12-6 306 137 1.68 22.12 36.99
bart-large-cnn (baseline) 406 381 1 21.06 30.63
distilbart-12-3-cnn 255 214 1.78 20.57 30.00
distilbart-12-6-cnn 306 307 1.24 21.26 30.59
distilbart-6-6-cnn 230 182 2.09 20.17 29.70