From f27b190eeac4c2302d24068eabf5e9d6044389ae Mon Sep 17 00:00:00 2001 From: Sylvain Gugger Date: Wed, 23 Nov 2022 12:55:26 +0000 Subject: [PATCH] Add note that this is the smallest version of the model (#18) - Add note that this is the smallest version of the model (611838ef095a5bb35bf2027d05e1194b7c9d37ac) Co-authored-by: helen --- README.md | 4 ++++ 1 file changed, 4 insertions(+) diff --git a/README.md b/README.md index 65fe7b5..a16a55f 100644 --- a/README.md +++ b/README.md @@ -34,6 +34,10 @@ This way, the model learns an inner representation of the English language that useful for downstream tasks. The model is best at what it was pretrained for however, which is generating texts from a prompt. +This is the **smallest** version of GPT-2, with 124M parameters. + +**Related Models:** [GPT-Large](https://huggingface.co/gpt2-large), [GPT-Medium](https://huggingface.co/gpt2-medium) and [GPT-XL](https://huggingface.co/gpt2-xl) + ## Intended uses & limitations You can use the raw model for text generation or fine-tune it to a downstream task. See the