From 80fcb577f994bdeac5f661b48545eb66c3f1fe18 Mon Sep 17 00:00:00 2001 From: patrickvonplaten Date: Thu, 26 May 2022 15:44:17 +0000 Subject: [PATCH] Change BOS token from 0 to 2 as BOS token is equal to EOS for OPT. See: https://github.com/huggingface/transformers/issues/17431 (#1) - Change BOS token from 0 to 2 as BOS token is equal to EOS for OPT. See: https://github.com/huggingface/transformers/issues/17431 (0a3f77286e1f0367b3b698f2b43696f36ac439a0) --- config.json | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/config.json b/config.json index 806bdf7..effa165 100644 --- a/config.json +++ b/config.json @@ -5,7 +5,7 @@ "OPTForCausalLM" ], "attention_dropout": 0.0, - "bos_token_id": 0, + "bos_token_id": 2, "hidden_size": 2048, "do_layer_norm_before": true, "dropout": 0.1,