Making the vocab_size match the tokenizer vocab_size.
This commit is contained in:
parent
10da33b5d6
commit
4be554b688
12
config.json
12
config.json
|
@ -37,7 +37,7 @@
|
|||
"LABEL_0": 0,
|
||||
"LABEL_1": 1
|
||||
},
|
||||
"layer_norm_eps": 1e-05,
|
||||
"layer_norm_eps": 1e-5,
|
||||
"length_penalty": 1.0,
|
||||
"max_length": 20,
|
||||
"max_position_embeddings": 512,
|
||||
|
@ -72,7 +72,7 @@
|
|||
"torchscript": false,
|
||||
"transformers_version": "4.11.0.dev0",
|
||||
"use_bfloat16": false,
|
||||
"vocab_size": 99
|
||||
"vocab_size": 1000
|
||||
},
|
||||
"text_config_dict": {
|
||||
"_name_or_path": "",
|
||||
|
@ -108,7 +108,7 @@
|
|||
"LABEL_0": 0,
|
||||
"LABEL_1": 1
|
||||
},
|
||||
"layer_norm_eps": 1e-05,
|
||||
"layer_norm_eps": 1e-5,
|
||||
"length_penalty": 1.0,
|
||||
"max_length": 20,
|
||||
"max_position_embeddings": 512,
|
||||
|
@ -143,7 +143,7 @@
|
|||
"torchscript": false,
|
||||
"transformers_version": "4.11.0.dev0",
|
||||
"use_bfloat16": false,
|
||||
"vocab_size": 99
|
||||
"vocab_size": 1000
|
||||
},
|
||||
"transformers_version": null,
|
||||
"vision_config": {
|
||||
|
@ -181,7 +181,7 @@
|
|||
"LABEL_0": 0,
|
||||
"LABEL_1": 1
|
||||
},
|
||||
"layer_norm_eps": 1e-05,
|
||||
"layer_norm_eps": 1e-5,
|
||||
"length_penalty": 1.0,
|
||||
"max_length": 20,
|
||||
"min_length": 0,
|
||||
|
@ -253,7 +253,7 @@
|
|||
"LABEL_0": 0,
|
||||
"LABEL_1": 1
|
||||
},
|
||||
"layer_norm_eps": 1e-05,
|
||||
"layer_norm_eps": 1e-5,
|
||||
"length_penalty": 1.0,
|
||||
"max_length": 20,
|
||||
"min_length": 0,
|
||||
|
|
Loading…
Reference in New Issue