Update config.json
This commit is contained in:
parent
a2ab452d41
commit
3727878143
|
@ -59,6 +59,7 @@
|
|||
"pad_token_id": 1,
|
||||
"prefix": null,
|
||||
"problem_type": null,
|
||||
"projection_dim" : 768,
|
||||
"pruned_heads": {},
|
||||
"remove_invalid_values": false,
|
||||
"repetition_penalty": 1.0,
|
||||
|
@ -83,7 +84,8 @@
|
|||
"hidden_size": 768,
|
||||
"intermediate_size": 3072,
|
||||
"num_attention_heads": 12,
|
||||
"num_hidden_layers": 12
|
||||
"num_hidden_layers": 12,
|
||||
"projection_dim": 768,
|
||||
},
|
||||
"torch_dtype": "float32",
|
||||
"transformers_version": null,
|
||||
|
@ -141,6 +143,7 @@
|
|||
"patch_size": 14,
|
||||
"prefix": null,
|
||||
"problem_type": null,
|
||||
"projection_dim" : 768,
|
||||
"pruned_heads": {},
|
||||
"remove_invalid_values": false,
|
||||
"repetition_penalty": 1.0,
|
||||
|
@ -166,6 +169,7 @@
|
|||
"intermediate_size": 4096,
|
||||
"num_attention_heads": 16,
|
||||
"num_hidden_layers": 24,
|
||||
"patch_size": 14
|
||||
"patch_size": 14,
|
||||
"projection_dim": 768,
|
||||
}
|
||||
}
|
||||
|
|
Loading…
Reference in New Issue