add tokenizer
This commit is contained in:
parent
d39f4f8a18
commit
d71cdce5b4
File diff suppressed because it is too large
Load Diff
|
@ -0,0 +1 @@
|
|||
{"bos_token": "<|endoftext|>", "eos_token": "<|endoftext|>", "unk_token": "<|endoftext|>", "pad_token": "<|endoftext|>"}
|
File diff suppressed because one or more lines are too long
|
@ -0,0 +1 @@
|
|||
{"unk_token": "<|endoftext|>", "bos_token": "<|endoftext|>", "eos_token": "<|endoftext|>", "add_prefix_space": false, "model_max_length": 1024, "special_tokens_map_file": null, "name_or_path": "./models/", "tokenizer_class": "GPT2Tokenizer"}
|
File diff suppressed because one or more lines are too long
Loading…
Reference in New Issue