Update README.md
This commit is contained in:
parent
44d3fdefa7
commit
e50d65b5a2
|
@ -25,13 +25,11 @@ This repository provides a 1.3B-parameter Japanese GPT model. The model was trai
|
||||||
|
|
||||||
# How to use the model
|
# How to use the model
|
||||||
|
|
||||||
*NOTE:* Use `T5Tokenizer` to initiate the tokenizer.
|
|
||||||
|
|
||||||
~~~~
|
~~~~
|
||||||
import torch
|
import torch
|
||||||
from transformers import T5Tokenizer, AutoModelForCausalLM
|
from transformers import AutoTokenizer, AutoModelForCausalLM
|
||||||
|
|
||||||
tokenizer = T5Tokenizer.from_pretrained("rinna/japanese-gpt-1b")
|
tokenizer = AutoTokenizer.from_pretrained("rinna/japanese-gpt-1b", use_fast=False)
|
||||||
model = AutoModelForCausalLM.from_pretrained("rinna/japanese-gpt-1b")
|
model = AutoModelForCausalLM.from_pretrained("rinna/japanese-gpt-1b")
|
||||||
|
|
||||||
if torch.cuda.is_available():
|
if torch.cuda.is_available():
|
||||||
|
@ -51,7 +49,7 @@ with torch.no_grad():
|
||||||
pad_token_id=tokenizer.pad_token_id,
|
pad_token_id=tokenizer.pad_token_id,
|
||||||
bos_token_id=tokenizer.bos_token_id,
|
bos_token_id=tokenizer.bos_token_id,
|
||||||
eos_token_id=tokenizer.eos_token_id,
|
eos_token_id=tokenizer.eos_token_id,
|
||||||
bad_word_ids=[[tokenizer.unk_token_id]]
|
bad_words_ids=[[tokenizer.unk_token_id]]
|
||||||
)
|
)
|
||||||
|
|
||||||
output = tokenizer.decode(output_ids.tolist()[0])
|
output = tokenizer.decode(output_ids.tolist()[0])
|
||||||
|
|
Loading…
Reference in New Issue