No padding for chat function
This commit is contained in:
parent
373fd6b9d4
commit
4b7ffbf070
|
@ -1243,7 +1243,7 @@ class ChatGLMForConditionalGeneration(ChatGLMPreTrainedModel):
|
||||||
for i, (old_query, response) in enumerate(history):
|
for i, (old_query, response) in enumerate(history):
|
||||||
prompt += "[Round {}]\n问:{}\n答:{}\n".format(i, old_query, response)
|
prompt += "[Round {}]\n问:{}\n答:{}\n".format(i, old_query, response)
|
||||||
prompt += "[Round {}]\n问:{}\n答:".format(len(history), query)
|
prompt += "[Round {}]\n问:{}\n答:".format(len(history), query)
|
||||||
inputs = tokenizer([prompt], return_tensors="pt", padding=True)
|
inputs = tokenizer([prompt], return_tensors="pt")
|
||||||
inputs = inputs.to(self.device)
|
inputs = inputs.to(self.device)
|
||||||
outputs = self.generate(**inputs, **gen_kwargs)
|
outputs = self.generate(**inputs, **gen_kwargs)
|
||||||
outputs = outputs.tolist()[0][len(inputs["input_ids"][0]):]
|
outputs = outputs.tolist()[0][len(inputs["input_ids"][0]):]
|
||||||
|
@ -1269,7 +1269,7 @@ class ChatGLMForConditionalGeneration(ChatGLMPreTrainedModel):
|
||||||
for i, (old_query, response) in enumerate(history):
|
for i, (old_query, response) in enumerate(history):
|
||||||
prompt += "[Round {}]\n问:{}\n答:{}\n".format(i, old_query, response)
|
prompt += "[Round {}]\n问:{}\n答:{}\n".format(i, old_query, response)
|
||||||
prompt += "[Round {}]\n问:{}\n答:".format(len(history), query)
|
prompt += "[Round {}]\n问:{}\n答:".format(len(history), query)
|
||||||
inputs = tokenizer([prompt], return_tensors="pt", padding=True)
|
inputs = tokenizer([prompt], return_tensors="pt")
|
||||||
inputs = inputs.to(self.device)
|
inputs = inputs.to(self.device)
|
||||||
for outputs in self.stream_generate(**inputs, **gen_kwargs):
|
for outputs in self.stream_generate(**inputs, **gen_kwargs):
|
||||||
outputs = outputs.tolist()[0][len(inputs["input_ids"][0]):]
|
outputs = outputs.tolist()[0][len(inputs["input_ids"][0]):]
|
||||||
|
|
Loading…
Reference in New Issue