update chatglm_llm.py
This commit is contained in:
parent
2224decef3
commit
54c983f4bc
|
|
@ -72,14 +72,14 @@ class ChatGLM(LLM):
|
|||
response, _ = self.model.chat(
|
||||
self.tokenizer,
|
||||
prompt,
|
||||
history=self.history[-self.history_len:] if self.history_len>0 else [],
|
||||
history=self.history[-self.history_len:] if self.history_len > 0 else [],
|
||||
max_length=self.max_token,
|
||||
temperature=self.temperature,
|
||||
)
|
||||
torch_gc()
|
||||
if stop is not None:
|
||||
response = enforce_stop_tokens(response, stop)
|
||||
self.history = self.history+[[None, response]]
|
||||
self.history = self.history + [[None, response]]
|
||||
return response
|
||||
|
||||
def chat(self,
|
||||
|
|
@ -87,12 +87,12 @@ class ChatGLM(LLM):
|
|||
response, _ = self.model.chat(
|
||||
self.tokenizer,
|
||||
prompt,
|
||||
history=[],#self.history[-self.history_len:] if self.history_len>0 else
|
||||
history=self.history[-self.history_len:] if self.history_len > 0 else [],
|
||||
max_length=self.max_token,
|
||||
temperature=self.temperature,
|
||||
)
|
||||
torch_gc()
|
||||
self.history = self.history+[[None, response]]
|
||||
self.history = self.history + [[None, response]]
|
||||
return response
|
||||
|
||||
def load_model(self,
|
||||
|
|
@ -159,7 +159,8 @@ class ChatGLM(LLM):
|
|||
new_prefix_state_dict[k[len("transformer.prefix_encoder."):]] = v
|
||||
self.model.transformer.prefix_encoder.load_state_dict(new_prefix_state_dict)
|
||||
self.model.transformer.prefix_encoder.float()
|
||||
except Exception:
|
||||
except Exception as e:
|
||||
print(e)
|
||||
print("加载PrefixEncoder模型参数失败")
|
||||
|
||||
self.model = self.model.eval()
|
||||
|
|
|
|||
Loading…
Reference in New Issue