diff --git a/chatglm_llm.py b/chatglm_llm.py index 3833e4f..776a8e1 100644 --- a/chatglm_llm.py +++ b/chatglm_llm.py @@ -52,6 +52,7 @@ class ChatGLM(LLM): max_length=self.max_token, temperature=self.temperature, ) + torch_gc() print("history: ", self.history) if stop is not None: response = enforce_stop_tokens(response, stop)