diff --git a/requirements.txt b/requirements.txt index 036489e..27e68e5 100644 --- a/requirements.txt +++ b/requirements.txt @@ -1,10 +1,10 @@ # API requirements -langchain>=0.0.334 +langchain>=0.0.342 langchain-experimental>=0.0.42 fschat[model_worker]>=0.2.33 xformers>=0.0.22.post7 -openai~=0.28.1 +openai>=1.3.5 sentence_transformers transformers>=4.35.2 torch==2.1.0 ##on win, install the cuda version manually if you want use gpu @@ -60,4 +60,4 @@ streamlit-antd-components>=0.2.3 streamlit-chatbox>=1.1.11 streamlit-aggrid>=0.3.4.post3 httpx[brotli,http2,socks]~=0.24.1 -watchdog +watchdog \ No newline at end of file diff --git a/requirements_api.txt b/requirements_api.txt index 85b2307..56059a2 100644 --- a/requirements_api.txt +++ b/requirements_api.txt @@ -1,10 +1,10 @@ # API requirements -langchain>=0.0.334 +langchain>=0.0.342 langchain-experimental>=0.0.42 fschat[model_worker]>=0.2.33 xformers>=0.0.22.post7 -openai~=0.28.1 +openai>=1.3.5 sentence_transformers transformers>=4.35.2 torch==2.1.0 @@ -49,4 +49,4 @@ qianfan # uncomment libs if you want to use corresponding vector store # pymilvus==2.1.3 # requires milvus==2.1.3 # psycopg2 -# pgvector +# pgvector \ No newline at end of file diff --git a/requirements_lite.txt b/requirements_lite.txt index 1cfcb81..16f4a51 100644 --- a/requirements_lite.txt +++ b/requirements_lite.txt @@ -1,8 +1,8 @@ -langchain>=0.0.334 -fschat>=0.2.32 -openai +langchain>=0.0.342 +fschat>=0.2.33 +openai>=1.3.5 # sentence_transformers -# transformers>=4.35.0 +# transformers>=4.35.2 # torch>=2.0.1 # torchvision # torchaudio @@ -59,4 +59,4 @@ einops duckduckgo-search metaphor-python strsimpy -markdownify +markdownify \ No newline at end of file diff --git a/server/chat/completion.py b/server/chat/completion.py index 31ae3ff..6e5827d 100644 --- a/server/chat/completion.py +++ b/server/chat/completion.py @@ -6,7 +6,8 @@ from langchain.chains import LLMChain from langchain.callbacks import AsyncIteratorCallbackHandler from typing import AsyncIterable, Optional import asyncio -from langchain.prompts.chat import PromptTemplate +from langchain.prompts import PromptTemplate + from server.utils import get_prompt_template