diff --git a/configs/server_config.py.example b/configs/server_config.py.example index 8de6b07..2e27c40 100644 --- a/configs/server_config.py.example +++ b/configs/server_config.py.example @@ -40,6 +40,8 @@ FSCHAT_MODEL_WORKERS = { "device": LLM_DEVICE, # False,'vllm',使用的推理加速框架,使用vllm如果出现HuggingFace通信问题,参见doc/FAQ # vllm对一些模型支持还不成熟,暂时默认关闭 + # fschat=0.2.33的代码有bug, 如需使用,源码修改fastchat.server.vllm_worker, + # 将103行中sampling_params = SamplingParams的参数stop=list(stop)修改为stop= [i for i in stop if i!=""] "infer_turbo": False, # model_worker多卡加载需要配置的参数 diff --git a/requirements.txt b/requirements.txt index ac18c33..036489e 100644 --- a/requirements.txt +++ b/requirements.txt @@ -7,9 +7,9 @@ xformers>=0.0.22.post7 openai~=0.28.1 sentence_transformers transformers>=4.35.2 -torch==2.1.0 -torchvision -torchaudio +torch==2.1.0 ##on win, install the cuda version manually if you want use gpu +torchvision #on win, install the cuda version manually if you want use gpu +torchaudio #on win, install the cuda version manually if you want use gpu fastapi>=0.104 nltk>=3.8.1 uvicorn~=0.23.1 @@ -54,7 +54,7 @@ vllm==0.2.2; sys_platform == "linux" # WebUI requirements -streamlit~=1.28.2 +streamlit~=1.28.2 # # on win, make sure write its path in environment variable streamlit-option-menu>=0.3.6 streamlit-antd-components>=0.2.3 streamlit-chatbox>=1.1.11