add some comment
This commit is contained in:
parent
f9f16edd13
commit
08530edb3d
|
|
@ -40,6 +40,8 @@ FSCHAT_MODEL_WORKERS = {
|
||||||
"device": LLM_DEVICE,
|
"device": LLM_DEVICE,
|
||||||
# False,'vllm',使用的推理加速框架,使用vllm如果出现HuggingFace通信问题,参见doc/FAQ
|
# False,'vllm',使用的推理加速框架,使用vllm如果出现HuggingFace通信问题,参见doc/FAQ
|
||||||
# vllm对一些模型支持还不成熟,暂时默认关闭
|
# vllm对一些模型支持还不成熟,暂时默认关闭
|
||||||
|
# fschat=0.2.33的代码有bug, 如需使用,源码修改fastchat.server.vllm_worker,
|
||||||
|
# 将103行中sampling_params = SamplingParams的参数stop=list(stop)修改为stop= [i for i in stop if i!=""]
|
||||||
"infer_turbo": False,
|
"infer_turbo": False,
|
||||||
|
|
||||||
# model_worker多卡加载需要配置的参数
|
# model_worker多卡加载需要配置的参数
|
||||||
|
|
|
||||||
|
|
@ -7,9 +7,9 @@ xformers>=0.0.22.post7
|
||||||
openai~=0.28.1
|
openai~=0.28.1
|
||||||
sentence_transformers
|
sentence_transformers
|
||||||
transformers>=4.35.2
|
transformers>=4.35.2
|
||||||
torch==2.1.0
|
torch==2.1.0 ##on win, install the cuda version manually if you want use gpu
|
||||||
torchvision
|
torchvision #on win, install the cuda version manually if you want use gpu
|
||||||
torchaudio
|
torchaudio #on win, install the cuda version manually if you want use gpu
|
||||||
fastapi>=0.104
|
fastapi>=0.104
|
||||||
nltk>=3.8.1
|
nltk>=3.8.1
|
||||||
uvicorn~=0.23.1
|
uvicorn~=0.23.1
|
||||||
|
|
@ -54,7 +54,7 @@ vllm==0.2.2; sys_platform == "linux"
|
||||||
|
|
||||||
# WebUI requirements
|
# WebUI requirements
|
||||||
|
|
||||||
streamlit~=1.28.2
|
streamlit~=1.28.2 # # on win, make sure write its path in environment variable
|
||||||
streamlit-option-menu>=0.3.6
|
streamlit-option-menu>=0.3.6
|
||||||
streamlit-antd-components>=0.2.3
|
streamlit-antd-components>=0.2.3
|
||||||
streamlit-chatbox>=1.1.11
|
streamlit-chatbox>=1.1.11
|
||||||
|
|
|
||||||
Loading…
Reference in New Issue