add some comment
This commit is contained in:
parent
f9f16edd13
commit
08530edb3d
|
|
@ -40,6 +40,8 @@ FSCHAT_MODEL_WORKERS = {
|
|||
"device": LLM_DEVICE,
|
||||
# False,'vllm',使用的推理加速框架,使用vllm如果出现HuggingFace通信问题,参见doc/FAQ
|
||||
# vllm对一些模型支持还不成熟,暂时默认关闭
|
||||
# fschat=0.2.33的代码有bug, 如需使用,源码修改fastchat.server.vllm_worker,
|
||||
# 将103行中sampling_params = SamplingParams的参数stop=list(stop)修改为stop= [i for i in stop if i!=""]
|
||||
"infer_turbo": False,
|
||||
|
||||
# model_worker多卡加载需要配置的参数
|
||||
|
|
|
|||
|
|
@ -7,9 +7,9 @@ xformers>=0.0.22.post7
|
|||
openai~=0.28.1
|
||||
sentence_transformers
|
||||
transformers>=4.35.2
|
||||
torch==2.1.0
|
||||
torchvision
|
||||
torchaudio
|
||||
torch==2.1.0 ##on win, install the cuda version manually if you want use gpu
|
||||
torchvision #on win, install the cuda version manually if you want use gpu
|
||||
torchaudio #on win, install the cuda version manually if you want use gpu
|
||||
fastapi>=0.104
|
||||
nltk>=3.8.1
|
||||
uvicorn~=0.23.1
|
||||
|
|
@ -54,7 +54,7 @@ vllm==0.2.2; sys_platform == "linux"
|
|||
|
||||
# WebUI requirements
|
||||
|
||||
streamlit~=1.28.2
|
||||
streamlit~=1.28.2 # # on win, make sure write its path in environment variable
|
||||
streamlit-option-menu>=0.3.6
|
||||
streamlit-antd-components>=0.2.3
|
||||
streamlit-chatbox>=1.1.11
|
||||
|
|
|
|||
Loading…
Reference in New Issue