From a83e03011dac384b21a8cd203a8ef7c7db8f38ce Mon Sep 17 00:00:00 2001 From: Zhi-guo Huang Date: Sat, 25 Nov 2023 19:41:21 +0800 Subject: [PATCH] Update server_config.py.example MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit 针对fschat=0.2.33关于vllm启动的bug给出修补方案 --- configs/server_config.py.example | 2 ++ 1 file changed, 2 insertions(+) diff --git a/configs/server_config.py.example b/configs/server_config.py.example index 714a343..6f5fb0d 100644 --- a/configs/server_config.py.example +++ b/configs/server_config.py.example @@ -40,6 +40,8 @@ FSCHAT_MODEL_WORKERS = { "device": LLM_DEVICE, # False,'vllm',使用的推理加速框架,使用vllm如果出现HuggingFace通信问题,参见doc/FAQ # vllm对一些模型支持还不成熟,暂时默认关闭 + # fschat=0.2.33的代码有bug, 如需使用,源码修改fastchat.server.vllm_worker, + # 将103行中sampling_params = SamplingParams的参数stop=list(stop)修改为stop= [i for i in stop if i!=""] "infer_turbo": False, # model_worker多卡加载需要配置的参数