remove xformers from requirements*.txt; check llm_model before change model in webui
This commit is contained in:
parent
0b25d7b079
commit
83e25f8011
|
|
@ -1,7 +1,6 @@
|
|||
langchain==0.0.317
|
||||
langchain-experimental==0.0.30
|
||||
fschat[model_worker]==0.2.31
|
||||
xformers==0.0.22.post4
|
||||
openai
|
||||
sentence_transformers
|
||||
transformers>=4.34
|
||||
|
|
|
|||
|
|
@ -1,7 +1,6 @@
|
|||
langchain==0.0.317
|
||||
langchain-experimental==0.0.30
|
||||
fschat[model_worker]==0.2.31
|
||||
xformers==0.0.22.post4
|
||||
openai
|
||||
sentence_transformers>=2.2.2
|
||||
transformers>=4.34
|
||||
|
|
|
|||
|
|
@ -88,10 +88,11 @@ def dialogue_page(api: ApiRequest):
|
|||
)
|
||||
|
||||
def on_llm_change():
|
||||
config = api.get_model_config(llm_model)
|
||||
if not config.get("online_api"): # 只有本地model_worker可以切换模型
|
||||
st.session_state["prev_llm_model"] = llm_model
|
||||
st.session_state["cur_llm_model"] = st.session_state.llm_model
|
||||
if llm_model:
|
||||
config = api.get_model_config(llm_model)
|
||||
if not config.get("online_api"): # 只有本地model_worker可以切换模型
|
||||
st.session_state["prev_llm_model"] = llm_model
|
||||
st.session_state["cur_llm_model"] = st.session_state.llm_model
|
||||
|
||||
def llm_model_format_func(x):
|
||||
if x in running_models:
|
||||
|
|
@ -118,9 +119,10 @@ def dialogue_page(api: ApiRequest):
|
|||
on_change=on_llm_change,
|
||||
key="llm_model",
|
||||
)
|
||||
if (st.session_state.get("prev_llm_model") != llm_model
|
||||
and not api.get_model_config(llm_model).get("online_api")
|
||||
and llm_model not in running_models):
|
||||
if (llm_model
|
||||
and st.session_state.get("prev_llm_model") != llm_model
|
||||
and not api.get_model_config(llm_model).get("online_api")
|
||||
and llm_model not in running_models):
|
||||
with st.spinner(f"正在加载模型: {llm_model},请勿进行操作或刷新页面"):
|
||||
prev_model = st.session_state.get("prev_llm_model")
|
||||
r = api.change_llm_model(prev_model, llm_model)
|
||||
|
|
|
|||
Loading…
Reference in New Issue