From 83e25f8011e3166199c5244686fa7cd8e2d52bb9 Mon Sep 17 00:00:00 2001 From: liunux4odoo Date: Fri, 20 Oct 2023 15:01:39 +0800 Subject: [PATCH] remove xformers from requirements*.txt; check llm_model before change model in webui --- requirements.txt | 1 - requirements_api.txt | 1 - webui_pages/dialogue/dialogue.py | 16 +++++++++------- 3 files changed, 9 insertions(+), 9 deletions(-) diff --git a/requirements.txt b/requirements.txt index 959003f..e1b0437 100644 --- a/requirements.txt +++ b/requirements.txt @@ -1,7 +1,6 @@ langchain==0.0.317 langchain-experimental==0.0.30 fschat[model_worker]==0.2.31 -xformers==0.0.22.post4 openai sentence_transformers transformers>=4.34 diff --git a/requirements_api.txt b/requirements_api.txt index f1d2f5f..3bcb1b5 100644 --- a/requirements_api.txt +++ b/requirements_api.txt @@ -1,7 +1,6 @@ langchain==0.0.317 langchain-experimental==0.0.30 fschat[model_worker]==0.2.31 -xformers==0.0.22.post4 openai sentence_transformers>=2.2.2 transformers>=4.34 diff --git a/webui_pages/dialogue/dialogue.py b/webui_pages/dialogue/dialogue.py index c7a029d..bdfdfa3 100644 --- a/webui_pages/dialogue/dialogue.py +++ b/webui_pages/dialogue/dialogue.py @@ -88,10 +88,11 @@ def dialogue_page(api: ApiRequest): ) def on_llm_change(): - config = api.get_model_config(llm_model) - if not config.get("online_api"): # 只有本地model_worker可以切换模型 - st.session_state["prev_llm_model"] = llm_model - st.session_state["cur_llm_model"] = st.session_state.llm_model + if llm_model: + config = api.get_model_config(llm_model) + if not config.get("online_api"): # 只有本地model_worker可以切换模型 + st.session_state["prev_llm_model"] = llm_model + st.session_state["cur_llm_model"] = st.session_state.llm_model def llm_model_format_func(x): if x in running_models: @@ -118,9 +119,10 @@ def dialogue_page(api: ApiRequest): on_change=on_llm_change, key="llm_model", ) - if (st.session_state.get("prev_llm_model") != llm_model - and not api.get_model_config(llm_model).get("online_api") - and llm_model not in running_models): + if (llm_model + and st.session_state.get("prev_llm_model") != llm_model + and not api.get_model_config(llm_model).get("online_api") + and llm_model not in running_models): with st.spinner(f"正在加载模型: {llm_model},请勿进行操作或刷新页面"): prev_model = st.session_state.get("prev_llm_model") r = api.change_llm_model(prev_model, llm_model)