From 9f4567865cd834ad3710657dc8f3a473f83533db Mon Sep 17 00:00:00 2001 From: imClumsyPanda Date: Tue, 1 Aug 2023 14:12:28 +0800 Subject: [PATCH] add chatglm2-6b-32k and make m3e default embedding model --- configs/model_config.py | 8 +++++++- 1 file changed, 7 insertions(+), 1 deletion(-) diff --git a/configs/model_config.py b/configs/model_config.py index f00b646..eb4c4a9 100644 --- a/configs/model_config.py +++ b/configs/model_config.py @@ -26,7 +26,7 @@ embedding_model_dict = { } # 选用的 Embedding 名称 -EMBEDDING_MODEL = "text2vec" +EMBEDDING_MODEL = "m3e-base" # Embedding 模型运行设备 EMBEDDING_DEVICE = "cuda" if torch.cuda.is_available() else "mps" if torch.backends.mps.is_available() else "cpu" @@ -51,6 +51,12 @@ llm_model_dict = { "api_key": "EMPTY" }, + "chatglm2-6b-32k": { + "local_model_path": "THUDM/chatglm2-6b-32k", # "THUDM/chatglm2-6b-32k", + "api_base_url": "http://localhost:8888/v1", # "name"修改为fastchat服务中的"api_base_url" + "api_key": "EMPTY" + }, + "vicuna-13b-hf": { "local_model_path": "", "api_base_url": "http://localhost:8000/v1", # "name"修改为fastchat服务中的"api_base_url"