适配远程LLM调用
This commit is contained in:
parent
24324563d6
commit
ab73f6ad93
|
|
@ -70,21 +70,19 @@ llm_model_dict = {
|
||||||
"provides": "LLamaLLM"
|
"provides": "LLamaLLM"
|
||||||
},
|
},
|
||||||
"fast-chat-chatglm-6b": {
|
"fast-chat-chatglm-6b": {
|
||||||
"name": "FastChatOpenAI",
|
"name": "chatglm-6b",
|
||||||
"pretrained_model_name": "FastChatOpenAI",
|
"pretrained_model_name": "FastChatOpenAI",
|
||||||
"local_model_path": None,
|
"local_model_path": None,
|
||||||
"provides": "FastChatOpenAILLM",
|
"provides": "FastChatOpenAILLM",
|
||||||
"api_base_url": "http://localhost:8000/v1",
|
"api_base_url": "http://localhost:8000/v1"
|
||||||
"model_name": "chatglm-6b"
|
|
||||||
},
|
},
|
||||||
|
|
||||||
"fast-chat-vicuna-13b-hf": {
|
"fast-chat-vicuna-13b-hf": {
|
||||||
"name": "FastChatOpenAI",
|
"name": "vicuna-13b-hf",
|
||||||
"pretrained_model_name": "vicuna-13b-hf",
|
"pretrained_model_name": "vicuna-13b-hf",
|
||||||
"local_model_path": None,
|
"local_model_path": None,
|
||||||
"provides": "FastChatOpenAILLM",
|
"provides": "FastChatOpenAILLM",
|
||||||
"api_base_url": "http://localhost:8000/v1",
|
"api_base_url": "http://localhost:8000/v1"
|
||||||
"model_name": "vicuna-13b-hf"
|
|
||||||
},
|
},
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
||||||
|
|
@ -34,7 +34,7 @@ def loaderLLM(llm_model: str = None, no_remote_model: bool = False, use_ptuning_
|
||||||
|
|
||||||
loaderCheckPoint.model_path = llm_model_info["local_model_path"]
|
loaderCheckPoint.model_path = llm_model_info["local_model_path"]
|
||||||
|
|
||||||
if 'FastChatOpenAILLM' in llm_model_info["local_model_path"]:
|
if 'FastChatOpenAILLM' in llm_model_info["provides"]:
|
||||||
loaderCheckPoint.unload_model()
|
loaderCheckPoint.unload_model()
|
||||||
else:
|
else:
|
||||||
loaderCheckPoint.reload_model()
|
loaderCheckPoint.reload_model()
|
||||||
|
|
@ -43,5 +43,5 @@ def loaderLLM(llm_model: str = None, no_remote_model: bool = False, use_ptuning_
|
||||||
modelInsLLM = provides_class(checkPoint=loaderCheckPoint)
|
modelInsLLM = provides_class(checkPoint=loaderCheckPoint)
|
||||||
if 'FastChatOpenAILLM' in llm_model_info["provides"]:
|
if 'FastChatOpenAILLM' in llm_model_info["provides"]:
|
||||||
modelInsLLM.set_api_base_url(llm_model_info['api_base_url'])
|
modelInsLLM.set_api_base_url(llm_model_info['api_base_url'])
|
||||||
modelInsLLM.call_model_name(llm_model_info['model_name'])
|
modelInsLLM.call_model_name(llm_model_info['name'])
|
||||||
return modelInsLLM
|
return modelInsLLM
|
||||||
|
|
|
||||||
Loading…
Reference in New Issue