优化天气查询时的健壮性

This commit is contained in:
weiweiw 2025-03-19 08:18:06 +08:00
parent bad876fc98
commit 28c87ac7a7
5 changed files with 9 additions and 9 deletions

View File

@ -13,6 +13,6 @@ print(f"cuDNN 版本: {cudnn_version}")
# 检查是否可以访问 CUDA
if torch.cuda.is_available():
print("CUDA is available. GPU name:", torch.cuda.get_device_name(0))
print("pip install sentence-transformers -i https://pypi.mirrors.ustc.edu.cn/simpleCUDA is available. GPU name:", torch.cuda.get_device_name(0))
else:
print("CUDA is not available. Please check your installation.")

View File

@ -104,8 +104,8 @@ def tavily_search(text, config, top_k):
raw_results = tavily_tool.run(text)
search_results = [{k: v for k, v in item.items() if k != 'url'} for item in raw_results]
print("=== 完整搜索返回值 ===")
print(search_results)
# print("=== 完整搜索返回值 ===")
# print(search_results)
return search_results
SEARCH_ENGINES = {
@ -158,7 +158,7 @@ def search_engine(query: str, top_k:int=0, engine_name: str="", config: dict={})
)
docs = [x for x in search_result2docs(results, engine_name) if x.page_content and x.page_content.strip()]
print(f"docs: {docs}")
print(f"len(docs): {len(docs)}")
return {"docs": docs, "search_engine": engine_name}
@ -167,7 +167,7 @@ def search_internet(query: str = Field(description="query for Internet search"))
"""用这个工具实现获取世界、历史、实时新闻、或除电力系统之外的信息查询"""
try:
print(f"search_internet: query: {query}")
return BaseToolOutput(search_engine(query=query), format=format_context)
return BaseToolOutput(data= search_engine(query=query), format=format_context)
except Exception as e:
logger.error(f"未知错误: {str(e)}")
return BaseToolOutput(f"搜索过程中发生未知错误,{str(e)}", format=format_context)

View File

@ -176,7 +176,7 @@ def format_context(self: BaseToolOutput) -> str:
doc = DocumentWithVSId.parse_obj(doc)
source_documents.append(doc.page_content)
print(f"format_context: doc.page_content: {doc.page_content}")
# print(f"format_context: doc.page_content: {doc.page_content}")
if len(source_documents) == 0:
context = "没有找到相关文档,请更换关键词重试"
else:

View File

@ -65,7 +65,7 @@ async def chat_completions(
# import rich
# rich.print(body)
# 当调用本接口且 body 中没有传入 "max_tokens" 参数时, 默认使用配置中定义的值
logger.info(f"body.model_config:{body.model_config},body.tools: {body.tools},body.messages:{body.messages}")
# logger.info(f"body.model_config:{body.model_config},body.tools: {body.tools},body.messages:{body.messages}")
if body.max_tokens in [None, 0]:
body.max_tokens = Settings.model_settings.MAX_TOKENS

View File

@ -517,7 +517,7 @@ class ToolSettings(BaseFileSettings):
},
"top_k": 5,
"verbose": "Origin",
"conclude_prompt": "<指令>这是搜索到的互联网信息,请你根据这些信息进行提取并有调理,简洁的回答问题。如果无法从中得到答案,请说 “无法搜索到能回答问题的内容”。 "
"conclude_prompt": "<指令>这是搜索到的互联网信息,请你根据这些信息进行提取并有调理,简洁的回答问题,不得包含有重复的词汇或句子。如果无法从中得到答案,请说 “无法搜索到能回答问题的内容”。 "
"</指令>\n<已知信息>{{ context }}</已知信息>\n"
"<问题>\n"
"{{ question }}\n"
@ -657,7 +657,7 @@ class PromptSettings(BaseFileSettings):
rag: dict = {
"default": (
"【指令】根据已知信息,简洁和专业的来回答问题"
"【指令】根据已知信息,简洁和专业的来回答问题,不得包含有重复的词汇或句子"
"如果无法从中得到答案,请说 “根据已知信息无法回答该问题”,不允许在答案中添加编造成分,答案请使用中文。\n\n"
"【已知信息】{{context}}\n\n"
"【问题】{{question}}\n"