依赖升级 (#2246)
大量依赖升级到最新版 测试环境:Ubuntu 22.04 LTS kernel 6.6.3 Python 3.10.12,Cuda12.3 update1
This commit is contained in:
parent
60f8e1d55e
commit
0cc1be224d
|
|
@ -15,9 +15,9 @@ EMBEDDING_DEVICE = "auto"
|
||||||
EMBEDDING_KEYWORD_FILE = "keywords.txt"
|
EMBEDDING_KEYWORD_FILE = "keywords.txt"
|
||||||
EMBEDDING_MODEL_OUTPUT_PATH = "output"
|
EMBEDDING_MODEL_OUTPUT_PATH = "output"
|
||||||
|
|
||||||
# 要运行的 LLM 名称,可以包括本地模型和在线模型。
|
# 要运行的 LLM 名称,可以包括本地模型和在线模型。第一个将作为 API 和 WEBUI 的默认模型
|
||||||
# 第一个将作为 API 和 WEBUI 的默认模型
|
# 在这里,我们使用目前主流的两个离线模型,其中,chatglm3-6b 为默认加载模型,如果你的显存不足,可使用 Qwen-1_8B-Chat, 该模型 FP16 仅需 3.8G显存。
|
||||||
LLM_MODELS = ["chatglm3-6b", "zhipu-api", "openai-api"]
|
LLM_MODELS = ["chatglm3-6b", "Qwen-1_8B-Chat", "zhipu-api", "openai-api"]
|
||||||
|
|
||||||
# AgentLM模型的名称 (可以不指定,指定之后就锁定进入Agent之后的Chain的模型,不指定就是LLM_MODELS[0])
|
# AgentLM模型的名称 (可以不指定,指定之后就锁定进入Agent之后的Chain的模型,不指定就是LLM_MODELS[0])
|
||||||
Agent_MODEL = None
|
Agent_MODEL = None
|
||||||
|
|
|
||||||
|
|
@ -93,7 +93,7 @@ FSCHAT_MODEL_WORKERS = {
|
||||||
|
|
||||||
},
|
},
|
||||||
# 可以如下示例方式更改默认配置
|
# 可以如下示例方式更改默认配置
|
||||||
# "Qwen-7B-Chat": { # 使用default中的IP和端口
|
# "Qwen-1_8B-Chat": { # 使用default中的IP和端口
|
||||||
# "device": "cpu",
|
# "device": "cpu",
|
||||||
# },
|
# },
|
||||||
"chatglm3-6b": { # 使用default中的IP和端口
|
"chatglm3-6b": { # 使用default中的IP和端口
|
||||||
|
|
|
||||||
Binary file not shown.
|
Before Width: | Height: | Size: 198 KiB |
Binary file not shown.
|
Before Width: | Height: | Size: 270 KiB |
Binary file not shown.
|
Before Width: | Height: | Size: 174 KiB |
|
|
@ -1,64 +1,69 @@
|
||||||
# API requirements
|
# API requirements
|
||||||
|
|
||||||
langchain==0.0.343
|
langchain==0.0.344
|
||||||
langchain-experimental>=0.0.42
|
langchain-experimental>=0.0.42
|
||||||
fschat[model_worker]>=0.2.33
|
pydantic==1.10.13
|
||||||
|
fschat>=0.2.33
|
||||||
xformers>=0.0.22.post7
|
xformers>=0.0.22.post7
|
||||||
openai>=1.3.5
|
openai>=1.3.6
|
||||||
sentence_transformers
|
sentence_transformers
|
||||||
transformers>=4.35.2
|
transformers>=4.35.2
|
||||||
torch==2.1.0 ##on win, install the cuda version manually if you want use gpu
|
torch==2.1.0 ##on Windows system, install the cuda version manually from https://pytorch.org/
|
||||||
torchvision #on win, install the cuda version manually if you want use gpu
|
torchvision #on Windows system, install the cuda version manually from https://pytorch.org/
|
||||||
torchaudio #on win, install the cuda version manually if you want use gpu
|
torchaudio #on Windows system, install the cuda version manually from https://pytorch.org/
|
||||||
fastapi>=0.104
|
fastapi>=0.104
|
||||||
nltk>=3.8.1
|
nltk>=3.8.1
|
||||||
uvicorn~=0.23.1
|
uvicorn>=0.24.0.post1
|
||||||
starlette~=0.27.0
|
starlette~=0.27.0
|
||||||
pydantic<2
|
|
||||||
unstructured[all-docs]==0.11.0
|
unstructured[all-docs]==0.11.0
|
||||||
python-magic-bin; sys_platform == 'win32'
|
python-magic-bin; sys_platform == 'win32'
|
||||||
SQLAlchemy==2.0.19
|
SQLAlchemy==2.0.19
|
||||||
faiss-cpu
|
faiss-cpu
|
||||||
accelerate
|
accelerate>=0.24.1
|
||||||
spacy
|
spacy>=3.7.2
|
||||||
PyMuPDF
|
PyMuPDF
|
||||||
rapidocr_onnxruntime
|
rapidocr_onnxruntime
|
||||||
|
requests>=2.31.0
|
||||||
requests
|
pathlib>=1.0.1
|
||||||
pathlib
|
pytest>=7.4.3
|
||||||
pytest
|
numexpr>=2.8.7
|
||||||
numexpr
|
strsimpy>=0.2.1
|
||||||
strsimpy
|
markdownify>=0.11.6
|
||||||
markdownify
|
tiktoken>=0.5.1
|
||||||
tiktoken
|
tqdm>=4.66.1
|
||||||
tqdm
|
|
||||||
websockets
|
websockets
|
||||||
numpy~=1.24.4
|
numpy~=1.24.4
|
||||||
pandas~=2.0.3
|
pandas~=2.0.3
|
||||||
einops
|
einops>=0.7.0
|
||||||
transformers_stream_generator==0.0.4
|
transformers_stream_generator==0.0.4
|
||||||
|
|
||||||
vllm==0.2.2; sys_platform == "linux"
|
vllm==0.2.2; sys_platform == "linux"
|
||||||
|
|
||||||
# online api libs dependencies
|
|
||||||
|
# Online api libs dependencies
|
||||||
|
|
||||||
# zhipuai>=1.0.7
|
# zhipuai>=1.0.7
|
||||||
# dashscope>=1.10.0
|
# dashscope>=1.10.0
|
||||||
# qianfan>=0.2.0
|
# qianfan>=0.2.0
|
||||||
# volcengine>=1.0.106
|
# volcengine>=1.0.106
|
||||||
|
# pymilvus>=2.3.3
|
||||||
# uncomment libs if you want to use corresponding vector store
|
|
||||||
# pymilvus==2.1.3 # requires milvus==2.1.3
|
|
||||||
# psycopg2
|
# psycopg2
|
||||||
# pgvector
|
# pgvector>=0.2.4
|
||||||
|
|
||||||
|
# Agent and Search Tools
|
||||||
|
|
||||||
|
arxiv>=2.0.0
|
||||||
|
youtube-search>=2.1.2
|
||||||
|
duckduckgo-search>=4.9.3
|
||||||
|
metaphor-python>=0.1.23
|
||||||
|
|
||||||
# WebUI requirements
|
# WebUI requirements
|
||||||
|
|
||||||
streamlit~=1.28.2 # # on win, make sure write its path in environment variable
|
streamlit>=1.29.0
|
||||||
streamlit-option-menu>=0.3.6
|
streamlit-option-menu>=0.3.6
|
||||||
streamlit-antd-components>=0.2.3
|
streamlit-antd-components>=0.2.3
|
||||||
streamlit-chatbox>=1.1.11
|
streamlit-chatbox>=1.1.11
|
||||||
streamlit-modal==0.1.0
|
streamlit-modal>=0.1.0
|
||||||
streamlit-aggrid>=0.3.4.post3
|
streamlit-aggrid>=0.3.4.post3
|
||||||
httpx[brotli,http2,socks]~=0.24.1
|
httpx[brotli,http2,socks]>=0.25.2
|
||||||
watchdog
|
watchdog>=3.0.0
|
||||||
|
|
||||||
|
|
|
||||||
|
|
@ -1,52 +1,57 @@
|
||||||
# API requirements
|
# API requirements
|
||||||
|
|
||||||
langchain==0.0.343
|
langchain==0.0.344
|
||||||
langchain-experimental>=0.0.42
|
langchain-experimental>=0.0.42
|
||||||
fschat[model_worker]>=0.2.33
|
pydantic==1.10.13
|
||||||
|
fschat>=0.2.33
|
||||||
xformers>=0.0.22.post7
|
xformers>=0.0.22.post7
|
||||||
openai>=1.3.5
|
openai>=1.3.6
|
||||||
sentence_transformers
|
sentence_transformers
|
||||||
transformers>=4.35.2
|
transformers>=4.35.2
|
||||||
torch==2.1.0
|
torch==2.1.0 ##on Windows system, install the cuda version manually from https://pytorch.org/
|
||||||
torchvision
|
torchvision #on Windows system, install the cuda version manually from https://pytorch.org/
|
||||||
torchaudio
|
torchaudio #on Windows system, install the cuda version manually from https://pytorch.org/
|
||||||
fastapi>=0.104
|
fastapi>=0.104
|
||||||
nltk>=3.8.1
|
nltk>=3.8.1
|
||||||
uvicorn~=0.23.1
|
uvicorn>=0.24.0.post1
|
||||||
starlette~=0.27.0
|
starlette~=0.27.0
|
||||||
pydantic<2
|
|
||||||
unstructured[all-docs]==0.11.0
|
unstructured[all-docs]==0.11.0
|
||||||
python-magic-bin; sys_platform == 'win32'
|
python-magic-bin; sys_platform == 'win32'
|
||||||
SQLAlchemy==2.0.19
|
SQLAlchemy==2.0.19
|
||||||
faiss-cpu
|
faiss-cpu
|
||||||
accelerate>=0.24.1
|
accelerate>=0.24.1
|
||||||
spacy
|
spacy>=3.7.2
|
||||||
PyMuPDF
|
PyMuPDF
|
||||||
rapidocr_onnxruntime
|
rapidocr_onnxruntime
|
||||||
|
requests>=2.31.0
|
||||||
requests
|
pathlib>=1.0.1
|
||||||
pathlib
|
pytest>=7.4.3
|
||||||
pytest
|
numexpr>=2.8.7
|
||||||
numexpr
|
strsimpy>=0.2.1
|
||||||
strsimpy
|
markdownify>=0.11.6
|
||||||
markdownify
|
tiktoken>=0.5.1
|
||||||
tiktoken
|
tqdm>=4.66.1
|
||||||
tqdm
|
|
||||||
websockets
|
websockets
|
||||||
numpy~=1.24.4
|
numpy~=1.24.4
|
||||||
pandas~=2.0.3
|
pandas~=2.0.3
|
||||||
einops
|
einops>=0.7.0
|
||||||
transformers_stream_generator>=0.0.4
|
transformers_stream_generator==0.0.4
|
||||||
|
vllm==0.2.2; sys_platform == "linux"
|
||||||
|
|
||||||
vllm>=0.2.0; sys_platform == "linux"
|
|
||||||
|
|
||||||
# online api libs
|
# Online api libs dependencies
|
||||||
zhipuai
|
|
||||||
dashscope>=1.10.0 # qwen
|
|
||||||
qianfan
|
|
||||||
# volcengine>=1.0.106 # fangzhou
|
|
||||||
|
|
||||||
# uncomment libs if you want to use corresponding vector store
|
# zhipuai>=1.0.7
|
||||||
# pymilvus==2.1.3 # requires milvus==2.1.3
|
# dashscope>=1.10.0
|
||||||
|
# qianfan>=0.2.0
|
||||||
|
# volcengine>=1.0.106
|
||||||
|
# pymilvus>=2.3.3
|
||||||
# psycopg2
|
# psycopg2
|
||||||
# pgvector
|
# pgvector>=0.2.4
|
||||||
|
|
||||||
|
# Agent and Search Tools
|
||||||
|
|
||||||
|
arxiv>=2.0.0
|
||||||
|
youtube-search>=2.1.2
|
||||||
|
duckduckgo-search>=4.9.3
|
||||||
|
metaphor-python>=0.1.23
|
||||||
|
|
@ -1,23 +1,21 @@
|
||||||
langchain==0.0.343
|
langchain==0.0.344
|
||||||
|
pydantic==1.10.13
|
||||||
fschat>=0.2.33
|
fschat>=0.2.33
|
||||||
openai>=1.3.5
|
openai>=1.3.6
|
||||||
# sentence_transformers
|
|
||||||
# transformers>=4.35.2
|
|
||||||
# torch>=2.0.1
|
|
||||||
# torchvision
|
|
||||||
# torchaudio
|
|
||||||
fastapi>=0.104.1
|
fastapi>=0.104.1
|
||||||
python-multipart
|
python-multipart
|
||||||
nltk~=3.8.1
|
nltk~=3.8.1
|
||||||
uvicorn~=0.23.1
|
uvicorn>=0.24.0.post1
|
||||||
starlette~=0.27.0
|
starlette~=0.27.0
|
||||||
pydantic~=1.10.11
|
|
||||||
unstructured[docx,csv]==0.11.0 # add pdf if need
|
unstructured[docx,csv]==0.11.0 # add pdf if need
|
||||||
python-magic-bin; sys_platform == 'win32'
|
python-magic-bin; sys_platform == 'win32'
|
||||||
SQLAlchemy==2.0.19
|
SQLAlchemy==2.0.19
|
||||||
|
numexpr>=2.8.7
|
||||||
|
strsimpy>=0.2.1
|
||||||
|
|
||||||
faiss-cpu
|
faiss-cpu
|
||||||
# accelerate
|
# accelerate>=0.24.1
|
||||||
# spacy
|
# spacy>=3.7.2
|
||||||
# PyMuPDF==1.22.5 # install if need pdf
|
# PyMuPDF==1.22.5 # install if need pdf
|
||||||
# rapidocr_onnxruntime>=1.3.2 # install if need pdf
|
# rapidocr_onnxruntime>=1.3.2 # install if need pdf
|
||||||
|
|
||||||
|
|
@ -26,18 +24,18 @@ pathlib
|
||||||
pytest
|
pytest
|
||||||
# scikit-learn
|
# scikit-learn
|
||||||
# numexpr
|
# numexpr
|
||||||
# vllm==0.1.7; sys_platform == "linux"
|
# vllm==0.2.2; sys_platform == "linux"
|
||||||
|
|
||||||
# online api libs
|
# online api libs
|
||||||
zhipuai
|
|
||||||
dashscope>=1.10.0 # qwen
|
zhipuai>=1.0.7 # zhipu
|
||||||
# qianfan
|
# dashscope>=1.10.0 # qwen
|
||||||
# volcengine>=1.0.106 # fangzhou
|
# volcengine>=1.0.106 # fangzhou
|
||||||
|
|
||||||
# uncomment libs if you want to use corresponding vector store
|
# uncomment libs if you want to use corresponding vector store
|
||||||
# pymilvus==2.1.3 # requires milvus==2.1.3
|
# pymilvus==2.1.3 # requires milvus==2.1.3
|
||||||
# psycopg2
|
# psycopg2
|
||||||
# pgvector
|
# pgvector>=0.2.4
|
||||||
|
|
||||||
numpy~=1.24.4
|
numpy~=1.24.4
|
||||||
pandas~=2.0.3
|
pandas~=2.0.3
|
||||||
|
|
@ -50,13 +48,15 @@ httpx~=0.24.1
|
||||||
watchdog
|
watchdog
|
||||||
tqdm
|
tqdm
|
||||||
websockets
|
websockets
|
||||||
|
einops>=0.7.0
|
||||||
|
|
||||||
# tiktoken
|
# tiktoken
|
||||||
einops
|
# scipy>=1.11.4
|
||||||
# scipy
|
|
||||||
# transformers_stream_generator==0.0.4
|
# transformers_stream_generator==0.0.4
|
||||||
|
|
||||||
# search engine libs
|
# Agent and Search Tools
|
||||||
duckduckgo-search
|
|
||||||
metaphor-python
|
arxiv>=2.0.0
|
||||||
strsimpy
|
youtube-search>=2.1.2
|
||||||
markdownify
|
duckduckgo-search>=4.9.3
|
||||||
|
metaphor-python>=0.1.23
|
||||||
|
|
@ -1,11 +1,10 @@
|
||||||
# WebUI requirements
|
# WebUI requirements
|
||||||
|
|
||||||
streamlit~=1.28.2
|
streamlit>=1.29.0
|
||||||
streamlit-option-menu>=0.3.6
|
streamlit-option-menu>=0.3.6
|
||||||
streamlit-antd-components>=0.2.3
|
streamlit-antd-components>=0.2.3
|
||||||
streamlit-chatbox>=1.1.11
|
streamlit-chatbox>=1.1.11
|
||||||
streamlit-modal==0.1.0
|
streamlit-modal>=0.1.0
|
||||||
streamlit-aggrid>=0.3.4.post3
|
streamlit-aggrid>=0.3.4.post3
|
||||||
httpx[brotli,http2,socks]~=0.24.1
|
httpx[brotli,http2,socks]>=0.25.2
|
||||||
watchdog
|
watchdog>=3.0.0
|
||||||
|
|
||||||
|
|
|
||||||
|
|
@ -170,7 +170,6 @@ class LLMKnowledgeChain(LLMChain):
|
||||||
queries = [(line.split(",")[0].strip(), line.split(",")[1].strip()) for line in lines]
|
queries = [(line.split(",")[0].strip(), line.split(",")[1].strip()) for line in lines]
|
||||||
except:
|
except:
|
||||||
queries = [(line.split(",")[0].strip(), line.split(",")[1].strip()) for line in lines]
|
queries = [(line.split(",")[0].strip(), line.split(",")[1].strip()) for line in lines]
|
||||||
print(queries)
|
|
||||||
run_manager.on_text("知识库查询询内容:\n\n" + str(queries) + " \n\n", color="blue", verbose=self.verbose)
|
run_manager.on_text("知识库查询询内容:\n\n" + str(queries) + " \n\n", color="blue", verbose=self.verbose)
|
||||||
output = self._evaluate_expression(queries)
|
output = self._evaluate_expression(queries)
|
||||||
run_manager.on_text("\nAnswer: ", verbose=self.verbose)
|
run_manager.on_text("\nAnswer: ", verbose=self.verbose)
|
||||||
|
|
|
||||||
Loading…
Reference in New Issue