From 6b734ced568cb37a2ec83f6a1419d63995737b70 Mon Sep 17 00:00:00 2001 From: hzg0601 Date: Thu, 14 Sep 2023 17:00:20 +0800 Subject: [PATCH] =?UTF-8?q?update=20readme,readme=5Fen:=E6=9B=B4=E6=96=B0f?= =?UTF-8?q?astchat=E4=BF=A1=E6=81=AF;=E5=88=A0=E9=99=A4=E8=AE=A2=E9=98=85?= =?UTF-8?q?=E5=8F=B7=E4=BA=8C=E7=BB=B4=E7=A0=81?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit --- README.md | 9 +++------ README_en.md | 6 ++---- 2 files changed, 5 insertions(+), 10 deletions(-) diff --git a/README.md b/README.md index 75c0536..50b16c0 100644 --- a/README.md +++ b/README.md @@ -112,6 +112,8 @@ docker run -d --gpus all -p 80:8501 registry.cn-beijing.aliyuncs.com/chatchat/ch - [FlagAlpha/Llama2-Chinese-13b-Chat](https://huggingface.co/FlagAlpha/Llama2-Chinese-13b-Chat) and others - [BAAI/AquilaChat-7B](https://huggingface.co/BAAI/AquilaChat-7B) - [all models of OpenOrca](https://huggingface.co/Open-Orca) +- [Spicyboros](https://huggingface.co/jondurbin/spicyboros-7b-2.2?not-for-all-audiences=true) + [airoboros 2.2](https://huggingface.co/jondurbin/airoboros-l2-13b-2.2) +- [VMware's OpenLLaMa OpenInstruct](https://huggingface.co/VMware/open-llama-7b-open-instruct) - 任何 [EleutherAI](https://huggingface.co/EleutherAI) 的 pythia 模型,如 [pythia-6.9b](https://huggingface.co/EleutherAI/pythia-6.9b) - 在以上模型基础上训练的任何 [Peft](https://github.com/huggingface/peft) 适配器。为了激活,模型路径中必须有 `peft` 。注意:如果加载多个peft模型,你可以通过在任何模型工作器中设置环境变量 `PEFT_SHARE_BASE_WEIGHTS=true` 来使它们共享基础模型的权重。 @@ -246,8 +248,7 @@ text_splitter_dict = { "tokenizer_name_or_path": "", ## 空格不填则默认使用大模型的分词器。 } } - ``` - +``` 如果你选择使用OpenAI的Embedding模型,请将模型的 ``key``写入 `embedding_model_dict`中。使用该模型,你需要能够访问OpenAI官的API,或设置代理。 @@ -525,7 +526,3 @@ $ streamlit run webui.py --server.port 666 二维码 🎉 langchain-ChatGLM 项目微信交流群,如果你也对本项目感兴趣,欢迎加入群聊参与讨论交流。 - -![1694251762513](image/README/1694251762513.png) - -🎉 langchain-ChatGLM 项目新开通了订阅号,欢迎订阅。 diff --git a/README_en.md b/README_en.md index 4f61aac..9e2c312 100644 --- a/README_en.md +++ b/README_en.md @@ -108,6 +108,8 @@ The project use [FastChat](https://github.com/lm-sys/FastChat) to provide the AP - [FlagAlpha/Llama2-Chinese-13b-Chat](https://huggingface.co/FlagAlpha/Llama2-Chinese-13b-Chat) and other models of FlagAlpha - [BAAI/AquilaChat-7B](https://huggingface.co/BAAI/AquilaChat-7B) - [all models of OpenOrca](https://huggingface.co/Open-Orca) +- [Spicyboros](https://huggingface.co/jondurbin/spicyboros-7b-2.2?not-for-all-audiences=true) + [airoboros 2.2](https://huggingface.co/jondurbin/airoboros-l2-13b-2.2) +- [VMware's OpenLLaMa OpenInstruct](https://huggingface.co/VMware/open-llama-7b-open-instruct) * Any [EleutherAI](https://huggingface.co/EleutherAI) pythia model such as [pythia-6.9b](https://huggingface.co/EleutherAI/pythia-6.9b)(任何 [EleutherAI](https://huggingface.co/EleutherAI) 的 pythia 模型,如 [pythia-6.9b](https://huggingface.co/EleutherAI/pythia-6.9b)) * Any [Peft](https://github.com/huggingface/peft) adapter trained on top of a model above. To activate, must have `peft` in the model path. Note: If loading multiple peft models, you can have them share the base model weights by setting the environment variable `PEFT_SHARE_BASE_WEIGHTS=true` in any model worker. @@ -356,7 +358,3 @@ Please refer to [FAQ](docs/FAQ.md) 二维码 **WeChat Group** - -![1694252029167](image/README_en/1694252029167.png) - -**WeChat Subscription**