liunux4odoo
|
9e2b411b01
|
cuda error with multiprocessing, change model_worker to main process
|
2023-07-31 11:18:57 +08:00 |
hzg0601
|
47dfb6cd8b
|
udpate llm_api
|
2023-07-31 11:00:33 +08:00 |
liunux4odoo
|
946f10e1f2
|
split api_start to create_app & run_api. user can run api with uvicorn in console: uvicorn server.api:app --port 7861
|
2023-07-31 10:05:19 +08:00 |
liunux4odoo
|
463659f0ba
|
llm_api升级到fschat==0.2.20,支持百川模型
|
2023-07-30 23:16:47 +08:00 |
liunux4odoo
|
51ea717606
|
Merge branch 'dev_fastchat_me' into dev_fastchat
|
2023-07-30 09:01:31 +08:00 |
liunux4odoo
|
179c2a9a92
|
修改server.chat.openai_chat中的参数定义,使其与openai中/v1/chat/completions接口的入参保持一致,并按照model_config提供默认值。
openai_chat中的接口还要修改:openai根据参数stream有不同的返回值,本接口要与其对应。
|
2023-07-30 08:56:49 +08:00 |
imClumsyPanda
|
05ccc0346e
|
update test code
|
2023-07-30 00:48:07 +08:00 |
imClumsyPanda
|
41444fd4b5
|
update requirements.txt and llm_api.py
|
2023-07-30 00:24:34 +08:00 |
liunux4odoo
|
1a7271e966
|
fix: model_worker need global variable: args
|
2023-07-29 23:22:25 +08:00 |
liunux4odoo
|
c880412300
|
use multiprocessing to run fastchat server
|
2023-07-29 23:01:24 +08:00 |
liunux4odoo
|
829ced398b
|
make api.py import model_config coreectlly
|
2023-07-29 23:00:50 +08:00 |
hzg0601
|
97ee4686a1
|
更新会议记录
|
2023-07-28 16:16:59 +08:00 |
hzg0601
|
154cad1b45
|
会议记录
|
2023-07-28 16:12:57 +08:00 |
imClumsyPanda
|
dcf49a59ef
|
v0.2.0 first commit
|
2023-07-27 23:22:07 +08:00 |