修改dockercompose,添加对阿里qwen的支持
This commit is contained in:
parent
5e48fc98ed
commit
4dc0f8e57a
@ -16,6 +16,7 @@ services:
|
|||||||
AVAIL_LLM_MODELS: ' ["gpt-3.5-turbo", "api2d-gpt-3.5-turbo", "gpt-4", "api2d-gpt-4", "newbing"] '
|
AVAIL_LLM_MODELS: ' ["gpt-3.5-turbo", "api2d-gpt-3.5-turbo", "gpt-4", "api2d-gpt-4", "newbing"] '
|
||||||
WEB_PORT: ' 22303 '
|
WEB_PORT: ' 22303 '
|
||||||
ADD_WAIFU: ' True '
|
ADD_WAIFU: ' True '
|
||||||
|
# THEME: ' Chuanhu-Small-and-Beautiful '
|
||||||
# DEFAULT_WORKER_NUM: ' 10 '
|
# DEFAULT_WORKER_NUM: ' 10 '
|
||||||
# AUTHENTICATION: ' [("username", "passwd"), ("username2", "passwd2")] '
|
# AUTHENTICATION: ' [("username", "passwd"), ("username2", "passwd2")] '
|
||||||
|
|
||||||
@ -28,7 +29,7 @@ services:
|
|||||||
|
|
||||||
|
|
||||||
### ===================================================
|
### ===================================================
|
||||||
### 【方案二】 如果需要运行ChatGLM本地模型
|
### 【方案二】 如果需要运行ChatGLM + Qwen + MOSS等本地模型
|
||||||
### ===================================================
|
### ===================================================
|
||||||
version: '3'
|
version: '3'
|
||||||
services:
|
services:
|
||||||
@ -36,11 +37,11 @@ services:
|
|||||||
image: ghcr.io/binary-husky/gpt_academic_chatglm_moss:master # (Auto Built by Dockerfile: docs/Dockerfile+ChatGLM)
|
image: ghcr.io/binary-husky/gpt_academic_chatglm_moss:master # (Auto Built by Dockerfile: docs/Dockerfile+ChatGLM)
|
||||||
environment:
|
environment:
|
||||||
# 请查阅 `config.py` 以查看所有的配置信息
|
# 请查阅 `config.py` 以查看所有的配置信息
|
||||||
API_KEY: ' sk-xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx,fkxxxxxx-xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx '
|
API_KEY: ' sk-xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx '
|
||||||
USE_PROXY: ' True '
|
USE_PROXY: ' True '
|
||||||
proxies: ' { "http": "socks5h://localhost:10880", "https": "socks5h://localhost:10880", } '
|
proxies: ' { "http": "socks5h://localhost:10880", "https": "socks5h://localhost:10880", } '
|
||||||
LLM_MODEL: ' gpt-3.5-turbo '
|
LLM_MODEL: ' gpt-3.5-turbo '
|
||||||
AVAIL_LLM_MODELS: ' ["chatglm", "moss", "gpt-3.5-turbo", "gpt-4", "newbing"] '
|
AVAIL_LLM_MODELS: ' ["chatglm", "qwen", "moss", "gpt-3.5-turbo", "gpt-4", "newbing"] '
|
||||||
LOCAL_MODEL_DEVICE: ' cuda '
|
LOCAL_MODEL_DEVICE: ' cuda '
|
||||||
DEFAULT_WORKER_NUM: ' 10 '
|
DEFAULT_WORKER_NUM: ' 10 '
|
||||||
WEB_PORT: ' 12303 '
|
WEB_PORT: ' 12303 '
|
||||||
@ -57,6 +58,10 @@ services:
|
|||||||
command: >
|
command: >
|
||||||
bash -c "python3 -u main.py"
|
bash -c "python3 -u main.py"
|
||||||
|
|
||||||
|
# P.S. 通过对 command 进行微调,可以便捷地安装额外的依赖
|
||||||
|
# command: >
|
||||||
|
# bash -c "pip install -r request_llm/requirements_qwen.txt && python3 -u main.py"
|
||||||
|
|
||||||
### ===================================================
|
### ===================================================
|
||||||
### 【方案三】 如果需要运行ChatGPT + LLAMA + 盘古 + RWKV本地模型
|
### 【方案三】 如果需要运行ChatGPT + LLAMA + 盘古 + RWKV本地模型
|
||||||
### ===================================================
|
### ===================================================
|
||||||
|
@ -18,6 +18,7 @@ WORKDIR /gpt/gpt_academic
|
|||||||
RUN git clone https://github.com/OpenLMLab/MOSS.git request_llm/moss
|
RUN git clone https://github.com/OpenLMLab/MOSS.git request_llm/moss
|
||||||
RUN python3 -m pip install -r requirements.txt
|
RUN python3 -m pip install -r requirements.txt
|
||||||
RUN python3 -m pip install -r request_llm/requirements_moss.txt
|
RUN python3 -m pip install -r request_llm/requirements_moss.txt
|
||||||
|
RUN python3 -m pip install -r request_llm/requirements_qwen.txt
|
||||||
RUN python3 -m pip install -r request_llm/requirements_chatglm.txt
|
RUN python3 -m pip install -r request_llm/requirements_chatglm.txt
|
||||||
RUN python3 -m pip install -r request_llm/requirements_newbing.txt
|
RUN python3 -m pip install -r request_llm/requirements_newbing.txt
|
||||||
|
|
||||||
|
Loading…
x
Reference in New Issue
Block a user