From 3073935e248b3172d0b7e7b6a6982d4c7a6326e1 Mon Sep 17 00:00:00 2001 From: binary-husky Date: Tue, 19 Sep 2023 16:49:33 +0800 Subject: [PATCH] =?UTF-8?q?=E4=BF=AE=E6=94=B9readme=20=E6=8E=A8=E9=80=81ve?= =?UTF-8?q?rsion=203.53?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit --- README.md | 22 ++++++++++----------- docker-compose.yml | 49 ++++++++++++++++++++++++++++++++++++++++++++++ version | 4 ++-- 3 files changed, 62 insertions(+), 13 deletions(-) diff --git a/README.md b/README.md index 04b2f63..01cb55c 100644 --- a/README.md +++ b/README.md @@ -149,25 +149,25 @@ python main.py ### 安装方法II:使用Docker +0. 部署项目的全部能力(这个是包含cuda和latex的大型镜像。如果您网速慢、硬盘小或没有显卡,则不推荐使用这个,建议使用方案1)(需要熟悉[Nvidia Docker](https://docs.nvidia.com/datacenter/cloud-native/container-toolkit/install-guide.html#installing-on-ubuntu-and-debian)运行时) + [![fullcapacity](https://github.com/binary-husky/gpt_academic/actions/workflows/build-with-all-capacity.yml/badge.svg?branch=master)](https://github.com/binary-husky/gpt_academic/actions/workflows/build-with-audio-assistant.yml) -1. 仅ChatGPT(推荐大多数人选择,等价于docker-compose方案1) +``` sh +# 修改docker-compose.yml,保留方案0并删除其他方案。修改docker-compose.yml中方案0的配置,参考其中注释即可 +docker-compose up +``` + +1. 仅ChatGPT(推荐大多数人选择) [![basic](https://github.com/binary-husky/gpt_academic/actions/workflows/build-without-local-llms.yml/badge.svg?branch=master)](https://github.com/binary-husky/gpt_academic/actions/workflows/build-without-local-llms.yml) [![basiclatex](https://github.com/binary-husky/gpt_academic/actions/workflows/build-with-latex.yml/badge.svg?branch=master)](https://github.com/binary-husky/gpt_academic/actions/workflows/build-with-latex.yml) [![basicaudio](https://github.com/binary-husky/gpt_academic/actions/workflows/build-with-audio-assistant.yml/badge.svg?branch=master)](https://github.com/binary-husky/gpt_academic/actions/workflows/build-with-audio-assistant.yml) - ``` sh -git clone --depth=1 https://github.com/binary-husky/gpt_academic.git # 下载项目 -cd gpt_academic # 进入路径 -nano config.py # 用任意文本编辑器编辑config.py, 配置 “Proxy”, “API_KEY” 以及 “WEB_PORT” (例如50923) 等 -docker build -t gpt-academic . # 安装 - -#(最后一步-Linux操作系统)用`--net=host`更方便快捷 -docker run --rm -it --net=host gpt-academic -#(最后一步-MacOS/Windows操作系统)只能用-p选项将容器上的端口(例如50923)暴露给主机上的端口 -docker run --rm -it -e WEB_PORT=50923 -p 50923:50923 gpt-academic +# 修改docker-compose.yml,保留方案1并删除其他方案。修改docker-compose.yml中方案1的配置,参考其中注释即可 +docker-compose up ``` + P.S. 如果需要依赖Latex的插件功能,请见Wiki。另外,您也可以直接使用docker-compose获取Latex功能(修改docker-compose.yml,保留方案4并删除其他方案)。 2. ChatGPT + ChatGLM2 + MOSS + LLAMA2 + 通义千问(需要熟悉[Nvidia Docker](https://docs.nvidia.com/datacenter/cloud-native/container-toolkit/install-guide.html#installing-on-ubuntu-and-debian)运行时) diff --git a/docker-compose.yml b/docker-compose.yml index 2387527..9e60c97 100644 --- a/docker-compose.yml +++ b/docker-compose.yml @@ -1,5 +1,54 @@ #【请修改完参数后,删除此行】请在以下方案中选择一种,然后删除其他的方案,最后docker-compose up运行 | Please choose from one of these options below, delete other options as well as This Line +## =================================================== +## 【方案零】 部署项目的全部能力(这个是包含cuda和latex的大型镜像。如果您网速慢、硬盘小或没有显卡,则不推荐使用这个) +## =================================================== +version: '3' +services: + gpt_academic_full_capability: + image: docker pull ghcr.io/binary-husky/gpt_academic_with_all_capacity:master + environment: + # 请查阅 `config.py`或者 github wiki 以查看所有的配置信息 + API_KEY: ' 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 ' + # USE_PROXY: ' True ' + # proxies: ' { "http": "http://localhost:10881", "https": "http://localhost:10881", } ' + LLM_MODEL: ' gpt-3.5-turbo ' + AVAIL_LLM_MODELS: ' ["gpt-3.5-turbo", "gpt-4", "qianfan", "sparkv2", "spark", "chatglm"] ' + BAIDU_CLOUD_API_KEY : ' bTUtwEAveBrQipEowUvDwYWq ' + BAIDU_CLOUD_SECRET_KEY : ' jqXtLvXiVw6UNdjliATTS61rllG8Iuni ' + XFYUN_APPID: ' 53a8d816 ' + XFYUN_API_SECRET: ' MjMxNDQ4NDE4MzM0OSNlNjQ2NTlhMTkx ' + XFYUN_API_KEY: ' 95ccdec285364869d17b33e75ee96447 ' + ENABLE_AUDIO: ' False ' + DEFAULT_WORKER_NUM: ' 20 ' + WEB_PORT: ' 12345 ' + ADD_WAIFU: ' False ' + ALIYUN_APPKEY: ' RxPlZrM88DnAFkZK ' + THEME: ' Chuanhu-Small-and-Beautiful ' + ALIYUN_ACCESSKEY: ' LTAI5t6BrFUzxRXVGUWnekh1 ' + ALIYUN_SECRET: ' eHmI20SVWIwQZxCiTD2bGQVspP9i68 ' + # LOCAL_MODEL_DEVICE: ' cuda ' + + # 加载英伟达显卡运行时 + # runtime: nvidia + # deploy: + # resources: + # reservations: + # devices: + # - driver: nvidia + # count: 1 + # capabilities: [gpu] + + # 与宿主的网络融合 + network_mode: "host" + + # 不使用代理网络拉取最新代码 + command: > + bash -c "python3 -u main.py" + + + + ## =================================================== ## 【方案一】 如果不需要运行本地模型(仅 chatgpt, azure, 星火, 千帆, claude 等在线大模型服务) ## =================================================== diff --git a/version b/version index 897b042..5562088 100644 --- a/version +++ b/version @@ -1,5 +1,5 @@ { - "version": 3.52, + "version": 3.53, "show_feature": true, - "new_feature": "提高稳定性&解决多用户冲突问题 <-> 支持插件分类和更多UI皮肤外观 <-> 支持用户使用自然语言调度各个插件(虚空终端) ! <-> 改进UI,设计新主题 <-> 支持借助GROBID实现PDF高精度翻译 <-> 接入百度千帆平台和文心一言 <-> 接入阿里通义千问、讯飞星火、上海AI-Lab书生 <-> 优化一键升级 <-> 提高arxiv翻译速度和成功率" + "new_feature": "支持动态选择不同界面主题 <-> 提高稳定性&解决多用户冲突问题 <-> 支持插件分类和更多UI皮肤外观 <-> 支持用户使用自然语言调度各个插件(虚空终端) ! <-> 改进UI,设计新主题 <-> 支持借助GROBID实现PDF高精度翻译 <-> 接入百度千帆平台和文心一言 <-> 接入阿里通义千问、讯飞星火、上海AI-Lab书生 <-> 优化一键升级 <-> 提高arxiv翻译速度和成功率" }