From 491174095a386167329dfa6ca6b1e218cedb41b4 Mon Sep 17 00:00:00 2001 From: binary-husky Date: Sat, 7 Oct 2023 11:59:06 +0800 Subject: [PATCH 01/13] =?UTF-8?q?=E6=9B=B4=E6=96=B0docker-compose=E8=AF=B4?= =?UTF-8?q?=E6=98=8E?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit --- docker-compose.yml | 34 +++++++++++++++++++++++++++++++--- 1 file changed, 31 insertions(+), 3 deletions(-) diff --git a/docker-compose.yml b/docker-compose.yml index bfe0c28..2bf9dfe 100644 --- a/docker-compose.yml +++ b/docker-compose.yml @@ -1,4 +1,28 @@ -#【请修改完参数后,删除此行】请在以下方案中选择一种,然后删除其他的方案,最后docker-compose up运行 | Please choose from one of these options below, delete other options as well as This Line +## =================================================== +# docker-compose.yml +## =================================================== +# 1. 请在以下方案中选择任意一种,然后删除其他的方案 +# 2. 修改你选择的方案中的environment环境变量,详情请见github wiki或者config.py +# 3. 选择一种暴露服务端口的方法,并对相应的配置做出修改: + # 【方法1: 适用于Linux,很方便,可惜windows不支持】与宿主的网络融合为一体,这个是默认配置 + # network_mode: "host" + # 【方法2: 适用于所有系统包括Windows和MacOS】端口映射,把容器的端口映射到宿主的端口(注意您需要先删除network_mode: "host",再追加以下内容) + # ports: + # - "12345:12345" # 注意!12345必须与WEB_PORT环境变量相互对应 +# 4. 最后`docker-compose up`运行 +# 5. 如果希望使用显卡,请关注 LOCAL_MODEL_DEVICE 和 英伟达显卡运行时 选项 +## =================================================== +# 1. Please choose one of the following options and delete the others. +# 2. Modify the environment variables in the selected option, see GitHub wiki or config.py for more details. +# 3. Choose a method to expose the server port and make the corresponding configuration changes: + # [Method 1: Suitable for Linux, convenient, but not supported for Windows] Fusion with the host network, this is the default configuration + # network_mode: "host" + # [Method 2: Suitable for all systems including Windows and MacOS] Port mapping, mapping the container port to the host port (note that you need to delete network_mode: "host" first, and then add the following content) + # ports: + # - "12345: 12345" # Note! 12345 must correspond to the WEB_PORT environment variable. +# 4. Finally, run `docker-compose up`. +# 5. If you want to use a graphics card, pay attention to the LOCAL_MODEL_DEVICE and Nvidia GPU runtime options. +## =================================================== ## =================================================== ## 【方案零】 部署项目的全部能力(这个是包含cuda和latex的大型镜像。如果您网速慢、硬盘小或没有显卡,则不推荐使用这个) @@ -39,10 +63,14 @@ services: # count: 1 # capabilities: [gpu] - # 与宿主的网络融合 + # 【WEB_PORT暴露方法1: 适用于Linux】与宿主的网络融合 network_mode: "host" - # 不使用代理网络拉取最新代码 + # 【WEB_PORT暴露方法2: 适用于所有系统】端口映射 + # ports: + # - "12345:12345" # 12345必须与WEB_PORT相互对应 + + # 启动容器后,运行main.py主程序 command: > bash -c "python3 -u main.py" From 1ef67303691862a7002d5ca2b33a55ccc3df52f1 Mon Sep 17 00:00:00 2001 From: binary-husky <96192199+binary-husky@users.noreply.github.com> Date: Sun, 8 Oct 2023 23:14:07 +0800 Subject: [PATCH 02/13] Update README.md --- README.md | 6 ++++-- 1 file changed, 4 insertions(+), 2 deletions(-) diff --git a/README.md b/README.md index 45048f6..61a1eb0 100644 --- a/README.md +++ b/README.md @@ -1,6 +1,6 @@ > **Note** > -> 2023.7.8: Gradio, Pydantic依赖调整,已修改 `requirements.txt`。请及时**更新代码**,安装依赖时,请严格选择`requirements.txt`中**指定的版本** +> 2023.10.8: Gradio, Pydantic依赖调整,已修改 `requirements.txt`。请及时**更新代码**,安装依赖时,请严格选择`requirements.txt`中**指定的版本** > > `pip install -r requirements.txt` @@ -310,6 +310,8 @@ Tip:不指定文件直接点击 `载入对话历史存档` 可以查看历史h ### II:版本: - version 3.60(todo): 优化虚空终端,引入code interpreter和更多插件 +- version 3.55: 重构前端界面,引入悬浮窗口与菜单栏 +- version 3.54: 新增动态代码解释器(Code Interpreter)(待完善) - version 3.53: 支持动态选择不同界面主题,提高稳定性&解决多用户冲突问题 - version 3.50: 使用自然语言调用本项目的所有函数插件(虚空终端),支持插件分类,改进UI,设计新主题 - version 3.49: 支持百度千帆平台和文心一言 @@ -331,7 +333,7 @@ Tip:不指定文件直接点击 `载入对话历史存档` 可以查看历史h - version 2.0: 引入模块化函数插件 - version 1.0: 基础功能 -gpt_academic开发者QQ群-2:610599535 +GPT Academic开发者QQ群:`610599535` - 已知问题 - 某些浏览器翻译插件干扰此软件前端的运行 From a5491b9199c834ce78eafdf677ec729cddf3e614 Mon Sep 17 00:00:00 2001 From: Skyzayre <120616113+Skyzayre@users.noreply.github.com> Date: Wed, 11 Oct 2023 00:26:16 +0800 Subject: [PATCH 03/13] Update Dockerfile MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit gradio已经更新到3.32.6,但是Dockerfile中仍然是3.32.2 --- Dockerfile | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/Dockerfile b/Dockerfile index ac47b8e..7371e45 100644 --- a/Dockerfile +++ b/Dockerfile @@ -17,7 +17,7 @@ WORKDIR /gpt # 安装大部分依赖,利用Docker缓存加速以后的构建 COPY requirements.txt ./ -COPY ./docs/gradio-3.32.2-py3-none-any.whl ./docs/gradio-3.32.2-py3-none-any.whl +COPY ./docs/gradio-3.32.6-py3-none-any.whl ./docs/gradio-3.32.6-py3-none-any.whl RUN pip3 install -r requirements.txt From 9ea0fe4de25409afedaf6b2e677367477f499c69 Mon Sep 17 00:00:00 2001 From: binary-husky Date: Thu, 12 Oct 2023 21:23:15 +0800 Subject: [PATCH 04/13] Update GithubAction+NoLocal+Latex --- docs/GithubAction+NoLocal+Latex | 6 ++---- 1 file changed, 2 insertions(+), 4 deletions(-) diff --git a/docs/GithubAction+NoLocal+Latex b/docs/GithubAction+NoLocal+Latex index 2f2608c..d460513 100644 --- a/docs/GithubAction+NoLocal+Latex +++ b/docs/GithubAction+NoLocal+Latex @@ -8,12 +8,10 @@ FROM fuqingxu/python311_texlive_ctex:latest # 指定路径 WORKDIR /gpt -RUN pip3 install gradio openai numpy arxiv rich +RUN pip3 install openai numpy arxiv rich RUN pip3 install colorama Markdown pygments pymupdf -RUN pip3 install python-docx moviepy pdfminer -RUN pip3 install zh_langchain==0.2.1 +RUN pip3 install python-docx pdfminer RUN pip3 install nougat-ocr -RUN pip3 install aliyun-python-sdk-core==2.13.3 pyOpenSSL scipy git+https://github.com/aliyun/alibabacloud-nls-python-sdk.git # 装载项目文件 COPY . . From c3abc46d4d8a18286089e78ea52023d7fd38eaa4 Mon Sep 17 00:00:00 2001 From: Menghuan1918 Date: Fri, 13 Oct 2023 15:23:06 +0800 Subject: [PATCH 05/13] =?UTF-8?q?=E5=9C=A8proxies=E8=BF=94=E5=9B=9E?= =?UTF-8?q?=E7=A9=BA=E6=97=B6=E4=BC=9A=E9=A6=96=E5=85=88=E5=B0=9D=E8=AF=95?= =?UTF-8?q?=E7=9B=B4=E6=8E=A5=E8=BF=9E=E6=8E=A5?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit --- crazy_functions/谷歌检索小助手.py | 8 +++++++- 1 file changed, 7 insertions(+), 1 deletion(-) diff --git a/crazy_functions/谷歌检索小助手.py b/crazy_functions/谷歌检索小助手.py index 05e80d2..dae8a2b 100644 --- a/crazy_functions/谷歌检索小助手.py +++ b/crazy_functions/谷歌检索小助手.py @@ -26,7 +26,13 @@ def get_meta_information(url, chatbot, history): 'Accept': 'text/html,application/xhtml+xml,application/xml;q=0.9,image/avif,image/webp,image/apng,*/*;q=0.8,application/signed-exchange;v=b3;q=0.7', 'Connection': 'keep-alive' } - session.proxies.update(proxies) + try: + session.proxies.update(proxies) + except: + report_execption(chatbot, history, + a=f"获取代理失败 无代理状态下很可能无法访问OpenAI家族的模型及谷歌学术 建议:检查USE_PROXY选项是否修改。", + b=f"尝试直接连接") + yield from update_ui(chatbot=chatbot, history=history) # 刷新界面 session.headers.update(headers) response = session.get(url) From 673faf8cefea374aae564132119444aa777ff049 Mon Sep 17 00:00:00 2001 From: binary-husky <96192199+binary-husky@users.noreply.github.com> Date: Sat, 14 Oct 2023 19:59:35 +0800 Subject: [PATCH 06/13] =?UTF-8?q?Grobid=E8=B4=9F=E8=BD=BD=E5=9D=87?= =?UTF-8?q?=E8=A1=A1?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit --- config.py | 3 ++- 1 file changed, 2 insertions(+), 1 deletion(-) diff --git a/config.py b/config.py index 22104f0..b8de793 100644 --- a/config.py +++ b/config.py @@ -177,7 +177,8 @@ HUGGINGFACE_ACCESS_TOKEN = "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" # 获取方法:复制以下空间https://huggingface.co/spaces/qingxu98/grobid,设为public,然后GROBID_URL = "https://(你的hf用户名如qingxu98)-(你的填写的空间名如grobid).hf.space" GROBID_URLS = [ "https://qingxu98-grobid.hf.space","https://qingxu98-grobid2.hf.space","https://qingxu98-grobid3.hf.space", - "https://shaocongma-grobid.hf.space","https://FBR123-grobid.hf.space", "https://yeku-grobid.hf.space", + "https://qingxu98-grobid4.hf.space","https://qingxu98-grobid5.hf.space", "https://qingxu98-grobid6.hf.space", + "https://qingxu98-grobid7.hf.space", "https://qingxu98-grobid8.hf.space", ] From 1bb437a5d04205a4b1264910af0a771c1e931fef Mon Sep 17 00:00:00 2001 From: binary-husky Date: Sun, 15 Oct 2023 21:17:00 +0800 Subject: [PATCH 07/13] =?UTF-8?q?=E5=BE=AE=E8=B0=83=E6=8F=90=E7=A4=BA?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit --- crazy_functional.py | 28 +++++++++------------------- 1 file changed, 9 insertions(+), 19 deletions(-) diff --git a/crazy_functional.py b/crazy_functional.py index 6a41600..1d8f5c7 100644 --- a/crazy_functional.py +++ b/crazy_functional.py @@ -190,10 +190,10 @@ def get_crazy_functions(): "Info": "多线程解析并翻译此项目的源码 | 不需要输入参数", "Function": HotReload(解析项目本身) }, - "[插件demo]历史上的今天": { + "历史上的今天": { "Group": "对话", "AsButton": True, - "Info": "查看历史上的今天事件 | 不需要输入参数", + "Info": "查看历史上的今天事件 (这是一个面向开发者的插件Demo) | 不需要输入参数", "Function": HotReload(高阶功能模板函数) }, "精准翻译PDF论文": { @@ -252,7 +252,7 @@ def get_crazy_functions(): "Function": HotReload(Latex中文润色) }, - # 被新插件取代 + # 已经被新插件取代 # "Latex项目全文中译英(输入路径或上传压缩包)": { # "Group": "学术", # "Color": "stop", @@ -260,6 +260,8 @@ def get_crazy_functions(): # "Info": "对Latex项目全文进行中译英处理 | 输入参数为路径或上传压缩包", # "Function": HotReload(Latex中译英) # }, + + # 已经被新插件取代 # "Latex项目全文英译中(输入路径或上传压缩包)": { # "Group": "学术", # "Color": "stop", @@ -395,7 +397,7 @@ def get_crazy_functions(): try: from crazy_functions.批量Markdown翻译 import Markdown翻译指定语言 function_plugins.update({ - "Markdown翻译(手动指定语言)": { + "Markdown翻译(指定翻译成何种语言)": { "Group": "编程", "Color": "stop", "AsButton": False, @@ -440,7 +442,7 @@ def get_crazy_functions(): try: from crazy_functions.交互功能函数模板 import 交互功能模板函数 function_plugins.update({ - "交互功能模板函数": { + "交互功能模板Demo函数(查找wallhaven.cc的壁纸)": { "Group": "对话", "Color": "stop", "AsButton": False, @@ -500,11 +502,11 @@ def get_crazy_functions(): if ENABLE_AUDIO: from crazy_functions.语音助手 import 语音助手 function_plugins.update({ - "实时音频采集": { + "实时语音对话": { "Group": "对话", "Color": "stop", "AsButton": True, - "Info": "开始语言对话 | 没有输入参数", + "Info": "这是一个时刻聆听着的语音对话助手 | 没有输入参数", "Function": HotReload(语音助手) } }) @@ -537,18 +539,6 @@ def get_crazy_functions(): except: print('Load function plugin failed') - # try: - # from crazy_functions.CodeInterpreter import 虚空终端CodeInterpreter - # function_plugins.update({ - # "CodeInterpreter(开发中,仅供测试)": { - # "Group": "编程|对话", - # "Color": "stop", - # "AsButton": False, - # "Function": HotReload(虚空终端CodeInterpreter) - # } - # }) - # except: - # print('Load function plugin failed') # try: # from crazy_functions.chatglm微调工具 import 微调数据集生成 From a4de91d0000a26d460188e85fcbd02af9c7e41b7 Mon Sep 17 00:00:00 2001 From: binary-husky <96192199+binary-husky@users.noreply.github.com> Date: Sun, 15 Oct 2023 22:53:57 +0800 Subject: [PATCH 08/13] =?UTF-8?q?=E4=BF=AE=E6=94=B9=E7=BC=A9=E8=BF=9B?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit --- docker-compose.yml | 8 ++++---- 1 file changed, 4 insertions(+), 4 deletions(-) diff --git a/docker-compose.yml b/docker-compose.yml index 2bf9dfe..dd40dd1 100644 --- a/docker-compose.yml +++ b/docker-compose.yml @@ -32,10 +32,10 @@ services: gpt_academic_full_capability: image: ghcr.io/binary-husky/gpt_academic_with_all_capacity:master environment: - # 请查阅 `config.py`或者 github wiki 以查看所有的配置信息 + # 请查阅 `config.py`或者 github wiki 以查看所有的配置信息 API_KEY: ' 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 ' - # USE_PROXY: ' True ' - # proxies: ' { "http": "http://localhost:10881", "https": "http://localhost:10881", } ' + # USE_PROXY: ' True ' + # proxies: ' { "http": "http://localhost:10881", "https": "http://localhost:10881", } ' LLM_MODEL: ' gpt-3.5-turbo ' AVAIL_LLM_MODELS: ' ["gpt-3.5-turbo", "gpt-4", "qianfan", "sparkv2", "spark", "chatglm"] ' BAIDU_CLOUD_API_KEY : ' bTUtwEAveBrQipEowUvDwYWq ' @@ -51,7 +51,7 @@ services: THEME: ' Chuanhu-Small-and-Beautiful ' ALIYUN_ACCESSKEY: ' LTAI5t6BrFUzxRXVGUWnekh1 ' ALIYUN_SECRET: ' eHmI20SVWIwQZxCiTD2bGQVspP9i68 ' - # LOCAL_MODEL_DEVICE: ' cuda ' + # LOCAL_MODEL_DEVICE: ' cuda ' # 加载英伟达显卡运行时 # runtime: nvidia From 98ef658307eea0cef2e9eb49e80b45408049de26 Mon Sep 17 00:00:00 2001 From: binary-husky Date: Mon, 16 Oct 2023 00:05:31 +0800 Subject: [PATCH 09/13] =?UTF-8?q?=E4=BF=AE=E5=A4=8Dwarmup=E6=A8=A1?= =?UTF-8?q?=E5=9D=97=E7=9A=84=E5=BB=B6=E8=BF=9F=E9=97=AE=E9=A2=98?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit --- main.py | 18 ++++++++++-------- 1 file changed, 10 insertions(+), 8 deletions(-) diff --git a/main.py b/main.py index 25b1f4a..e7fa3c5 100644 --- a/main.py +++ b/main.py @@ -409,19 +409,21 @@ def main(): demo.load(None, inputs=[gr.Textbox(LAYOUT, visible=False)], outputs=None, _js='(LAYOUT)=>{GptAcademicJavaScriptInit(LAYOUT);}') # gradio的inbrowser触发不太稳定,回滚代码到原始的浏览器打开函数 - def auto_opentab_delay(): + def run_delayed_tasks(): import threading, webbrowser, time print(f"如果浏览器没有自动打开,请复制并转到以下URL:") if DARK_MODE: print(f"\t「暗色主题已启用(支持动态切换主题)」: http://localhost:{PORT}") else: print(f"\t「亮色主题已启用(支持动态切换主题)」: http://localhost:{PORT}") - def open(): - time.sleep(2) # 打开浏览器 - webbrowser.open_new_tab(f"http://localhost:{PORT}") - threading.Thread(target=open, name="open-browser", daemon=True).start() - threading.Thread(target=auto_update, name="self-upgrade", daemon=True).start() - threading.Thread(target=warm_up_modules, name="warm-up", daemon=True).start() - auto_opentab_delay() + def auto_updates(): time.sleep(0); auto_update() + def open_browser(): time.sleep(2); webbrowser.open_new_tab(f"http://localhost:{PORT}") + def warm_up_mods(): time.sleep(4); warm_up_modules() + + threading.Thread(target=auto_updates, name="self-upgrade", daemon=True).start() # 查看自动更新 + threading.Thread(target=open_browser, name="open-browser", daemon=True).start() # 打开浏览器页面 + threading.Thread(target=warm_up_mods, name="warm-up", daemon=True).start() # 预热tiktoken模块 + + run_delayed_tasks() demo.queue(concurrency_count=CONCURRENT_COUNT).launch( quiet=True, server_name="0.0.0.0", From 560d4e2cb14e4372be8674270503071e4e0e6617 Mon Sep 17 00:00:00 2001 From: binary-husky <96192199+binary-husky@users.noreply.github.com> Date: Wed, 18 Oct 2023 11:10:38 +0800 Subject: [PATCH 10/13] =?UTF-8?q?=E4=BF=AE=E6=AD=A3Dockerfile=E4=B8=AD?= =?UTF-8?q?=E7=9A=84=E9=94=99=E8=AF=AF?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit --- docs/GithubAction+NoLocal+Latex | 3 --- 1 file changed, 3 deletions(-) diff --git a/docs/GithubAction+NoLocal+Latex b/docs/GithubAction+NoLocal+Latex index cbed50f..d460513 100644 --- a/docs/GithubAction+NoLocal+Latex +++ b/docs/GithubAction+NoLocal+Latex @@ -5,9 +5,6 @@ FROM fuqingxu/python311_texlive_ctex:latest -# 删除文档文件以节约空间 -rm -rf /usr/local/texlive/2023/texmf-dist/doc - # 指定路径 WORKDIR /gpt From 775b07dbccbbfe68ecb2d7944f471596bd6ecd8b Mon Sep 17 00:00:00 2001 From: binary-husky <96192199+binary-husky@users.noreply.github.com> Date: Wed, 18 Oct 2023 11:15:35 +0800 Subject: [PATCH 11/13] =?UTF-8?q?=E4=B8=BADockerfile=E6=B7=BB=E5=8A=A0?= =?UTF-8?q?=E6=9B=B4=E5=A4=9A=E6=B3=A8=E9=87=8A?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit --- Dockerfile | 21 +++++++++++---------- 1 file changed, 11 insertions(+), 10 deletions(-) diff --git a/Dockerfile b/Dockerfile index 7371e45..f51befa 100644 --- a/Dockerfile +++ b/Dockerfile @@ -1,34 +1,35 @@ -# 此Dockerfile适用于“无本地模型”的环境构建,如果需要使用chatglm等本地模型或者latex运行依赖,请参考 docker-compose.yml -# 如何构建: 先修改 `config.py`, 然后 `docker build -t gpt-academic . ` -# 如何运行(Linux下): `docker run --rm -it --net=host gpt-academic ` -# 如何运行(其他操作系统,选择任意一个固定端口50923): `docker run --rm -it -e WEB_PORT=50923 -p 50923:50923 gpt-academic ` +# 此Dockerfile适用于“无本地模型”的迷你运行环境构建 +# 如果需要使用chatglm等本地模型或者latex运行依赖,请参考 docker-compose.yml +# - 如何构建: 先修改 `config.py`, 然后 `docker build -t gpt-academic . ` +# - 如何运行(Linux下): `docker run --rm -it --net=host gpt-academic ` +# - 如何运行(其他操作系统,选择任意一个固定端口50923): `docker run --rm -it -e WEB_PORT=50923 -p 50923:50923 gpt-academic ` FROM python:3.11 -# 非必要步骤,更换pip源 +# 非必要步骤,更换pip源 (以下三行,可以删除) RUN echo '[global]' > /etc/pip.conf && \ echo 'index-url = https://mirrors.aliyun.com/pypi/simple/' >> /etc/pip.conf && \ echo 'trusted-host = mirrors.aliyun.com' >> /etc/pip.conf -# 进入工作路径 +# 进入工作路径(必要) WORKDIR /gpt -# 安装大部分依赖,利用Docker缓存加速以后的构建 +# 安装大部分依赖,利用Docker缓存加速以后的构建 (以下三行,可以删除) COPY requirements.txt ./ COPY ./docs/gradio-3.32.6-py3-none-any.whl ./docs/gradio-3.32.6-py3-none-any.whl RUN pip3 install -r requirements.txt -# 装载项目文件,安装剩余依赖 +# 装载项目文件,安装剩余依赖(必要) COPY . . RUN pip3 install -r requirements.txt -# 非必要步骤,用于预热模块 +# 非必要步骤,用于预热模块(可以删除) RUN python3 -c 'from check_proxy import warm_up_modules; warm_up_modules()' -# 启动 +# 启动(必要) CMD ["python3", "-u", "main.py"] From 20e3eee6e7a4a382645c7f4cd31177afd6116fff Mon Sep 17 00:00:00 2001 From: binary-husky <96192199+binary-husky@users.noreply.github.com> Date: Wed, 18 Oct 2023 16:23:28 +0800 Subject: [PATCH 12/13] Update GithubAction+NoLocal+Latex --- docs/GithubAction+NoLocal+Latex | 5 +++++ 1 file changed, 5 insertions(+) diff --git a/docs/GithubAction+NoLocal+Latex b/docs/GithubAction+NoLocal+Latex index d460513..be9fda2 100644 --- a/docs/GithubAction+NoLocal+Latex +++ b/docs/GithubAction+NoLocal+Latex @@ -4,6 +4,11 @@ # - 3 运行 docker run -v /home/fuqingxu/arxiv_cache:/root/arxiv_cache --rm -it --net=host gpt-academic-nolocal-latex FROM fuqingxu/python311_texlive_ctex:latest +ENV PATH "$PATH:/usr/local/texlive/2022/bin/x86_64-linux" +ENV PATH "$PATH:/usr/local/texlive/2023/bin/x86_64-linux" +ENV PATH "$PATH:/usr/local/texlive/2024/bin/x86_64-linux" +ENV PATH "$PATH:/usr/local/texlive/2025/bin/x86_64-linux" +ENV PATH "$PATH:/usr/local/texlive/2026/bin/x86_64-linux" # 指定路径 WORKDIR /gpt From 3410bd9b1d087fa5f5eb856c0dd3a2d444d44955 Mon Sep 17 00:00:00 2001 From: binary-husky <96192199+binary-husky@users.noreply.github.com> Date: Thu, 19 Oct 2023 16:05:12 +0800 Subject: [PATCH 13/13] Update README.md --- README.md | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/README.md b/README.md index 61a1eb0..77ff15e 100644 --- a/README.md +++ b/README.md @@ -1,6 +1,6 @@ > **Note** > -> 2023.10.8: Gradio, Pydantic依赖调整,已修改 `requirements.txt`。请及时**更新代码**,安装依赖时,请严格选择`requirements.txt`中**指定的版本** +> 2023.10.8: Gradio, Pydantic依赖调整,已修改 `requirements.txt`。请及时**更新代码**,安装依赖时,请严格选择`requirements.txt`中**指定的版本**。 > > `pip install -r requirements.txt` @@ -16,7 +16,7 @@ To translate this project to arbitrary language with GPT, read and run [`multi_l > > 1.请注意只有 **高亮** 标识的函数插件(按钮)才支持读取文件,部分插件位于插件区的**下拉菜单**中。另外我们以**最高优先级**欢迎和处理任何新插件的PR。 > -> 2.本项目中每个文件的功能都在[自译解报告`self_analysis.md`](https://github.com/binary-husky/gpt_academic/wiki/GPT‐Academic项目自译解报告)详细说明。随着版本的迭代,您也可以随时自行点击相关函数插件,调用GPT重新生成项目的自我解析报告。常见问题[`wiki`](https://github.com/binary-husky/gpt_academic/wiki)。[安装方法](#installation) | [配置说明](https://github.com/binary-husky/gpt_academic/wiki/%E9%A1%B9%E7%9B%AE%E9%85%8D%E7%BD%AE%E8%AF%B4%E6%98%8E)。 +> 2.本项目中每个文件的功能都在[自译解报告`self_analysis.md`](https://github.com/binary-husky/gpt_academic/wiki/GPT‐Academic项目自译解报告)详细说明。随着版本的迭代,您也可以随时自行点击相关函数插件,调用GPT重新生成项目的自我解析报告。常见问题[`wiki`](https://github.com/binary-husky/gpt_academic/wiki)。[常规安装方法](#installation) | [一键安装脚本](https://github.com/binary-husky/gpt_academic/releases) | [配置说明](https://github.com/binary-husky/gpt_academic/wiki/%E9%A1%B9%E7%9B%AE%E9%85%8D%E7%BD%AE%E8%AF%B4%E6%98%8E)。 > > 3.本项目兼容并鼓励尝试国产大语言模型ChatGLM和Moss等等。支持多个api-key共存,可在配置文件中填写如`API_KEY="openai-key1,openai-key2,azure-key3,api2d-key4"`。需要临时更换`API_KEY`时,在输入区输入临时的`API_KEY`然后回车键提交后即可生效。