From 0ac7734c7d039b0663bf2468b6c58683087b6e90 Mon Sep 17 00:00:00 2001 From: binary-husky <96192199+binary-husky@users.noreply.github.com> Date: Thu, 13 Apr 2023 16:36:08 +0800 Subject: [PATCH] Delete Dockerfile+ChatGLM --- Dockerfile+ChatGLM | 50 ---------------------------------------------- 1 file changed, 50 deletions(-) delete mode 100644 Dockerfile+ChatGLM diff --git a/Dockerfile+ChatGLM b/Dockerfile+ChatGLM deleted file mode 100644 index f99f2a6..0000000 --- a/Dockerfile+ChatGLM +++ /dev/null @@ -1,50 +0,0 @@ -# How to build | 如何构建: docker build -t gpt-academic --network=host -f Dockerfile+ChatGLM . -# How to run | 如何运行 (1) 直接运行: docker run --rm -it --net=host --gpus=all gpt-academic -# How to run | 如何运行 (2) 我想运行之前进容器做一些调整: docker run --rm -it --net=host --gpus=all gpt-academic bash - -# 从NVIDIA源,从而支持显卡运损(检查宿主的nvidia-smi中的cuda版本必须>=11.3) -FROM nvidia/cuda:11.3.1-runtime-ubuntu20.04 -ARG useProxyNetwork='' -RUN apt-get update -RUN apt-get install -y curl proxychains curl -RUN apt-get install -y git python python3 python-dev python3-dev --fix-missing - -# 配置代理网络(构建Docker镜像时使用) -# # comment out below if you do not need proxy network | 如果不需要翻墙 - 从此行向下删除 -RUN $useProxyNetwork curl cip.cc -RUN sed -i '$ d' /etc/proxychains.conf -RUN sed -i '$ d' /etc/proxychains.conf -RUN echo "socks5 127.0.0.1 10880" >> /etc/proxychains.conf -ARG useProxyNetwork=proxychains -# # comment out above if you do not need proxy network | 如果不需要翻墙 - 从此行向上删除 - - -# use python3 as the system default python -RUN curl -sS https://bootstrap.pypa.io/get-pip.py | python3.8 - -# 下载分支 -WORKDIR /gpt -RUN $useProxyNetwork git clone https://github.com/binary-husky/chatgpt_academic.git -b v3.0 -WORKDIR /gpt/chatgpt_academic -RUN $useProxyNetwork python3 -m pip install -r requirements.txt -RUN $useProxyNetwork python3 -m pip install -r request_llm/requirements_chatglm.txt -RUN $useProxyNetwork python3 -m pip install torch --extra-index-url https://download.pytorch.org/whl/cu113 - -# 预热CHATGLM参数(非必要 可选步骤) -RUN echo ' \n\ -from transformers import AutoModel, AutoTokenizer \n\ -chatglm_tokenizer = AutoTokenizer.from_pretrained("THUDM/chatglm-6b", trust_remote_code=True) \n\ -chatglm_model = AutoModel.from_pretrained("THUDM/chatglm-6b", trust_remote_code=True).float() ' >> warm_up_chatglm.py -RUN python3 -u warm_up_chatglm.py -RUN $useProxyNetwork git pull - -# 为chatgpt-academic配置代理和API-KEY (非必要 可选步骤) -RUN echo ' \n\ -API_KEY = "sk-xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx" \n\ -USE_PROXY = True \n\ -LLM_MODEL = "chatglm" \n\ -LOCAL_MODEL_DEVICE = "cuda" \n\ -proxies = { "http": "socks5h://localhost:10880", "https": "socks5h://localhost:10880", } ' >> config_private.py - -# 启动 -CMD ["python3", "-u", "main.py"]