Compare commits

...

18 Commits
master ... test

Author SHA1 Message Date
qingxu fu
3103817990 add private conf 2023-03-22 17:54:15 +08:00
qingxu fu
88aaf310c4 upload 2023-03-22 17:48:25 +08:00
qingxu fu
e8d86a3242 代理位置 2023-03-22 17:45:10 +08:00
qingxu fu
37fd1b1c97 upload 2023-03-22 17:35:23 +08:00
qingxu fu
e3e313beab logging 2023-03-22 17:32:48 +08:00
qingxu fu
380a8a9d4d fix logging encoding 2023-03-22 17:30:30 +08:00
qingxu fu
d00f6bb1a6 add proxy debug funtion 2023-03-22 17:25:37 +08:00
binary-husky
f8a44a82a9
Update README.md 2023-03-22 16:09:37 +08:00
binary-husky
2be45c386d
Update predict.py 2023-03-21 21:24:38 +08:00
binary-husky
2b7e8b9278
Update README.md 2023-03-21 17:53:40 +08:00
binary-husky
8590014462
Update README.md 2023-03-21 17:53:04 +08:00
binary-husky
8215caddb2
Update README.md 2023-03-21 15:49:52 +08:00
505030475
93a9d27b1e ok 2023-03-21 13:53:24 +08:00
binary-husky
e128523103
Update README.md 2023-03-21 13:45:08 +08:00
505030475
da0caab4cf add deploy method for windows 2023-03-21 13:35:53 +08:00
binary-husky
6e9813565a
Update README.md 2023-03-20 18:56:09 +08:00
binary-husky
01cc409f99
Update README.md 2023-03-20 18:55:06 +08:00
Your Name
3438f8f291 readme 2023-03-20 18:39:48 +08:00
10 changed files with 149 additions and 71 deletions

View File

@ -2,7 +2,42 @@
**如果喜欢这个项目请给它一个Star**
## 使用docker
- 支持GPT输出的markdown表格
<div align="center">
<img src="demo2.jpg" width="500" >
</div>
- 如果输出包含公式会同时以tex形式和渲染形式显示方便复制和阅读
<div align="center">
<img src="demo.jpg" width="500" >
</div>
- 所有按钮都通过读取functional.py动态生成可随意加自定义功能解放粘贴板
<div align="center">
<img src="公式.gif" width="700" >
</div>
- 代码的显示自然也不在话下 https://www.bilibili.com/video/BV1F24y147PD/
<div align="center">
<img src="润色.gif" width="700" >
</div>
## 直接运行 (Windows or Linux)
```
# 下载项目
git clone https://github.com/binary-husky/chatgpt_academic.git
cd chatgpt_academic
# 配置 海外Proxy 和 OpenAI API KEY
config.py
# 安装依赖
python -m pip install -r requirements.txt
# 运行
python main.py
```
## 使用docker (Linux)
``` sh
# 下载项目
@ -17,6 +52,25 @@ docker run --rm -it --net=host gpt-academic
```
## 自定义新的便捷按钮
打开functional.py只需看一眼就知道怎么弄了
例如
```
"英译中": {
"Prefix": "请翻译成中文:\n\n",
"Button": None,
"Suffix": "",
},
```
## 配置代理
在```config.py```中修改端口与代理软件对应
![image](https://user-images.githubusercontent.com/96192199/226571294-37a47cd9-4d40-4c16-97a2-d360845406f7.png)
![image](https://user-images.githubusercontent.com/96192199/226838985-e5c95956-69c2-4c23-a4dd-cd7944eeb451.png)
## 参考项目
```
https://github.com/Python-Markdown/markdown

26
check_proxy.py Normal file
View File

@ -0,0 +1,26 @@
"""
用python的requests库查询本机ip地址所在地
ChatGPT:
"""
def check_proxy(proxies):
import requests
proxies_https = proxies['https'] if proxies is not None else ''
try:
response = requests.get("https://ipapi.co/json/", proxies=proxies, timeout=4)
data = response.json()
print(f'查询代理的地理位置,返回的结果是{data}')
country = data['country_name']
result = f"代理配置 {proxies_https}, 代理所在地:{country}"
print(result)
return result
except:
result = f"代理配置 {proxies_https}, 代理所在地查询超时,代理可能无效"
print(result)
return result
if __name__ == '__main__':
try: from config_private import proxies # 放自己的秘密如API和代理网址 os.path.exists('config_private.py')
except: from config import proxies
check_proxy(proxies)

View File

@ -1,11 +1,23 @@
# my_api_key = "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"
# API_KEY = "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" 此key无效
API_KEY = "sk-此处填API秘钥"
API_URL = "https://api.openai.com/v1/chat/completions"
# 改为True应用代理
USE_PROXY = False
if USE_PROXY:
# 代理网络的地址,打开你的科学上网软件查看代理的协议(socks5/http)、地址(localhost)和端口(11284)
proxies = { "http": "socks5h://localhost:11284", "https": "socks5h://localhost:11284", }
print('网络代理状态:运行。')
else:
proxies = None
print('网络代理状态:未配置。无代理状态下很可能无法访问。')
# 发送请求到OpenAI后等待多久判定为超时
TIMEOUT_SECONDS = 20
# 网页的端口, -1代表随机端口
WEB_PORT = -1
# 检查一下是不是忘了改config
if API_KEY == "sk-此处填API秘钥":
assert False, "请在config文件中修改API密钥, 添加海外代理之后再运行"

BIN
demo.jpg Normal file

Binary file not shown.

After

Width:  |  Height:  |  Size: 262 KiB

BIN
demo2.jpg Normal file

Binary file not shown.

After

Width:  |  Height:  |  Size: 264 KiB

View File

@ -10,19 +10,16 @@ def get_functionals():
"英语学术润色": {
"Prefix": "Below is a paragraph from an academic paper. Polish the writing to meet the academic style, \
improve the spelling, grammar, clarity, concision and overall readability. When neccessary, rewrite the whole sentence. \
Furthermore, list all modification and explain the reasons to do so in markdown table.\n\n",
"Button": None,
"Suffix": "",
"Color": "stop",
Furthermore, list all modification and explain the reasons to do so in markdown table.\n\n", # 前言
"Suffix": "", # 后语
"Color": "stop", # 按钮颜色
},
"中文学术润色": {
"Prefix": "作为一名中文学术论文写作改进助理,你的任务是改进所提供文本的拼写、语法、清晰、简洁和整体可读性,同时分解长句,减少重复,并提供改进建议。请只提供文本的更正版本,避免包括解释。请编辑以下文本:\n\n",
"Button": None,
"Suffix": "",
},
"查找语法错误": {
"Prefix": "Below is a paragraph from an academic paper. Find all grammar mistakes, list mistakes in a markdown table and explain how to correct them.\n\n",
"Button": None,
"Suffix": "",
},
"中英互译": {
@ -37,28 +34,23 @@ When providing translations, please use Chinese to explain each sentences ten
For phrases or individual words that require translation, provide the source (dictionary) for each one.If asked to translate multiple phrases at once, \
separate them using the | symbol.Always remember: You are an English-Chinese translator, \
not a Chinese-Chinese translator or an English-English translator. Below is the text you need to translate: \n\n",
"Button": None,
"Suffix": "",
"Color": "stop",
},
"中译英": {
"Prefix": "Please translate following sentence to English: \n\n",
"Button": None,
"Suffix": "",
},
"学术中译英": {
"Prefix": "Please translate following sentence to English with academic writing, and provide some related authoritative examples: \n\n",
"Button": None,
"Suffix": "",
},
"英译中": {
"Prefix": "请翻译成中文:\n\n",
"Button": None,
"Suffix": "",
},
"解释代码": {
"Prefix": "请解释以下代码:\n```\n",
"Button": None,
"Suffix": "\n```\n",
"Color": "stop",
},

18
main.py
View File

@ -4,6 +4,9 @@ import markdown, mdtex2html
from predict import predict
from show_math import convert as convert_math
try: from config_private import proxies, WEB_PORT # 放自己的秘密如API和代理网址 os.path.exists('config_private.py')
except: from config import proxies, WEB_PORT
def find_free_port():
import socket
from contextlib import closing
@ -11,16 +14,16 @@ def find_free_port():
s.bind(('', 0))
s.setsockopt(socket.SOL_SOCKET, socket.SO_REUSEADDR, 1)
return s.getsockname()[1]
PORT = find_free_port()
PORT = find_free_port() if WEB_PORT <= 0 else WEB_PORT
initial_prompt = "Serve me as a writing and programming assistant."
title_html = """<h1 align="center">ChatGPT 学术优化</h1>"""
import logging
os.makedirs('gpt_log', exist_ok=True)
logging.basicConfig(filename='gpt_log/predict.log', level=logging.INFO)
logging.basicConfig(filename='gpt_log/chat_secrets.log', level=logging.INFO, encoding='utf-8')
print('所有问询记录将自动保存在本地目录./gpt_log/chat_secrets.log请注意自我隐私保护哦')
from functional import get_functionals
functional = get_functionals()
@ -46,9 +49,6 @@ def markdown_convertion(txt):
else:
return markdown.markdown(txt,extensions=['fenced_code','tables'])
# math_config = {'mdx_math': {'enable_dollar_delimiter': True}}
# markdown.markdown(txt, extensions=['fenced_code', 'tables', 'mdx_math'], extension_configs=math_config)
def format_io(self,y):
if y is None:
@ -84,8 +84,8 @@ with gr.Blocks() as demo:
for k in functional:
variant = functional[k]["Color"] if "Color" in functional[k] else "secondary"
functional[k]["Button"] = gr.Button(k, variant=variant)
statusDisplay = gr.Markdown("status: ready")
from check_proxy import check_proxy
statusDisplay = gr.Markdown(f"{check_proxy(proxies)}")
systemPromptTxt = gr.Textbox(show_label=True, placeholder=f"System Prompt", label="System prompt", value=initial_prompt).style(container=True)
#inputs, top_p, temperature, top_k, repetition_penalty
with gr.Accordion("arguments", open=False):

View File

@ -4,23 +4,11 @@ import logging
import traceback
import requests
import importlib
import os
if os.path.exists('config_private.py'):
# 放自己的秘密如API和代理网址
from config_private import proxies, API_URL, API_KEY
else:
from config import proxies, API_URL, API_KEY
def compose_system(system_prompt):
return {"role": "system", "content": system_prompt}
def compose_user(user_input):
return {"role": "user", "content": user_input}
try: from config_private import proxies, API_URL, API_KEY, TIMEOUT_SECONDS # 放自己的秘密如API和代理网址 os.path.exists('config_private.py')
except: from config import proxies, API_URL, API_KEY, TIMEOUT_SECONDS
timeout_bot_msg = 'Request timeout, network error. please check proxy settings in config.py.'
def predict(inputs, top_p, temperature, chatbot=[], history=[], system_prompt='', retry=False,
stream = True, additional_fn=None):
@ -35,7 +23,7 @@ def predict(inputs, top_p, temperature, chatbot=[], history=[], system_prompt=''
raw_input = inputs
logging.info(f'[raw_input] {raw_input}')
chatbot.append((inputs, ""))
yield chatbot, history, "Waiting"
yield chatbot, history, "等待响应"
headers = {
"Content-Type": "application/json",
@ -46,29 +34,32 @@ def predict(inputs, top_p, temperature, chatbot=[], history=[], system_prompt=''
print(f"chat_counter - {chat_counter}")
messages = [compose_system(system_prompt)]
messages = [{"role": "system", "content": system_prompt}]
if chat_counter:
for index in range(0, 2*chat_counter, 2):
d1 = {}
d1["role"] = "user"
d1["content"] = history[index]
d2 = {}
d2["role"] = "assistant"
d2["content"] = history[index+1]
if d1["content"] != "":
if d2["content"] != "" or retry:
messages.append(d1)
messages.append(d2)
what_i_have_asked = {}
what_i_have_asked["role"] = "user"
what_i_have_asked["content"] = history[index]
what_gpt_answer = {}
what_gpt_answer["role"] = "assistant"
what_gpt_answer["content"] = history[index+1]
if what_i_have_asked["content"] != "":
if not (what_gpt_answer["content"] != "" or retry): continue
if what_gpt_answer["content"] == timeout_bot_msg: continue
messages.append(what_i_have_asked)
messages.append(what_gpt_answer)
else:
messages[-1]['content'] = d2['content']
messages[-1]['content'] = what_gpt_answer['content']
if retry and chat_counter:
messages.pop()
else:
temp3 = {}
temp3["role"] = "user"
temp3["content"] = inputs
messages.append(temp3)
what_i_ask_now = {}
what_i_ask_now["role"] = "user"
what_i_ask_now["content"] = inputs
messages.append(what_i_ask_now)
chat_counter += 1
# messages
payload = {
"model": "gpt-3.5-turbo",
@ -87,10 +78,10 @@ def predict(inputs, top_p, temperature, chatbot=[], history=[], system_prompt=''
try:
# make a POST request to the API endpoint using the requests.post method, passing in stream=True
response = requests.post(API_URL, headers=headers, proxies=proxies,
json=payload, stream=True, timeout=15)
json=payload, stream=True, timeout=TIMEOUT_SECONDS)
except:
chatbot[-1] = ((chatbot[-1][0], 'Requests Timeout, Network Error.'))
yield chatbot, history, "Requests Timeout"
chatbot[-1] = ((chatbot[-1][0], timeout_bot_msg))
yield chatbot, history, "请求超时"
raise TimeoutError
token_counter = 0
@ -101,8 +92,6 @@ def predict(inputs, top_p, temperature, chatbot=[], history=[], system_prompt=''
stream_response = response.iter_lines()
while True:
chunk = next(stream_response)
# print(chunk)
if chunk == b'data: [DONE]':
break
@ -119,16 +108,21 @@ def predict(inputs, top_p, temperature, chatbot=[], history=[], system_prompt=''
break
except Exception as e:
traceback.print_exc()
print(chunk.decode())
chunkjson = json.loads(chunk.decode()[6:])
status_text = f"id: {chunkjson['id']}, finish_reason: {chunkjson['choices'][0]['finish_reason']}"
partial_words = partial_words + \
json.loads(chunk.decode()[6:])[
'choices'][0]["delta"]["content"]
if token_counter == 0:
history.append(" " + partial_words)
else:
history[-1] = partial_words
chatbot[-1] = (history[-2], history[-1])
token_counter += 1
yield chatbot, history, status_text
try:
chunkjson = json.loads(chunk.decode()[6:])
status_text = f"finish_reason: {chunkjson['choices'][0]['finish_reason']}"
partial_words = partial_words + json.loads(chunk.decode()[6:])['choices'][0]["delta"]["content"]
if token_counter == 0:
history.append(" " + partial_words)
else:
history[-1] = partial_words
chatbot[-1] = (history[-2], history[-1])
token_counter += 1
yield chatbot, history, status_text
except Exception as e:
traceback.print_exc()
print(chunk.decode())
yield chatbot, history, "Json解析不合常规"

BIN
公式.gif Normal file

Binary file not shown.

After

Width:  |  Height:  |  Size: 11 MiB

BIN
润色.gif Normal file

Binary file not shown.

After

Width:  |  Height:  |  Size: 12 MiB