better traceback
This commit is contained in:
parent
ea4cd5a75a
commit
031232c986
@ -1,8 +1,4 @@
|
|||||||
|
|
||||||
"""
|
|
||||||
我:用python的requests库查询本机ip地址所在地
|
|
||||||
ChatGPT:
|
|
||||||
"""
|
|
||||||
def check_proxy(proxies):
|
def check_proxy(proxies):
|
||||||
import requests
|
import requests
|
||||||
proxies_https = proxies['https'] if proxies is not None else '无'
|
proxies_https = proxies['https'] if proxies is not None else '无'
|
||||||
|
@ -1,25 +1,8 @@
|
|||||||
from functools import wraps
|
|
||||||
from predict import predict_no_ui
|
from predict import predict_no_ui
|
||||||
|
from toolbox import CatchException, report_execption
|
||||||
fast_debug = False
|
fast_debug = False
|
||||||
|
|
||||||
|
|
||||||
def report_execption(chatbot, history, a, b):
|
|
||||||
chatbot.append((a, b))
|
|
||||||
history.append(a); history.append(b)
|
|
||||||
|
|
||||||
# 捕获不能预料的异常
|
|
||||||
def CatchException(f):
|
|
||||||
@wraps(f)
|
|
||||||
def decorated(txt, top_p, temperature, chatbot, history, systemPromptTxt, WEB_PORT):
|
|
||||||
try:
|
|
||||||
yield from f(txt, top_p, temperature, chatbot, history, systemPromptTxt, WEB_PORT)
|
|
||||||
except Exception as e:
|
|
||||||
import traceback
|
|
||||||
tb_str = traceback.format_exc()
|
|
||||||
chatbot[-1] = (chatbot[-1][0], f"[Local Message] something error occured: \n {tb_str}")
|
|
||||||
yield chatbot, history, f'异常 {e}'
|
|
||||||
return decorated
|
|
||||||
|
|
||||||
def 解析Paper(file_manifest, project_folder, top_p, temperature, chatbot, history, systemPromptTxt):
|
def 解析Paper(file_manifest, project_folder, top_p, temperature, chatbot, history, systemPromptTxt):
|
||||||
import time, glob, os
|
import time, glob, os
|
||||||
print('begin analysis on:', file_manifest)
|
print('begin analysis on:', file_manifest)
|
||||||
|
@ -1,18 +1,7 @@
|
|||||||
from functools import wraps
|
|
||||||
from predict import predict_no_ui
|
from predict import predict_no_ui
|
||||||
|
from toolbox import CatchException, report_execption
|
||||||
fast_debug = False
|
fast_debug = False
|
||||||
|
|
||||||
def CatchException(f):
|
|
||||||
@wraps(f)
|
|
||||||
def decorated(txt, top_p, temperature, chatbot, history, systemPromptTxt, WEB_PORT):
|
|
||||||
try:
|
|
||||||
yield from f(txt, top_p, temperature, chatbot, history, systemPromptTxt, WEB_PORT)
|
|
||||||
except Exception as e:
|
|
||||||
import traceback
|
|
||||||
tb_str = traceback.format_exc()
|
|
||||||
chatbot[-1] = (chatbot[-1][0], f"[Local Message] something error occured: \n {tb_str}")
|
|
||||||
yield chatbot, history, f'异常 {e}'
|
|
||||||
return decorated
|
|
||||||
|
|
||||||
|
|
||||||
@CatchException
|
@CatchException
|
||||||
@ -66,12 +55,6 @@ def 解析项目本身(txt, top_p, temperature, chatbot, history, systemPromptTx
|
|||||||
history.append(i_say); history.append(gpt_say)
|
history.append(i_say); history.append(gpt_say)
|
||||||
yield chatbot, history, '正常'
|
yield chatbot, history, '正常'
|
||||||
|
|
||||||
def report_execption(chatbot, history, a, b):
|
|
||||||
chatbot.append((a, b))
|
|
||||||
history.append(a); history.append(b)
|
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
def 解析源代码(file_manifest, project_folder, top_p, temperature, chatbot, history, systemPromptTxt):
|
def 解析源代码(file_manifest, project_folder, top_p, temperature, chatbot, history, systemPromptTxt):
|
||||||
|
51
main.py
51
main.py
@ -1,20 +1,11 @@
|
|||||||
import os; os.environ['no_proxy'] = '*'
|
import os; os.environ['no_proxy'] = '*'
|
||||||
import gradio as gr
|
import gradio as gr
|
||||||
import markdown, mdtex2html
|
|
||||||
from predict import predict
|
from predict import predict
|
||||||
from show_math import convert as convert_math
|
from toolbox import format_io, find_free_port
|
||||||
|
|
||||||
try: from config_private import proxies, WEB_PORT # 放自己的秘密如API和代理网址 os.path.exists('config_private.py')
|
try: from config_private import proxies, WEB_PORT # 放自己的秘密如API和代理网址 os.path.exists('config_private.py')
|
||||||
except: from config import proxies, WEB_PORT
|
except: from config import proxies, WEB_PORT
|
||||||
|
|
||||||
def find_free_port():
|
|
||||||
import socket
|
|
||||||
from contextlib import closing
|
|
||||||
with closing(socket.socket(socket.AF_INET, socket.SOCK_STREAM)) as s:
|
|
||||||
s.bind(('', 0))
|
|
||||||
s.setsockopt(socket.SOL_SOCKET, socket.SO_REUSEADDR, 1)
|
|
||||||
return s.getsockname()[1]
|
|
||||||
|
|
||||||
PORT = find_free_port() if WEB_PORT <= 0 else WEB_PORT
|
PORT = find_free_port() if WEB_PORT <= 0 else WEB_PORT
|
||||||
|
|
||||||
initial_prompt = "Serve me as a writing and programming assistant."
|
initial_prompt = "Serve me as a writing and programming assistant."
|
||||||
@ -33,41 +24,6 @@ functional = get_functionals()
|
|||||||
from functional_crazy import get_crazy_functionals
|
from functional_crazy import get_crazy_functionals
|
||||||
crazy_functional = get_crazy_functionals()
|
crazy_functional = get_crazy_functionals()
|
||||||
|
|
||||||
def reset_textbox(): return gr.update(value='')
|
|
||||||
|
|
||||||
def text_divide_paragraph(text):
|
|
||||||
if '```' in text:
|
|
||||||
# careful input
|
|
||||||
return text
|
|
||||||
else:
|
|
||||||
# wtf input
|
|
||||||
lines = text.split("\n")
|
|
||||||
for i, line in enumerate(lines):
|
|
||||||
if i!=0: lines[i] = "<p>"+lines[i].replace(" ", " ")+"</p>"
|
|
||||||
text = "".join(lines)
|
|
||||||
return text
|
|
||||||
|
|
||||||
def markdown_convertion(txt):
|
|
||||||
if ('$' in txt) and ('```' not in txt):
|
|
||||||
math_config = {'mdx_math': {'enable_dollar_delimiter': True}}
|
|
||||||
return markdown.markdown(txt,extensions=['fenced_code','tables']) + '<br><br>' + \
|
|
||||||
markdown.markdown(convert_math(txt, splitParagraphs=False),extensions=['fenced_code','tables'])
|
|
||||||
else:
|
|
||||||
return markdown.markdown(txt,extensions=['fenced_code','tables'])
|
|
||||||
|
|
||||||
|
|
||||||
def format_io(self,y):
|
|
||||||
if y is None:
|
|
||||||
return []
|
|
||||||
i_ask, gpt_reply = y[-1]
|
|
||||||
|
|
||||||
i_ask = text_divide_paragraph(i_ask) # 输入部分太自由,预处理一波
|
|
||||||
|
|
||||||
y[-1] = (
|
|
||||||
None if i_ask is None else markdown.markdown(i_ask, extensions=['fenced_code','tables']),
|
|
||||||
None if gpt_reply is None else markdown_convertion(gpt_reply)
|
|
||||||
)
|
|
||||||
return y
|
|
||||||
gr.Chatbot.postprocess = format_io
|
gr.Chatbot.postprocess = format_io
|
||||||
|
|
||||||
with gr.Blocks() as demo:
|
with gr.Blocks() as demo:
|
||||||
@ -103,7 +59,6 @@ with gr.Blocks() as demo:
|
|||||||
|
|
||||||
txt.submit(predict, [txt, top_p, temperature, chatbot, history, systemPromptTxt], [chatbot, history, statusDisplay])
|
txt.submit(predict, [txt, top_p, temperature, chatbot, history, systemPromptTxt], [chatbot, history, statusDisplay])
|
||||||
submitBtn.click(predict, [txt, top_p, temperature, chatbot, history, systemPromptTxt], [chatbot, history, statusDisplay], show_progress=True)
|
submitBtn.click(predict, [txt, top_p, temperature, chatbot, history, systemPromptTxt], [chatbot, history, statusDisplay], show_progress=True)
|
||||||
# submitBtn.click(reset_textbox, [], [txt])
|
|
||||||
for k in functional:
|
for k in functional:
|
||||||
functional[k]["Button"].click(predict,
|
functional[k]["Button"].click(predict,
|
||||||
[txt, top_p, temperature, chatbot, history, systemPromptTxt, TRUE, gr.State(k)], [chatbot, history, statusDisplay], show_progress=True)
|
[txt, top_p, temperature, chatbot, history, systemPromptTxt, TRUE, gr.State(k)], [chatbot, history, statusDisplay], show_progress=True)
|
||||||
@ -111,15 +66,15 @@ with gr.Blocks() as demo:
|
|||||||
crazy_functional[k]["Button"].click(crazy_functional[k]["Function"],
|
crazy_functional[k]["Button"].click(crazy_functional[k]["Function"],
|
||||||
[txt, top_p, temperature, chatbot, history, systemPromptTxt, gr.State(PORT)], [chatbot, history, statusDisplay])
|
[txt, top_p, temperature, chatbot, history, systemPromptTxt, gr.State(PORT)], [chatbot, history, statusDisplay])
|
||||||
|
|
||||||
print(f"URL http://localhost:{PORT}")
|
|
||||||
demo.title = "ChatGPT 学术优化"
|
|
||||||
|
|
||||||
def auto_opentab_delay():
|
def auto_opentab_delay():
|
||||||
import threading, webbrowser, time
|
import threading, webbrowser, time
|
||||||
|
print(f"URL http://localhost:{PORT}")
|
||||||
def open(): time.sleep(2)
|
def open(): time.sleep(2)
|
||||||
webbrowser.open_new_tab(f'http://localhost:{PORT}')
|
webbrowser.open_new_tab(f'http://localhost:{PORT}')
|
||||||
t = threading.Thread(target=open)
|
t = threading.Thread(target=open)
|
||||||
t.daemon = True; t.start()
|
t.daemon = True; t.start()
|
||||||
|
|
||||||
auto_opentab_delay()
|
auto_opentab_delay()
|
||||||
|
demo.title = "ChatGPT 学术优化"
|
||||||
demo.queue().launch(server_name="0.0.0.0", share=True, server_port=PORT)
|
demo.queue().launch(server_name="0.0.0.0", share=True, server_port=PORT)
|
||||||
|
68
toolbox.py
Normal file
68
toolbox.py
Normal file
@ -0,0 +1,68 @@
|
|||||||
|
import markdown, mdtex2html
|
||||||
|
from show_math import convert as convert_math
|
||||||
|
from functools import wraps
|
||||||
|
|
||||||
|
|
||||||
|
def regular_txt_to_markdown(text):
|
||||||
|
text = text.replace('\n', '\n\n')
|
||||||
|
text = text.replace('\n\n\n', '\n\n')
|
||||||
|
text = text.replace('\n\n\n', '\n\n')
|
||||||
|
return text
|
||||||
|
|
||||||
|
def CatchException(f):
|
||||||
|
@wraps(f)
|
||||||
|
def decorated(txt, top_p, temperature, chatbot, history, systemPromptTxt, WEB_PORT):
|
||||||
|
try:
|
||||||
|
yield from f(txt, top_p, temperature, chatbot, history, systemPromptTxt, WEB_PORT)
|
||||||
|
except Exception as e:
|
||||||
|
import traceback
|
||||||
|
from check_proxy import check_proxy
|
||||||
|
try: from config_private import proxies
|
||||||
|
except: from config import proxies
|
||||||
|
tb_str = regular_txt_to_markdown(traceback.format_exc())
|
||||||
|
chatbot[-1] = (chatbot[-1][0], f"[Local Message] 实验性函数调用出错: \n\n {tb_str} \n\n 当前代理可用性: \n\n {check_proxy(proxies)}")
|
||||||
|
yield chatbot, history, f'异常 {e}'
|
||||||
|
return decorated
|
||||||
|
|
||||||
|
def report_execption(chatbot, history, a, b):
|
||||||
|
chatbot.append((a, b))
|
||||||
|
history.append(a); history.append(b)
|
||||||
|
|
||||||
|
def text_divide_paragraph(text):
|
||||||
|
if '```' in text:
|
||||||
|
# careful input
|
||||||
|
return text
|
||||||
|
else:
|
||||||
|
# wtf input
|
||||||
|
lines = text.split("\n")
|
||||||
|
for i, line in enumerate(lines):
|
||||||
|
if i!=0: lines[i] = "<p>"+lines[i].replace(" ", " ")+"</p>"
|
||||||
|
text = "".join(lines)
|
||||||
|
return text
|
||||||
|
|
||||||
|
def markdown_convertion(txt):
|
||||||
|
if ('$' in txt) and ('```' not in txt):
|
||||||
|
return markdown.markdown(txt,extensions=['fenced_code','tables']) + '<br><br>' + \
|
||||||
|
markdown.markdown(convert_math(txt, splitParagraphs=False),extensions=['fenced_code','tables'])
|
||||||
|
else:
|
||||||
|
return markdown.markdown(txt,extensions=['fenced_code','tables'])
|
||||||
|
|
||||||
|
|
||||||
|
def format_io(self, y):
|
||||||
|
if y is None: return []
|
||||||
|
i_ask, gpt_reply = y[-1]
|
||||||
|
i_ask = text_divide_paragraph(i_ask) # 输入部分太自由,预处理一波
|
||||||
|
y[-1] = (
|
||||||
|
None if i_ask is None else markdown.markdown(i_ask, extensions=['fenced_code','tables']),
|
||||||
|
None if gpt_reply is None else markdown_convertion(gpt_reply)
|
||||||
|
)
|
||||||
|
return y
|
||||||
|
|
||||||
|
|
||||||
|
def find_free_port():
|
||||||
|
import socket
|
||||||
|
from contextlib import closing
|
||||||
|
with closing(socket.socket(socket.AF_INET, socket.SOCK_STREAM)) as s:
|
||||||
|
s.bind(('', 0))
|
||||||
|
s.setsockopt(socket.SOL_SOCKET, socket.SO_REUSEADDR, 1)
|
||||||
|
return s.getsockname()[1]
|
Loading…
x
Reference in New Issue
Block a user