diff --git a/crazy_functions/Latex全文润色.py b/crazy_functions/Latex全文润色.py index 46026fb..da03686 100644 --- a/crazy_functions/Latex全文润色.py +++ b/crazy_functions/Latex全文润色.py @@ -13,7 +13,7 @@ class PaperFileGroup(): # count_token import tiktoken from toolbox import get_conf - enc = tiktoken.encoding_for_model(*get_conf('LLM_MODEL')) + enc = tiktoken.encoding_for_model("gpt-3.5-turbo") def get_token_num(txt): return len(enc.encode(txt, disallowed_special=())) self.get_token_num = get_token_num diff --git a/crazy_functions/Latex全文翻译.py b/crazy_functions/Latex全文翻译.py index 06b4502..de25e62 100644 --- a/crazy_functions/Latex全文翻译.py +++ b/crazy_functions/Latex全文翻译.py @@ -13,7 +13,7 @@ class PaperFileGroup(): # count_token import tiktoken from toolbox import get_conf - enc = tiktoken.encoding_for_model(*get_conf('LLM_MODEL')) + enc = tiktoken.encoding_for_model("gpt-3.5-turbo") def get_token_num(txt): return len(enc.encode(txt, disallowed_special=())) self.get_token_num = get_token_num diff --git a/crazy_functions/crazy_utils.py b/crazy_functions/crazy_utils.py index 5aeb8ca..babad27 100644 --- a/crazy_functions/crazy_utils.py +++ b/crazy_functions/crazy_utils.py @@ -4,7 +4,7 @@ from toolbox import update_ui, get_conf def input_clipping(inputs, history, max_token_limit): import tiktoken import numpy as np - enc = tiktoken.encoding_for_model(*get_conf('LLM_MODEL')) + enc = tiktoken.encoding_for_model("gpt-3.5-turbo") def get_token_num(txt): return len(enc.encode(txt, disallowed_special=())) mode = 'input-and-history' diff --git a/crazy_functions/代码重写为全英文_多线程.py b/crazy_functions/代码重写为全英文_多线程.py index a09a3f5..f8ecec9 100644 --- a/crazy_functions/代码重写为全英文_多线程.py +++ b/crazy_functions/代码重写为全英文_多线程.py @@ -61,7 +61,7 @@ def 全项目切换英文(txt, llm_kwargs, plugin_kwargs, chatbot, history, sys_ MAX_TOKEN = 3000 import tiktoken from toolbox import get_conf - enc = tiktoken.encoding_for_model(*get_conf('LLM_MODEL')) + enc = tiktoken.encoding_for_model("gpt-3.5-turbo") def get_token_fn(txt): return len(enc.encode(txt, disallowed_special=())) diff --git a/crazy_functions/批量Markdown翻译.py b/crazy_functions/批量Markdown翻译.py index 119d3a5..823ca3f 100644 --- a/crazy_functions/批量Markdown翻译.py +++ b/crazy_functions/批量Markdown翻译.py @@ -13,7 +13,7 @@ class PaperFileGroup(): # count_token import tiktoken from toolbox import get_conf - enc = tiktoken.encoding_for_model(*get_conf('LLM_MODEL')) + enc = tiktoken.encoding_for_model("gpt-3.5-turbo") def get_token_num(txt): return len(enc.encode(txt, disallowed_special=())) self.get_token_num = get_token_num diff --git a/crazy_functions/批量翻译PDF文档_多线程.py b/crazy_functions/批量翻译PDF文档_多线程.py index 79c1253..0d2f1d5 100644 --- a/crazy_functions/批量翻译PDF文档_多线程.py +++ b/crazy_functions/批量翻译PDF文档_多线程.py @@ -69,7 +69,7 @@ def 解析PDF(file_manifest, project_folder, llm_kwargs, plugin_kwargs, chatbot, # 递归地切割PDF文件 from .crazy_utils import breakdown_txt_to_satisfy_token_limit_for_pdf from toolbox import get_conf - enc = tiktoken.encoding_for_model(*get_conf('LLM_MODEL')) + enc = tiktoken.encoding_for_model("gpt-3.5-turbo") def get_token_num(txt): return len(enc.encode(txt, disallowed_special=())) paper_fragments = breakdown_txt_to_satisfy_token_limit_for_pdf( txt=file_content, get_token_fn=get_token_num, limit=TOKEN_LIMIT_PER_FRAGMENT) diff --git a/crazy_functions/理解PDF文档内容.py b/crazy_functions/理解PDF文档内容.py index 168b2c9..05fbd49 100644 --- a/crazy_functions/理解PDF文档内容.py +++ b/crazy_functions/理解PDF文档内容.py @@ -18,7 +18,7 @@ def 解析PDF(file_name, llm_kwargs, plugin_kwargs, chatbot, history, system_pro from .crazy_utils import breakdown_txt_to_satisfy_token_limit_for_pdf from toolbox import get_conf - enc = tiktoken.encoding_for_model(*get_conf('LLM_MODEL')) + enc = tiktoken.encoding_for_model("gpt-3.5-turbo") def get_token_num(txt): return len(enc.encode(txt, disallowed_special=())) paper_fragments = breakdown_txt_to_satisfy_token_limit_for_pdf( txt=file_content, get_token_fn=get_token_num, limit=TOKEN_LIMIT_PER_FRAGMENT) diff --git a/request_llm/bridge_chatgpt.py b/request_llm/bridge_chatgpt.py index e9dfc6b..f37abe8 100644 --- a/request_llm/bridge_chatgpt.py +++ b/request_llm/bridge_chatgpt.py @@ -22,8 +22,8 @@ import importlib # config_private.py放自己的秘密如API和代理网址 # 读取时首先看是否存在私密的config_private配置文件(不受git管控),如果有,则覆盖原config文件 from toolbox import get_conf, update_ui -proxies, API_URL, API_KEY, TIMEOUT_SECONDS, MAX_RETRY, LLM_MODEL = \ - get_conf('proxies', 'API_URL', 'API_KEY', 'TIMEOUT_SECONDS', 'MAX_RETRY', 'LLM_MODEL') +proxies, API_URL, API_KEY, TIMEOUT_SECONDS, MAX_RETRY = \ + get_conf('proxies', 'API_URL', 'API_KEY', 'TIMEOUT_SECONDS', 'MAX_RETRY') timeout_bot_msg = '[Local Message] Request timeout. Network error. Please check proxy settings in config.py.' + \ '网络错误,检查代理服务器是否可用,以及代理设置的格式是否正确,格式须是[协议]://[地址]:[端口],缺一不可。'