From 06f8094a0a9f13349cfe05d5cfb2375d412d37b1 Mon Sep 17 00:00:00 2001 From: qingxu fu <505030475@qq.com> Date: Thu, 6 Apr 2023 17:23:26 +0800 Subject: [PATCH] fix error --- request_llm/bridge_chatgpt.py | 2 +- request_llm/bridge_tgui.py | 2 +- 2 files changed, 2 insertions(+), 2 deletions(-) diff --git a/request_llm/bridge_chatgpt.py b/request_llm/bridge_chatgpt.py index 237db04..c27a4ec 100644 --- a/request_llm/bridge_chatgpt.py +++ b/request_llm/bridge_chatgpt.py @@ -146,7 +146,7 @@ def predict(inputs, top_p, temperature, chatbot=[], history=[], system_prompt='' if additional_fn is not None: import core_functional importlib.reload(core_functional) # 热更新prompt - core_functional = core_functional.get_functions() + core_functional = core_functional.get_core_functions() if "PreProcess" in core_functional[additional_fn]: inputs = core_functional[additional_fn]["PreProcess"](inputs) # 获取预处理函数(如果有的话) inputs = core_functional[additional_fn]["Prefix"] + inputs + core_functional[additional_fn]["Suffix"] diff --git a/request_llm/bridge_tgui.py b/request_llm/bridge_tgui.py index 15f60b7..fceaa56 100644 --- a/request_llm/bridge_tgui.py +++ b/request_llm/bridge_tgui.py @@ -103,7 +103,7 @@ def predict_tgui(inputs, top_p, temperature, chatbot=[], history=[], system_prom if additional_fn is not None: import core_functional importlib.reload(core_functional) # 热更新prompt - core_functional = core_functional.get_functions() + core_functional = core_functional.get_core_functions() if "PreProcess" in core_functional[additional_fn]: inputs = core_functional[additional_fn]["PreProcess"](inputs) # 获取预处理函数(如果有的话) inputs = core_functional[additional_fn]["Prefix"] + inputs + core_functional[additional_fn]["Suffix"]