diff --git a/request_llm/bridge_chatgpt.py b/request_llm/bridge_chatgpt.py index 237db04..c27a4ec 100644 --- a/request_llm/bridge_chatgpt.py +++ b/request_llm/bridge_chatgpt.py @@ -146,7 +146,7 @@ def predict(inputs, top_p, temperature, chatbot=[], history=[], system_prompt='' if additional_fn is not None: import core_functional importlib.reload(core_functional) # 热更新prompt - core_functional = core_functional.get_functions() + core_functional = core_functional.get_core_functions() if "PreProcess" in core_functional[additional_fn]: inputs = core_functional[additional_fn]["PreProcess"](inputs) # 获取预处理函数(如果有的话) inputs = core_functional[additional_fn]["Prefix"] + inputs + core_functional[additional_fn]["Suffix"] diff --git a/request_llm/bridge_tgui.py b/request_llm/bridge_tgui.py index 15f60b7..fceaa56 100644 --- a/request_llm/bridge_tgui.py +++ b/request_llm/bridge_tgui.py @@ -103,7 +103,7 @@ def predict_tgui(inputs, top_p, temperature, chatbot=[], history=[], system_prom if additional_fn is not None: import core_functional importlib.reload(core_functional) # 热更新prompt - core_functional = core_functional.get_functions() + core_functional = core_functional.get_core_functions() if "PreProcess" in core_functional[additional_fn]: inputs = core_functional[additional_fn]["PreProcess"](inputs) # 获取预处理函数(如果有的话) inputs = core_functional[additional_fn]["Prefix"] + inputs + core_functional[additional_fn]["Suffix"]