handle local llm dependency error properly
This commit is contained in:
parent
c17fc2a9b5
commit
184e417fec
@ -58,8 +58,8 @@ class GetONNXGLMHandle(LocalLLMHandle):
|
|||||||
def try_to_import_special_deps(self, **kwargs):
|
def try_to_import_special_deps(self, **kwargs):
|
||||||
# import something that will raise error if the user does not install requirement_*.txt
|
# import something that will raise error if the user does not install requirement_*.txt
|
||||||
# 🏃♂️🏃♂️🏃♂️ 主进程执行
|
# 🏃♂️🏃♂️🏃♂️ 主进程执行
|
||||||
# from modelscope import AutoModelForCausalLM, AutoTokenizer, GenerationConfig
|
import importlib
|
||||||
pass
|
importlib.import_module('modelscope')
|
||||||
|
|
||||||
|
|
||||||
# ------------------------------------------------------------------------------------------------------------------------
|
# ------------------------------------------------------------------------------------------------------------------------
|
||||||
|
@ -124,6 +124,7 @@ def get_local_llm_predict_fns(LLMSingletonClass, model_name):
|
|||||||
"""
|
"""
|
||||||
_llm_handle = LLMSingletonClass()
|
_llm_handle = LLMSingletonClass()
|
||||||
if len(observe_window) >= 1: observe_window[0] = load_message + "\n\n" + _llm_handle.info
|
if len(observe_window) >= 1: observe_window[0] = load_message + "\n\n" + _llm_handle.info
|
||||||
|
if not _llm_handle.running: raise RuntimeError(_llm_handle.info)
|
||||||
|
|
||||||
# chatglm 没有 sys_prompt 接口,因此把prompt加入 history
|
# chatglm 没有 sys_prompt 接口,因此把prompt加入 history
|
||||||
history_feedin = []
|
history_feedin = []
|
||||||
@ -152,6 +153,7 @@ def get_local_llm_predict_fns(LLMSingletonClass, model_name):
|
|||||||
_llm_handle = LLMSingletonClass()
|
_llm_handle = LLMSingletonClass()
|
||||||
chatbot[-1] = (inputs, load_message + "\n\n" + _llm_handle.info)
|
chatbot[-1] = (inputs, load_message + "\n\n" + _llm_handle.info)
|
||||||
yield from update_ui(chatbot=chatbot, history=[])
|
yield from update_ui(chatbot=chatbot, history=[])
|
||||||
|
if not _llm_handle.running: raise RuntimeError(_llm_handle.info)
|
||||||
|
|
||||||
if additional_fn is not None:
|
if additional_fn is not None:
|
||||||
from core_functional import handle_core_functionality
|
from core_functional import handle_core_functionality
|
||||||
|
Loading…
x
Reference in New Issue
Block a user