Merge pull request #746 from Rid7/claude
接入Claude in Slack服务,暂时不支持历史消息设置(单个slack实例,多人使用请谨慎隐私风险)
This commit is contained in:
		
						commit
						e7080e671d
					
				@ -44,9 +44,10 @@ WEB_PORT = -1
 | 
				
			|||||||
# 如果OpenAI不响应(网络卡顿、代理失败、KEY失效),重试的次数限制
 | 
					# 如果OpenAI不响应(网络卡顿、代理失败、KEY失效),重试的次数限制
 | 
				
			||||||
MAX_RETRY = 2
 | 
					MAX_RETRY = 2
 | 
				
			||||||
 | 
					
 | 
				
			||||||
# OpenAI模型选择是(gpt4现在只对申请成功的人开放,体验gpt-4可以试试api2d)
 | 
					# 模型选择是
 | 
				
			||||||
LLM_MODEL = "gpt-3.5-turbo" # 可选 ↓↓↓
 | 
					LLM_MODEL = "gpt-3.5-turbo" # 可选 ↓↓↓
 | 
				
			||||||
AVAIL_LLM_MODELS = ["gpt-3.5-turbo", "api2d-gpt-3.5-turbo", "gpt-4", "api2d-gpt-4", "chatglm", "moss", "newbing"]
 | 
					AVAIL_LLM_MODELS = ["gpt-3.5-turbo", "api2d-gpt-3.5-turbo", "gpt-4", "api2d-gpt-4", "chatglm", "moss", "newbing", "stack-claude"]
 | 
				
			||||||
 | 
					# P.S. 其他可用的模型还包括 ["jittorllms_rwkv", "jittorllms_pangualpha", "jittorllms_llama"]
 | 
				
			||||||
 | 
					
 | 
				
			||||||
# 本地LLM模型如ChatGLM的执行方式 CPU/GPU
 | 
					# 本地LLM模型如ChatGLM的执行方式 CPU/GPU
 | 
				
			||||||
LOCAL_MODEL_DEVICE = "cpu" # 可选 "cuda"
 | 
					LOCAL_MODEL_DEVICE = "cpu" # 可选 "cuda"
 | 
				
			||||||
@ -75,3 +76,7 @@ NEWBING_STYLE = "creative"  # ["creative", "balanced", "precise"]
 | 
				
			|||||||
NEWBING_COOKIES = """
 | 
					NEWBING_COOKIES = """
 | 
				
			||||||
your bing cookies here
 | 
					your bing cookies here
 | 
				
			||||||
"""
 | 
					"""
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					# Slack Claude bot, 使用教程详情见 request_llm/README.md
 | 
				
			||||||
 | 
					SLACK_CLAUDE_BOT_ID = ''   
 | 
				
			||||||
 | 
					SLACK_CLAUDE_USER_TOKEN = ''
 | 
				
			||||||
 | 
				
			|||||||
@ -13,6 +13,31 @@ LLM_MODEL = "chatglm"
 | 
				
			|||||||
`python main.py`
 | 
					`python main.py`
 | 
				
			||||||
``` 
 | 
					``` 
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					## Claude-Stack
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					- 请参考此教程获取  https://zhuanlan.zhihu.com/p/627485689
 | 
				
			||||||
 | 
					    - 1、SLACK_CLAUDE_BOT_ID 
 | 
				
			||||||
 | 
					    - 2、SLACK_CLAUDE_USER_TOKEN
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					- 把token加入config.py
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					## Newbing
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					- 使用cookie editor获取cookie(json)
 | 
				
			||||||
 | 
					- 把cookie(json)加入config.py (NEWBING_COOKIES)
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					## Moss
 | 
				
			||||||
 | 
					- 使用docker-compose
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					## RWKV
 | 
				
			||||||
 | 
					- 使用docker-compose
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					## LLAMA
 | 
				
			||||||
 | 
					- 使用docker-compose
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					## 盘古
 | 
				
			||||||
 | 
					- 使用docker-compose
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					
 | 
				
			||||||
---
 | 
					---
 | 
				
			||||||
## Text-Generation-UI (TGUI,调试中,暂不可用)
 | 
					## Text-Generation-UI (TGUI,调试中,暂不可用)
 | 
				
			||||||
 | 
				
			|||||||
@ -130,6 +130,7 @@ model_info = {
 | 
				
			|||||||
        "tokenizer": tokenizer_gpt35,
 | 
					        "tokenizer": tokenizer_gpt35,
 | 
				
			||||||
        "token_cnt": get_token_num_gpt35,
 | 
					        "token_cnt": get_token_num_gpt35,
 | 
				
			||||||
    },
 | 
					    },
 | 
				
			||||||
 | 
					
 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					
 | 
				
			||||||
@ -186,8 +187,20 @@ if "moss" in AVAIL_LLM_MODELS:
 | 
				
			|||||||
            "token_cnt": get_token_num_gpt35,
 | 
					            "token_cnt": get_token_num_gpt35,
 | 
				
			||||||
        },
 | 
					        },
 | 
				
			||||||
    })
 | 
					    })
 | 
				
			||||||
 | 
					if "stack-claude" in AVAIL_LLM_MODELS:
 | 
				
			||||||
 | 
					    from .bridge_stackclaude import predict_no_ui_long_connection as claude_noui
 | 
				
			||||||
 | 
					    from .bridge_stackclaude import predict as claude_ui
 | 
				
			||||||
 | 
					    # claude
 | 
				
			||||||
 | 
					    model_info.update({
 | 
				
			||||||
 | 
					        "stack-claude": {
 | 
				
			||||||
 | 
					            "fn_with_ui": claude_ui,
 | 
				
			||||||
 | 
					            "fn_without_ui": claude_noui,
 | 
				
			||||||
 | 
					            "endpoint": None,
 | 
				
			||||||
 | 
					            "max_token": 8192,
 | 
				
			||||||
 | 
					            "tokenizer": tokenizer_gpt35,
 | 
				
			||||||
 | 
					            "token_cnt": get_token_num_gpt35,
 | 
				
			||||||
 | 
					        }
 | 
				
			||||||
 | 
					    })
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					
 | 
				
			||||||
def LLM_CATCH_EXCEPTION(f):
 | 
					def LLM_CATCH_EXCEPTION(f):
 | 
				
			||||||
 | 
				
			|||||||
@ -153,7 +153,7 @@ class NewBingHandle(Process):
 | 
				
			|||||||
            # 进入任务等待状态
 | 
					            # 进入任务等待状态
 | 
				
			||||||
            asyncio.run(self.async_run())
 | 
					            asyncio.run(self.async_run())
 | 
				
			||||||
        except Exception:
 | 
					        except Exception:
 | 
				
			||||||
            tb_str = '```\n' + trimmed_format_exc() + '```'
 | 
					            tb_str = '\n```\n' + trimmed_format_exc() + '\n```\n'
 | 
				
			||||||
            self.child.send(f'[Local Message] Newbing失败 {tb_str}.')
 | 
					            self.child.send(f'[Local Message] Newbing失败 {tb_str}.')
 | 
				
			||||||
            self.child.send('[Fail]')
 | 
					            self.child.send('[Fail]')
 | 
				
			||||||
            self.child.send('[Finish]')
 | 
					            self.child.send('[Finish]')
 | 
				
			||||||
 | 
				
			|||||||
							
								
								
									
										296
									
								
								request_llm/bridge_stackclaude.py
									
									
									
									
									
										Normal file
									
								
							
							
						
						
									
										296
									
								
								request_llm/bridge_stackclaude.py
									
									
									
									
									
										Normal file
									
								
							@ -0,0 +1,296 @@
 | 
				
			|||||||
 | 
					from .bridge_newbing import preprocess_newbing_out, preprocess_newbing_out_simple
 | 
				
			||||||
 | 
					from multiprocessing import Process, Pipe
 | 
				
			||||||
 | 
					from toolbox import update_ui, get_conf, trimmed_format_exc
 | 
				
			||||||
 | 
					import threading
 | 
				
			||||||
 | 
					import importlib
 | 
				
			||||||
 | 
					import logging
 | 
				
			||||||
 | 
					import time
 | 
				
			||||||
 | 
					from toolbox import get_conf
 | 
				
			||||||
 | 
					import asyncio
 | 
				
			||||||
 | 
					load_message = "正在加载Claude组件,请稍候..."
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					try:
 | 
				
			||||||
 | 
					    """
 | 
				
			||||||
 | 
					    ========================================================================
 | 
				
			||||||
 | 
					    第一部分:Slack API Client
 | 
				
			||||||
 | 
					    https://github.com/yokonsan/claude-in-slack-api
 | 
				
			||||||
 | 
					    ========================================================================
 | 
				
			||||||
 | 
					    """
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					    from slack_sdk.errors import SlackApiError
 | 
				
			||||||
 | 
					    from slack_sdk.web.async_client import AsyncWebClient
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					    class SlackClient(AsyncWebClient):
 | 
				
			||||||
 | 
					        """SlackClient类用于与Slack API进行交互,实现消息发送、接收等功能。
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					            属性:
 | 
				
			||||||
 | 
					            - CHANNEL_ID:str类型,表示频道ID。
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					            方法:
 | 
				
			||||||
 | 
					            - open_channel():异步方法。通过调用conversations_open方法打开一个频道,并将返回的频道ID保存在属性CHANNEL_ID中。
 | 
				
			||||||
 | 
					            - chat(text: str):异步方法。向已打开的频道发送一条文本消息。
 | 
				
			||||||
 | 
					            - get_slack_messages():异步方法。获取已打开频道的最新消息并返回消息列表,目前不支持历史消息查询。
 | 
				
			||||||
 | 
					            - get_reply():异步方法。循环监听已打开频道的消息,如果收到"Typing…_"结尾的消息说明Claude还在继续输出,否则结束循环。
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					        """
 | 
				
			||||||
 | 
					        CHANNEL_ID = None
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					        async def open_channel(self):
 | 
				
			||||||
 | 
					            response = await self.conversations_open(users=get_conf('SLACK_CLAUDE_BOT_ID')[0])
 | 
				
			||||||
 | 
					            self.CHANNEL_ID = response["channel"]["id"]
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					        async def chat(self, text):
 | 
				
			||||||
 | 
					            if not self.CHANNEL_ID:
 | 
				
			||||||
 | 
					                raise Exception("Channel not found.")
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					            resp = await self.chat_postMessage(channel=self.CHANNEL_ID, text=text)
 | 
				
			||||||
 | 
					            self.LAST_TS = resp["ts"]
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					        async def get_slack_messages(self):
 | 
				
			||||||
 | 
					            try:
 | 
				
			||||||
 | 
					                # TODO:暂时不支持历史消息,因为在同一个频道里存在多人使用时历史消息渗透问题
 | 
				
			||||||
 | 
					                resp = await self.conversations_history(channel=self.CHANNEL_ID, oldest=self.LAST_TS, limit=1)
 | 
				
			||||||
 | 
					                msg = [msg for msg in resp["messages"]
 | 
				
			||||||
 | 
					                    if msg.get("user") == get_conf('SLACK_CLAUDE_BOT_ID')[0]]
 | 
				
			||||||
 | 
					                return msg
 | 
				
			||||||
 | 
					            except (SlackApiError, KeyError) as e:
 | 
				
			||||||
 | 
					                raise RuntimeError(f"获取Slack消息失败。")
 | 
				
			||||||
 | 
					        
 | 
				
			||||||
 | 
					        async def get_reply(self):
 | 
				
			||||||
 | 
					            while True:
 | 
				
			||||||
 | 
					                slack_msgs = await self.get_slack_messages()
 | 
				
			||||||
 | 
					                if len(slack_msgs) == 0:
 | 
				
			||||||
 | 
					                    await asyncio.sleep(0.5)
 | 
				
			||||||
 | 
					                    continue
 | 
				
			||||||
 | 
					                
 | 
				
			||||||
 | 
					                msg = slack_msgs[-1]
 | 
				
			||||||
 | 
					                if msg["text"].endswith("Typing…_"):
 | 
				
			||||||
 | 
					                    yield False, msg["text"]
 | 
				
			||||||
 | 
					                else:
 | 
				
			||||||
 | 
					                    yield True, msg["text"]
 | 
				
			||||||
 | 
					                    break
 | 
				
			||||||
 | 
					except:
 | 
				
			||||||
 | 
					    pass
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					"""
 | 
				
			||||||
 | 
					========================================================================
 | 
				
			||||||
 | 
					第二部分:子进程Worker(调用主体)
 | 
				
			||||||
 | 
					========================================================================
 | 
				
			||||||
 | 
					"""
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					class ClaudeHandle(Process):
 | 
				
			||||||
 | 
					    def __init__(self):
 | 
				
			||||||
 | 
					        super().__init__(daemon=True)
 | 
				
			||||||
 | 
					        self.parent, self.child = Pipe()
 | 
				
			||||||
 | 
					        self.claude_model = None
 | 
				
			||||||
 | 
					        self.info = ""
 | 
				
			||||||
 | 
					        self.success = True
 | 
				
			||||||
 | 
					        self.local_history = []
 | 
				
			||||||
 | 
					        self.check_dependency()
 | 
				
			||||||
 | 
					        if self.success: 
 | 
				
			||||||
 | 
					            self.start()
 | 
				
			||||||
 | 
					            self.threadLock = threading.Lock()
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					    def check_dependency(self):
 | 
				
			||||||
 | 
					        try:
 | 
				
			||||||
 | 
					            self.success = False
 | 
				
			||||||
 | 
					            import slack_sdk
 | 
				
			||||||
 | 
					            self.info = "依赖检测通过,等待Claude响应。注意目前不能多人同时调用Claude接口(有线程锁),否则将导致每个人的Claude问询历史互相渗透。调用Claude时,会自动使用已配置的代理。"
 | 
				
			||||||
 | 
					            self.success = True
 | 
				
			||||||
 | 
					        except:
 | 
				
			||||||
 | 
					            self.info = "缺少的依赖,如果要使用Claude,除了基础的pip依赖以外,您还需要运行`pip install -r request_llm/requirements_slackclaude.txt`安装Claude的依赖,然后重启程序。"
 | 
				
			||||||
 | 
					            self.success = False
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					    def ready(self):
 | 
				
			||||||
 | 
					        return self.claude_model is not None    
 | 
				
			||||||
 | 
					    
 | 
				
			||||||
 | 
					    async def async_run(self):
 | 
				
			||||||
 | 
					        await self.claude_model.open_channel()
 | 
				
			||||||
 | 
					        while True:
 | 
				
			||||||
 | 
					            # 等待
 | 
				
			||||||
 | 
					            kwargs = self.child.recv()
 | 
				
			||||||
 | 
					            question = kwargs['query']
 | 
				
			||||||
 | 
					            history = kwargs['history']
 | 
				
			||||||
 | 
					            # system_prompt=kwargs['system_prompt']
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					            # 是否重置
 | 
				
			||||||
 | 
					            if len(self.local_history) > 0 and len(history) == 0:
 | 
				
			||||||
 | 
					                # await self.claude_model.reset()
 | 
				
			||||||
 | 
					                self.local_history = []
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					            # 开始问问题
 | 
				
			||||||
 | 
					            prompt = ""
 | 
				
			||||||
 | 
					            # Slack API最好不要添加系统提示
 | 
				
			||||||
 | 
					            # if system_prompt not in self.local_history:
 | 
				
			||||||
 | 
					            #     self.local_history.append(system_prompt)
 | 
				
			||||||
 | 
					            #     prompt += system_prompt + '\n'
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					            # 追加历史
 | 
				
			||||||
 | 
					            for ab in history:
 | 
				
			||||||
 | 
					                a, b = ab
 | 
				
			||||||
 | 
					                if a not in self.local_history:
 | 
				
			||||||
 | 
					                    self.local_history.append(a)
 | 
				
			||||||
 | 
					                    prompt += a + '\n'
 | 
				
			||||||
 | 
					                # if b not in self.local_history:
 | 
				
			||||||
 | 
					                #     self.local_history.append(b)
 | 
				
			||||||
 | 
					                #     prompt += b + '\n'
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					            # 问题
 | 
				
			||||||
 | 
					            prompt += question
 | 
				
			||||||
 | 
					            self.local_history.append(question)
 | 
				
			||||||
 | 
					            print('question:', prompt)
 | 
				
			||||||
 | 
					            # 提交
 | 
				
			||||||
 | 
					            await self.claude_model.chat(prompt)
 | 
				
			||||||
 | 
					            # 获取回复
 | 
				
			||||||
 | 
					            # async for final, response in self.claude_model.get_reply():
 | 
				
			||||||
 | 
					            #     await self.handle_claude_response(final, response)
 | 
				
			||||||
 | 
					            async for final, response in self.claude_model.get_reply():                
 | 
				
			||||||
 | 
					                if not final:
 | 
				
			||||||
 | 
					                    print(response)
 | 
				
			||||||
 | 
					                    self.child.send(str(response))
 | 
				
			||||||
 | 
					                else:
 | 
				
			||||||
 | 
					                    # 防止丢失最后一条消息
 | 
				
			||||||
 | 
					                    slack_msgs = await self.claude_model.get_slack_messages()
 | 
				
			||||||
 | 
					                    last_msg = slack_msgs[-1]["text"] if slack_msgs and len(slack_msgs) > 0 else ""
 | 
				
			||||||
 | 
					                    if last_msg:
 | 
				
			||||||
 | 
					                        self.child.send(last_msg)
 | 
				
			||||||
 | 
					                    print('-------- receive final ---------')
 | 
				
			||||||
 | 
					                    self.child.send('[Finish]')
 | 
				
			||||||
 | 
					                    
 | 
				
			||||||
 | 
					    def run(self):
 | 
				
			||||||
 | 
					        """
 | 
				
			||||||
 | 
					        这个函数运行在子进程
 | 
				
			||||||
 | 
					        """
 | 
				
			||||||
 | 
					        # 第一次运行,加载参数
 | 
				
			||||||
 | 
					        self.success = False
 | 
				
			||||||
 | 
					        self.local_history = []
 | 
				
			||||||
 | 
					        if (self.claude_model is None) or (not self.success):
 | 
				
			||||||
 | 
					            # 代理设置
 | 
				
			||||||
 | 
					            proxies, = get_conf('proxies')
 | 
				
			||||||
 | 
					            if proxies is None:
 | 
				
			||||||
 | 
					                self.proxies_https = None
 | 
				
			||||||
 | 
					            else:
 | 
				
			||||||
 | 
					                self.proxies_https = proxies['https']
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					            try:
 | 
				
			||||||
 | 
					                SLACK_CLAUDE_USER_TOKEN, = get_conf('SLACK_CLAUDE_USER_TOKEN')
 | 
				
			||||||
 | 
					                self.claude_model = SlackClient(token=SLACK_CLAUDE_USER_TOKEN, proxy=self.proxies_https)
 | 
				
			||||||
 | 
					                print('Claude组件初始化成功。')
 | 
				
			||||||
 | 
					            except:
 | 
				
			||||||
 | 
					                self.success = False
 | 
				
			||||||
 | 
					                tb_str = '\n```\n' + trimmed_format_exc() + '\n```\n'
 | 
				
			||||||
 | 
					                self.child.send(f'[Local Message] 不能加载Claude组件。{tb_str}')
 | 
				
			||||||
 | 
					                self.child.send('[Fail]')
 | 
				
			||||||
 | 
					                self.child.send('[Finish]')
 | 
				
			||||||
 | 
					                raise RuntimeError(f"不能加载Claude组件。")
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					        self.success = True
 | 
				
			||||||
 | 
					        try:
 | 
				
			||||||
 | 
					            # 进入任务等待状态
 | 
				
			||||||
 | 
					            asyncio.run(self.async_run())
 | 
				
			||||||
 | 
					        except Exception:
 | 
				
			||||||
 | 
					            tb_str = '\n```\n' + trimmed_format_exc() + '\n```\n'
 | 
				
			||||||
 | 
					            self.child.send(f'[Local Message] Claude失败 {tb_str}.')
 | 
				
			||||||
 | 
					            self.child.send('[Fail]')
 | 
				
			||||||
 | 
					            self.child.send('[Finish]')
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					    def stream_chat(self, **kwargs):
 | 
				
			||||||
 | 
					        """
 | 
				
			||||||
 | 
					        这个函数运行在主进程
 | 
				
			||||||
 | 
					        """
 | 
				
			||||||
 | 
					        self.threadLock.acquire()
 | 
				
			||||||
 | 
					        self.parent.send(kwargs)    # 发送请求到子进程
 | 
				
			||||||
 | 
					        while True:
 | 
				
			||||||
 | 
					            res = self.parent.recv()    # 等待Claude回复的片段
 | 
				
			||||||
 | 
					            if res == '[Finish]':
 | 
				
			||||||
 | 
					                break       # 结束
 | 
				
			||||||
 | 
					            elif res == '[Fail]':
 | 
				
			||||||
 | 
					                self.success = False
 | 
				
			||||||
 | 
					                break
 | 
				
			||||||
 | 
					            else:
 | 
				
			||||||
 | 
					                yield res   # Claude回复的片段
 | 
				
			||||||
 | 
					        self.threadLock.release()
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					"""
 | 
				
			||||||
 | 
					========================================================================
 | 
				
			||||||
 | 
					第三部分:主进程统一调用函数接口
 | 
				
			||||||
 | 
					========================================================================
 | 
				
			||||||
 | 
					"""
 | 
				
			||||||
 | 
					global claude_handle
 | 
				
			||||||
 | 
					claude_handle = None
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					def predict_no_ui_long_connection(inputs, llm_kwargs, history=[], sys_prompt="", observe_window=None, console_slience=False):
 | 
				
			||||||
 | 
					    """
 | 
				
			||||||
 | 
					        多线程方法
 | 
				
			||||||
 | 
					        函数的说明请见 request_llm/bridge_all.py
 | 
				
			||||||
 | 
					    """
 | 
				
			||||||
 | 
					    global claude_handle
 | 
				
			||||||
 | 
					    if (claude_handle is None) or (not claude_handle.success):
 | 
				
			||||||
 | 
					        claude_handle = ClaudeHandle()
 | 
				
			||||||
 | 
					        observe_window[0] = load_message + "\n\n" + claude_handle.info
 | 
				
			||||||
 | 
					        if not claude_handle.success:
 | 
				
			||||||
 | 
					            error = claude_handle.info
 | 
				
			||||||
 | 
					            claude_handle = None
 | 
				
			||||||
 | 
					            raise RuntimeError(error)
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					    # 没有 sys_prompt 接口,因此把prompt加入 history
 | 
				
			||||||
 | 
					    history_feedin = []
 | 
				
			||||||
 | 
					    for i in range(len(history)//2):
 | 
				
			||||||
 | 
					        history_feedin.append([history[2*i], history[2*i+1]])
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					    watch_dog_patience = 5  # 看门狗 (watchdog) 的耐心, 设置5秒即可
 | 
				
			||||||
 | 
					    response = ""
 | 
				
			||||||
 | 
					    observe_window[0] = "[Local Message]: 等待Claude响应中 ..."
 | 
				
			||||||
 | 
					    for response in claude_handle.stream_chat(query=inputs, history=history_feedin, system_prompt=sys_prompt, max_length=llm_kwargs['max_length'], top_p=llm_kwargs['top_p'], temperature=llm_kwargs['temperature']):
 | 
				
			||||||
 | 
					        observe_window[0] = preprocess_newbing_out_simple(response)
 | 
				
			||||||
 | 
					        if len(observe_window) >= 2:
 | 
				
			||||||
 | 
					            if (time.time()-observe_window[1]) > watch_dog_patience:
 | 
				
			||||||
 | 
					                raise RuntimeError("程序终止。")
 | 
				
			||||||
 | 
					    return preprocess_newbing_out_simple(response)
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					def predict(inputs, llm_kwargs, plugin_kwargs, chatbot, history=[], system_prompt='', stream=True, additional_fn=None):
 | 
				
			||||||
 | 
					    """
 | 
				
			||||||
 | 
					        单线程方法
 | 
				
			||||||
 | 
					        函数的说明请见 request_llm/bridge_all.py
 | 
				
			||||||
 | 
					    """
 | 
				
			||||||
 | 
					    chatbot.append((inputs, "[Local Message]: 等待Claude响应中 ..."))
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					    global claude_handle
 | 
				
			||||||
 | 
					    if (claude_handle is None) or (not claude_handle.success):
 | 
				
			||||||
 | 
					        claude_handle = ClaudeHandle()
 | 
				
			||||||
 | 
					        chatbot[-1] = (inputs, load_message + "\n\n" + claude_handle.info)
 | 
				
			||||||
 | 
					        yield from update_ui(chatbot=chatbot, history=[])
 | 
				
			||||||
 | 
					        if not claude_handle.success:
 | 
				
			||||||
 | 
					            claude_handle = None
 | 
				
			||||||
 | 
					            return
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					    if additional_fn is not None:
 | 
				
			||||||
 | 
					        import core_functional
 | 
				
			||||||
 | 
					        importlib.reload(core_functional)    # 热更新prompt
 | 
				
			||||||
 | 
					        core_functional = core_functional.get_core_functions()
 | 
				
			||||||
 | 
					        if "PreProcess" in core_functional[additional_fn]:
 | 
				
			||||||
 | 
					            inputs = core_functional[additional_fn]["PreProcess"](
 | 
				
			||||||
 | 
					                inputs)  # 获取预处理函数(如果有的话)
 | 
				
			||||||
 | 
					        inputs = core_functional[additional_fn]["Prefix"] + \
 | 
				
			||||||
 | 
					            inputs + core_functional[additional_fn]["Suffix"]
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					    history_feedin = []
 | 
				
			||||||
 | 
					    for i in range(len(history)//2):
 | 
				
			||||||
 | 
					        history_feedin.append([history[2*i], history[2*i+1]])
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					    chatbot[-1] = (inputs, "[Local Message]: 等待Claude响应中 ...")
 | 
				
			||||||
 | 
					    response = "[Local Message]: 等待Claude响应中 ..."
 | 
				
			||||||
 | 
					    yield from update_ui(chatbot=chatbot, history=history, msg="Claude响应缓慢,尚未完成全部响应,请耐心完成后再提交新问题。")
 | 
				
			||||||
 | 
					    for response in claude_handle.stream_chat(query=inputs, history=history_feedin, system_prompt=system_prompt):
 | 
				
			||||||
 | 
					        chatbot[-1] = (inputs, preprocess_newbing_out(response))
 | 
				
			||||||
 | 
					        yield from update_ui(chatbot=chatbot, history=history, msg="Claude响应缓慢,尚未完成全部响应,请耐心完成后再提交新问题。")
 | 
				
			||||||
 | 
					    if response == "[Local Message]: 等待Claude响应中 ...":
 | 
				
			||||||
 | 
					        response = "[Local Message]: Claude响应异常,请刷新界面重试 ..."
 | 
				
			||||||
 | 
					    history.extend([inputs, response])
 | 
				
			||||||
 | 
					    logging.info(f'[raw_input] {inputs}')
 | 
				
			||||||
 | 
					    logging.info(f'[response] {response}')
 | 
				
			||||||
 | 
					    yield from update_ui(chatbot=chatbot, history=history, msg="完成全部响应,请提交新问题。")
 | 
				
			||||||
							
								
								
									
										1
									
								
								request_llm/requirements_slackclaude.txt
									
									
									
									
									
										Normal file
									
								
							
							
						
						
									
										1
									
								
								request_llm/requirements_slackclaude.txt
									
									
									
									
									
										Normal file
									
								
							@ -0,0 +1 @@
 | 
				
			|||||||
 | 
					slack-sdk==3.21.3
 | 
				
			||||||
		Loading…
	
	
			
			x
			
			
		
	
		Reference in New Issue
	
	Block a user