diff --git a/.gitignore b/.gitignore
index 4abd059..dfd7895 100644
--- a/.gitignore
+++ b/.gitignore
@@ -147,4 +147,5 @@ private*
crazy_functions/test_project/pdf_and_word
crazy_functions/test_samples
request_llm/jittorllms
-multi-language
\ No newline at end of file
+multi-language
+request_llm/moss
diff --git a/README.md b/README.md
index b2cddba..3e16f0b 100644
--- a/README.md
+++ b/README.md
@@ -267,6 +267,12 @@ Tip:不指定文件直接点击 `载入对话历史存档` 可以查看历史h
+9. OpenAI音频解析与总结
+
+

+
+
+
## 版本:
- version 3.5(Todo): 使用自然语言调用本项目的所有函数插件(高优先级)
diff --git a/config.py b/config.py
index 37d5b66..6dbee96 100644
--- a/config.py
+++ b/config.py
@@ -44,9 +44,10 @@ WEB_PORT = -1
# 如果OpenAI不响应(网络卡顿、代理失败、KEY失效),重试的次数限制
MAX_RETRY = 2
-# OpenAI模型选择是(gpt4现在只对申请成功的人开放,体验gpt-4可以试试api2d)
+# 模型选择是
LLM_MODEL = "gpt-3.5-turbo" # 可选 ↓↓↓
-AVAIL_LLM_MODELS = ["gpt-3.5-turbo", "api2d-gpt-3.5-turbo", "gpt-4", "api2d-gpt-4", "chatglm", "moss", "newbing"]
+AVAIL_LLM_MODELS = ["gpt-3.5-turbo", "api2d-gpt-3.5-turbo", "gpt-4", "api2d-gpt-4", "chatglm", "moss", "newbing", "stack-claude"]
+# P.S. 其他可用的模型还包括 ["jittorllms_rwkv", "jittorllms_pangualpha", "jittorllms_llama"]
# 本地LLM模型如ChatGLM的执行方式 CPU/GPU
LOCAL_MODEL_DEVICE = "cpu" # 可选 "cuda"
@@ -75,3 +76,7 @@ NEWBING_STYLE = "creative" # ["creative", "balanced", "precise"]
NEWBING_COOKIES = """
your bing cookies here
"""
+
+# 如果需要使用Slack Claude,使用教程详情见 request_llm/README.md
+SLACK_CLAUDE_BOT_ID = ''
+SLACK_CLAUDE_USER_TOKEN = ''
diff --git a/core_functional.py b/core_functional.py
index 536ccb6..e126b57 100644
--- a/core_functional.py
+++ b/core_functional.py
@@ -68,4 +68,11 @@ def get_core_functions():
"Prefix": r"请解释以下代码:" + "\n```\n",
"Suffix": "\n```\n",
},
+ "参考文献转Bib": {
+ "Prefix": r"Here are some bibliography items, please transform them into bibtex style." +
+ r"Note that, reference styles maybe more than one kind, you should transform each item correctly." +
+ r"Items need to be transformed:",
+ "Suffix": r"",
+ "Visible": False,
+ }
}
diff --git a/crazy_functional.py b/crazy_functional.py
index 3e7b12f..462000e 100644
--- a/crazy_functional.py
+++ b/crazy_functional.py
@@ -246,5 +246,15 @@ def get_crazy_functions():
"Function": HotReload(图片生成)
},
})
+ from crazy_functions.总结音视频 import 总结音视频
+ function_plugins.update({
+ "批量总结音视频(输入路径或上传压缩包)": {
+ "Color": "stop",
+ "AsButton": False,
+ "AdvancedArgs": True,
+ "ArgsReminder": "调用openai api 使用whisper-1模型, 目前支持的格式:mp4, m4a, wav, mpga, mpeg, mp3。此处可以输入解析提示,例如:解析为简体中文(默认)。",
+ "Function": HotReload(总结音视频)
+ }
+ })
###################### 第n组插件 ###########################
return function_plugins
diff --git a/crazy_functions/图片生成.py b/crazy_functions/图片生成.py
index ecb75cd..5bf8bc4 100644
--- a/crazy_functions/图片生成.py
+++ b/crazy_functions/图片生成.py
@@ -55,6 +55,7 @@ def 图片生成(prompt, llm_kwargs, plugin_kwargs, chatbot, history, system_pro
history = [] # 清空历史,以免输入溢出
chatbot.append(("这是什么功能?", "[Local Message] 生成图像, 请先把模型切换至gpt-xxxx或者api2d-xxxx。如果中文效果不理想, 尝试Prompt。正在处理中 ....."))
yield from update_ui(chatbot=chatbot, history=history) # 刷新界面 # 由于请求gpt需要一段时间,我们先及时地做一次界面更新
+ if ("advanced_arg" in plugin_kwargs) and (plugin_kwargs["advanced_arg"] == ""): plugin_kwargs.pop("advanced_arg")
resolution = plugin_kwargs.get("advanced_arg", '256x256')
image_url, image_path = gen_image(llm_kwargs, prompt, resolution)
chatbot.append([prompt,
diff --git a/crazy_functions/总结音视频.py b/crazy_functions/总结音视频.py
new file mode 100644
index 0000000..62f05d3
--- /dev/null
+++ b/crazy_functions/总结音视频.py
@@ -0,0 +1,184 @@
+from toolbox import CatchException, report_execption, select_api_key, update_ui, write_results_to_file, get_conf
+from .crazy_utils import request_gpt_model_in_new_thread_with_ui_alive
+
+def split_audio_file(filename, split_duration=1000):
+ """
+ 根据给定的切割时长将音频文件切割成多个片段。
+
+ Args:
+ filename (str): 需要被切割的音频文件名。
+ split_duration (int, optional): 每个切割音频片段的时长(以秒为单位)。默认值为1000。
+
+ Returns:
+ filelist (list): 一个包含所有切割音频片段文件路径的列表。
+
+ """
+ from moviepy.editor import AudioFileClip
+ import os
+ os.makedirs('gpt_log/mp3/cut/', exist_ok=True) # 创建存储切割音频的文件夹
+
+ # 读取音频文件
+ audio = AudioFileClip(filename)
+
+ # 计算文件总时长和切割点
+ total_duration = audio.duration
+ split_points = list(range(0, int(total_duration), split_duration))
+ split_points.append(int(total_duration))
+ filelist = []
+
+ # 切割音频文件
+ for i in range(len(split_points) - 1):
+ start_time = split_points[i]
+ end_time = split_points[i + 1]
+ split_audio = audio.subclip(start_time, end_time)
+ split_audio.write_audiofile(f"gpt_log/mp3/cut/{filename[0]}_{i}.mp3")
+ filelist.append(f"gpt_log/mp3/cut/{filename[0]}_{i}.mp3")
+
+ audio.close()
+ return filelist
+
+def AnalyAudio(parse_prompt, file_manifest, llm_kwargs, chatbot, history):
+ import os, requests
+ from moviepy.editor import AudioFileClip
+ from request_llm.bridge_all import model_info
+
+ # 设置OpenAI密钥和模型
+ api_key = select_api_key(llm_kwargs['api_key'], llm_kwargs['llm_model'])
+ chat_endpoint = model_info[llm_kwargs['llm_model']]['endpoint']
+
+ whisper_endpoint = chat_endpoint.replace('chat/completions', 'audio/transcriptions')
+ url = whisper_endpoint
+ headers = {
+ 'Authorization': f"Bearer {api_key}"
+ }
+
+ os.makedirs('gpt_log/mp3/', exist_ok=True)
+ for index, fp in enumerate(file_manifest):
+ audio_history = []
+ # 提取文件扩展名
+ ext = os.path.splitext(fp)[1]
+ # 提取视频中的音频
+ if ext not in [".mp3", ".wav", ".m4a", ".mpga"]:
+ audio_clip = AudioFileClip(fp)
+ audio_clip.write_audiofile(f'gpt_log/mp3/output{index}.mp3')
+ fp = f'gpt_log/mp3/output{index}.mp3'
+ # 调用whisper模型音频转文字
+ voice = split_audio_file(fp)
+ for j, i in enumerate(voice):
+ with open(i, 'rb') as f:
+ file_content = f.read() # 读取文件内容到内存
+ files = {
+ 'file': (os.path.basename(i), file_content),
+ }
+ data = {
+ "model": "whisper-1",
+ "prompt": parse_prompt,
+ 'response_format': "text"
+ }
+
+ chatbot.append([f"将 {i} 发送到openai音频解析终端 (whisper),当前参数:{parse_prompt}", "正在处理 ..."])
+ yield from update_ui(chatbot=chatbot, history=history) # 刷新界面
+ proxies, = get_conf('proxies')
+ response = requests.post(url, headers=headers, files=files, data=data, proxies=proxies).text
+
+ chatbot.append(["音频解析结果", response])
+ history.extend(["音频解析结果", response])
+ yield from update_ui(chatbot=chatbot, history=history) # 刷新界面
+
+ i_say = f'请对下面的音频片段做概述,音频内容是 ```{response}```'
+ i_say_show_user = f'第{index + 1}段音频的第{j + 1} / {len(voice)}片段。'
+ gpt_say = yield from request_gpt_model_in_new_thread_with_ui_alive(
+ inputs=i_say,
+ inputs_show_user=i_say_show_user,
+ llm_kwargs=llm_kwargs,
+ chatbot=chatbot,
+ history=[],
+ sys_prompt=f"总结音频。音频文件名{fp}"
+ )
+
+ chatbot[-1] = (i_say_show_user, gpt_say)
+ history.extend([i_say_show_user, gpt_say])
+ audio_history.extend([i_say_show_user, gpt_say])
+
+ # 已经对该文章的所有片段总结完毕,如果文章被切分了
+ result = "".join(audio_history)
+ if len(audio_history) > 1:
+ i_say = f"根据以上的对话,使用中文总结音频“{result}”的主要内容。"
+ i_say_show_user = f'第{index + 1}段音频的主要内容:'
+ gpt_say = yield from request_gpt_model_in_new_thread_with_ui_alive(
+ inputs=i_say,
+ inputs_show_user=i_say_show_user,
+ llm_kwargs=llm_kwargs,
+ chatbot=chatbot,
+ history=audio_history,
+ sys_prompt="总结文章。"
+ )
+
+ history.extend([i_say, gpt_say])
+ audio_history.extend([i_say, gpt_say])
+
+ res = write_results_to_file(history)
+ chatbot.append((f"第{index + 1}段音频完成了吗?", res))
+ yield from update_ui(chatbot=chatbot, history=history) # 刷新界面
+
+ # 删除中间文件夹
+ import shutil
+ shutil.rmtree('gpt_log/mp3')
+ res = write_results_to_file(history)
+ chatbot.append(("所有音频都总结完成了吗?", res))
+ yield from update_ui(chatbot=chatbot, history=history)
+
+
+@CatchException
+def 总结音视频(txt, llm_kwargs, plugin_kwargs, chatbot, history, system_prompt, WEB_PORT):
+ import glob, os
+
+ # 基本信息:功能、贡献者
+ chatbot.append([
+ "函数插件功能?",
+ "总结音视频内容,函数插件贡献者: dalvqw & BinaryHusky"])
+ yield from update_ui(chatbot=chatbot, history=history) # 刷新界面
+
+ try:
+ from moviepy.editor import AudioFileClip
+ except:
+ report_execption(chatbot, history,
+ a=f"解析项目: {txt}",
+ b=f"导入软件依赖失败。使用该模块需要额外依赖,安装方法```pip install --upgrade moviepy```。")
+ yield from update_ui(chatbot=chatbot, history=history) # 刷新界面
+ return
+
+ # 清空历史,以免输入溢出
+ history = []
+
+ # 检测输入参数,如没有给定输入参数,直接退出
+ if os.path.exists(txt):
+ project_folder = txt
+ else:
+ if txt == "": txt = '空空如也的输入栏'
+ report_execption(chatbot, history, a=f"解析项目: {txt}", b=f"找不到本地项目或无权访问: {txt}")
+ yield from update_ui(chatbot=chatbot, history=history) # 刷新界面
+ return
+
+ # 搜索需要处理的文件清单
+ extensions = ['.mp4', '.m4a', '.wav', '.mpga', '.mpeg', '.mp3', '.avi', '.mkv', '.flac', '.aac']
+
+ if txt.endswith(tuple(extensions)):
+ file_manifest = [txt]
+ else:
+ file_manifest = []
+ for extension in extensions:
+ file_manifest.extend(glob.glob(f'{project_folder}/**/*{extension}', recursive=True))
+
+ # 如果没找到任何文件
+ if len(file_manifest) == 0:
+ report_execption(chatbot, history, a=f"解析项目: {txt}", b=f"找不到任何音频或视频文件: {txt}")
+ yield from update_ui(chatbot=chatbot, history=history) # 刷新界面
+ return
+
+ # 开始正式执行任务
+ if ("advanced_arg" in plugin_kwargs) and (plugin_kwargs["advanced_arg"] == ""): plugin_kwargs.pop("advanced_arg")
+ parse_prompt = plugin_kwargs.get("advanced_arg", '将音频解析为简体中文')
+ yield from AnalyAudio(parse_prompt, file_manifest, llm_kwargs, chatbot, history)
+
+ yield from update_ui(chatbot=chatbot, history=history) # 刷新界面
diff --git a/crazy_functions/解析JupyterNotebook.py b/crazy_functions/解析JupyterNotebook.py
index 95a3d69..b4bcd56 100644
--- a/crazy_functions/解析JupyterNotebook.py
+++ b/crazy_functions/解析JupyterNotebook.py
@@ -67,6 +67,7 @@ def parseNotebook(filename, enable_markdown=1):
def ipynb解释(file_manifest, project_folder, llm_kwargs, plugin_kwargs, chatbot, history, system_prompt):
from .crazy_utils import request_gpt_model_multi_threads_with_very_awesome_ui_and_high_efficiency
+ if ("advanced_arg" in plugin_kwargs) and (plugin_kwargs["advanced_arg"] == ""): plugin_kwargs.pop("advanced_arg")
enable_markdown = plugin_kwargs.get("advanced_arg", "1")
try:
enable_markdown = int(enable_markdown)
diff --git a/crazy_functions/询问多个大语言模型.py b/crazy_functions/询问多个大语言模型.py
index 2939d04..ec9fd4a 100644
--- a/crazy_functions/询问多个大语言模型.py
+++ b/crazy_functions/询问多个大语言模型.py
@@ -45,6 +45,7 @@ def 同时问询_指定模型(txt, llm_kwargs, plugin_kwargs, chatbot, history,
chatbot.append((txt, "正在同时咨询ChatGPT和ChatGLM……"))
yield from update_ui(chatbot=chatbot, history=history) # 刷新界面 # 由于请求gpt需要一段时间,我们先及时地做一次界面更新
+ if ("advanced_arg" in plugin_kwargs) and (plugin_kwargs["advanced_arg"] == ""): plugin_kwargs.pop("advanced_arg")
# llm_kwargs['llm_model'] = 'chatglm&gpt-3.5-turbo&api2d-gpt-3.5-turbo' # 支持任意数量的llm接口,用&符号分隔
llm_kwargs['llm_model'] = plugin_kwargs.get("advanced_arg", 'chatglm&gpt-3.5-turbo') # 'chatglm&gpt-3.5-turbo' # 支持任意数量的llm接口,用&符号分隔
gpt_say = yield from request_gpt_model_in_new_thread_with_ui_alive(
diff --git a/main.py b/main.py
index 4de8015..d9888f8 100644
--- a/main.py
+++ b/main.py
@@ -74,6 +74,7 @@ def main():
with gr.Accordion("基础功能区", open=True) as area_basic_fn:
with gr.Row():
for k in functional:
+ if ("Visible" in functional[k]) and (not functional[k]["Visible"]): continue
variant = functional[k]["Color"] if "Color" in functional[k] else "secondary"
functional[k]["Button"] = gr.Button(k, variant=variant)
with gr.Accordion("函数插件区", open=True) as area_crazy_fn:
@@ -144,6 +145,7 @@ def main():
clearBtn2.click(lambda: ("",""), None, [txt, txt2])
# 基础功能区的回调函数注册
for k in functional:
+ if ("Visible" in functional[k]) and (not functional[k]["Visible"]): continue
click_handle = functional[k]["Button"].click(fn=ArgsGeneralWrapper(predict), inputs=[*input_combo, gr.State(True), gr.State(k)], outputs=output_combo)
cancel_handles.append(click_handle)
# 文件上传区,接收文件后与chatbot的互动
diff --git a/request_llm/README.md b/request_llm/README.md
index 4a912d1..545bc1f 100644
--- a/request_llm/README.md
+++ b/request_llm/README.md
@@ -13,6 +13,31 @@ LLM_MODEL = "chatglm"
`python main.py`
```
+## Claude-Stack
+
+- 请参考此教程获取 https://zhuanlan.zhihu.com/p/627485689
+ - 1、SLACK_CLAUDE_BOT_ID
+ - 2、SLACK_CLAUDE_USER_TOKEN
+
+- 把token加入config.py
+
+## Newbing
+
+- 使用cookie editor获取cookie(json)
+- 把cookie(json)加入config.py (NEWBING_COOKIES)
+
+## Moss
+- 使用docker-compose
+
+## RWKV
+- 使用docker-compose
+
+## LLAMA
+- 使用docker-compose
+
+## 盘古
+- 使用docker-compose
+
---
## Text-Generation-UI (TGUI,调试中,暂不可用)
diff --git a/request_llm/bridge_all.py b/request_llm/bridge_all.py
index 9dbcf79..0c46812 100644
--- a/request_llm/bridge_all.py
+++ b/request_llm/bridge_all.py
@@ -130,6 +130,7 @@ model_info = {
"tokenizer": tokenizer_gpt35,
"token_cnt": get_token_num_gpt35,
},
+
}
@@ -186,8 +187,20 @@ if "moss" in AVAIL_LLM_MODELS:
"token_cnt": get_token_num_gpt35,
},
})
-
-
+if "stack-claude" in AVAIL_LLM_MODELS:
+ from .bridge_stackclaude import predict_no_ui_long_connection as claude_noui
+ from .bridge_stackclaude import predict as claude_ui
+ # claude
+ model_info.update({
+ "stack-claude": {
+ "fn_with_ui": claude_ui,
+ "fn_without_ui": claude_noui,
+ "endpoint": None,
+ "max_token": 8192,
+ "tokenizer": tokenizer_gpt35,
+ "token_cnt": get_token_num_gpt35,
+ }
+ })
def LLM_CATCH_EXCEPTION(f):
diff --git a/request_llm/bridge_newbing.py b/request_llm/bridge_newbing.py
index dca7485..2136f01 100644
--- a/request_llm/bridge_newbing.py
+++ b/request_llm/bridge_newbing.py
@@ -153,7 +153,7 @@ class NewBingHandle(Process):
# 进入任务等待状态
asyncio.run(self.async_run())
except Exception:
- tb_str = '```\n' + trimmed_format_exc() + '```'
+ tb_str = '\n```\n' + trimmed_format_exc() + '\n```\n'
self.child.send(f'[Local Message] Newbing失败 {tb_str}.')
self.child.send('[Fail]')
self.child.send('[Finish]')
diff --git a/request_llm/bridge_stackclaude.py b/request_llm/bridge_stackclaude.py
new file mode 100644
index 0000000..f9f3e84
--- /dev/null
+++ b/request_llm/bridge_stackclaude.py
@@ -0,0 +1,296 @@
+from .bridge_newbing import preprocess_newbing_out, preprocess_newbing_out_simple
+from multiprocessing import Process, Pipe
+from toolbox import update_ui, get_conf, trimmed_format_exc
+import threading
+import importlib
+import logging
+import time
+from toolbox import get_conf
+import asyncio
+load_message = "正在加载Claude组件,请稍候..."
+
+try:
+ """
+ ========================================================================
+ 第一部分:Slack API Client
+ https://github.com/yokonsan/claude-in-slack-api
+ ========================================================================
+ """
+
+ from slack_sdk.errors import SlackApiError
+ from slack_sdk.web.async_client import AsyncWebClient
+
+ class SlackClient(AsyncWebClient):
+ """SlackClient类用于与Slack API进行交互,实现消息发送、接收等功能。
+
+ 属性:
+ - CHANNEL_ID:str类型,表示频道ID。
+
+ 方法:
+ - open_channel():异步方法。通过调用conversations_open方法打开一个频道,并将返回的频道ID保存在属性CHANNEL_ID中。
+ - chat(text: str):异步方法。向已打开的频道发送一条文本消息。
+ - get_slack_messages():异步方法。获取已打开频道的最新消息并返回消息列表,目前不支持历史消息查询。
+ - get_reply():异步方法。循环监听已打开频道的消息,如果收到"Typing…_"结尾的消息说明Claude还在继续输出,否则结束循环。
+
+ """
+ CHANNEL_ID = None
+
+ async def open_channel(self):
+ response = await self.conversations_open(users=get_conf('SLACK_CLAUDE_BOT_ID')[0])
+ self.CHANNEL_ID = response["channel"]["id"]
+
+ async def chat(self, text):
+ if not self.CHANNEL_ID:
+ raise Exception("Channel not found.")
+
+ resp = await self.chat_postMessage(channel=self.CHANNEL_ID, text=text)
+ self.LAST_TS = resp["ts"]
+
+ async def get_slack_messages(self):
+ try:
+ # TODO:暂时不支持历史消息,因为在同一个频道里存在多人使用时历史消息渗透问题
+ resp = await self.conversations_history(channel=self.CHANNEL_ID, oldest=self.LAST_TS, limit=1)
+ msg = [msg for msg in resp["messages"]
+ if msg.get("user") == get_conf('SLACK_CLAUDE_BOT_ID')[0]]
+ return msg
+ except (SlackApiError, KeyError) as e:
+ raise RuntimeError(f"获取Slack消息失败。")
+
+ async def get_reply(self):
+ while True:
+ slack_msgs = await self.get_slack_messages()
+ if len(slack_msgs) == 0:
+ await asyncio.sleep(0.5)
+ continue
+
+ msg = slack_msgs[-1]
+ if msg["text"].endswith("Typing…_"):
+ yield False, msg["text"]
+ else:
+ yield True, msg["text"]
+ break
+except:
+ pass
+
+"""
+========================================================================
+第二部分:子进程Worker(调用主体)
+========================================================================
+"""
+
+
+class ClaudeHandle(Process):
+ def __init__(self):
+ super().__init__(daemon=True)
+ self.parent, self.child = Pipe()
+ self.claude_model = None
+ self.info = ""
+ self.success = True
+ self.local_history = []
+ self.check_dependency()
+ if self.success:
+ self.start()
+ self.threadLock = threading.Lock()
+
+ def check_dependency(self):
+ try:
+ self.success = False
+ import slack_sdk
+ self.info = "依赖检测通过,等待Claude响应。注意目前不能多人同时调用Claude接口(有线程锁),否则将导致每个人的Claude问询历史互相渗透。调用Claude时,会自动使用已配置的代理。"
+ self.success = True
+ except:
+ self.info = "缺少的依赖,如果要使用Claude,除了基础的pip依赖以外,您还需要运行`pip install -r request_llm/requirements_slackclaude.txt`安装Claude的依赖,然后重启程序。"
+ self.success = False
+
+ def ready(self):
+ return self.claude_model is not None
+
+ async def async_run(self):
+ await self.claude_model.open_channel()
+ while True:
+ # 等待
+ kwargs = self.child.recv()
+ question = kwargs['query']
+ history = kwargs['history']
+ # system_prompt=kwargs['system_prompt']
+
+ # 是否重置
+ if len(self.local_history) > 0 and len(history) == 0:
+ # await self.claude_model.reset()
+ self.local_history = []
+
+ # 开始问问题
+ prompt = ""
+ # Slack API最好不要添加系统提示
+ # if system_prompt not in self.local_history:
+ # self.local_history.append(system_prompt)
+ # prompt += system_prompt + '\n'
+
+ # 追加历史
+ for ab in history:
+ a, b = ab
+ if a not in self.local_history:
+ self.local_history.append(a)
+ prompt += a + '\n'
+ # if b not in self.local_history:
+ # self.local_history.append(b)
+ # prompt += b + '\n'
+
+ # 问题
+ prompt += question
+ self.local_history.append(question)
+ print('question:', prompt)
+ # 提交
+ await self.claude_model.chat(prompt)
+ # 获取回复
+ # async for final, response in self.claude_model.get_reply():
+ # await self.handle_claude_response(final, response)
+ async for final, response in self.claude_model.get_reply():
+ if not final:
+ print(response)
+ self.child.send(str(response))
+ else:
+ # 防止丢失最后一条消息
+ slack_msgs = await self.claude_model.get_slack_messages()
+ last_msg = slack_msgs[-1]["text"] if slack_msgs and len(slack_msgs) > 0 else ""
+ if last_msg:
+ self.child.send(last_msg)
+ print('-------- receive final ---------')
+ self.child.send('[Finish]')
+
+ def run(self):
+ """
+ 这个函数运行在子进程
+ """
+ # 第一次运行,加载参数
+ self.success = False
+ self.local_history = []
+ if (self.claude_model is None) or (not self.success):
+ # 代理设置
+ proxies, = get_conf('proxies')
+ if proxies is None:
+ self.proxies_https = None
+ else:
+ self.proxies_https = proxies['https']
+
+ try:
+ SLACK_CLAUDE_USER_TOKEN, = get_conf('SLACK_CLAUDE_USER_TOKEN')
+ self.claude_model = SlackClient(token=SLACK_CLAUDE_USER_TOKEN, proxy=self.proxies_https)
+ print('Claude组件初始化成功。')
+ except:
+ self.success = False
+ tb_str = '\n```\n' + trimmed_format_exc() + '\n```\n'
+ self.child.send(f'[Local Message] 不能加载Claude组件。{tb_str}')
+ self.child.send('[Fail]')
+ self.child.send('[Finish]')
+ raise RuntimeError(f"不能加载Claude组件。")
+
+ self.success = True
+ try:
+ # 进入任务等待状态
+ asyncio.run(self.async_run())
+ except Exception:
+ tb_str = '\n```\n' + trimmed_format_exc() + '\n```\n'
+ self.child.send(f'[Local Message] Claude失败 {tb_str}.')
+ self.child.send('[Fail]')
+ self.child.send('[Finish]')
+
+ def stream_chat(self, **kwargs):
+ """
+ 这个函数运行在主进程
+ """
+ self.threadLock.acquire()
+ self.parent.send(kwargs) # 发送请求到子进程
+ while True:
+ res = self.parent.recv() # 等待Claude回复的片段
+ if res == '[Finish]':
+ break # 结束
+ elif res == '[Fail]':
+ self.success = False
+ break
+ else:
+ yield res # Claude回复的片段
+ self.threadLock.release()
+
+
+"""
+========================================================================
+第三部分:主进程统一调用函数接口
+========================================================================
+"""
+global claude_handle
+claude_handle = None
+
+
+def predict_no_ui_long_connection(inputs, llm_kwargs, history=[], sys_prompt="", observe_window=None, console_slience=False):
+ """
+ 多线程方法
+ 函数的说明请见 request_llm/bridge_all.py
+ """
+ global claude_handle
+ if (claude_handle is None) or (not claude_handle.success):
+ claude_handle = ClaudeHandle()
+ observe_window[0] = load_message + "\n\n" + claude_handle.info
+ if not claude_handle.success:
+ error = claude_handle.info
+ claude_handle = None
+ raise RuntimeError(error)
+
+ # 没有 sys_prompt 接口,因此把prompt加入 history
+ history_feedin = []
+ for i in range(len(history)//2):
+ history_feedin.append([history[2*i], history[2*i+1]])
+
+ watch_dog_patience = 5 # 看门狗 (watchdog) 的耐心, 设置5秒即可
+ response = ""
+ observe_window[0] = "[Local Message]: 等待Claude响应中 ..."
+ for response in claude_handle.stream_chat(query=inputs, history=history_feedin, system_prompt=sys_prompt, max_length=llm_kwargs['max_length'], top_p=llm_kwargs['top_p'], temperature=llm_kwargs['temperature']):
+ observe_window[0] = preprocess_newbing_out_simple(response)
+ if len(observe_window) >= 2:
+ if (time.time()-observe_window[1]) > watch_dog_patience:
+ raise RuntimeError("程序终止。")
+ return preprocess_newbing_out_simple(response)
+
+
+def predict(inputs, llm_kwargs, plugin_kwargs, chatbot, history=[], system_prompt='', stream=True, additional_fn=None):
+ """
+ 单线程方法
+ 函数的说明请见 request_llm/bridge_all.py
+ """
+ chatbot.append((inputs, "[Local Message]: 等待Claude响应中 ..."))
+
+ global claude_handle
+ if (claude_handle is None) or (not claude_handle.success):
+ claude_handle = ClaudeHandle()
+ chatbot[-1] = (inputs, load_message + "\n\n" + claude_handle.info)
+ yield from update_ui(chatbot=chatbot, history=[])
+ if not claude_handle.success:
+ claude_handle = None
+ return
+
+ if additional_fn is not None:
+ import core_functional
+ importlib.reload(core_functional) # 热更新prompt
+ core_functional = core_functional.get_core_functions()
+ if "PreProcess" in core_functional[additional_fn]:
+ inputs = core_functional[additional_fn]["PreProcess"](
+ inputs) # 获取预处理函数(如果有的话)
+ inputs = core_functional[additional_fn]["Prefix"] + \
+ inputs + core_functional[additional_fn]["Suffix"]
+
+ history_feedin = []
+ for i in range(len(history)//2):
+ history_feedin.append([history[2*i], history[2*i+1]])
+
+ chatbot[-1] = (inputs, "[Local Message]: 等待Claude响应中 ...")
+ response = "[Local Message]: 等待Claude响应中 ..."
+ yield from update_ui(chatbot=chatbot, history=history, msg="Claude响应缓慢,尚未完成全部响应,请耐心完成后再提交新问题。")
+ for response in claude_handle.stream_chat(query=inputs, history=history_feedin, system_prompt=system_prompt):
+ chatbot[-1] = (inputs, preprocess_newbing_out(response))
+ yield from update_ui(chatbot=chatbot, history=history, msg="Claude响应缓慢,尚未完成全部响应,请耐心完成后再提交新问题。")
+ if response == "[Local Message]: 等待Claude响应中 ...":
+ response = "[Local Message]: Claude响应异常,请刷新界面重试 ..."
+ history.extend([inputs, response])
+ logging.info(f'[raw_input] {inputs}')
+ logging.info(f'[response] {response}')
+ yield from update_ui(chatbot=chatbot, history=history, msg="完成全部响应,请提交新问题。")
diff --git a/request_llm/requirements_slackclaude.txt b/request_llm/requirements_slackclaude.txt
new file mode 100644
index 0000000..472d58c
--- /dev/null
+++ b/request_llm/requirements_slackclaude.txt
@@ -0,0 +1 @@
+slack-sdk==3.21.3
\ No newline at end of file
diff --git a/version b/version
index e833fda..81729fe 100644
--- a/version
+++ b/version
@@ -1,5 +1,5 @@
{
- "version": 3.34,
+ "version": 3.35,
"show_feature": true,
- "new_feature": "修复新版gradio(3.28.3)的暗色主题适配 <-> 提供复旦MOSS模型适配(启用需额外依赖) <-> 提供docker-compose方案兼容LLAMA盘古RWKV等模型的后端 <-> 新增Live2D WAIFU装饰 <-> 完善对话历史的保存/载入/删除 <-> ChatGLM加线程锁提高并发稳定性 <-> 支持NewBing <-> Markdown翻译功能支持直接输入Readme文件网址 <-> 保存对话功能 <-> 解读任意语言代码+同时询问任意的LLM组合 <-> 添加联网(Google)回答问题插件"
+ "new_feature": "添加了OpenAI图片生成插件 <-> 添加了OpenAI音频转文本总结插件 <-> 通过Slack添加对Claude的支持 <-> 提供复旦MOSS模型适配(启用需额外依赖) <-> 提供docker-compose方案兼容LLAMA盘古RWKV等模型的后端 <-> 新增Live2D装饰 <-> 完善对话历史的保存/载入/删除 <-> 保存对话功能"
}