From ba0a8b7072e3047b67c06fdd2979c21631761e91 Mon Sep 17 00:00:00 2001 From: awwaawwa <8493196+awwaawwa@users.noreply.github.com> Date: Thu, 11 Apr 2024 22:02:40 +0800 Subject: [PATCH] integrate gpt-4-turbo-2024-04-09 (#1698) MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit * 接入 gpt-4-turbo-2024-04-09 模型 * add gpt-4-turbo and change to vision * add gpt-4-turbo to avail llm models * 暂时将gpt-4-turbo接入至普通版本 --- config.py | 2 +- request_llms/bridge_all.py | 19 +++++++++++++++++++ 2 files changed, 20 insertions(+), 1 deletion(-) diff --git a/config.py b/config.py index 4aab6cc..380a64b 100644 --- a/config.py +++ b/config.py @@ -32,7 +32,7 @@ else: # [step 3]>> 模型选择是 (注意: LLM_MODEL是默认选中的模型, 它*必须*被包含在AVAIL_LLM_MODELS列表中 ) LLM_MODEL = "gpt-3.5-turbo-16k" # 可选 ↓↓↓ -AVAIL_LLM_MODELS = ["gpt-4-1106-preview", "gpt-4-turbo-preview", "gpt-4-vision-preview", +AVAIL_LLM_MODELS = ["gpt-4-1106-preview", "gpt-4-turbo-preview", "gpt-4-vision-preview", "gpt-4-turbo", "gpt-4-turbo-2024-04-09", "gpt-3.5-turbo-1106", "gpt-3.5-turbo-16k", "gpt-3.5-turbo", "azure-gpt-3.5", "gpt-4", "gpt-4-32k", "azure-gpt-4", "glm-4", "glm-3-turbo", "gemini-pro", "chatglm3" diff --git a/request_llms/bridge_all.py b/request_llms/bridge_all.py index d187ef1..d20e033 100644 --- a/request_llms/bridge_all.py +++ b/request_llms/bridge_all.py @@ -200,6 +200,25 @@ model_info = { "token_cnt": get_token_num_gpt4, }, + "gpt-4-turbo": { + "fn_with_ui": chatgpt_ui, + "fn_without_ui": chatgpt_noui, + "endpoint": openai_endpoint, + "max_token": 128000, + "tokenizer": tokenizer_gpt4, + "token_cnt": get_token_num_gpt4, + }, + + "gpt-4-turbo-2024-04-09": { + "fn_with_ui": chatgpt_ui, + "fn_without_ui": chatgpt_noui, + "endpoint": openai_endpoint, + "max_token": 128000, + "tokenizer": tokenizer_gpt4, + "token_cnt": get_token_num_gpt4, + }, + + "gpt-3.5-random": { "fn_with_ui": chatgpt_ui, "fn_without_ui": chatgpt_noui,