From 1f02c4b63c0387c09835c3865654d87c6686faf2 Mon Sep 17 00:00:00 2001 From: LL Date: Wed, 14 Jun 2023 04:17:43 +0900 Subject: [PATCH] [feat] Update AI model settings for GPT-3.5 and add GPT-3.5 Turbo 16k option The commit updates AI model settings for GPT-3.5 by allowing up to 4000 tokens for context size. It also adds a new option for GPT-3.5 Turbo 16k with a maximum context of 16000 tokens. Additionally, the commit limits the context size when exceeding the maximum limit for each model. --- src/lib/Setting/Pages/BotSettings.svelte | 4 +++- src/lib/UI/ModelList.svelte | 3 +++ src/ts/process/index.ts | 5 +++++ src/ts/process/request.ts | 2 ++ 4 files changed, 13 insertions(+), 1 deletion(-) diff --git a/src/lib/Setting/Pages/BotSettings.svelte b/src/lib/Setting/Pages/BotSettings.svelte index 4650f3fd..9a016673 100644 --- a/src/lib/Setting/Pages/BotSettings.svelte +++ b/src/lib/Setting/Pages/BotSettings.svelte @@ -110,7 +110,7 @@ Claude {language.apiKey} {/if} -{#if $DataBase.aiModel === 'gpt35' || $DataBase.aiModel === 'gpt4' || $DataBase.subModel === 'gpt4' || $DataBase.subModel === 'gpt35'|| $DataBase.aiModel === 'gpt4_32k' || $DataBase.subModel === 'gpt4_32k'} +{#if $DataBase.aiModel === 'gpt35' || $DataBase.aiModel === 'gpt35_16k' || $DataBase.subModel === 'gpt35_16k' || $DataBase.aiModel === 'gpt4' || $DataBase.subModel === 'gpt4' || $DataBase.subModel === 'gpt35'|| $DataBase.aiModel === 'gpt4_32k' || $DataBase.subModel === 'gpt4_32k'} OpenAI {language.apiKey}
@@ -166,6 +166,8 @@ {language.maxContextSize} {#if $DataBase.aiModel === 'gpt35'} +{:else if $DataBase.aiModel === 'gpt35_16k'} + {:else if $DataBase.aiModel === 'gpt4' || $DataBase.aiModel === 'textgen_webui'} {:else if $DataBase.aiModel === 'custom'} diff --git a/src/lib/UI/ModelList.svelte b/src/lib/UI/ModelList.svelte index ddcd9d8a..8e2352e8 100644 --- a/src/lib/UI/ModelList.svelte +++ b/src/lib/UI/ModelList.svelte @@ -13,6 +13,8 @@ switch(name){ case "gpt35": return "GPT-3.5 Turbo" + case "gpt35_16k": + return "GPT-3.5 Turbo 16k" case "gpt4": return "GPT-4" case "gpt4_32k": @@ -52,6 +54,7 @@
+ diff --git a/src/ts/process/index.ts b/src/ts/process/index.ts index 690825a8..b8366a49 100644 --- a/src/ts/process/index.ts +++ b/src/ts/process/index.ts @@ -121,6 +121,11 @@ export async function sendChat(chatProcessIndex = -1,arg:{chatAdditonalTokens?:n maxContextTokens = 4000 } } + if(db.aiModel === 'gpt35_16k'){ + if(maxContextTokens > 16000){ + maxContextTokens = 16000 + } + } if(db.aiModel === 'gpt4'){ if(maxContextTokens > 8000){ maxContextTokens = 8000 diff --git a/src/ts/process/request.ts b/src/ts/process/request.ts index e4ae002d..cea4dbb6 100644 --- a/src/ts/process/request.ts +++ b/src/ts/process/request.ts @@ -57,6 +57,7 @@ export async function requestChatDataMain(arg:requestDataArgument, model:'model' switch(aiModel){ case 'gpt35': + case 'gpt35_16k': case 'gpt4': case 'gpt4_32k':{ @@ -69,6 +70,7 @@ export async function requestChatDataMain(arg:requestDataArgument, model:'model' const body = ({ model: aiModel === 'gpt35' ? 'gpt-3.5-turbo' + : aiModel === 'gpt35_16k' ? 'gpt-3.5-turbo-16k' : aiModel === 'gpt4' ? 'gpt-4' : 'gpt-4-32k', messages: formated, temperature: temperature,