[feat] Update AI model settings for GPT-3.5 and add GPT-3.5 Turbo 16k option
The commit updates AI model settings for GPT-3.5 by allowing up to 4000 tokens for context size. It also adds a new option for GPT-3.5 Turbo 16k with a maximum context of 16000 tokens. Additionally, the commit limits the context size when exceeding the maximum limit for each model.
This commit is contained in:
@@ -121,6 +121,11 @@ export async function sendChat(chatProcessIndex = -1,arg:{chatAdditonalTokens?:n
|
||||
maxContextTokens = 4000
|
||||
}
|
||||
}
|
||||
if(db.aiModel === 'gpt35_16k' || db.aiModel === 'gpt35_16k_0613'){
|
||||
if(maxContextTokens > 16000){
|
||||
maxContextTokens = 16000
|
||||
}
|
||||
}
|
||||
if(db.aiModel === 'gpt4'){
|
||||
if(maxContextTokens > 8000){
|
||||
maxContextTokens = 8000
|
||||
|
||||
Reference in New Issue
Block a user