Add geminiThinking flag and enhance thoughts handling in chat processing

This commit is contained in:
Kwaroran
2024-12-20 02:53:58 +09:00
parent bd14dcd038
commit 4afe32a2c2
3 changed files with 43 additions and 30 deletions

View File

@@ -17,6 +17,7 @@ export enum LLMFlags{
hasVideoInput,
OAICompletionTokens,
DeveloperRole,
geminiThinking
}
export enum LLMProvider{
@@ -814,7 +815,7 @@ export const LLMModels: LLMModel[] = [
id: 'gemini-2.0-flash-thinking-exp-1219',
provider: LLMProvider.GoogleCloud,
format: LLMFormat.GoogleCloud,
flags: [LLMFlags.hasImageInput, LLMFlags.hasFirstSystemPrompt, LLMFlags.poolSupported, LLMFlags.hasAudioInput, LLMFlags.hasVideoInput, LLMFlags.hasStreaming],
flags: [LLMFlags.hasImageInput, LLMFlags.hasFirstSystemPrompt, LLMFlags.poolSupported, LLMFlags.hasAudioInput, LLMFlags.hasVideoInput, LLMFlags.hasStreaming, LLMFlags.geminiThinking],
parameters: ['temperature', 'top_k', 'top_p', 'presence_penalty', 'frequency_penalty'],
tokenizer: LLMTokenizer.GoogleCloud,
recommended: true