Fix mistral top_p and fix seperateParameters
This commit is contained in:
@@ -708,7 +708,7 @@ export const LLMModels: LLMModel[] = [
|
||||
format: LLMFormat.Mistral,
|
||||
flags: [LLMFlags.hasFirstSystemPrompt, LLMFlags.mustStartWithUserInput, LLMFlags.requiresAlternateRole],
|
||||
recommended: true,
|
||||
parameters: ['temperature', 'presence_penalty', 'frequency_penalty'],
|
||||
parameters: ['temperature', 'presence_penalty', 'frequency_penalty', 'top_p'],
|
||||
tokenizer: LLMTokenizer.Mistral
|
||||
},
|
||||
{
|
||||
@@ -719,7 +719,7 @@ export const LLMModels: LLMModel[] = [
|
||||
format: LLMFormat.Mistral,
|
||||
flags: [LLMFlags.hasFirstSystemPrompt, LLMFlags.mustStartWithUserInput, LLMFlags.requiresAlternateRole],
|
||||
recommended: true,
|
||||
parameters: ['temperature', 'presence_penalty', 'frequency_penalty'],
|
||||
parameters: ['temperature', 'presence_penalty', 'frequency_penalty', 'top_p'],
|
||||
tokenizer: LLMTokenizer.Mistral
|
||||
},
|
||||
{
|
||||
@@ -729,7 +729,7 @@ export const LLMModels: LLMModel[] = [
|
||||
provider: LLMProvider.Mistral,
|
||||
format: LLMFormat.Mistral,
|
||||
flags: [LLMFlags.hasFirstSystemPrompt, LLMFlags.mustStartWithUserInput, LLMFlags.requiresAlternateRole],
|
||||
parameters: ['temperature', 'presence_penalty', 'frequency_penalty'],
|
||||
parameters: ['temperature', 'presence_penalty', 'frequency_penalty', 'top_p'],
|
||||
tokenizer: LLMTokenizer.Mistral
|
||||
},
|
||||
{
|
||||
@@ -739,7 +739,7 @@ export const LLMModels: LLMModel[] = [
|
||||
provider: LLMProvider.Mistral,
|
||||
format: LLMFormat.Mistral,
|
||||
flags: [LLMFlags.hasFirstSystemPrompt, LLMFlags.mustStartWithUserInput, LLMFlags.requiresAlternateRole],
|
||||
parameters: ['temperature', 'presence_penalty', 'frequency_penalty'],
|
||||
parameters: ['temperature', 'presence_penalty', 'frequency_penalty', 'top_p'],
|
||||
tokenizer: LLMTokenizer.Mistral
|
||||
},
|
||||
{
|
||||
@@ -749,7 +749,7 @@ export const LLMModels: LLMModel[] = [
|
||||
provider: LLMProvider.Mistral,
|
||||
format: LLMFormat.Mistral,
|
||||
flags: [LLMFlags.hasFirstSystemPrompt, LLMFlags.mustStartWithUserInput, LLMFlags.requiresAlternateRole],
|
||||
parameters: ['temperature', 'presence_penalty', 'frequency_penalty'],
|
||||
parameters: ['temperature', 'presence_penalty', 'frequency_penalty', 'top_p'],
|
||||
recommended: true,
|
||||
tokenizer: LLMTokenizer.Mistral
|
||||
},
|
||||
|
||||
@@ -520,10 +520,9 @@ async function requestOpenAI(arg:RequestDataArgumentExtended):Promise<requestDat
|
||||
body: applyParameters({
|
||||
model: requestModel,
|
||||
messages: reformatedChat,
|
||||
top_p: db.top_p,
|
||||
safe_prompt: false,
|
||||
max_tokens: arg.maxTokens,
|
||||
}, ['temperature', 'presence_penalty', 'frequency_penalty'], {}, arg.mode ),
|
||||
}, ['temperature', 'presence_penalty', 'frequency_penalty', 'top_p'], {}, arg.mode ),
|
||||
headers: {
|
||||
"Authorization": "Bearer " + db.mistralKey,
|
||||
},
|
||||
|
||||
@@ -453,7 +453,7 @@ export function setDatabase(data:Database){
|
||||
data.vertexClientEmail ??= ''
|
||||
data.vertexPrivateKey ??= ''
|
||||
data.seperateParametersEnabled ??= false
|
||||
data.seperateParameters = {
|
||||
data.seperateParameters ??= {
|
||||
memory: {},
|
||||
emotion: {},
|
||||
translate: {},
|
||||
|
||||
Reference in New Issue
Block a user