[feat] supaMemory modified

This commit is contained in:
kwaroran
2023-05-16 03:58:42 +09:00
parent 087d8dcf0c
commit e88eb48a1e
4 changed files with 41 additions and 23 deletions

View File

@@ -8,7 +8,7 @@
}, },
"package": { "package": {
"productName": "RisuAI", "productName": "RisuAI",
"version": "0.9.4" "version": "0.9.5"
}, },
"tauri": { "tauri": {
"allowlist": { "allowlist": {

View File

@@ -7,7 +7,7 @@ import { cloneDeep } from 'lodash';
export const DataBase = writable({} as any as Database) export const DataBase = writable({} as any as Database)
export const loadedStore = writable(false) export const loadedStore = writable(false)
export let appVer = '0.9.4' export let appVer = '0.9.5'
export function setDatabase(data:Database){ export function setDatabase(data:Database){

View File

@@ -87,35 +87,53 @@ export async function supaMemory(chats:OpenAIChat[],currentTokens:number,maxCont
chunkSize += tokens chunkSize += tokens
} }
const promptbody:OpenAIChat[] = [ const supaPrompt = db.supaMemoryPrompt === '' ?
{
role: "user",
content: stringlizedChat
},
{
role: "system",
content: db.supaMemoryPrompt === '' ?
"[Summarize the ongoing role story. It must also remove redundancy and unnecessary content from the prompt so that gpt3 and other sublanguage models]\n" "[Summarize the ongoing role story. It must also remove redundancy and unnecessary content from the prompt so that gpt3 and other sublanguage models]\n"
: db.supaMemoryPrompt : db.supaMemoryPrompt
}
]
const da = await requestChatData({ const promptbody = stringlizedChat + '\n\n' + supaPrompt + "\n\nOutput:"
formated: promptbody,
bias: {}
}, 'submodel')
if(da.type === 'fail'){ const da = await fetch("https://api.openai.com/v1/completions",{
headers: {
"Content-Type": "application/json",
"Authorization": "Bearer " + db.openAIKey
},
method: "POST",
body: JSON.stringify({
"model": "text-davinci-003",
"prompt": promptbody,
"max_tokens": 500,
"temperature": 0
})
})
// const promptbody:OpenAIChat[] = [
// {
// role: "user",
// content: stringlizedChat
// },
// {
// role: "system",
// content: supaPrompt
// }
// ]
// const da = await requestChatData({
// formated: promptbody,
// bias: {}
// }, 'submodel')
const result = (await da.json()).choices[0].text.trim()
if(da.status < 200 || da.status >= 300){
return { return {
currentTokens: currentTokens, currentTokens: currentTokens,
chats: chats, chats: chats,
error: "SupaMemory: HTTP: " + da.result error: "SupaMemory: HTTP: " + await da.text()
} }
} }
const tokenz = await tokenize(da.result + '\n\n') + 5 const tokenz = await tokenize(result + '\n\n') + 5
currentTokens += tokenz currentTokens += tokenz
supaMemory += da.result + '\n\n' supaMemory += result + '\n\n'
console.log(tokenz) console.log(tokenz)
} }

View File

@@ -1 +1 @@
{"version":"0.9.4"} {"version":"0.9.5"}