From e03e7f9c4bd0fbd694d60d338a1968f3f7a28546 Mon Sep 17 00:00:00 2001 From: LightningHyperBlaze45654 <73149145+LightningHyperBlaze45654@users.noreply.github.com> Date: Sun, 16 Jun 2024 10:28:31 -0700 Subject: [PATCH] Resolved all buggy issues why no json? Idk, blame me --- src/ts/process/memory/hypav2.ts | 21 +++++++++++---------- 1 file changed, 11 insertions(+), 10 deletions(-) diff --git a/src/ts/process/memory/hypav2.ts b/src/ts/process/memory/hypav2.ts index 950b04bf..59435ed8 100644 --- a/src/ts/process/memory/hypav2.ts +++ b/src/ts/process/memory/hypav2.ts @@ -20,7 +20,7 @@ export interface HypaV2Data { async function summary(stringlizedChat: string): Promise<{ success: boolean; data: string }> { const db = get(DataBase); - console.log(db.supaModelType, " is being used for Summarization task on HypaV2Memory"); + console.log("Summarization actively called"); if (db.supaModelType === 'distilbart') { try { @@ -48,15 +48,16 @@ async function summary(stringlizedChat: string): Promise<{ success: boolean; dat "Authorization": "Bearer " + db.supaMemoryKey }, method: "POST", - body: JSON.stringify({ + body: { "model": db.supaModelType === 'curie' ? "text-curie-001" - : db.supaModelType === 'instruct35' ? 'gpt-3.5-turbo-instruct' : "text-davinci-003", + : db.supaModelType === 'instruct35' ? 'gpt-3.5-turbo-instruct' + : "text-davinci-003", "prompt": promptbody, "max_tokens": 600, "temperature": 0 - }) - }); - console.log("Using Chatgpt 3.5 instruct for SupaMemory"); + } + }) + console.log("Using openAI instruct 3.5 for SupaMemory"); try { if (!da.ok) { @@ -138,7 +139,6 @@ export async function hypaMemoryV2( for (const chat of removedChats) { currentTokens -= await tokenizer.tokenizeChat(chat); } - chats = chats.slice(ind); mainPrompt = chunk.text; const mpToken = await tokenizer.tokenizeChat({ role: 'system', content: mainPrompt }); allocatedTokens -= mpToken; @@ -239,11 +239,12 @@ export async function hypaMemoryV2( }); // Add the remaining chats after the last mainChunk's targetId - if (data.mainChunks.length > 0) { - const lastTargetId = data.mainChunks[0].targetId; + const lastTargetId = data.mainChunks.length > 0 ? data.mainChunks[0].targetId : null; + if (lastTargetId) { const lastIndex = chats.findIndex(chat => chat.memo === lastTargetId); if (lastIndex !== -1) { - chats.push(...chats.slice(lastIndex + 1)); + const remainingChats = chats.slice(lastIndex + 1); + chats = chats.slice(0, 1).concat(remainingChats); } }