# PR Checklist
- [x] Did you check if it works normally in all models? *ignore this
when it dosen't uses models*
- [x] Did you check if it works normally in all of web, local and node
hosted versions? if it dosen't, did you blocked it in those versions?
- [x] Did you added a type def?
# Description
~No bug, work as intended. But, haven't checked actual summary with it.
Very unstable, need confirmation/check.~
### **NO bug, work as intended, summarizes correctly, moving mainChunks
to chunks works well, no re-summarization issues. It works!**
Added new variable to database.ts, which indicates if HypaV2 Memory is
activated or not.
hypav2: boolean
This was added to stop overlapping of memoryType, when HypaV2 Memory is
selected and supaMemoryModel is changed.
Added supaMemoryType selection, HypaMemoryType selection,
supaMemoryPrompt changing section on OtherBotSettings.svelte, and
implemented on hypav2.ts.
Also added OpenAI key when summarization model is GPT 3.5 instruct.
Also suggested memoryAlgorithmType:string variable on database.ts, to
further add more memory types.

fixed minor punctuations, and changed summarize function of
hypav2.ts(same as the one on supaMemory.ts)
Everything works as intended.
when the first chat limit is reached, it succesfully summarize the previous chats according to the chunk size, and then returns the chat with summarized text as a system prompt, with other chats after that summarized point correctly appended.
Automatic sub-chunking also works perfectly.
Tested with sub-model.