[feat] llama tokenzier
This commit is contained in:
BIN
public/token/llama/llama.model
Normal file
BIN
public/token/llama/llama.model
Normal file
Binary file not shown.
@@ -17,10 +17,14 @@ async function encode(data:string):Promise<(number[]|Uint32Array|Int32Array)>{
|
|||||||
if(db.aiModel.startsWith('novelai')){
|
if(db.aiModel.startsWith('novelai')){
|
||||||
return await tokenizeWebTokenizers(data, 'novelai')
|
return await tokenizeWebTokenizers(data, 'novelai')
|
||||||
}
|
}
|
||||||
|
if(db.aiModel.startsWith('local_') || db.aiModel === 'mancer' || db.aiModel === 'textgen_webui'){
|
||||||
|
return await tokenizeWebTokenizers(data, 'llama')
|
||||||
|
|
||||||
|
}
|
||||||
return await tikJS(data)
|
return await tikJS(data)
|
||||||
}
|
}
|
||||||
|
|
||||||
type tokenizerType = 'novellist'|'claude'|'novelai'
|
type tokenizerType = 'novellist'|'claude'|'novelai'|'llama'
|
||||||
|
|
||||||
let tikParser:Tiktoken = null
|
let tikParser:Tiktoken = null
|
||||||
let tokenizersTokenizer:Tokenizer = null
|
let tokenizersTokenizer:Tokenizer = null
|
||||||
@@ -60,6 +64,11 @@ async function tokenizeWebTokenizers(text:string, type:tokenizerType) {
|
|||||||
).arrayBuffer())
|
).arrayBuffer())
|
||||||
|
|
||||||
break
|
break
|
||||||
|
case 'llama':
|
||||||
|
tokenizersTokenizer = await webTokenizer.Tokenizer.fromSentencePiece(
|
||||||
|
await (await fetch("/token/llama/llama.model")
|
||||||
|
).arrayBuffer())
|
||||||
|
|
||||||
}
|
}
|
||||||
tokenizersType = type
|
tokenizersType = type
|
||||||
}
|
}
|
||||||
|
|||||||
Reference in New Issue
Block a user