[fix] ooba to use llama tokenizer
This commit is contained in:
@@ -19,7 +19,11 @@ async function encode(data:string):Promise<(number[]|Uint32Array|Int32Array)>{
|
|||||||
if(db.aiModel.startsWith('novelai')){
|
if(db.aiModel.startsWith('novelai')){
|
||||||
return await tokenizeWebTokenizers(data, 'novelai')
|
return await tokenizeWebTokenizers(data, 'novelai')
|
||||||
}
|
}
|
||||||
if(db.aiModel.startsWith('local_') || db.aiModel === 'mancer' || db.aiModel === 'textgen_webui' || (db.aiModel === 'reverse_proxy' && db.reverseProxyOobaMode)){
|
if(db.aiModel.startsWith('local_') ||
|
||||||
|
db.aiModel === 'mancer' ||
|
||||||
|
db.aiModel === 'textgen_webui' ||
|
||||||
|
(db.aiModel === 'reverse_proxy' && db.reverseProxyOobaMode ||
|
||||||
|
db.aiModel === 'ooba')){
|
||||||
return await tokenizeWebTokenizers(data, 'llama')
|
return await tokenizeWebTokenizers(data, 'llama')
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|||||||
Reference in New Issue
Block a user