| | import type { Model } from "$lib/types/Model"; |
| | import { AutoTokenizer, PreTrainedTokenizer } from "@huggingface/transformers"; |
| |
|
| | export async function getTokenizer(_modelTokenizer: Exclude<Model["tokenizer"], undefined>) { |
| | if (typeof _modelTokenizer === "string") { |
| | |
| | return await AutoTokenizer.from_pretrained(_modelTokenizer); |
| | } else { |
| | |
| | const { tokenizerUrl, tokenizerConfigUrl } = _modelTokenizer satisfies { |
| | tokenizerUrl: string; |
| | tokenizerConfigUrl: string; |
| | }; |
| | const tokenizerJSON = await (await fetch(tokenizerUrl)).json(); |
| | const tokenizerConfig = await (await fetch(tokenizerConfigUrl)).json(); |
| | return new PreTrainedTokenizer(tokenizerJSON, tokenizerConfig); |
| | } |
| | } |
| |
|