LFM2-MoE-WebGPU / src /model-config.ts
mlabonne's picture
Add demo files (#1)
4755edd
export interface ExamplePrompt {
label: string;
prompt: string;
}
export interface ModelOption {
modelId: string;
label: string;
sublabel: string;
description: string;
}
export interface ModelConfig {
models: ModelOption[];
defaultModelId: string;
examplePrompts: ExamplePrompt[];
}
export const MODEL_CONFIG: ModelConfig = {
models: [
{
modelId: "LiquidAI/LFM2-8B-A1B-ONNX",
label: "LFM2 8B-A1B",
sublabel: "8B 路 Faster",
description: "Fast and lightweight mixture-of-experts model.",
},
{
modelId: "LiquidAI/LFM2-24B-A2B-ONNX",
label: "LFM2 24B-A2B",
sublabel: "24B 路 Smarter",
description: "Larger model with more experts for improved performance on complex tasks.",
},
],
defaultModelId: "LiquidAI/LFM2-8B-A1B-ONNX",
examplePrompts: [
{
label: "Solve a logic puzzle",
prompt:
"Five people finished a race. Ava finished ahead of Ben but behind Chloe. Diego finished behind Ben but ahead of Emma. What was the finishing order?",
},
{
label: "Summarize a meeting",
prompt:
"Turn these notes into a crisp follow-up email with action items:\n- launch moved to next Wednesday\n- Priya owns final QA\n- Marcus will update docs\n- follow-up check-in on Monday at 10 AM",
},
{
label: "Write TypeScript code",
prompt:
"Write a clean TypeScript function that groups an array of objects by a given key and returns a Record. Include a short usage example.",
},
{
label: "Explain a concept",
prompt: 'Explain the term "Mixture of Experts"',
},
],
};