Merged Model: llama_onto_qwen_lambda-0

This model is a result of parameter averaging (Model Soup) across 1 models.

Merged Models

The following models were included in the merge:

  • toksuite/meta-llama-Llama-3.2-1B

Merging Configuration

  • Method: Weighted Parameter Averaging
  • Weights: Simple average with merging lambda = 0.0.
  • Excluded Layers: Embeddings and LM Head were kept from the host model (toksuite/Qwen-Qwen3-8B).

Usage

from transformers import AutoModelForCausalLM, AutoTokenizer

model = AutoModelForCausalLM.from_pretrained("flexitok/llama_onto_qwen_lambda-0")
tokenizer = AutoTokenizer.from_pretrained("flexitok/llama_onto_qwen_lambda-0")
Downloads last month
-
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support