deepseek-7b-math-hendrycksmath-lambda09

2モデルの線形補間マージモデル。

Merge Configuration

Parameter Value
Model A deepseek-ai/deepseek-math-7b-instruct
Model B jahyungu/deepseek-math-7b-instruct_hendrycks_math
λ_a 0.90
λ_b 0.10
Formula θ* = 0.90 × θ_a + 0.10 × θ_b
dtype torch.float16

Tokenizer

Union tokenizer (mergekit-style): vocabularies of both models are merged.

  • Union vocab size: 100002
  • Tokens added from Model B: 0
  • Tokens only in Model A: 0

For tokens missing from a model, the other model's embedding is used as fallback before linear interpolation.

Description

This model was created by linearly interpolating the parameters of two models:

  • Model A (deepseek-ai/deepseek-math-7b-instruct): weight = 0.90
  • Model B (jahyungu/deepseek-math-7b-instruct_hendrycks_math): weight = 0.10
Downloads last month
7
Safetensors
Model size
7B params
Tensor type
F16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for lejelly/deepseek-7b-math-hendrycksmath-lambda09

Finetuned
(33)
this model