deepseek-7b-math-code-lambda085
2モデルの線形補間マージモデル。
Merge Configuration
| Parameter | Value |
|---|---|
| Model A | deepseek-ai/deepseek-math-7b-instruct |
| Model B | deepseek-ai/deepseek-coder-7b-instruct-v1.5 |
| λ_a | 0.85 |
| λ_b | 0.15 |
| Formula | θ* = 0.85 × θ_a + 0.15 × θ_b |
| dtype | torch.float16 |
Tokenizer
Union tokenizer (mergekit-style): vocabularies of both models are merged.
- Union vocab size: 100016
- Tokens added from Model B: 14
- Tokens only in Model A: 0
For tokens missing from a model, the other model's embedding is used as fallback before linear interpolation.
Description
This model was created by linearly interpolating the parameters of two models:
- Model A (
deepseek-ai/deepseek-math-7b-instruct): weight = 0.85 - Model B (
deepseek-ai/deepseek-coder-7b-instruct-v1.5): weight = 0.15
- Downloads last month
- 9
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
🙋
Ask for provider support
Model tree for lejelly/deepseek-7b-math-code-lambda085
Base model
deepseek-ai/deepseek-coder-7b-instruct-v1.5