| --- |
| license: apache-2.0 |
| base_model: |
| - IlyaGusev/vikhr_nemo_orpo_dostoevsky_12b_slerp |
| - DavidAU/Mistral-Nemo-2407-12B-Thinking-Claude-Gemini-GPT5.2-Uncensored-HERETIC |
| - Naphula/MN-12B-Mag-Mell-R1-Uncensored |
| - Aleteian/Pathfinder-RP-12B-RU |
| library_name: transformers |
| language: |
| - en |
| - ru |
| tags: |
| - mergekit |
| - merge |
| - russian |
| - uncensored |
| - roleplay |
| - mixtral-nemo |
| --- |
| |
| # MN-12B-Hydra-RP-RU |
|
|
| <details> |
| <summary>🇷🇺 Нажмите, чтобы развернуть описание на русском</summary> |
|
|
| ## 🌟 О модели |
|
|
| **MN-12B-Hydra-RP-RU** — экспериментальный merge на базе Mistral Nemo 12B, сочетающий: |
|
|
| * 🎭 Сильные ролевые способности |
| * 📚 Глубокий литературный русский язык |
| * 🔓 Снятую цензуру |
|
|
| Модель собрана методом TIES-merging, что позволяет объединять веса нескольких моделей с минимальными конфликтами между параметрами. |
|
|
| ## 🎯 Особенности |
|
|
| * Основной язык — русский |
| * Хорошо держит персонажей и контекст |
| * Следует инструкциям |
| * Сохраняет возможности базового Nemo |
| * Не проходила дополнительного обучения после слияния |
|
|
| ## ⚠️ Важно |
|
|
| Uncensored-характер модели означает, что она может генерировать контент, который некоторые пользователи сочтут неподобающим. |
|
|
| </details> |
|
|
| High-quality TIES merge based on **Mistral Nemo 12B**, optimized for roleplay, strong Russian language capabilities, and uncensored behavior. |
|
|
| --- |
|
|
| ## 🌍 Overview |
|
|
| **MN-12B-Hydra-RP-RU** is an experimental merge built on top of [Mistral Nemo 12B](https://huggingface.co/mistralai/Mistral-Nemo-Base-2407), combining strengths from multiple fine-tuned models: |
|
|
| * 🎭 Advanced roleplay capability from Pathfinder-RP |
| * 📚 Deep Russian language fluency inspired by Vikhr + Dostoevsky-style tuning |
| * 🔓 Reduced safety filtering via uncensored components |
|
|
| The merge was created using **TIES merging**, which allows combining model deltas while minimizing destructive interference between weights. |
|
|
| --- |
|
|
| ## 🎯 Key Features |
|
|
| | Feature | Description | |
| | ------------------------- | ------------------------------------------------ | |
| | **Languages** | Russian, English | |
| | **Censorship** | Uncensored behavior | |
| | **Roleplay** | Strong character consistency and narrative depth | |
| | **Instruction Following** | Reliable prompt adherence | |
| | **Tool Calling** | Retains base Nemo capabilities | |
| | **Architecture** | Mistral Nemo 12B | |
|
|
| --- |
|
|
| ## 🧩 Model Composition |
|
|
| The merge combines the following models: |
|
|
| | Model | Role in merge | Weight | |
| | ------------------------------ | ------------------------- | ------ | |
| | **Pathfinder-RP-12B-RU** | Base model, RP backbone | 0.60 | |
| | **Vikhr Nemo ORPO Dostoevsky** | Literary Russian depth | 0.25 | |
| | **HERETIC Uncensored** | Safety removal | 0.30 | |
| | **Mag-Mell R1 Uncensored** | Additional uncensor delta | 0.20 | |
|
|
| *Weights shown before normalization (final weights are normalized to sum = 1).* |
|
|
| --- |
|
|
| ## 💡 Usage Example |
|
|
| ```python |
| from transformers import AutoTokenizer, AutoModelForCausalLM |
| import torch |
| |
| model_name = "limloop/MN-12B-Hydra-RP-RU" |
| tokenizer = AutoTokenizer.from_pretrained(model_name) |
| model = AutoModelForCausalLM.from_pretrained( |
| model_name, |
| torch_dtype=torch.bfloat16, |
| device_map="auto" |
| ) |
| |
| prompt = "You are a medieval innkeeper. Greet the traveler!" |
| messages = [{"role": "user", "content": prompt}] |
| inputs = tokenizer.apply_chat_template(messages, return_tensors="pt").to(model.device) |
| |
| outputs = model.generate(inputs, max_new_tokens=512, temperature=0.7) |
| response = tokenizer.decode(outputs[0], skip_special_tokens=True) |
| print(response) |
| ``` |
|
|
| --- |
|
|
| ## ⚙️ Merge Details |
|
|
| Built using [mergekit](https://github.com/cg123/mergekit) with the **TIES** method (Trim, Elect Sign, Merge). |
|
|
| Core mechanism: |
|
|
| 1. Trim low-magnitude deltas via `density` |
| 2. Resolve sign conflicts |
| 3. Weighted averaging of aligned parameters |
|
|
| ### Merge Configuration |
|
|
| ```yaml |
| models: |
| - model: Aleteian/Pathfinder-RP-12B-RU |
| weight: 0.6 |
| - model: IlyaGusev/vikhr_nemo_orpo_dostoevsky_12b_slerp |
| weight: 0.25 |
| density: 0.9 |
| - model: DavidAU/Mistral-Nemo-2407-12B-Thinking-Claude-Gemini-GPT5.2-Uncensored-HERETIC |
| weight: 0.3 |
| density: 0.9 |
| - model: Naphula/MN-12B-Mag-Mell-R1-Uncensored |
| weight: 0.2 |
| density: 0.9 |
| |
| merge_method: ties |
| parameters: |
| epsilon: 0.01 |
| normalize: true |
| base_model: Aleteian/Pathfinder-RP-12B-RU |
| dtype: bfloat16 |
| tokenizer: |
| source: base |
| ``` |
|
|
| --- |
|
|
| ## ⚠️ Known Characteristics |
|
|
| * No additional post-merge fine-tuning |
| * May switch to English on complex reasoning tasks |
| * Uncensored components allow generation of explicit or controversial content |
|
|