| | --- |
| | license: apache-2.0 |
| | tags: |
| | - unsloth |
| | - Uncensored |
| | - text-generation-inference |
| | - transformers |
| | - unsloth |
| | - llama |
| | - trl |
| | - roleplay |
| | - conversational |
| | datasets: |
| | - iamketan25/roleplay-instructions-dataset |
| | - N-Bot-Int/Iris-Uncensored-R1 |
| | - N-Bot-Int/Moshpit-Combined-R2-Uncensored |
| | - N-Bot-Int/Mushed-Dataset-Uncensored |
| | - N-Bot-Int/Muncher-R1-Uncensored |
| | - N-Bot-Int/Millia-R1_DPO |
| | language: |
| | - en |
| | base_model: |
| | - N-Bot-Int/MiniMaid-L1 |
| | pipeline_tag: text-generation |
| | library_name: peft |
| | metrics: |
| | - character |
| | - bleu |
| | - rouge |
| | --- |
| | # THIS IS THE FINAL MiniMaid-L Series, This is because we've hit the final Ceiling for a 1B model! Thank you so much for your Support! |
| | - If you loved our Models, then please consider donating and supporting us through Ko-fi! |
| | - [](https://ko-fi.com/J3J61D8NHV) |
| | |
| |  |
| | # MiniMaid-L3 |
| | - Introducing MiniMaid-L3 model! Our brand new finetuned MiniMaid-L2 Architecture, allowing for an Even More Coherent and |
| | Immersive Roleplay through the Use of Knowledge distillation! |
| |
|
| | - MiniMaid-L3 is a Small Update to L2, Which uses Knowledge distillation to combine our L2 Architecture, and A Popular |
| | Roleplaying Model named MythoMax, which also uses a Combanant Technology to Combine models and create MythoMax-7B, |
| | MiniMaid-L3 on the other hand is a distillation of MiniMaid-L2, combined with using MythoMax Knowledge Distillation, |
| | which created MiniMaid-L3, a More Capable Model that Outcompete its descendance in both roleplaying scenarios |
| | And even Knock MiniMaid-L2's BLEU scoring! |
| |
|
| |
|
| | # MiniMaid-L1 Base-Model Card Procedure: |
| | - **MiniMaid-L1** achieve a good Performance through process of DPO and Combined Heavy Finetuning, To Prevent Overfitting, |
| | We used high LR decays, And Introduced Randomization techniques to prevent the AI from learning and memorizing, |
| | However since training this on Google Colab is difficult, the Model might underperform or underfit on specific tasks |
| | Or overfit on knowledge it manage to latched on! However please be guided that we did our best, and it will improve as we move onwards! |
| |
|
| | - MiniMaid-L3 is Another Instance of Our Smallest Model Yet! if you find any issue, then please don't hesitate to email us at: |
| | [nexus.networkinteractives@gmail.com](mailto:nexus.networkinteractives@gmail.com) |
| | about any overfitting, or improvements for the future Model **V4**, |
| | Once again feel free to Modify the LORA to your likings, However please consider Adding this Page |
| | for credits and if you'll increase its **Dataset**, then please handle it with care and ethical considerations |
| |
|
| |
|
| | |
| | - MiniMaid-L3 is |
| | - **Developed by:** N-Bot-Int |
| | - **License:** apache-2.0 |
| | - **Parent Model from model:** unsloth/llama-3.2-3b-instruct-unsloth-bnb-1bit |
| | - **Dataset Combined Using:** NKDProtoc(Propietary Software) |
| |
|
| | - MiniMaid-L3 Official Metric Score |
| |  |
| | - Metrics Made By **ItsMeDevRoland** |
| | Which compares: |
| | - **MiniMaid-L2 GGUFF** |
| | - **MiniMaid-L3 GGUFF** |
| | Which are All Ranked with the Same Prompt, Same Temperature, Same Hardware(Google Colab), |
| | To Properly Showcase the differences and strength of the Models |
| | |
| | - **Visit Below to See details!** |
| |
|
| | --- |
| | # 🧵 MiniMaid-L3: Slower Steps, Deeper Stories — The Immersive Upgrade |
| | > "She’s more grounded, more convincing — and when it comes to roleplay, she’s in a league of her own." |
| |  |
| | --- |
| |
|
| | # MiniMaid-L3 doesn’t just iterate — she elevates. Built on L2’s disciplined architecture, L3 doubles down on character immersion and emotional coherence, refining every line she delivers. |
| | - 💬 Roleplay Evaluation (v2) |
| | - 🧠 Character Consistency: 0.54 → 0.55 (+) |
| | - 🌊 Immersion: 0.59 → 0.66 (↑) |
| | - 🎭 Overall RP Score: 0.72 → 0.75 |
| | > L3’s immersive depth marks a new high in believability and emotional traction — she's not just playing a part, she becomes it. |
| |
|
| | # 📊 Slower, But Smarter |
| | - 🕒 Inference Time: 39.1s (↑ from 34.5s) |
| | - ⚡ Tokens/sec: 6.61 (slight dip) |
| | - 📏 BLEU/ROUGE-L: Mixed — slight BLEU gain, ROUGE-L softened |
| | > Sure, she takes her time — but it’s worth it. L3 trades a few milliseconds for measured, thoughtful outputs that stick the landing every time. |
| |
|
| | # 🎯 Refined Roleplay, Recalibrated Goals |
| | - MiniMaid-L3 isn’t trying to be the fastest. She’s here to be real — holding character, deepening immersion, and generating stories that linger. |
| | - 🛠️ Designed For: |
| | - Narrative-focused deployments |
| | - Long-form interaction and memory retention |
| | - Low-size, high-fidelity simulation |
| | --- |
| | > “MiniMaid-L3 sacrifices a bit of speed to speak with soul. She’s no longer just reacting — she’s inhabiting. It’s not about talking faster — it’s about meaning more.” |
| | # MiniMaid-L3 is the slow burn that brings the fire. |
| | --- |
| |
|
| | - # Notice |
| | - **For a Good Experience, Please use** |
| | - Low temperature 1.5, min_p = 0.1 and max_new_tokens = 128 |
| | |
| | |
| | - # Detail card: |
| | - Parameter |
| | - 1 Billion Parameters |
| | - (Please visit your GPU Vendor if you can Run 1B models) |
| | |
| | - Finetuning tool: |
| | - Unsloth AI |
| | - This llama model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library. |
| | [<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth) |
| | - Fine-tuned Using: |
| | - Google Colab |