Maxime Labonne PRO
AI & ML interests
Post-training, model editing, quantization
Recent Activity
repliedto their post 1 day ago
Big update to llm-datasets, my curated list of datasets and tools for post-training LLMs.
> Added many new datasets
> New "thinking" column
> Refreshed recommended tools.
Thanks to everyone who told me they used it for their research at ICLR, you motivated this update! liked a dataset 4 days ago
aladinDJ/ultramix-DPO-annotatedOrganizations
replied to their post 1 day ago
reacted to Tonic's post with 🔥 2 months ago
Post
3672
🤔 Who would win ?
- a fully subsidized ai lab
- 3 random students named
kurakurai ?
demo : Tonic/fr-on-device
if you like it give the demo a little star and send a shoutout to : @MaxLSB @jddqd and @GAD-cell for absolutely obliterating the pareto frontier of the french language understanding .
- a fully subsidized ai lab
OR - 3 random students named
demo : Tonic/fr-on-device
if you like it give the demo a little star and send a shoutout to : @MaxLSB @jddqd and @GAD-cell for absolutely obliterating the pareto frontier of the french language understanding .
replied to their post 4 months ago
32k here. 1B models are really not good with long context (yet), unfortunately.
Post
10315
New family of 1B models just dropped!
> LiquidAI/LFM2.5-1.2B-Base: 10T → 28T tokens
> LiquidAI/LFM2.5-1.2B-Instruct: new large-scale multi-stage RL
> LiquidAI/LFM2.5-1.2B-JP: our most polite model
> LiquidAI/LFM2.5-VL-1.6B: multi-image multilingual
> LiquidAI/LFM2.5-Audio-1.5B: 8x times faster, no quality loss
Super proud of this release 🤗
> LiquidAI/LFM2.5-1.2B-Base: 10T → 28T tokens
> LiquidAI/LFM2.5-1.2B-Instruct: new large-scale multi-stage RL
> LiquidAI/LFM2.5-1.2B-JP: our most polite model
> LiquidAI/LFM2.5-VL-1.6B: multi-image multilingual
> LiquidAI/LFM2.5-Audio-1.5B: 8x times faster, no quality loss
Super proud of this release 🤗
replied to their post 4 months ago
posted an update 4 months ago
Post
10315
New family of 1B models just dropped!
> LiquidAI/LFM2.5-1.2B-Base: 10T → 28T tokens
> LiquidAI/LFM2.5-1.2B-Instruct: new large-scale multi-stage RL
> LiquidAI/LFM2.5-1.2B-JP: our most polite model
> LiquidAI/LFM2.5-VL-1.6B: multi-image multilingual
> LiquidAI/LFM2.5-Audio-1.5B: 8x times faster, no quality loss
Super proud of this release 🤗
> LiquidAI/LFM2.5-1.2B-Base: 10T → 28T tokens
> LiquidAI/LFM2.5-1.2B-Instruct: new large-scale multi-stage RL
> LiquidAI/LFM2.5-1.2B-JP: our most polite model
> LiquidAI/LFM2.5-VL-1.6B: multi-image multilingual
> LiquidAI/LFM2.5-Audio-1.5B: 8x times faster, no quality loss
Super proud of this release 🤗
posted an update 7 months ago
Post
8429
LiquidAI/LFM2-8B-A1B just dropped!
8.3B params with only 1.5B active/token 🚀
> Quality ≈ 3–4B dense, yet faster than Qwen3-1.7B
> MoE designed to run on phones/laptops (llama.cpp / vLLM)
> Pre-trained on 12T tokens → strong math/code/IF
8.3B params with only 1.5B active/token 🚀
> Quality ≈ 3–4B dense, yet faster than Qwen3-1.7B
> MoE designed to run on phones/laptops (llama.cpp / vLLM)
> Pre-trained on 12T tokens → strong math/code/IF
posted an update 7 months ago
Post
3883
⚛️ New drop of tiny task-specific models!
Want to do data extraction, translation, RAG, tool use, or math on a Raspberry Pi? We got you covered! ✅
These tiny models were fine-tuned to perform narrow tasks extremely well, making them competitive with much larger models.
You can deploy them today on-device or even on GPUs for big data operations!
LiquidAI/liquid-nanos-68b98d898414dd94d4d5f99a
Want to do data extraction, translation, RAG, tool use, or math on a Raspberry Pi? We got you covered! ✅
These tiny models were fine-tuned to perform narrow tasks extremely well, making them competitive with much larger models.
You can deploy them today on-device or even on GPUs for big data operations!
LiquidAI/liquid-nanos-68b98d898414dd94d4d5f99a
posted an update 9 months ago
Post
6983
Liquid just released two 450M and 1.6B param VLMs!
They're super fast and leverage SigLIP2 NaFlex encoders to handle native resolutions without distortion. It's ideal for on-device deployment in constrained environments like phones.
It's available today on Hugging Face, with an inference and a fine-tuning Colab notebooks.
LiquidAI/LFM2-VL-450M
LiquidAI/LFM2-VL-1.6B
They're super fast and leverage SigLIP2 NaFlex encoders to handle native resolutions without distortion. It's ideal for on-device deployment in constrained environments like phones.
It's available today on Hugging Face, with an inference and a fine-tuning Colab notebooks.
LiquidAI/LFM2-VL-450M
LiquidAI/LFM2-VL-1.6B
reacted to sergiopaniego's post with 🤗 10 months ago
Post
1878
Test SmolLM3, the newest fully open model released by @HuggingFaceTB !
It's smol (3B), multilingual (6 languages), comes with dual mode reasoning (think/no_think modes) and supports long-context (128k).
Try it now in the notebook below!! ⬇️
Colab notebook: https://colab.research.google.com/github/sergiopaniego/samples/blob/main/smollm3_3b_inference.ipynb
notebook: https://github.com/sergiopaniego/samples/blob/main/smollm3_3b_inference.ipynb
blog: https://huggingface.co/blog/smollm3
It's smol (3B), multilingual (6 languages), comes with dual mode reasoning (think/no_think modes) and supports long-context (128k).
Try it now in the notebook below!! ⬇️
Colab notebook: https://colab.research.google.com/github/sergiopaniego/samples/blob/main/smollm3_3b_inference.ipynb
notebook: https://github.com/sergiopaniego/samples/blob/main/smollm3_3b_inference.ipynb
blog: https://huggingface.co/blog/smollm3
posted an update 10 months ago
Post
5762
Based on a new hybrid architecture, these 350M, 700M, and 1.2B models are both fast and performant, ideal for on-device deployment.
I recommend fine-tuning them to power your next edge application. We already provide Colab notebooks to guide you. More to come soon!
📝 Blog post: https://www.liquid.ai/blog/liquid-foundation-models-v2-our-second-series-of-generative-ai-models
🤗 Models: LiquidAI/lfm2-686d721927015b2ad73eaa38
That's super cool, congrats! :)
reacted to drwlf's post with ❤️🤗 11 months ago
Post
5861
Having an insanely good medical LLM is pointless if it won’t answer your questions!
So we’ve made 2 notebook for abliterating any model in order to achieve a good model that will actually help you!
The notebooks are made using @mlabonne ‘s abliteration logic and datasets!
Feel free to use them and happy training 😊
https://github.com/dralexlup/LLM-Abliteration
So we’ve made 2 notebook for abliterating any model in order to achieve a good model that will actually help you!
The notebooks are made using @mlabonne ‘s abliteration logic and datasets!
Feel free to use them and happy training 😊
https://github.com/dralexlup/LLM-Abliteration
replied to their post 11 months ago
Will do at some point, but I don't have time to write this down at the moment.
reacted to burtenshaw's post with 🚀❤️🤗 about 1 year ago
Post
3530
NEW UNIT in the Hugging Face Reasoning course. We dive deep into the algorithm behind DeepSeek R1 with an advanced and hands-on guide to interpreting GRPO.
🔗
reasoning-course
This unit is super useful if you’re tuning models with reinforcement learning. It will help with:
- interpreting loss and reward progression during training runs
- selecting effective parameters for training
- reviewing and defining effective reward functions
This unit also works up smoothly toward the existing practical exercises form @mlabonne and Unsloth.
📣 Shout out to @ShirinYamani who wrote the unit. Follow for more great content.
🔗
This unit is super useful if you’re tuning models with reinforcement learning. It will help with:
- interpreting loss and reward progression during training runs
- selecting effective parameters for training
- reviewing and defining effective reward functions
This unit also works up smoothly toward the existing practical exercises form @mlabonne and Unsloth.
📣 Shout out to @ShirinYamani who wrote the unit. Follow for more great content.
posted an update about 1 year ago
Post
19524
✂️ AutoAbliteration
I made a Colab notebook to automatically abliterate models.
It's quite general, so you can do interesting stuff like blocking a given language in the model outputs.
💻 Colab: https://colab.research.google.com/drive/1RmLv-pCMBBsQGXQIM8yF-OdCNyoylUR1?usp=sharing
I made a Colab notebook to automatically abliterate models.
It's quite general, so you can do interesting stuff like blocking a given language in the model outputs.
💻 Colab: https://colab.research.google.com/drive/1RmLv-pCMBBsQGXQIM8yF-OdCNyoylUR1?usp=sharing