Hugging Face
Models
Datasets
Spaces
Buckets
new
Docs
Enterprise
Pricing
Log In
Sign Up
36.9
TFLOPS
16
6
27
ManniX
PRO
ManniX-ITA
Follow
sebdg's profile picture
Fernanduli's profile picture
kingsenglishvietnam's profile picture
33 followers
ยท
18 following
https://github.com/mann1x
mann1x
AI & ML interests
None yet
Recent Activity
reacted
to
danielhanchen
's
post
with ๐ค
about 18 hours ago
We collaborated with NVIDIA to teach you how we made LLM training ~25% faster! ๐ Learn how 3 optimizations help your home GPU train models faster: 1. Packed-sequence metadata caching 2. Double-buffered checkpoint reloads 3. Faster MoE routing Guide: https://unsloth.ai/blog/nvidia-collab GitHub: https://github.com/unslothai/unsloth
posted
an
update
2 days ago
After the Feb/Mar '26 collapse of Claude Code I started building my own framework. The token crunch is still only mitigated, but the reasoning and quality are back โ better than before. For research and LLM training recipes, though, diverse knowledge and a second point of view are crucial. Pairing my Claude Max sub with an Ollama Pro sub has already saved me from days of botched trainings โ multiple frontier models helping Claude is next level. Acting as the middleman myself was interesting but inefficient, so I shipped skills that let Claude talk to Ollama models directly. ๐ claude-hooks v1.1.0 ships two LLM-to-LLM skills. ๐ฌ /get-advice โ single-shot second opinion. Claude runs a multi-turn conversation with a configured Ollama advisor; the advisor grounds in your project through read_file / grep / glob / list_files / recall_memory tools. Effort tiers cap fresh-session retries. ๐ค /consultants โ multi-agent council for cross-cutting questions: ๐งฉ planner โ researcher โ critic โ synthesizer Each role runs its own Ollama model. ๐พ Sessions persist to disk (summary.md + transcript.md + SQLite per-role message threads); closed sessions reopen and produce follow-ups ๐ indistinguishable from warm ones. ๐ฏ x-tier effort multiplies diversity: โข xmedium / xhigh โ researcher fans across N models in parallel โข xmax โ + multi-critic + meta-critic combine; critics anonymized as "Critic 1/2/3" to avoid model-bias ๐ก๏ธ Cloud-flap recovery, three layers: 15-attempt / ~15min retry budget; synthesizer failure-fallback model chain; degraded-answer composer surfaces researcher + critic work even when synthesis fails. ๐ 7 cloud models benchmarked & Claude-graded on locked queries: โข PROD-READY (P:A R:A C:A S:A): kimi-k2.6, gemma4:31b, glm-5.1 โข Role specialists: minimax-m2.7 (critic), qwen3.5 (planner) ๐ง Linux/macOS/Windows. No per-project setup. ๐ github.com/mann1x/claude-hooks
reacted
to
mlabonne
's
post
with ๐
5 days ago
Big update to llm-datasets, my curated list of datasets and tools for post-training LLMs. > Added many new datasets > New "thinking" column > Refreshed recommended tools. Thanks to everyone who told me they used it for their research at ICLR, you motivated this update!
View all activity
Organizations
None yet
ManniX-ITA
's datasets
1
Sort:ย Recently updated
ManniX-ITA/osync-code
Viewer
โข
Updated
Jan 12
โข
1
โข
9