DistilBERT release Collection Original DistilBERT model, checkpoints obtained from using teacher-student learning from the original BERT checkpoints. • 6 items • Updated Apr 17, 2024 • 43
view post Post 2784 Qwen3.6-35B-A3B can now be run locally! 💜The model is the strongest mid-sized LLM on nearly all benchmarks.Run on 23GB RAM via Unsloth Dynamic GGUFs.GGUFs to run: unsloth/Qwen3.6-35B-A3B-GGUFGuide: https://unsloth.ai/docs/models/qwen3.6 See translation 13 replies · 😎 8 8 🔥 7 7 🤗 1 1 + Reply
ChartNet: A Million-Scale, High-Quality Multimodal Dataset for Robust Chart Understanding Paper • 2603.27064 • Published Mar 28 • 27
The Past Is Not Past: Memory-Enhanced Dynamic Reward Shaping Paper • 2604.11297 • Published 16 days ago • 138
QuanBench+: A Unified Multi-Framework Benchmark for LLM-Based Quantum Code Generation Paper • 2604.08570 • Published Mar 25 • 124
Running Featured 206 Gemma 4 WebGPU 🚀 206 Run Gemma 4 locally in-browser on WebGPU w/ Transformers.js
Scaling Test-Time Compute with Open Models Collection Models and datasets used in our blog post: https://huggingface.co/spaces/HuggingFaceH4/blogpost-scaling-test-time-compute • 10 items • Updated Jan 6, 2025 • 30
view article Article Welcome Gemma 4: Frontier multimodal intelligence on device +5 27 days ago • 882
Jackrong/Qwen3.5-27B-Claude-4.6-Opus-Reasoning-Distilled Image-Text-to-Text • 28B • Updated 23 days ago • 470k • 2.8k