KVAE 2.0 Collection KVAE 2.0 is a family of video tokenizers with a time compression ratio of 4 and spacial compression ratio of 8 and 16 β’ 2 items β’ Updated 7 days ago β’ 2
Interpreting CLIP with Hierarchical Sparse Autoencoders Paper β’ 2502.20578 β’ Published Feb 27, 2025 β’ 1
SOM Directions are Better than One: Multi-Directional Refusal Suppression in Language Models Paper β’ 2511.08379 β’ Published Nov 11, 2025 β’ 4
Effective Reasoning Chains Reduce Intrinsic Dimensionality Paper β’ 2602.09276 β’ Published Feb 9 β’ 11
LLaDA2.1: Speeding Up Text Diffusion via Token Editing Paper β’ 2602.08676 β’ Published Feb 9 β’ 71
AudioSAE: Towards Understanding of Audio-Processing Models with Sparse AutoEncoders Paper β’ 2602.05027 β’ Published Feb 4 β’ 63
GDPO: Group reward-Decoupled Normalization Policy Optimization for Multi-reward RL Optimization Paper β’ 2601.05242 β’ Published Jan 8 β’ 230
Cross-Frame Representation Alignment for Fine-Tuning Video Diffusion Models Paper β’ 2506.09229 β’ Published Jun 10, 2025 β’ 7
LTX-2: Efficient Joint Audio-Visual Foundation Model Paper β’ 2601.03233 β’ Published Jan 6 β’ 177
Gamayun's Path to Multilingual Mastery: Cost-Efficient Training of a 1.5B-Parameter LLM Paper β’ 2512.21580 β’ Published Dec 25, 2025 β’ 8
PanGu-Ξ£: Towards Trillion Parameter Language Model with Sparse Heterogeneous Computing Paper β’ 2303.10845 β’ Published Mar 20, 2023 β’ 3
Unveiling Intrinsic Dimension of Texts: from Academic Abstract to Creative Story Paper β’ 2511.15210 β’ Published Nov 19, 2025 β’ 91
One-Step Residual Shifting Diffusion for Image Super-Resolution via Distillation Paper β’ 2503.13358 β’ Published Mar 17, 2025 β’ 95
Feature-Level Insights into Artificial Text Detection with Sparse Autoencoders Paper β’ 2503.03601 β’ Published Mar 5, 2025 β’ 233
LLM-Microscope: Uncovering the Hidden Role of Punctuation in Context Memory of Transformers Paper β’ 2502.15007 β’ Published Feb 20, 2025 β’ 175
Grass: Compute Efficient Low-Memory LLM Training with Structured Sparse Gradients Paper β’ 2406.17660 β’ Published Jun 25, 2024 β’ 5
An Image is Worth 32 Tokens for Reconstruction and Generation Paper β’ 2406.07550 β’ Published Jun 11, 2024 β’ 60
Make It Count: Text-to-Image Generation with an Accurate Number of Objects Paper β’ 2406.10210 β’ Published Jun 14, 2024 β’ 78