Language Models Without a Trainable Input Embedding Table Collection This collection is provided for reproducibility of the paper's main claim • 3 items • Updated 1 day ago
Bochkov/llm-fix-min-affine-recoded-minimal-code-table-free Text Generation • 0.5B • Updated 1 day ago • 13
Bochkov/llm-fix-min-affine-recoded-minimal-code-table-free Text Generation • 0.5B • Updated 1 day ago • 13
Bochkov/llm-fix-min-baseline-learned-input-table-model-classic Text Generation • 0.5B • Updated 1 day ago • 20
Bochkov/llm-fix-min-baseline-learned-input-table-model-classic Text Generation • 0.5B • Updated 1 day ago • 20
view post Post 3112 New REPL environment in OpenEnv available! ✨Used in the Recursive Language Models (RLM) paper by Alex Zhang.Ready for inference & post-training using trajectories. Handles long contexts:> Run Python code in a sandbox> Make recursive calls to LMs> Explore data programmatically> Return final resultDocs: https://meta-pytorch.org/OpenEnv/environments/repl/Inference script: https://github.com/meta-pytorch/OpenEnv/blob/main/examples/repl_oolong_simple.py See translation 🔥 10 10 + Reply
NVIDIA Nemotron Nano 2: An Accurate and Efficient Hybrid Mamba-Transformer Reasoning Model Paper • 2508.14444 • Published Aug 20, 2025 • 48
Bochkov/growing-transformers-model-frozen-16-bit-baseline-monolyth-181m Text Generation • Updated Jan 9 • 33
Bochkov/growing-transformers-model-unfrozen-baseline-monolyth-247m Text Generation • Updated Jan 9 • 5
Bochkov/growing-transformers-model-frozen-unicode-baseline-monolyth-247m Text Generation • Updated Jan 9 • 1