CoIn: Counting the Invisible Reasoning Tokens in Commercial Opaque LLM APIs Paper • 2505.13778 • Published 15 days ago • 4
view article Article From DeepSpeed to FSDP and Back Again with Hugging Face Accelerate By muellerzr and 3 others • Jun 13, 2024 • 54
LLM-Drop Collection Model weights of paper "What Matters in Transformers? Not All Attention is Needed" (https://arxiv.org/abs/2406.15786) • 14 items • Updated Oct 23, 2024 • 4
What Matters in Transformers? Not All Attention is Needed Paper • 2406.15786 • Published Jun 22, 2024 • 32
🪐 SmolLM Collection A series of smol LLMs: 135M, 360M and 1.7B. We release base and Instruct models as well as the training corpus and some WebGPU demos • 12 items • Updated 29 days ago • 227