Running on CPU Upgrade 1.16k 1.16k The Smol Training Playbook: The Secrets to Building World-Class LLMs 📝
view article Article ⚡ nano-vLLM: Lightweight, Low-Latency LLM Inference from Scratch By zamal • Jun 28 • 20
mmBERT: a modern multilingual encoder Collection mmBERT is trained on 3T tokens from over 1800 languages, showing SoTA scores on benchmarks and exceptional low-resource performance • 16 items • Updated Sep 9 • 48
mmBERT: A Modern Multilingual Encoder with Annealed Language Learning Paper • 2509.06888 • Published Sep 8 • 12
DINOv3 Collection DINOv3: foundation models producing excellent dense features, outperforming SotA w/o fine-tuning - https://arxiv.org/abs/2508.10104 • 13 items • Updated Aug 21 • 367
MobileLLM-R1 Collection MobileLLM-R1, a series of sub-billion parameter reasoning models • 7 items • Updated 21 days ago • 21