view article Article ⚡ nano-vLLM: Lightweight, Low-Latency LLM Inference from Scratch By zamal • Jun 28 • 18
mmBERT: a modern multilingual encoder Collection mmBERT is trained on 3T tokens from over 1800 languages, showing SoTA scores on benchmarks and exceptional low-resource performance • 16 items • Updated Sep 9 • 47
mmBERT: A Modern Multilingual Encoder with Annealed Language Learning Paper • 2509.06888 • Published Sep 8 • 12
DINOv3 Collection DINOv3: foundation models producing excellent dense features, outperforming SotA w/o fine-tuning - https://arxiv.org/abs/2508.10104 • 13 items • Updated Aug 21 • 365
MobileLLM-R1 Collection MobileLLM-R1, a series of sub-billion parameter reasoning models • 7 items • Updated 16 days ago • 19