Does Reinforcement Learning Really Incentivize Reasoning Capacity in LLMs Beyond the Base Model? Paper β’ 2504.13837 β’ Published 20 days ago β’ 119
Gemma 3 Collection All versions of Google's new multimodal models including QAT in 1B, 4B, 12B, and 27B sizes. In GGUF, dynamic 4-bit and 16-bit formats. β’ 50 items β’ Updated 6 days ago β’ 58
Running 2.56k 2.56k The Ultra-Scale Playbook π The ultimate guide to training LLM on large GPU Clusters
view article Article Fine-tuning LLMs to 1.58bit: extreme quantization made easy Sep 18, 2024 β’ 242
Running 118 118 Open-LLM performances are plateauing, letβs make the leaderboard steep again π Update leaderboard for fair model evaluation