Efficient Hybrid Language Model Compression through Group-Aware SSM Pruning Paper • 2504.11409 • Published 23 days ago • 10
Hymba: A Hybrid-head Architecture for Small Language Models Paper • 2411.13676 • Published Nov 20, 2024 • 45
MaskLLM: Learnable Semi-Structured Sparsity for Large Language Models Paper • 2409.17481 • Published Sep 26, 2024 • 48
Minitron Collection A family of compressed models obtained via pruning and knowledge distillation • 12 items • Updated 3 days ago • 61
LLM Pruning and Distillation in Practice: The Minitron Approach Paper • 2408.11796 • Published Aug 21, 2024 • 59