An implementation of the MoDeGPT LLM compression from the ICLR 2025 Conference paper: Modular Decomposition For Large Language Model Compression.
-
Updated
Apr 3, 2026 - Python
An implementation of the MoDeGPT LLM compression from the ICLR 2025 Conference paper: Modular Decomposition For Large Language Model Compression.
Týr-the-Pruner: Structural Pruning LLMs via Global Sparsity Distribution Optimization (NeurIPS25)
Add a description, image, and links to the sparsity-allocation topic page so that developers can more easily learn about it.
To associate your repository with the sparsity-allocation topic, visit your repo's landing page and select "manage topics."