Abstract
We present a comprehensive report on compressing the Llama 3.1 8B and Mistral NeMo 12B models to 4B and 8B parameters, respectively, using pruning and distillation. We explore two distinct pruning strategies: (1) depth pruning and (2) joint hidden/attention/MLP (width) pruning, and evaluate the results on common benchmarks from the LM Evaluation Harness. The models are then aligned with NeMo Aligner and tested in instruct-tuned versions. This approach produces a compelling 4B model from Llama 3.1 8B and a state-of-the-art Mistral-NeMo-Minitron-8B (MN-Minitron-8B for brevity) model from Mistral NeMo 12B. We found that with no access to the original data, it is beneficial to slightly fine-tune teacher models on the distillation dataset. We open-source our base model weights on Hugging Face with a permissive license.
Community
Models are at https://huggingface.co/collections/nvidia/minitron-669ac727dc9c86e6ab7f0f3e
Quick demo of Base models (not Instruct): https://huggingface.co/spaces/nvidia/minitron
Examples and tutorial (WIP): https://github.com/NVlabs/Minitron
Authors are happy to answer your questions.
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- Compact Language Models via Pruning and Knowledge Distillation (2024)
- Gemma 2: Improving Open Language Models at a Practical Size (2024)
- MINI-LLM: Memory-Efficient Structured Pruning for Large Language Models (2024)
- Pruning Large Language Models to Intra-module Low-rank Architecture with Transitional Activations (2024)
- Reconstruct the Pruned Model without Any Retraining (2024)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment:
@librarian-bot
recommend
I have showcased this paper on my blog : https://ajithp.com/2024/08/25/minitron-nvidias-breakthrough-in-llm-efficiency-pruning-and-distillation-for-smaller-faster-ai-models/
We wrote a summary about this paper and a few other papers here. Please do give it a read.
Contents:
- Nvidia MiniTron
- 1.5 Pints
- Jamba-1.5
- FocusLLM
https://datta0.substack.com/p/ai-unplugged-18-minitron-and-llama
Models citing this paper 9
Browse 9 models citing this paperDatasets citing this paper 0
No dataset linking this paper