Search:
Match:
1 results
Paper#LLM🔬 ResearchAnalyzed: Jan 3, 2026 16:22

Width Pruning in Llama-3: Enhancing Instruction Following by Reducing Factual Knowledge

Published:Dec 27, 2025 18:09
1 min read
ArXiv

Analysis

This paper challenges the common understanding of model pruning by demonstrating that width pruning, guided by the Maximum Absolute Weight (MAW) criterion, can selectively improve instruction-following capabilities while degrading performance on tasks requiring factual knowledge. This suggests that pruning can be used to trade off knowledge for improved alignment and truthfulness, offering a novel perspective on model optimization and alignment.
Reference

Instruction-following capabilities improve substantially (+46% to +75% in IFEval for Llama-3.2-1B and 3B models).