Large language models (LLMs) are essential for various applications but require significant computational power. Traditional pruning methods don't maintain performance across diverse tasks. Researchers from Apple AI and UC Santa Barbara developed Instruction-Following Pruning (IFPruning), a dynamic approach that adapts LLMs to
Sort: