LLM Redundancy

ShortGPT: Layers in Large Language Models are More Redundant Than You Expect
We investigate the redundancy within Transformer layers and propose an effective layer-based pruning method.
ShortGPT: Layers in Large Language Models are More Redundant Than You Expect
ShortGPT: Layers in Large Language Models are More Redundant Than You Expect
We investigate the redundancy within Transformer layers and propose an effective layer-based pruning method.
ShortGPT: Layers in Large Language Models are More Redundant Than You Expect