Publications

ShortGPT: Layers in Large Language Models are More Redundant Than You Expect
ShortGPT: Layers in Large Language Models are More Redundant Than You Expect

We investigate the redundancy within Transformer layers and propose an effective layer-based pruning method.

Base of RoPE Bounds Context Length
Base of RoPE Bounds Context Length

This work contributes to the investigation of the lower bounds of the Base in RoPE, providing a theoretical foundation for the long-context extrapolation of models.