Qingyu Zhang

Qingyu Zhang

Master Student of Computer Science and Technology

Institute of Software Chinese Academy of Sciences

Biography

I’m Qingyu Zhang, a first-year master’s student at Chinese Information Processing Laboratory in the Institute of Software Chinese Academy of Sciences. My research focuses on large language models, including long-text capabilities and multi-turn dialogue abilities.

Interests
  • LLM Long Context
  • LLM Compression & Efficiency
  • LLM Post-training
  • LLM Reinforcement Learning
Education
  • M.S. in Computer Science and Technology, 2024 - Present

    Institute of Software Chinese Academy of Sciences

  • B.S. in Computer Science and Technology, 2020 - 2024

    College of Computer and Data Science, Fuzhou University

News

  • May, 2025 One paper “ShortGPT” is accepted by ACL Findings 2025.
  • Sep, 2024 One paper “Base of RoPE Bounds Context Length” is accepted by NeurIPS 2024.
  • Jun, 2024 Honored as an Outstanding Graduate at Fuzhou University.
  • May, 2023 Won the First Prize in the 10th ASC Student Supercomputer Challenge.
  • Nov, 2022 Won the First Prize in the 13th National College Student Mathematics Competition.

Experience

 
 
 
 
 
Algorithm Intern
December 2024 – Present Beijing, China
  • Led the R&D of an RL-based dialogue optimization system for large models.
  • Deployed in a live business environment, increasing core business conversion rate by ~20%.
  • Research submitted to AAAI 2026.
 
 
 
 
 
Foundation Model Intern
January 2024 – October 2024 Beijing, China
  • Investigated Transformer redundancy and proposed a layer-based pruning method (ShortGPT, ACL Findings, 2025).
  • Researched the lower bounds of RoPE Base (Base of RoPE Bounds Context Length, NeurIPS, 2024).
  • Proposed a variant of the “Needle in a Haystack” evaluation method (Patent Granted).
 
 
 
 
 
Research Intern
October 2023 – September 2024 Beijing, China
  • Adapted and optimized SFT/DPO algorithms for the Megatron framework (ACL Demo, 2025).
  • Implemented large-scale distributed training on Ascend 910b using the ModelLink framework.

Recent Publications

Here are some of my recent publications. You can find the full list in my CV.
(2025). ShortGPT: Layers in Large Language Models are More Redundant Than You Expect. In ACL Findings 2025.

PDF Cite

(2024). Base of RoPE Bounds Context Length. In NeurIPS 2024.

PDF Cite

Contact