Jiaqi Leng (冷家祺)

prof_pic.jpg

I am a final-year undergraduate student in Computer Science and Technology at Fudan University. During Fall 2024, I was an exchange student at The University of Texas at Austin.

Currently, I am working as a research intern at NYU Shanghai with Prof. Yucheng Lu, focusing on efficient byte-level modeling. Previously, I worked as a research intern at Ant Group, collaborating with Xiang Hu on efficient attention mechanisms for large language models.

My research interests mainly lie in:

  • Efficient deep learning and model architectures
  • Long-context modeling and length extrapolation
  • Sparse attention mechanisms

Please feel free to reach out! 👋

selected publications

  1. Preprint
    Understanding and Improving Length Generalization in Hierarchical Sparse Attention Models
    arXiv preprint, 2025
  2. NeurIPS
    Random Long-Context Access for Mamba via Hardware-aligned Hierarchical Sparse Attention
    Xiang Hu, Jiaqi Leng, Jun Zhao, Kewei Tu, and Wei Wu
    2025