Fanxu Meng(孟繁续)

I am a fourth-year Ph.D. student at the Institute for Artificial Intelligence, Peking University, advised by Prof. Muhan Zhang. My research focuses on parameter-efficient fine-tuning of large language models (LLMs) and efficient inference for long-context LLMs. I have served as a reviewer for leading conferences and journals, including NeurIPS, ICML, ICLR, CVPR, TPAMI, COLM, AAAI, and IJCAI. Prior to joining Peking University, I received my Master’s degree from Harbin Institute of Technology, Shenzhen, where I was advised by Prof. Guangming Lu. I also spent over two years at Tencent YouTu as an intern and later as a full-time researcher, collaborating with Xing Sun, Hao Cheng, Ke Li, and Di Yin.
[ASPLOS'26 - Summer cycle] TPLA: Tensor Parallel Latent Attention
An Attention Mechanism Friendly to Tensor Parallelism and PD Decoupling, (Paper)
Xiaojuan Tang*, Fanxu Meng*, Pingzhi Tang, Yuxuan Wang, Di Yin, Xing Sun, Muhan Zhang.