About me

Hi! I’m Luning Wang (王麓宁), currently a Master’s student at the University of Michigan. Before that, I got my Bachelor’s degree at Tsinghua University.

I’m now actively looking for full time work opportunities in the field of LLMs, MLSys, and potentially other AI-related fields. I’m expected to graduate in 05/2026 and planning to work in China Mainland or Hongkong or Singapore. Feel free to contact with me via Email if there’s an opportunity!

🎓 Education

  • [08/2024~05/2026] M.S. Department of Electrical and Computer Engineering, University of Michigan
  • [09/2020~06/2024] B.Eng. Department of Electronic Engineering, Tsinghua University

💻 Internship

I’ve had several internship experiences. See my CV for more details of my work.

📖 Research

I mainly focused on infrastructure and efficiency optimization of Large Language Models (a.k.a AI-Infra / MLSys) in my past research, including the compression and acceleration techniques of LLMs.

Aside from that, I also have some experiences on LLM reasoning, LLM agents, Biomedical LLMs, etc.

I’m open to discussion and collaboration. Feel free to drop me an Email or send me a message on LinkedIn!

📝 Selected Publications

Here are some of my representative works:

  • [NeurIPS ENLSP Workshop’24] CSKV: Training-Efficient Channel Shrinking for KV Cache in Long-Context Scenarios. Luning Wang, Shiyao Li, Xuefei Ning, Zhihang Yuan, Shengen Yan, Guohao Dai, Yu Wang. [pdf] [github]
  • [ICML’24] Evaluating Quantized Large Language Models. Shiyao Li, Xuefei Ning, Luning Wang, Tengxuan Liu, Xiangsheng Shi, Shengen Yan, Guohao Dai, Huazhong Yang, Yu Wang. [pdf] [github]
  • [NeurIPS ENLSP Workshop’23] LLM-MQ: Mixed-precision Quantization for Efficient LLM Deployment. Shiyao Li, Xuefei Ning, Ke Hong, Tengxuan Liu, Luning Wang, Xiuhong Li, Kai Zhong, Guohao Dai, Huazhong Yang, Yu Wang. [pdf]

See my Google Scholar for the full list of my publications.

⚙️ Service

  • [02/2025] Paper reviewer at the ICLR 2025 Workshop on Reasoning and Planning for LLMs.