πŸ‘‹ About Me

Hi! I am a second-year master’s student at Tsinghua University, under the supervision of Prof. Xiu Li. I received my bachelor’s degree with honors from Shandong University in June 2023.

I have been fortunate to collaborate with exceptional researchers who have generously shared their guidance and insights. Currently, I am a research intern at Large Model Center, Shanghai AI Laboratory, advised by Dr. Biqing Qi. Previously, I interned at Intelligent Photonics and Electronics Center (IPEC), Shanghai AI Laboratory, advised by Dr. Chenjia Bai. Before that, I was a research intern at Peking University, advised by Prof. Yali Du and Prof. Yaodong Yang.

Research Interests: My research centers around Large Language Models (LLMs) and Reinforcement Learning (RL). Specifically, I am interested in:

  • Reasoning Capabilities: Enhancing the reasoning and generalization abilities of LLMs and Multi-modal LLMs (MLLMs), from both training-time and test-time perspectives.
  • Efficiency: Improving the training and inference efficiency of LLMs while maintaining or enhancing performance.
  • Evaluation: Developing more reliable and comprehensive evaluation methods to better assess LLM performance across diverse scenarios.
  • LLM4RL: Leveraging the power of LLMs/MLLMs to improve RL algorithms in embodied AI tasks, particularly in the context of reward design and RL from Human/AI Feedback (RLHF/RLAIF).

If you are interested in collaboration, please feel free to reach out via e-mail!

🌟 News

  • [2025.03] Β πŸŽ‰ One paper accepted by Reasoning and Planning for LLMs Workshop @ ICLR 2025
  • [2025.02] Β πŸ”₯ Preprint Can 1B LLM Surpass 405B LLM? Rethinking Compute-Optimal Test-Time Scaling released on arXiv (Project Page)
  • [2025.01] Β πŸŽ‰ One paper accepted by ICLR 2025
  • [2024.12] Β πŸŽ‰ One paper accepted by AAAI 2025 and selected for oral presentation (Top 4.6%)
  • [2024.05] Β πŸŽ‰ One paper accepted by ICML 2024
  • [2024.01] Β πŸŽ‰ One paper accepted by ICLR 2024
  • [2023.10] Β πŸŽ‰ One paper accepted by OTML Workshop @ NeurIPS 2023
  • [2022.09] Β πŸŽ‰ One paper accepted by NeurIPS 2022

πŸ“ Publications

(* indicates equal contribution)

Preprints

Conference Papers

Workshop Papers

πŸŽ“ Education

πŸŽ– Honors and Awards

  • National Scholarship (Top 1%), 2022.12
  • National Scholarship (Top 1%), 2021.12
  • First Prize in China Undergraduate Mathematical Contest in Modeling (CUMCM) (Top 0.65%), 2021.11
  • Outstanding Student of Shandong Province (Top 0.6%), 2022.05
  • Outstanding Graduate of Shandong Province (Top 6%), 2023.04
  • Dishang Scholarship, 2022.10

πŸ’» Internships

πŸŽ™ Invited Talks

  • Can 1B LLM Surpass 405B LLM? Rethinking Compute-Optimal Test-Time Scaling. Xiaohongshu. 2025.02.
  • Can 1B LLM Surpass 405B LLM? Rethinking Compute-Optimal Test-Time Scaling. Huawei Noah’s Ark Lab. 2025.03.

πŸ› οΈ Services

  • Conference Reviewer: NeurIPS (2024), ICLR (2025), ICML (2025), AAMAS (2024), AISTATS (2025), ECAI (2024)
  • Workshop Reviewer: NeurIPS OTML (2023)