ℹ️ Short Bio

Hi, I’m Yuanzhe, a second year master student at University of Califronia, San Diego 🔱.

I have the great honor of being collaborating with Prof. Yaoqing Yang from CS@Dartmouth College, Julian McAuley from CSE@UC San Diego, Zhiting Hu HDSI@UC San Diego.

My current research is focused on

My research leverages mathematical insights into LLMs to develop efficient algorithms, while simultaneously unlocking advanced memory and reasoning capabilities in LLMs and Agents. Currently, I am actively seeking for 26 Fall PhD Positions, industrial research internship after M.S graduation (about six months), and research collobration opportunities . Feel free to reach out!

🔥 News

  • 2026.01:  🎉🎉 Our paper “Evaluating Memory in LLM Agents via Incremental Multi-Turn Interactions” was accepted by ICLR 2026.
  • 2025.09:  😁 Excited to share that our recent work “K2-Think: A Parameter-Efficient Reasoning System”.
  • 2025.07:  😁 We open-sourced the MemoryAgentBench. Thanks for the great help from Yu Wang!
  • 2025.05:  🎉🎉 Two papers are accepted by ICML 2025 as Poster! See you at Vancouver.
  • 2024.09:  🎉🎉 Excited to share that our work “Model Balancing Helps Low-data Training and Fine-tuning” is accepted by EMNLP 2024 as Oral Presentation!
  • 2024.06:  😁 I graduated from HUST!
  • 2024.06:  😄 I created my account on OpenReview!

📖 Educations

University of California, San Diego (UCSD)
M.S. in Computer Science and Engineering
2024.09 - 2026.03 (Expected)
Huazhong University of Science and Technology (HUST)
B.S. in Artificial Intelligence, Innovation Experimental Honor Class, Qiming School
GPA: 3.91/4.0
2020.09 - 2024.06

⚙️ Research Project

📖 Mathematical Analysis and Optimization on LLMs and SciML Models

# denotes equal contribution

ICML 2025
sym

Eigenspectrum Analysis of Neural Networks without Aspect Ratio Bias

Yuanzhe Hu, Kinshuk Goel, Vlad Killiakov, Yaoqing Yang

ICML 2025

Short Summary: A layer-wise LLM pruning method inspired by Marchenko–Pastur (MP) law.

Paper | Video | Review

Star Count

EMNLP 2024
sym

Model Balancing Helps Low-data Training and Fine-tuning

{Zihang Liu#, Yuanzhe Hu#}, Tianyu Pang, Yefan Zhou, Pu Ren, Yaoqing Yang

EMNLP 2024 , Oral (168/6105=2.75%), Meta Review OA=5.0

Short Summary: Learning rate scheduler for LLM fine-tuning on low-source dataset.

Paper | Video | Review

Star Count

Under Review
sym

Unveiling Multi-regime Patterns in SciML: Distinct Failure Modes and Regime-specific Optimization

{Yuanzhe Hu#, Xiaopeng Wang#, Yuxin Wang#, Xiaokun Zhong#}, Haiquan Lu, Tianyu Pang, Michael W. Mahoney,Yujun Yan, Pu Ren, Yaoqing Yang

Under Review

Short Summary: We propose a regime-based diagnostic framework using loss-landscape analysis to characterize the unique, pathological failure modes of scientific machine learning (SciML) models and identify regime-specific optimization strategies.

Under Review
sym

Spectral Signatures of Large Language Models

{Zhuoying Zhang#, Ishan Verma Prasad#, Yuanzhe Hu#}, Zihang Liu, Hengrui Luo, Pu Ren, Yaoqing Yang

Under Review

Short Summary: We use the shape information of the weight empirical spectral density as a compact spectral signature of each model.

🤔 Enhancing Memory and Reasoning in LLMs and Agents

MIRIX
sym

MIRIX: Multi-Agent Memory System for LLM-Based Agents

My Contribution: Designed the framework for MIRIX’s Evaluation, project maintenance and bug solving.

Open-Source Project, 3K+ 🌟 stars

Website Star Count Fork Count

ICLR 2026
sym

Evaluating Memory in LLM Agents via Incremental Multi-Turn Interactions

{Yuanzhe Hu#, Yu Wang#}, Julian McAuley

ICLR 2026

Short Summary: MemoryAgentBench is a new benchmark designed to comprehensively evaluate memory agents in LLMs.

Paper

Star Count HF Dataset Dataset Downloads

Under Review
sym

Benchmarking Agent Memory in Interdependent Multi-Session Agentic Tasks

{Zexue He#, Yu Wang#, Churan Zhi#, Yuanzhe Hu#, Tzu-Ping Chen#}, Lang Yin, Ze Chen, Tong Arthur Wu, Siru Ouyang, Zihan Wang, Jiaxin Pei, Julian McAuley, Yejin Choi, Alex Pentland

Under Review

Short Summary: We present MemoryAreana, a new evaluation gym designed to bridge the gap between isolated recall and execution by benchmarking agents on tasks where memory acquisition and action are tightly coupled.

ICML 2025
sym

M+: Extending MemoryLLM with Scalable Long-Term Memory

Yu Wang, Dmitry Krotov, Yuanzhe Hu, Yifan Gao, Wangchunshu Zhou, Julian McAuley, Dan Gutfreund, Rogerio Feris, Zexue He

ICML 2025

Short Summary: M+ enhances long-term information retention in LLMs by integrating a retriever-based long-term memory mechanism.

Paper | Review

Star Count Model Model Downloads 机器之心

Under Review
sym

Mem-$\alpha$: Learning Memory Construction via Reinforcement Learning

Yu Wang, Ryuichi Takanobu, Zhiqi Liang, Yuzhen Mao, Yuanzhe Hu, Julian McAuley, Xiaojian Wu

Under Review

Short Summary: Mem-alpha, a reinforcement learning framework, enhances memory management in LLMs through interaction and feedback.

Paper

Star Count 量子位 机器之心

Tech Report
sym

K2-Think: A Parameter-Efficient Reasoning System

Zhoujun Cheng, Richard Fan, Shibo Hao, Taylor W. Killian, Haonan Li, Suqi Sun, Hector Ren, Alexander Moreno, Daqian Zhang, Tianjun Zhong, Yuxin Xiong, Yuanzhe Hu, Yutao Xie, Xudong Han, Yuqi Wang, Varad Pimpalkhute, Yonghao Zhuang, Aaryamonvikram Singh, Xuezhi Liang, Anze Xie, Jianshu She, Desai Fan, Chengqian Gao, Liqun Ma, Mikhail Yurochkin, John Maggs, Xuezhe Ma, Guowei He, Zhiting Hu, Zhengzhong Liu, Eric P. Xing

MBZUAI IFM / LLM 360 Tech Report

Short Summary: K2-Think is a parameter-efficient reasoning system based on a 32B model.

Paper

Model Model Downloads NY Times Forbes