ℹ️ Short Bio

Hi, I’m Yuanzhe, a master student at University of Califronia, San Diego. I have the great honor of being collaborating with Prof. Yaoqing Yang, Prof. Julian McAuley, Prof. Zhiting Hu and Dr. Ren Pu.

My current research is focused on

  • Understanding the mechanisms, dynamics and generalization of LLM and SciML models via mathematical analysis.
  • Memory and Reasoning in LLM & Agent.

📖 Educations

University of California, San Diego (UCSD)
M.S. in Computer Science and Engineering
2024.09 - 2026.03 (Expected)
Huazhong University of Science and Technology (HUST)
B.S. in Artificial Intelligence, Innovation Experimental Honor Class, Qiming School
GPA: 3.91/4.0
2020.09 - 2024.06

📝 Writing Samples

(# denotes equal contribution)

First Authored

ICML 2025
sym

[1] Eigenspectrum Analysis of Neural Networks without Aspect Ratio Bias

Yuanzhe Hu, Kinshuk Goel, Vlad Killiakov, Yaoqing Yang

ICML 2025

Short Summary: FARMS, a method for normalizing weight matrices by subsampling with a fixed aspect ratio based on M-P Law, improves the accuracy of eigenspectrum analysis and layer-wise hyperparameter assignment across various domains, including image classification, scientific machine learning, and large language model pruning.

Code | Paper and Review | ICML Website | Star Count

EMNLP 2024
sym

[2] Model Balancing Helps Low-data Training and Fine-tuning

{Zihang Liu#, Yuanzhe Hu#}, Tianyu Pang, Yefan Zhou, Pu Ren, Yaoqing Yang

EMNLP 2024 , Oral (168/6105=2.75%), Meta Review OA=5.0

Short Summary: This paper introduce a modified layer-wise learning rate scheduler, improves low-data training and fine-tuning performance in both NLP and SciML by balancing training quality across model layers.

Code | Paper | Video | Review | Star Count

Under Review
sym

[3] Evaluating Memory in LLM Agents via Incremental Multi-Turn Interactions

{Yuanzhe Hu#, Yu Wang#}, Julian McAuley

Under Review / ICML 2025 LCFM Workshop

Short Summary: MemoryAgentBench is a new benchmark designed to evaluate four core competencies of memory agents in Large Language Models, highlighting the need for improved memory mechanisms.

HF Dataset | Code | Paper | Star Count

Contributed

ICML 2025
sym

[4] M+: Extending MemoryLLM with Scalable Long-Term Memory

Yu Wang, Dmitry Krotov, Yuanzhe Hu, Yifan Gao, Wangchunshu Zhou, Julian McAuley, Dan Gutfreund, Rogerio Feris, Zexue He

ICML 2025

Short Summary: M+ enhances long-term information retention in large language models by integrating a retriever-based long-term memory mechanism, outperforming MemoryLLM and other baselines in knowledge retention tasks.

机器之心 | HF Model | Code | Paper | Review | Star Count

🔥 News

  • 2025.07:  😁 We open-sourced the MemoryAgentBench. Thanks for the great help from Yu Wang!
  • 2025.05:  🎉🎉 Two papers are accepted by ICML 2025 as Poster! See you at Vancouver.
  • 2024.09:  🎉🎉 Excited to share that our work “Model Balancing Helps Low-data Training and Fine-tuning” is accepted by EMNLP 2024 as Oral Presentation!
  • 2024.06:  😁 I graduated from HUST!
  • 2024.06:  😄 I created my account on OpenReview!

Last Update: 09/2025