ℹ️ Short Bio
Hi, I’m Yuanzhe, a master student at University of Califronia, San Diego. I have the great honor of being collaborating with Prof. Yaoqing Yang, Prof. Julian McAuley, Prof. Zhiting Hu and Dr. Ren Pu.
My current research is focused on
- Understanding the mechanisms, dynamics and generalization of LLM and SciML models via mathematical analysis.
- Memory and Reasoning in LLM & Agent.
📖 Educations
![]() |
University of California, San Diego (UCSD) M.S. in Computer Science and Engineering |
2024.09 - 2026.03 (Expected) |
![]() |
Huazhong University of Science and Technology (HUST) B.S. in Artificial Intelligence, Innovation Experimental Honor Class, Qiming School GPA: 3.91/4.0 |
2020.09 - 2024.06 |
📝 Writing Samples
(# denotes equal contribution)
First Authored

[1] Eigenspectrum Analysis of Neural Networks without Aspect Ratio Bias
Yuanzhe Hu, Kinshuk Goel, Vlad Killiakov, Yaoqing Yang
ICML 2025
Short Summary: FARMS, a method for normalizing weight matrices by subsampling with a fixed aspect ratio based on M-P Law, improves the accuracy of eigenspectrum analysis and layer-wise hyperparameter assignment across various domains, including image classification, scientific machine learning, and large language model pruning.
Code | Paper and Review | ICML Website |

[2] Model Balancing Helps Low-data Training and Fine-tuning
{Zihang Liu#, Yuanzhe Hu#}, Tianyu Pang, Yefan Zhou, Pu Ren, Yaoqing Yang
EMNLP 2024 , Oral (168/6105=2.75%), Meta Review OA=5.0
Short Summary: This paper introduce a modified layer-wise learning rate scheduler, improves low-data training and fine-tuning performance in both NLP and SciML by balancing training quality across model layers.

[3] Evaluating Memory in LLM Agents via Incremental Multi-Turn Interactions
{Yuanzhe Hu#, Yu Wang#}, Julian McAuley
Under Review / ICML 2025 LCFM Workshop
Short Summary: MemoryAgentBench is a new benchmark designed to evaluate four core competencies of memory agents in Large Language Models, highlighting the need for improved memory mechanisms.
HF Dataset | Code | Paper |
Contributed

[4] M+: Extending MemoryLLM with Scalable Long-Term Memory
Yu Wang, Dmitry Krotov, Yuanzhe Hu, Yifan Gao, Wangchunshu Zhou, Julian McAuley, Dan Gutfreund, Rogerio Feris, Zexue He
ICML 2025
Short Summary: M+ enhances long-term information retention in large language models by integrating a retriever-based long-term memory mechanism, outperforming MemoryLLM and other baselines in knowledge retention tasks.
🔥 News
- 2025.07: 😁 We open-sourced the MemoryAgentBench. Thanks for the great help from Yu Wang!
- 2025.05: 🎉🎉 Two papers are accepted by ICML 2025 as Poster! See you at Vancouver.
- 2024.09: 🎉🎉 Excited to share that our work “Model Balancing Helps Low-data Training and Fine-tuning” is accepted by EMNLP 2024 as Oral Presentation!
- 2024.06: 😁 I graduated from HUST!
- 2024.06: 😄 I created my account on OpenReview!
Last Update: 09/2025