I am a PhD student at Princeton University and a Princeton AI Lab Fellow, working with Prof. Mengdi Wang, Prof. Andrew Yao, and Prof. Quanquan Gu, where my research focuses on building scalable and capable large language models (LLMs) and multimodal foundation models. My work explores methods to improve LLM reasoning, data curation and algorithms for foundation models, as well as the development of new attention mechanisms, positional encodings, and model architectures.
Previously, I was a visiting PhD student at the UCLA AGI Lab, and as a Top Seed research intern with the Seed Foundation Model Team, working on LLM and MLLM pretraining and scaling. I earned my PhD candidacy and Master of Science in Computer Science from IIIS at Tsinghua University, and a Bachelor of Science in Mathematics and Computer Science from Yuanpei College at Peking University.
I am currently exploring opportunities at Frontier AI Labs and would be pleased to discuss potential collaborations via .