I am a PhD student at Shanghai Jiao Tong University, advised by Prof. Pengfei Liu. I work on foundation models and agentic systems; recently, I have been focusing on and thinking about self-evolving and long-horizon agents.
Selected Projects
View full list
Qwen3.5: Towards Native Multimodal Agents
Qwen Team
Blog ·
Models
A native multimodal model family with powerful agent capabilities.
Qwen Team
Blog ·
A native multimodal model family with powerful agent capabilities.
Qwen3-Coder: Agentic Coding in the World
Qwen Team (Core Contributor)
Blog · Code ·
Models
Focused on improving agentic coding capabilities.
Qwen Team (Core Contributor)
Blog · Code ·
Focused on improving agentic coding capabilities.
Qwen3-Coder-Next: Pushing Small Hybrid Models on Agentic Coding
Qwen Team (Core Contributor)
Blog · PDF · Code ·
Model
A local-first agentic coding model based on Qwen3-Next.
Qwen Team (Core Contributor)
Blog · PDF · Code ·
A local-first agentic coding model based on Qwen3-Next.
Revisiting Reinforcement Learning for LLM Reasoning from A Cross-Domain Perspective
Zhoujun Cheng*, Shibo Hao*, Tianyang Liu*, Fan Zhou, Feng Yao, Yuexin Bian, Yutao Xie, Zhengzhong Liu, Zhiting Hu (*=equal contribution) [more authors]
NeurIPS 2025.
PDF · Code ·
Dataset
GURU: An open RL suite for developing general-purpose reasoning models.
Zhoujun Cheng*, Shibo Hao*, Tianyang Liu*, Fan Zhou, Feng Yao, Yuexin Bian, Yutao Xie, Zhengzhong Liu, Zhiting Hu (*=equal contribution) [more authors]
NeurIPS 2025.
PDF · Code ·
GURU: An open RL suite for developing general-purpose reasoning models.
OctoThinker: Mid-Training Incentivizes RL Scaling
Zengzhi Wang*, Fan Zhou*, Xuefeng Li, Pengfei Liu
ICML 2025, AI4Math Workshop.
PDF · Blog · Code ·
Resources
A mid-training ablation study in the era of RL scaling, with a 70+B token mid-training dataset.
Zengzhi Wang*, Fan Zhou*, Xuefeng Li, Pengfei Liu
ICML 2025, AI4Math Workshop.
PDF · Blog · Code ·
A mid-training ablation study in the era of RL scaling, with a 70+B token mid-training dataset.
MegaMath: Pushing the Limits of Open Math Corpora
Fan Zhou*, Zengzhi Wang*, Nikhil Ranjan, Zhoujun Cheng, Liping Tang, Guowei He, Zhengzhong Liu, Eric P. Xing
COLM 2025.
PDF · Code ·
Dataset
(>70K Downloads, >350B Tokens)
The largest open math pre-training dataset with 370B tokens.
Fan Zhou*, Zengzhi Wang*, Nikhil Ranjan, Zhoujun Cheng, Liping Tang, Guowei He, Zhengzhong Liu, Eric P. Xing
COLM 2025.
PDF · Code ·
The largest open math pre-training dataset with 370B tokens.
Programming Every Example: Lifting Pre-training Data Quality like Experts at Scale
Fan Zhou*, Zengzhi Wang*, Qian Liu, Junlong Li, Pengfei Liu, (*=equal contribution)
ICML 2025
PDF · Code ·
Dataset
(>10K Downloads, >500B Tokens)
·
Project Page
A small-LLM-based pre-training data refining framework via seamless program generation.
Fan Zhou*, Zengzhi Wang*, Qian Liu, Junlong Li, Pengfei Liu, (*=equal contribution)
ICML 2025
PDF · Code ·
A small-LLM-based pre-training data refining framework via seamless program generation.
OpenAgents: An Open Platform for Language Agents in the Wild
Tianbao Xie*, Fan Zhou*, Zhoujun Cheng*, Peng Shi*, Luoxuan Weng*, Yitao Liu*, Toh Jing Hua, Junning Zhao, Qian Liu, Che Liu, Leo Z. Liu, Yiheng Xu, Hongjin Su, Dongchan Shin, Caiming Xiong, Tao Yu, (*=equal contribution)
COLM 2024
PDF · Code · Blog (7.5K Users)
An open platform for using, hosting, and building language agents.
Tianbao Xie*, Fan Zhou*, Zhoujun Cheng*, Peng Shi*, Luoxuan Weng*, Yitao Liu*, Toh Jing Hua, Junning Zhao, Qian Liu, Che Liu, Leo Z. Liu, Yiheng Xu, Hongjin Su, Dongchan Shin, Caiming Xiong, Tao Yu, (*=equal contribution)
COLM 2024
PDF · Code · Blog (7.5K Users)
An open platform for using, hosting, and building language agents.
Lemur: Harmonizing Natural Language and Code for Language Agents
Yiheng Xu*, Hongjin Su*, Chen Xing*, Boyu Mi, Qian Liu, Weijia Shi, Binyuan Hui, Fan Zhou, Yitao Liu, Tianbao Xie, Zhoujun Cheng, Siheng Zhao, Lingpeng Kong, Bailin Wang, Caiming Xiong, Tao Yu, (*=equal contribution)
ICLR 2024, Spotlight
PDF · Code ·
Models
·
Blog
A 70B language agent model pre-trained on balanced code and text corpora.
Yiheng Xu*, Hongjin Su*, Chen Xing*, Boyu Mi, Qian Liu, Weijia Shi, Binyuan Hui, Fan Zhou, Yitao Liu, Tianbao Xie, Zhoujun Cheng, Siheng Zhao, Lingpeng Kong, Bailin Wang, Caiming Xiong, Tao Yu, (*=equal contribution)
ICLR 2024, Spotlight
PDF · Code ·
A 70B language agent model pre-trained on balanced code and text corpora.
Experience
- 2021.09 - 2024.03 M.S. at SJTU, Computer Science.
- 2017.09 - 2021.06 B.S. at SJTU, Computer Science (IEEE Honor Class).
Service and Awards
- Service Reviewer: ICLR, NeurIPS, COLM, ACL, IJCAI, COLING, ...
- 2022 MSRA: Award of Excellent Intern.