About me
I’m a Research Scientist at Snowflake AI Research, where I train reasoning models and LLM agents, especially for structured data. Before that, I was a Research Scientist on Bytedance’s Seed‑Foundation‑Code team.
I earned my Ph.D. from the University of Illinois at Urbana‑Champaign under Prof. Ruoyu Sun, with co‑advising from Prof. Naira Hovakimyan. In my first year I collaborated with Prof. Justin Sirignano on deep‑learning for computational fluid dynamics (CFD), and during my master’s studies I worked with Prof. Kyle Smith on numerical simulation for energy storage systems.
Research interests: Reasoning Models; Agent; Efficient Deep Learning; Natural Language Processing; Computer Vision; Numerical Methods.
Hiring: Snowflake AI Research is seeking research scientist interns for remote, long-term positions (3-6 months, part-time) starting August 2025. Projects include:
Post‑training tool‑augmented LLMs
Development of personalized Text‑to‑SQL models
Deep Research on structured data
Developing and training of data‑science agents
If you are interested, don’t hesitate to send your resume to yite.wang@snowflake.com .
News
[April. 2025] I have joined Snowflake AI Research working on reasoning models and LLM Agents.
[May. 2024] I have joined ByteDance Seed-Foundation-Code as a Research Scientist working on LLM for code.
[Mar. 2024] I have successfully defended my Ph.D. thesis!
[Jan. 2024] Our paper on model expansion is accepted to ICLR’2024!
[Oct. 2023] I have received the NeurIPS 2023 Scholar Award!
[Sept. 2023] Our paper about dynamic sparse training for GANs is accepted to NeurIPS’2023!
[Jan. 2023] Our paper about foresight pruning is accepted to ICLR’2023!
[March. 2022] Our paper about neural architecture search for meta learning is accpeted to CVPR’2022!
Publications
( $\dagger$ denotes to equal contribution)
Fullstack bench: Evaluating llms as full stack coder [arXiv]
Seed-Foundation-Code Team, ByteDance
Designing Large Foundation Models for Efficient Training and Inference: A Survey [arxiv]
Dong Liu, Yanxuan Yu, Yite Wang, Jing Wu, Zhongwei Wan, Sina Alinejad, Benjamin Lengerich, Ying Nian Wu
LEMON: Lossless model expansion [arXiv, code]
Yite Wang, Jiahao Su, Hanlin Lu, Cong Xie, Tianyi Liu, Jianbo Yuan, Haibin Lin, Ruoyu Sun, Hongxia Yang
ICLR’2024:International Conference on Learning Representations
Balanced Training for Sparse GANs [arXiv, page, code]
Yite Wang$\dagger$, Jing Wu$\dagger$, Naira Hovakimyan, Ruoyu Sun
NeurIPS’2023: Neural Information Processing Systems
NTK-SAP: Improving neural network pruning by aligning training dynamics [arXiv, page, code]
Yite Wang, Dawei Li, Ruoyu Sun
ICLR’2023: International Conference on Learning Representations
Global Convergence of MAML and Theory-Inspired Neural Architecture Search for Few-Shot Learning [arXiv, code]
Haoxiang Wang$\dagger$, Yite Wang$\dagger$, Ruoyu Sun, Bo Li
CVPR’2022: The IEEE/CVF Computer Vision and Pattern Recognition Conference
Numerical investigation of convective transport in redox flow battery tanks: Using baffles to increase utilization [paper]
Yite Wang, Kyle Smith
Journal of Energy Storage
