Yizhong Wang

Research Scientist, ByteDance Seed
Incoming Assistant Professor, CS Department, University of Texas at Austin

Yizhong Wang

I am a Research Scientist at ByteDance Seed and an incoming Assistant Professor at the University of Texas at Austin. I received my PhD from the Paul G. Allen School of Computer Science & Engineering at the University of Washington, where I was co-advised by Hannaneh Hajishirzi and Noah Smith.

I study how (natural/human) language can help AI understand, reason, learn, communicate, and interact with the world. This has led to my past work on instruction tuning, synthetic data generation, RLVR, and open language models. Recently, I am thinking about learning algorithms that can enable more autonomy in AI, and how to use this in high-value but challenging domains (e.g., scientific discovery).

Prospective students and collaborators, please see the Prospective Students section below.

News

July 27, 2025 — We had a tutorial on Synthetic Data in the Era of LLMs at ACL 2025.
June 16, 2025 — I joined ByteDance Seed as a Research Scientist.
May 27, 2025 — I have defended my PhD dissertation!

Selected Publications

* indicates equal contribution. For a full list, see my Google Scholar page.

Tülu 3: Pushing Frontiers in Open Language Model Post-Training
Nathan Lambert, Jacob Morrison, Valentina Pyatkin, Shengyi Huang, Hamish Ivison, Faeze Brahman, Lj Miranda, ..., Luca Soldaini, Noah A. Smith, Yizhong Wang, Pradeep Dasigi, Hannaneh Hajishirzi
COLM 2025
Hybrid Preferences: Learning to Route Instances for Human vs. AI Feedback
Lj Miranda*, Yizhong Wang*, Yanai Elazar, Sachin Kumar, Valentina Pyatkin, Faeze Brahman, Noah A. Smith, Hannaneh Hajishirzi, Pradeep Dasigi
ACL 2025
Unpacking DPO and PPO: Disentangling Best Practices for Learning from Preference Feedback
Hamish Ivison, Yizhong Wang, Jiacheng Liu, Zeqiu Wu, Valentina Pyatkin, Nathan Lambert, Noah A. Smith, Yejin Choi, Hannaneh Hajishirzi
NeurIPS 2024
OLMo: Accelerating the Science of Language Models
Dirk Groeneveld, Iz Beltagy, Pete Walsh, Akshita Bhagia, Rodney Kinney, Oyvind Tafjord, Ananya Harsh Jha, Hamish Ivison, Ian Magnusson, Yizhong Wang, et al.
ACL 2024 (Best Theme Paper)
How Far Can Camels Go? Exploring the State of Instruction Tuning on Open Resources
Yizhong Wang*, Hamish Ivison*, Pradeep Dasigi, Jack Hessel, Tushar Khot, Khyathi Raghavi Chandu, David Wadden, Kelsey MacMillan, Noah A. Smith, Iz Beltagy, Hannaneh Hajishirzi
NeurIPS 2023
Self-Instruct: Aligning Language Models with Self-Generated Instructions
Yizhong Wang, Yeganeh Kordi, Swaroop Mishra, Alisa Liu, Noah A Smith, Daniel Khashabi, Hannaneh Hajishirzi
ACL 2023
Super-NaturalInstructions: Generalization via Declarative Instructions on 1600+ NLP Tasks
Yizhong Wang*, Swaroop Mishra*, Pegah Alipoormolabashi, Yeganeh Kordi et al.
EMNLP 2022
DROP: A Reading Comprehension Benchmark Requiring Discrete Reasoning Over Paragraphs
Dheeru Dua, Yizhong Wang, Pradeep Dasigi, Gabriel Stanovsky, Sameer Singh and Matt Gardner
NAACL 2019
A Two-Stage Parsing Method for Text-level Discourse Analysis
Yizhong Wang, Sujian Li and Houfeng Wang
ACL 2017 (Outstanding Paper Award)

Prospective Students

I plan to recruit multiple CS PhD students to start from Fall 2026 at the the University of Texas at Austin. If your research interests align with mine (or something new about AI really excites you), I strongly encourage you to directly apply to the UT Austin CS PhD program. Please mention my name as a potential advisor in your application.

You are also welcome to email me with your CV and a brief description of your research interests if you are interested in joining my group as a Postdoc, PhD student, or research assistant. I will try my best to respond.