Yizhong Wang

PhD candidate
Paul G. Allen School of Computer Science & Engineering
University of Washington, Seattle, WA

Email: yizhongw [at] cs.washington.edu
CV Short Bio X GitHub Google Scholar Semantic Scholar

Hello! I am a final-year PhD candidate at the Paul G. Allen School of Computer Science & Engineering, University of Washington. I am fortunate to be co-advised by Hannaneh Hajishirzi and Noah Smith. I am also a student researcher at Ai2. I have previously interned at Meta AI and Microsoft Research Asia. Prior to UW, I did a master at Peking University and an undergraduate at Shanghai Jiao Tong University.

I work broadly on natural language processing, machine learning, and artificial intelligence. In particular, I enjoy studying fundamental data challenges in AI development and how to advance AI through data creation. My work has led to the unification of NLP tasks with instruction tuning (e.g., Super-NaturalInstructions), pioneered the use of synthetic data for language model training (e.g., Self-Instruct), and scientified the development of fully open language models (e.g., OLMo and Tülu). I believe that data serves as the foundation that defines the behavior and upper limits of AI models. It also provides an effective, interpretable, and beneficial ground to advance AI as a community. I have been thinking about these topics recently:

I am on the job market now! Feel free to reach out if you would like to share opportunities, collaborate, or just chat :)

News

  • Nov. 21, 2024
  • Tülu has evolved to v3, with SoTA performance & fully open post-training recipes & a playground!
  • Oct. 29, 2024
  • We released Hybrid Preferences, a framework for combining human and AI feedbakc for better RLHF.
  • Sep. 25, 2024
  • Tulu 2.5 got accepted to NeurIPS 2024!
  • July 10, 2024
  • OLMo won the Best Theme Paper Award at ACL 2024!
  • July 10, 2024
  • Two papers about proxy tuning and hallucination detection were accepted to the first COLM conference!
  • May 16, 2024
  • OLMo was accepted to ACL 2024 main conference, and temporal alignment was accepted to the findings. See people in Bangkok!
  • Feb. 12, 2024
  • I have passed my PhD general exam! 🏃
  • Feb. 1, 2024
  • I am excited to be part of the OLMo first release. Check out the blog post and tech report.

    Selected Publications

    * indicates equal contribution. For a full list of publications, please refer to my Google Scholar page.

    Tülu 3: Pushing Frontiers in Open Language Model Post-Training

    Nathan Lambert, Jacob Morrison, Valentina Pyatkin, Shengyi Huang, Hamish Ivison, Faeze Brahman, Lj Miranda, ..., Luca Soldaini, Noah A. Smith, Yizhong Wang, Pradeep Dasigi, Hannaneh Hajishirzi

    Arxiv
    Media coverage: Geekwire, TechCrunch, VentureBeat, MSN, and more.
    Hybrid Preferences: Learning to Route Instances for Human vs. AI Feedback

    Lj Miranda*, Yizhong Wang*, Yanai Elazar, Sachin Kumar, Valentina Pyatkin, Faeze Brahman, Noah A. Smith, Hannaneh Hajishirzi, Pradeep Dasigi

    Arxiv
    Unpacking DPO and PPO: Disentangling Best Practices for Learning from Preference Feedback

    Hamish Ivison, Yizhong Wang, Jiacheng Liu, Zeqiu Wu, Valentina Pyatkin, Nathan Lambert, Noah A. Smith, Yejin Choi, Hannaneh Hajishirzi

    NeurIPS 2024
    Set the Clock: Temporal Alignment of Pretrained Language Models

    Bowen Zhao*, Zander Brumbaugh*, Yizhong Wang*, Hannaneh Hajishirzi, Noah A. Smith

    ACL 2024 Findings
    OLMo: Accelerating the Science of Language Models

    Dirk Groeneveld, Iz Beltagy, Pete Walsh, Akshita Bhagia, Rodney Kinney, Oyvind Tafjord, Ananya Harsh Jha, Hamish Ivison, Ian Magnusson, Yizhong Wang, et al.

    ACL 2024 (Best Theme Paper)
    Media coverage: Forbes, GeekWire, TechCrunch, Axios, and more.
    How Far Can Camels Go? Exploring the State of Instruction Tuning on Open Resources

    Yizhong Wang*, Hamish Ivison*, Pradeep Dasigi, Jack Hessel, Tushar Khot, Khyathi Raghavi Chandu, David Wadden, Kelsey MacMillan, Noah A. Smith, Iz Beltagy, Hannaneh Hajishirzi

    NeurIPS 2023 (Spotlight)
    Self-Instruct: Aligning Language Models with Self-Generated Instructions

    Yizhong Wang, Yeganeh Kordi, Swaroop Mishra, Alisa Liu, Noah A Smith, Daniel Khashabi, Hannaneh Hajishirzi

    ACL 2023 (Most Influential Paper #1 by Paper Digest)
    Super-NaturalInstructions: Generalization via Declarative Instructions on 1600+ NLP Tasks

    Yizhong Wang*, Swaroop Mishra*, Pegah Alipoormolabashi, Yeganeh Kordi et al.

    EMNLP 2022 (Most Influential Paper #2 by Paper Digest)
    Probing Across Time: What Does RoBERTa Know and When?

    Leo Z. Liu*, Yizhong Wang*, Jungo Kasai, Hannaneh Hajishirzi, Noah A. Smith

    EMNLP 2021 Findings
    Do Neural NLP Models Know Numbers? Probing Numeracy in Embeddings

    Eric Wallace*, Yizhong Wang*, Sujian Li, Sameer Singh and Matt Gardner

    EMNLP-IJCNLP 2019
    DROP: A Reading Comprehension Benchmark Requiring Discrete Reasoning Over Paragraphs

    Dheeru Dua, Yizhong Wang, Pradeep Dasigi, Gabriel Stanovsky, Sameer Singh and Matt Gardner

    NAACL 2019 (Most Influential Paper #8 by Paper Digest)
    Multi-Passage Machine Reading Comprehension with Cross-Passage Answer Verification

    Yizhong Wang, Kai Liu, Jing Liu, Wei He, Yajuan Lyu, Hua Wu, Sujian Li and Haifeng Wang.

    ACL 2018
    A Two-Stage Parsing Method for Text-level Discourse Analysis

    Yizhong Wang, Sujian Li and Houfeng Wang

    ACL 2017 (Outstanding Paper Award)