Joshua P. Gardner |
I received a PhD in computer science from the University of Washington's Paul G. Allen School of Computer Science & Engineering, where I was fortunate to be advised by Ludwig Schmidt. I hold an M.S. in Applied Statistics and an M.S. in Information Science from the University of Michigan. I also hold a B.A. with Highest Honors in Philosophy from the University of Michigan.
My research focuses on empirical machine learning: designing reliable controlled experiments to understand artificial intelligence/machine learning systems, and using the insights from this experimentation to develop improved methods. In particular, my recent research applies this lens to the training and fine-tuning of large "foundation" models. I have studied a diverse set of domains and applications under this general theme, including tabular data, large language models, multimodal learning, music and audio, and federated and collaborative learning.
Selected Publications
For a full list of publications see my research page or Google Scholar profile.
-
Large Scale Transfer Learning for Tabular Data via Language Modeling
Josh Gardner, Juan C. Perdomo, Ludwig Schmidt.
Neural Information Processing Systems (NeurIPS) 2024.
[arxiv] [code] [model + data]
-
DataComp-LM: In search of the next generation of training sets for language models
Neural Information Processing Systems (NeurIPS) 2024 (Datasets & Benchmarks Track).
Jeffrey Li, Alex Fang, Georgios Smyrnis, Maor Ivgi, [...] Josh Gardner, [...], Achal Dave, Ludwig Schmidt, Vaishaal Shankar. (59 total authors)
[arxiv] [web]
-
LLark: A Multimodal Instruction-Following Language Model for Music
Josh Gardner, Simon Durand, Daniel Stoller, Rachel Bittner.
International Conference on Machine Learning (ICML) 2024.
[arxiv] [code] [web] [blog]
-
Benchmarking Distribution Shift in Tabular Data with TableShift
Josh Gardner, Zoran Popović, Ludwig Schmidt.
Neural Information Processing Systems (NeurIPS) 2023 (Datasets & Benchmarks Track).
[arxiv] [code] [web]
-
Cross-Institutional Transfer Learning for Educational Models: Implications for Model Performance, Fairness, and Equity
Josh Gardner, Renzhe Yu, Quan Nguyen, Christopher Brooks, Rene Kizilcec.
ACM Conference on Fairness, Accountability, and Transparency (ACM FAccT) 2023.
[pdf] [arxiv] [code]
-
Subgroup Robustness Grows on Trees: An Empirical Baseline Study
Josh Gardner, Zoran Popović, Ludwig Schmidt.
Neural Information Processing Systems (NeurIPS) 2022.
[arxiv] [code]
-
OpenFlamingo: An Open-Source Framework for Training Vision-Language Models with In-Context Learning
Anas Awadalla, Irena Gao, Josh Gardner, Jack Hessel, Yusuf Hanafy, Wanrong Zhu, Kalyani Marathe, Yonatan Bitton, Samir Gadre, Jenia Jitsev, Simon Kornblith, Pang Wei Koh, Gabriel Ilharco, Mitchell Wortsman, Ludwig Schmidt.
[arxiv] [blog] [code]
-
MT3: Multi-Task Multitrack Music Transcription
Josh Gardner, Ian Simon, Ethan Manilow, Curtis Hawthorne, Jesse Engel.
International Conference on Learning Representations (ICLR) 2022.
Spotlight Presentation (top 6.7% of submissions)
[arxiv] [web] [blog] [code]
-
Evaluating the Fairness of Predictive Student Models Through Slicing Analysis
Josh Gardner, Christopher Brooks, and Ryan Baker.
International Conference on Learning Analytics and Knowledge (LAK) 2019.
Best Paper Award
[pdf]