
Zihao Ye (叶子豪)
Ph.D. student @ SAMPL,
UW CSE
Bill &
Melinda Gates Center, Room 330
Email : zhye [at] cs [dot] washington [dot] edu
Google Scholar : Zihao
Ye
Github : yzh119
About Me
This is Zihao Ye, a third-year Ph.D. student at the University of Washington’s Paul G. Allen School of Computer Science and Engineering, advised by Luis Ceze in the SAMPL research group. I also work closely with Tianqi Chen on Apache TVM project.
Prior to joining UW, I spent two years at AWS where I worked with Minjie Wang and Zheng Zhang. I obtained my bachelor’s degree from ACM Honors Class at Shanghai Jiao Tong University.
We are organizing talks at SAMPL, topics include Systems, Architecture, Compilers, Verification and Machine Learning.
Besides research, I enjoy diving into software/hardware details, as well as working on open-source projects.
Research
I have broad interests in Computer Systems, Compiler, Programming Languages, and Computer Architecture. My current research centers around sparse computation:
- Programming Abstraction/Compilers for Sparse Workloads on Heterogeneous Hardware.
- Domain Specific Accelerator and Memory Architecture for Sparsity.
- Scalable and Affordable Sparse Computation in Data Centers.
Feel free to drop me an email if we have aligned interests, and I’m open to collaborations.
News
- Jun 2023 I’ll give a talk on Sparsity in LLMs at CTSTA @ PLDI 2023, see you all in Orlando!
- May 2023 Please take a look at the MLC-LLM project, which enables the deployment of LLM on a wide range of hardware platforms. I am honored to be a part of this project and collaborate with an extraordinary team!
- Mar 2023 SparseTIR has been awarded Distinguished Artifact at ASPLOS 2023!
- Feb 2023 I’m going to be a TA for CSE 599M in Spring 2023, which is ML for ML Systems taught by Luis Ceze. I’m really excited to be a part of this new course.
- Jan 2023 We are proud to announce that SparseTIR will be featured at ASPLOS 2023. We’ll be heading to Vancouver and can’t wait to see everyone there!
Current Projects
Compiler for Sparsity in Deep Learning
SparseTIR is a tensor-level abstraction for representing and optimizing sparse/irregular operators in Deep Learning. The project is in close collaboration with Ruihang and Tianqi from CMU Catalyst and Junru from OctoML, we thank the support and advice from advisors TQ and Luis, and assistance from the TensorIR team.
- [ Documentation ], [ Paper(PDF) ], [ Code ]
Talks
Composable Abstractions for Sparse Compilation in Deep Learning
- Mar 2023 @ ASPLOS 2023, Vancouver, Canada
- Mar 2023 @ TVMCon
- Nov 2022 @ Amazon AI
- Oct 2022 @ CRISP Liaison Meeting
- Aug 2022 @ Tsinghua NISF-EFC Group
- Aug 2022 @ Cornell Zhang Research Group
- July 2022 @ Google MLIR Reading Group
- Dec 2021 @ TVMCon
Selected Publications
ASPLOS 2023 |
![]() The 28th ACM International Conference on Architectural Support for Programming Languages and Operating Systems, 2023. Distinguished Artifact Award. |
Activity and Service
- External Review Committee: MLSys 2023
- Artifact Evaluation Committee: MLSys 2022, CGO 2023, OSDI/ATC 2023
- Reviewer: IEEE CAL