Learning workload-aware cost model for sparse tensor program
This talk presents WACO, a novel method of co-optimizing the format and the schedule of a given sparsity pattern in a sparse tensor program. A core challenge in this paper is the design of a lightweight cost model that accurately predicts the runtime of a sparse tensor program by considering the sparsity pattern, the format, and the schedule. The key idea in addressing this is exploiting a sparse convolutional network to learn meaningful features of the sparsity pattern and embedding a coupled behavior between the format and the schedule using a specially designed schedule template. We evaluated WACO for four different algorithms (SpMV, SpMM, SDDMM, and MTTKRP) on a CPU using 726 different sparsity patterns. Our experimental results showed that WACO outperformed four state-of-the-art baselines, Intel MKL, BestFormat, TACO with a default schedule, and ASpT. Compared to the best of four baselines, WACO achieved 1.43×, 1.18×, 1.14×, and 1.27× average speedups on SpMV, SpMM, SDDMM, and MTTKRP, respectively.
Sun 18 JunDisplayed time zone: Eastern Time (US & Canada) change
14:00 - 15:30 | |||
14:00 15mTalk | Learning workload-aware cost model for sparse tensor program CTSTA Jaeyeon Won Massachusetts Institute of Technology | ||
14:15 15mTalk | Autoscheduling for Sparse Tensor Contraction CTSTA Kirshanthan Sundararajah Purdue University | ||
14:30 10mPanel | Discussion CTSTA | ||
14:40 15mTalk | Fantastic Sparse Masks and Where to Find Them CTSTA Shiwei Liu The University of Texas at Austin, Texas, USA | ||
14:55 15mTalk | Moving the MLIR Sparse Compilation Pipeline into ProductionVirtual CTSTA | ||
15:10 15mPanel | Discussion CTSTA | ||
15:25 5mDay closing | Closing CTSTA |