Fantastic Sparse Masks and Where to Find Them
Sparsity has widely shown its versatility in model compression, robustness improvement, and overfitting mitigation by selectively masking out a portion of parameters. However, traditional methods to obtain such masks usually involve pre-training a dense model. As powerful foundation models become prevailing, the cost of the pre-training step can be prohibitive. In this talk, I will present our recent work on efficient methods to obtain such fantastic masks by training sparse neural networks from scratch, without the need for any dense pre-training steps.
Shiwei Liu is a Postdoctoral Fellow at the University of Texas at Austin. He obtained his Ph.D. from the Eindhoven University of Technology in 2022. His research interests cover sparsity in neural networks and efficient ML. He has over 30 publications in top-tier machine learning conferences, such as IJCAI, ICLR, ICML, NeurIPS, IJCV, UAI, and LoG. Shiwei won the best paper award at the LoG’22 conference and the Cum Laude (distinguished Ph.D. thesis) at the Eindhoven University of Technology. He has served as an area chair in ICIP‘22 and ICIP’23; and a PC member of almost all top-tier ML/CV conferences. Shiwei has co-organized two tutorials in IJCAI and ECMLPKDD, which were widely acclaimed by the audience. He has also provided more than 20 invited talks at many universities, companies, research labs, and conferences.
Sun 18 JunDisplayed time zone: Eastern Time (US & Canada) change
14:00 - 15:30 | |||
14:00 15mTalk | Learning workload-aware cost model for sparse tensor program CTSTA Jaeyeon Won Massachusetts Institute of Technology | ||
14:15 15mTalk | Autoscheduling for Sparse Tensor Contraction CTSTA Kirshanthan Sundararajah Purdue University | ||
14:30 10mPanel | Discussion CTSTA | ||
14:40 15mTalk | Fantastic Sparse Masks and Where to Find Them CTSTA Shiwei Liu The University of Texas at Austin, Texas, USA | ||
14:55 15mTalk | Moving the MLIR Sparse Compilation Pipeline into ProductionVirtual CTSTA | ||
15:10 15mPanel | Discussion CTSTA | ||
15:25 5mDay closing | Closing CTSTA |