Abstract
Achieving efficient execution of machine learning models on mobile/edge devices has attracted significant attention recently. A key challenge is to generate high-performance tensor programs for each operator inside a DNN model efficiently. To this end, deep learning compilers have adopted auto-Tuning approaches such as Ansor. However, it is challenging to optimize tensor codes for mobile/edge devices by auto-Tuning due to limited time budgets and on-device resources. A key component of DNN compilers is the cost model that can predict the performance of each configuration on specific devices. However, current design of cost models cannot provide transferable features among different hardware accelerators efficiently and effectively. In this paper, we propose Moses, a simple yet efficient design based on the lottery ticket hypothesis, which fully takes advantage of the hardware-Agnostic features transferable to the target device via domain adaptation to optimize the time-consuming auto-Tuning process of DNN compiling on a new hardware platform. Compared with state-of-The-Art approaches, Moses achieves up to 1.53X efficiency gain in the search stage and 1.41X inference speedup on challenging DNN benchmarks. © 2023 ACM.
| Original language | English |
|---|---|
| Title of host publication | HotMobile '23: Proceedings of the 24th International Workshop on Mobile Computing Systems and Applications |
| Publisher | Association for Computing Machinery |
| Pages | 22-28 |
| ISBN (Print) | 9798400700170 |
| DOIs | |
| Publication status | Published - Feb 2023 |
| Event | 24th International Workshop on Mobile Computing Systems and Applications (HotMobile 2023) - Newport Beach, United States Duration: 22 Feb 2023 → 23 Feb 2023 |
Publication series
| Name | HotMobile - Proceedings of the International Workshop on Mobile Computing Systems and Applications |
|---|
Conference
| Conference | 24th International Workshop on Mobile Computing Systems and Applications (HotMobile 2023) |
|---|---|
| Place | United States |
| City | Newport Beach |
| Period | 22/02/23 → 23/02/23 |
Funding
The work described in this article was supported by the Research Grants Council (RGC)-General Research Fund under Grant No. 14209619.
Research Keywords
- DNN compiler
- efficient DNN processing
- transfer learning
RGC Funding Information
- RGC-funded