Poster: Energy-Aware Partitioning for Edge AI
D. Katare (TU Delft - Information and Communication Technology)
M. Zhou (Fudan University, TU Delft - Information and Communication Technology)
Yang Chen (Fudan University)
M.F.W.H.A. Janssen (TU Delft - Engineering, Systems and Services)
Aaron Yi Ding (TU Delft - Information and Communication Technology)
More Info
expand_more
Other than for strictly personal use, it is not permitted to download, forward or distribute the text or part of it, without the consent of the author(s) and/or copyright holder(s), unless the work is under an open content license such as Creative Commons.
Abstract
Model partitioning is a promising solution to reduce the high computation load and transmission of high-volume data. Within the scope of Edge AI, the fundamentals of model partitioning involve splitting the model for local computing at the edge and offloading heavy computation tasks to the cloud or server. This approach benefits scenarios with limited computing and battery capacity with low latency requirements, such as connected autonomous vehicles. However, while model partitioning offers advantages in reducing the onboard computation, memory requirements and inference time, it also introduces challenges such as increased energy consumption for partitioned computations and overhead for transferring partitioned data/model. In this work, we explore hybrid model partitioning to optimize computational and communication energy consumption. Our results provide an initial analysis of the tradeoff between energy and accuracy, focusing on the energy-aware model partitioning for future Edge AI applications.