Abstract
Today's intelligent applications can achieve high performance accuracy using machine learning (ML) techniques, such as deep neural networks (DNNs). Traditionally, in a remote DNN inference problem, an edge device transmits raw data to a remote node that performs the inference task. However, this may incur high transmission energy costs and puts data privacy at risk. In this paper, we propose a technique to reduce the total energy bill at the edge device by utilizing model compression and time-varying model split between the edge and remote nodes. The time-varying representation accounts for time-varying channels and can significantly reduce the total energy at the edge device while maintaining high accuracy (low loss). We implement our approach in an image classification task using the MNIST dataset, and the system environment is simulated as a trajectory navigation scenario to emulate different channel conditions. Numerical simulations show that our proposed solution results in minimal energy consumption and CO2 emission compared to the considered baselines while exhibiting robust performance across different channel conditions and bandwidth regime choices.
| Original language | English |
|---|---|
| Title of host publication | 2021 IEEE 32nd Annual International Symposium on Personal, Indoor and Mobile Radio Communications, PIMRC 2021 |
| Publisher | Institute of Electrical and Electronics Engineers Inc. |
| Pages | 1173-1178 |
| Number of pages | 6 |
| ISBN (Electronic) | 9781728175867 |
| DOIs | |
| State | Published - 13 Sep 2021 |
| Externally published | Yes |
Publication series
| Name | IEEE International Symposium on Personal, Indoor and Mobile Radio Communications, PIMRC |
|---|---|
| Volume | 2021-September |
Bibliographical note
Publisher Copyright:© 2021 IEEE.
Keywords
- Deep learning
- edge computing
- energy efficiency
- model compression
- remote inference
- split learning
ASJC Scopus subject areas
- Electrical and Electronic Engineering