Articles with "dnn inference" as a keyword



Photo by avasol from unsplash

Throughput Maximization of Delay-Aware DNN Inference in Edge Computing by Exploring DNN Model Partitioning and Inference Parallelism

Sign Up to like & get
recommendations!
Published in 2023 at "IEEE Transactions on Mobile Computing"

DOI: 10.1109/tmc.2021.3125949

Abstract: Mobile Edge Computing (MEC) has emerged as a promising paradigm catering to overwhelming explosions of mobile applications, by offloading compute-intensive tasks to MEC networks for processing. The surging of deep learning brings new vigor and… read more here.

Keywords: delay aware; aware dnn; dnn inference; edge ... See more keywords
Photo by acfb5071 from unsplash

Coordinated Batching and DVFS for DNN Inference on GPU Accelerators

Sign Up to like & get
recommendations!
Published in 2022 at "IEEE Transactions on Parallel and Distributed Systems"

DOI: 10.1109/tpds.2022.3144614

Abstract: Employing hardware accelerators to improve the performance and energy-efficiency of DNN applications is on the rise. One challenge of using hardware accelerators, including the GPU-based ones, is that their performance is limited by internal and… read more here.

Keywords: power; dnn inference; dvfs; dnn ... See more keywords