Exploring the use of data compression for accelerating machine learning in the edge with remote virtual graphics processing units

Cristian Peñaranda*, Carlos Reaño, Federico Silla

*Corresponding author for this work

Research output: Contribution to journalArticlepeer-review

5 Citations (Scopus)
22 Downloads (Pure)

Abstract

Internet of Things (IoT) devices are usually low performance nodes connected by low bandwidth networks. To improve performance in such scenarios, some computations could be done at the edge of the network. However, edge devices may not have enough computing power to accelerate applications such as the popular machine learning ones. Using remote virtual graphics processing units (GPUs) can address this concern by accelerating applications leveraging a GPU installed in a remote device. However, this requires exchanging data with the remote GPU across the slow network. To address the problem with the slow network, the data to be exchanged with the remote GPU could be compressed. In this article, we explore the suitability of using data compression in the context of remote GPU virtualization frameworks in edge scenarios executing machine learning applications. We use popular machine learning applications to carry out such exploration. After characterizing the GPU data transfers of these applications, we analyze the usage of existing compression libraries for compressing those data transfers to/from the remote GPU. Our exploration shows that transferring compressed data becomes more beneficial as networks get slower, reducing transfer time by up to 10 times. Our analysis also reveals that efficient integration of compression into remote GPU virtualization frameworks is strongly required.

Original languageEnglish
Article numbere7328
Number of pages19
JournalConcurrency and Computation: Practice and Experience
Volume35
Issue number20
Early online date17 Oct 2022
DOIs
Publication statusPublished - Sept 2023
Externally publishedYes

Bibliographical note

Funding Information:
This work was supported in part by the project “AI in Secure Privacy‐Preserving Computing Continuum (AI‐SPRINT)” through the European Union's Horizon 2020 Research and Innovation Programme under Grant 101016577, and in part by the European Union's Horizon 2020 research and innovation programme under Grant agreement No. 101017861.

Publisher Copyright:
© 2022 The Authors. Concurrency and Computation: Practice and Experience published by John Wiley & Sons Ltd.

Keywords

  • data compression
  • edge computing
  • GPU virtualization
  • machine learning

ASJC Scopus subject areas

  • Theoretical Computer Science
  • Software
  • Computer Science Applications
  • Computer Networks and Communications
  • Computational Theory and Mathematics

Fingerprint

Dive into the research topics of 'Exploring the use of data compression for accelerating machine learning in the edge with remote virtual graphics processing units'. Together they form a unique fingerprint.

Cite this