The effects of weight quantization on online federated learning for the IoT: a case study

Nil Llisterri Gimenez, Junkyu Lee, Felix Freitag*, Hans Vandierendonck

*Corresponding author for this work

Research output: Contribution to journalArticlepeer-review

6 Citations (Scopus)
7 Downloads (Pure)

Abstract

Many weight quantization approaches were explored to save the communication bandwidth between the clients and the server in federated learning using high-end computing machines. However, there is a lack of weight quantization research for online federated learning using TinyML devices which are restricted by the mini-batch size, the neural network size, and the communication method due to their severe hardware resource constraints and power budgets. We name Tiny Online Federated Learning (TinyOFL) for online federated learning using TinyML devices in the Internet of Things (IoT). This paper performs a comprehensive analysis of the effects of weight quantization in TinyOFL in terms of accuracy, stability, overfitting, communication efficiency, energy consumption, and delivery time, and extracts practical guidelines on how to apply the weight quantization to TinyOFL. Our analysis is supported by a TinyOFL case study with three Arduino Portenta H7 boards running federated learning clients for a keyword spotting task. Our findings include that in TinyOFL, a more aggressive weight quantization can be allowed than in online learning without FL, without affecting the accuracy thanks to TinyOFL's quasi-batch training property. For example, using 7-bit weights achieved the equivalent accuracy to 32-bit floating point weights, while saving communication bandwidth by 4.6×. Overfitting by increasing network width rarely occurs in TinyOFL, but may occur if strong weight quantization is applied. The experiments also showed that there is a design space for TinyOFL applications by compensating for the accuracy loss due to weight quantization with an increase of the neural network size.

Original languageEnglish
Pages (from-to)5490-5502
Number of pages13
JournalIEEE Access
Volume12
DOIs
Publication statusPublished - 04 Jan 2024

Bibliographical note

Publisher Copyright:
© 2013 IEEE.

Keywords

  • approximate computing
  • federated learning
  • IoT
  • TinyML

ASJC Scopus subject areas

  • General Computer Science
  • General Materials Science
  • General Engineering

Fingerprint

Dive into the research topics of 'The effects of weight quantization on online federated learning for the IoT: a case study'. Together they form a unique fingerprint.

Cite this