Performance-Aware Client and Quantization Level Selection Algorithm for Fast Federated Learning

Sangwon Seo, Jaewook Lee, Haneul Ko, Sangheon Pack

Research output: Chapter in Book/Report/Conference proceedingConference contribution

Abstract

In federated learning (FL), which clients are selected and which quantization levels are chosen for the deep model parameters have significant impacts on the learning time as well as the learning accuracy. In this paper, we formulate a joint optimization problem on the client and quantization level selections. As a low complexity solution to the formulated problem, we develop a performance-aware client and quantization level selection (PA-CQLS) algorithm where the FL server estimates the individual round times of clients based on their computing power and channel quality, and determines the most appropriate clients and quantization levels accordingly. Simulation results show that PA-CQLS can reduce the round time by up to 70% compared to conventional algorithms.

Original languageEnglish
Title of host publication2022 IEEE Wireless Communications and Networking Conference, WCNC 2022
PublisherInstitute of Electrical and Electronics Engineers Inc.
Pages1892-1897
Number of pages6
ISBN (Electronic)9781665442664
DOIs
Publication statusPublished - 2022
Event2022 IEEE Wireless Communications and Networking Conference, WCNC 2022 - Austin, United States
Duration: 2022 Apr 102022 Apr 13

Publication series

NameIEEE Wireless Communications and Networking Conference, WCNC
Volume2022-April
ISSN (Print)1525-3511

Conference

Conference2022 IEEE Wireless Communications and Networking Conference, WCNC 2022
Country/TerritoryUnited States
CityAustin
Period22/4/1022/4/13

ASJC Scopus subject areas

  • Engineering(all)

Fingerprint

Dive into the research topics of 'Performance-Aware Client and Quantization Level Selection Algorithm for Fast Federated Learning'. Together they form a unique fingerprint.

Cite this