Title
Device Sampling for Heterogeneous Federated Learning: Theory, Algorithms, and Implementation
Abstract
The conventional federated learning (FedL) architecture distributes machine learning (ML) across worker devices by having them train local models that are periodically aggregated by a server. FedL ignores two important characteristics of contemporary wireless networks, however: (i) the network may contain heterogeneous communication/computation resources, while (ii) there may be significant overlaps in devices' local data distributions. In this work, we develop a novel optimization methodology that jointly accounts for these factors via intelligent device sampling complemented by device-to-device (D2D) offloading. Our optimization aims to select the best combination of sampled nodes and data offloading configuration to maximize FedL training accuracy subject to realistic constraints on the network topology and device capabilities. Theoretical analysis of the D2D offloading subproblem leads to new FedL convergence bounds and an efficient sequential convex optimizer. Using this result, we develop a sampling methodology based on graph convolutional networks (GCNs) which learns the relationship between network attributes, sampled nodes, and resulting offloading that maximizes FedL accuracy. Through evaluation on real-world datasets and network measurements from our IoT testbed, we find that our methodology while sampling less than 5% of all devices outperforms conventional FedL substantially both in terms of trained model accuracy and required resource utilization.
Year
DOI
Venue
2021
10.1109/INFOCOM42981.2021.9488906
IEEE CONFERENCE ON COMPUTER COMMUNICATIONS (IEEE INFOCOM 2021)
DocType
ISSN
Citations 
Conference
0743-166X
0
PageRank 
References 
Authors
0.34
0
6
Name
Order
Citations
PageRank
Su Wang100.34
Mengyuan Lee2143.57
Seyyedali Hosseinalipour3365.35
Roberto Morabito400.68
Mung Chiang57303486.32
Christopher G. Brinton611815.23