Client Selection for Wireless Federated Learning With Data and Latency Heterogeneity

Document Type

Article

Publication Date

1-1-2024

Abstract

Federated learning is a distributed machine learning paradigm that allows multiple edge devices to collaboratively train a shared model without exchanging raw data. However, the training efficiency of federated learning is highly dependent on client selection. Moreover, due to the varying wireless communication environments and various computation latencies among the clients, selecting clients randomly or uniformly may not be optimal for balancing the data diversity and training efficiency. In this article, we formulate a new latency-minimization problem that simultaneously optimizes client selection and training procedures in federated learning, which takes into account the data and latency heterogeneity among the clients. Given the nonconvexity of the problem, we derive a new convergence upper bound for federated learning with probabilistic client selection. To solve the mixed integer nonlinear programming problem, we introduce a hybrid solution that integrates grid search techniques with the polyhedral active set algorithm. Numerical analyses and experiments on real-world data demonstrate that our scheme outperforms the existing ones in terms of overall training latency and achieves up to three times acceleration over random client selection, especially in scenarios with highly heterogeneous data and latencies among the clients.

Publication Source (Journal or Book title)

IEEE Internet of Things Journal

First Page

32183

Last Page

32196

This document is currently not available here.

Share

COinS