In this article, we consider the problem of federated learning (FL) with training data that are non independent and identically distributed (non-IID) across the clients. To cope with data heterogeneity,… Click to show full abstract
In this article, we consider the problem of federated learning (FL) with training data that are non independent and identically distributed (non-IID) across the clients. To cope with data heterogeneity, an iterative federated clustering algorithm (IFCA) has been proposed. IFCA partitions the clients into a number of clusters and lets the clients in the same cluster optimize a shared model. However, in IFCA, the clusters are nonoverlapping, which leads to an inefficient utilization of the local information since the knowledge of a client is used by only one cluster during each round. To capture the complex nature of real-world data, soft clustering methods with overlapping clusters have been proposed that attain superior performance over the hard ones. Motivated by this, we propose a new algorithm named FL with soft clustering (FLSC) by combining the strengths of soft clustering and IFCA, where the clients are partitioned into overlapping clusters and the information of each participating client is used by multiple clusters simultaneously during each round. The experimental results show that FLSC achieves better learning performance on the classification tasks on the MNIST and Fashion-MNIST data sets, compared with the state-of-the-art baseline methods, i.e., the global model method and IFCA.
               
Click one of the above tabs to view related content.