Skip to content

Latest commit

 

History

History
64 lines (42 loc) · 7.26 KB

File metadata and controls

64 lines (42 loc) · 7.26 KB

Back to https://github.com/Kwangkee/FL


■ 공정한 연합학습 (Fair Federated Learning)

[Ditto: Fair and Robust Federated Learning Through Personalization], https://proceedings.mlr.press/v139/li21h.html
[InclusiveFL: No One Left Behind: Inclusive Federated Learning over Heterogeneous Devices], https://dl.acm.org/doi/10.1145/3534678.3539086

Improving Fairness via Federated Learning, https://arxiv.org/abs/2110.15545?context=cs
KDD 2021 Tutorial on Towards Fair Federated Learning, www.cas.mcmaster.ca/~chul9/Contents/KDD_2021_Tutorial.html

Ditto

Ditto: Fair and Robust Federated Learning Through Personalization, https://proceedings.mlr.press/v139/li21h.html
Github: https://github.com/litian96/ditto

Fairness and robustness are two important concerns for federated learning systems. In this work, we identify that

  • robustness to data and model poisoning attacks and
  • fairness, measured as the uniformity of performance across devices,

are competing constraints in statistically heterogeneous networks. To address these constraints, we propose employing a simple, general framework for personalized federated learning, Ditto, that can inherently provide fairness and robustness benefits, and develop a scalable solver for it. Theoretically, we analyze the ability of Ditto to achieve fairness and robustness simultaneously on a class of linear problems. Empirically, across a suite of federated datasets, we show that Ditto not only achieves competitive performance relative to recent personalization methods, but also enables more accurate, robust, and fair models relative to state-of-the-art fair or robust baselines.

Simultaneously satisfying these varied constraints can be exceptionally difficult (Kairouz et al., 2019).

Many prior efforts have separately considered fairness or robustness in federated learning.

we show that the constraints of fairness and robustness can directly compete with one another when training a single global model, and that simultaneously optimizing for accuracy, fairness, and robustness requires careful consideration.

For example, as we empirically demonstrate (Section 4), current fairness approaches can render FL systems highly susceptible to training time attacks from malicious devices. On the other hand, robust methods may filter out rare but informative updates, causing unfairness (Wang et al., 2020).

Our work differs from these approaches by simultaneously learning local and global models via a global-regularized MTL framework, which applies to non-convex ML objectives.

Finally, a key contribution of our work is jointly exploring the robustness and fairness benefits of personalized FL.

The benefits of personalization for fairness alone have been demonstrated empirically in prior work (Wang et al., 2019; Hao et al., 2020). Connections between personalization and robustness have also been explored in Yu et al. (2020), although the authors propose using personalization methods on top of robust mechanisms. Our work differs from these works by arguing that MTL itself offers inherent robustness and fairness benefits, and exploring the challenges that exist when attempting to satisfy both constraints simultaneously.

Here the hyperparameter λ controls the interpolation between local and global models.

  • When λ is set to 0, Ditto is reduced to training local models;
  • as λ grows large, it recovers global model objective (Global Obj) (λ → +∞).

Ditto with an appropriate value of λ offers a tradeoff between these two extremes:

the smaller λ, the more the personalized models vk can deviate from the (corrupted) global model w, potentially providing robustness at the expense of generalization. In the heterogeneous case (which can lead to issues of unfairness as described in Section 2), a finite λ exists to offer robustness and fairness jointly. We explore these ideas more rigorously in Section 3.3 by analyzing the tradeoffs between accuracy, fairness, and robustness in terms of λ for a class of linear regression problems, and demonstrate fairness/robustness benefits of Ditto empirically in Section 4.

InclusiveFL

No One Left Behind: Inclusive Federated Learning over Heterogeneous Devices, https://dl.acm.org/doi/10.1145/3534678.3539086

The straightforward solutions like removing the weak clients or using a small model to fit all clients will lead to some problems, such as under-representation of dropped clients and inferior accuracy due to data loss or limited model representation ability. In this work, we propose InclusiveFL, a client-inclusive federated learning method to handle this problem.

  • The core idea of InclusiveFL is to assign models of different sizes to clients with different computing capabilities, bigger models for powerful clients and smaller ones for weak clients.
  • We also propose an effective method to share the knowledge among local models with different sizes.

In this way, all the clients can participate in FL training, and the final model can be big and powerful enough. Besides, we propose a momentum knowledge distillation method to better transfer knowledge in big models on powerful clients to the small models on weak clients. Extensive experiments on many real-world benchmark datasets demonstrate the effectiveness of InclusiveFL in learning accurate models from clients with heterogeneous devices under the FL framework.

Hence, conventional federated learning methods [25, 31] make an essential assumption that all clients have sufficient local resources to train models with the same architecture.

  • A straightforward way [1] is to drop weak clients and only aggregate parameters from powerful clients with sufficient resources. But excluding data on weak devices in training leads to fairness issues [13, 28] because weak clients would be underrepresented in the final global model. Furthermore, the loss of data on weak clients causes an inferior accuracy [29], especially when the amount of weak clients is enormous (e.g., in less developed areas).
  • Alternatively, a client-inclusive baseline is choosing a small global model to fit the minimum capability that all clients can offer. However, the representation ability of the final global model would be largely limited by the small model architecture [24].

In this paper, we propose a client-inclusive federated training solution InclusiveFL that can train a large model over devices with heterogeneous capabilities, and address the above key challenges. InclusiveFL assigns models of different sizes to clients with different computing capabilities,

  • bigger models for powerful clients and
  • smaller ones for weak clients.

We propose a layer-wise heterogeneous aggregation method to update the parameters of shared bottom layers. Furthermore, since the small model on weak clients may not be strong enough, we propose to transfer the knowledge learned by large models on powerful clients to small models on weak clients via a momentum distillation method [46]. Intuitively, by encouraging the top encoder layer in small models to imitate the behavior of top encoder layers in a larger model, we can achieve effective knowledge transfer in InclusiveFL.


Back to Papers
Back to https://github.com/Kwangkee/FL