https://scholars.lib.ntu.edu.tw/handle/123456789/629192
標題: | Robust Aggregation for Federated Learning by Minimum γ-Divergence Estimation | 作者: | Li, Cen-Jhih Huang, Pin-Han Ma, Yi-Ting HUNG HUNG Huang, Su-Yun |
關鍵字: | byzantine problem; density power divergence; federated learning; influence function; robustness; γ-divergence | 公開日期: | 13-五月-2022 | 出版社: | MDPI | 卷: | 24 | 期: | 5 | 來源出版物: | Entropy | 摘要: | Federated learning is a framework for multiple devices or institutions, called local clients, to collaboratively train a global model without sharing their data. For federated learning with a central server, an aggregation algorithm integrates model information sent from local clients to update the parameters for a global model. Sample mean is the simplest and most commonly used aggregation method. However, it is not robust for data with outliers or under the Byzantine problem, where Byzantine clients send malicious messages to interfere with the learning process. Some robust aggregation methods were introduced in literature including marginal median, geometric median and trimmed-mean. In this article, we propose an alternative robust aggregation method, named γ-mean, which is the minimum divergence estimation based on a robust density power divergence. This γ-mean aggregation mitigates the influence of Byzantine clients by assigning fewer weights. This weighting scheme is data-driven and controlled by the γ value. Robustness from the viewpoint of the influence function is discussed and some numerical results are presented. |
URI: | https://scholars.lib.ntu.edu.tw/handle/123456789/629192 | ISSN: | 1099-4300 | DOI: | 10.3390/e24050686 |
顯示於: | 流行病學與預防醫學研究所 |
在 IR 系統中的文件,除了特別指名其著作權條款之外,均受到著作權保護,並且保留所有的權利。