• Login
  • Register
  • Search

Guiding Federated Learning with Low Rank Metrics

Toby Elkoushy

Abstract


Federated learning (FL) is a distributed machine learning method that protects the privacy of local data, and thus is used in scenarios like medical applications. However, FL algorithms are known to demonstrate deteriorated performance when the local data of each client is not independently and identically distributed (Non- IID). In this work, we explore the possibility of guiding the aggregation process in FL using low-rank based layer-wise metrics, and propose an aggregation framework that allows the use of any layer-wise metric to guide the aggregation process. specifically, a layer-wise metric is used as the weight for the weighted averaging process in the aggregation, and the metric is optionally smoothed by exponential averaging.
In this work, we propose two low-rank layer-wise metrics. However, experiments do not show signifi cant improvement of the proposed methods over the baseline method (FedAvg), thus this paper mainly serves as a record of our exploration and the experimental results.

Keywords


Federated learning (FL); Low rank; Non-IID data; Explainable metrics

Full Text:

PDF

Included Database


References


[1] Daniel J. Beutel, Taner Topal, Akhil Mathur, Xinchi Qiu, Titouan Parcollet, and Nicholas D. Lane. Flower: A friendly federated learning research framework. CoRR, abs/2007.14390, 2020.

[2] Chaoyang He, Songze Li, Jinhyun So, Mi Zhang, Hongyi Wang, Xiaoyang Wang, Praneeth Vepakomma, Abhishek Singh, Hang Qiu, Li Shen, Peilin Zhao, Yan Kang, Yang Liu, Ramesh Raskar, Qiang Yang, Murali Annavaram, and Salman Avestimehr. Fedml: A research library and benchmark for federated machine learning. CoRR, abs/2007.13518, 2020.

[3] Mahdi S Hosseini, Mathieu Tuli, and Konstantinos N Plataniotis. Exploiting explainable metrics for augmented sgd. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 10296–10306, 2022.

[4] Tzu-Ming Harry Hsu, Hang Qi, and Matthew Brown. Measuring the effects of non-identical data distribution for federated visual classification. CoRR, abs/1909.06335, 2019.

[5] Jakub Konecˇný, H. Brendan McMahan, Daniel Ramage, and Peter Richtárik. Federated op- timization: Distributed machine learning for on-device intelligence. CoRR, abs/1610.02527, 2016.

[6] Gihun Lee, Yongjin Shin, Minchan Jeong, and Se-Young Yun. Preservation of the global knowledge by not-true self knowledge distillation in federated learning. CoRR, abs/2106.03097, 2021.

[7] Qinbin Li, Yiqun Diao, Quan Chen, and Bingsheng He. Federated learning on non-iid data silos: An experimental study. CoRR, abs/2102.02079, 2021.

[8] H. Brendan McMahan, Eider Moore, Daniel Ramage, Seth Hampson, and Blaise Agüera y Arcas. Communication-efficient learning of deep networks from decentralized data. 2016.

[9] Bjarne Pfitzner, Nico Steckhan, and Bert Arnrich. Federated learning in a medical context: A systematic literature review, 07 2020.




DOI: http://dx.doi.org/10.18686/ahe.v7i29.10721

Refbacks