Personalized Federated Learning Algorithm Based on Decoupled Self-distillation
Author:
  • Article
  • | |
  • Metrics
  • |
  • Reference [25]
  • | |
  • Cited by
  • | |
  • Comments
    Abstract:

    Federated learning (FL) is an emerging distributed machine learning framework aimed at addressing issues of data privacy protection and efficient distributed computing. It allows multiple clients to collaboratively train a global model without sharing their data. However, due to the heterogeneity in the data distribution of each client, a single global model often fails to meet the personalized needs of different clients. To address this issue, this paper proposes a federated learning algorithm that combines self-distillation and decoupled knowledge distillation. The algorithm retains the client’s historical model as a teacher model to distill and guide the training of the local model, and after obtaining a new local model, it is uploaded to the server for weighted averaging and aggregation. In the knowledge distillation process, the decoupled distillation of target class knowledge and non-target class knowledge allows for a more thorough transmission of personalized knowledge. Experimental results show that the proposed method outperforms existing federated learning methods in classification accuracy on the CIFAR-10 and CIFAR-100 datasets.

    Reference
    [1] McMahan B, Moore E, Ramage D, et al. Communication-efficient learning of deep networks from decentralized data. Proceedings of the 20th International Conference on Artificial Intelligence and Statistics. Fort Lauderdale: PMLR, 2017. 1273–1282.
    [2] Sui DB, Chen YB, Zhao J, et al. FedED: Federated learning via ensemble distillation for medical relation extraction. Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP). ACL, 2020. 2118–2128.
    [3] Kaleem S, Sohail A, Tariq MU, et al. An improved big data analytics architecture using federated learning for IoT-enabled urban intelligent transportation systems. Sustainability, 2023, 15(21): 15333.
    [4] Liu M, Ho S, Wang MQ, et al. Federated learning meets natural language processing: A survey. arXiv:2107.12603, 2021.
    [5] Lin BY, He CY, Zeng ZH, et al. FedNLP: Benchmarking federated learning methods for natural language processing tasks. Findings of the Association for Computational Linguistics: NAACL 2022. Seattle: ACL, 2022. 157–175.
    [6] Zhao Y, Li M, Lai LZ, et al. Federated learning with non-IID data. arXiv:1806.00582, 2018.
    [7] Li T, Sahu AK, Zaheer M, et al. Federated optimization in heterogeneous networks. Proceedings of the 3rd Conference on Machine Learning and Systems. Austin, 2020. 429–450.
    [8] Tang XY, Guo S, Guo JC. Personalized federated learning with contextualized generalization. Proceedings of the 31st International Joint Conference on Artificial Intelligence. Vienna: IJCAI, 2022. 2241–2247.
    [9] Aljahdali M, Abdelmoniem AM, Canini M, et al. Flashback: Understanding and mitigating forgetting in federated learning. arXiv:2402.05558, 2024.
    [10] Xu YC, Ma WB, Dai CF, et al. Generalized federated learning via gradient norm-aware minimization and control variables. Mathematics, 2024, 12(17): 2644.
    [11] Hinton G, Vinyals O, Dean J. Distilling the knowledge in a neural network. arXiv:1503.02531, 2015.
    [12] Zhao BR, Cui Q, Song RJ, et al. Decoupled knowledge distillation. Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition. New Orleans: IEEE, 2022. 11943–11952.
    [13] Smith V, Chiang CK, Sanjabi M, et al. Federated multi-task learning. Proceedings of the 31st International Conference on Neural Information Processing Systems. Long Beach: Curran Associates Inc., 2017. 4427–4437.
    [14] Mills J, Hu J, Min GY. Multi-task federated learning for personalised deep neural networks in edge computing. IEEE Transactions on Parallel and Distributed Systems, 2022, 33(3): 630–641.
    [15] Finn C, Abbeel P, Levine S. Model-agnostic meta-learning for fast adaptation of deep networks. Proceedings of the 34th International Conference on Machine Learning. Sydney: PMLR, 2017. 1126–1135.
    [16] Jiang YH, Konečný J, Rush K, et al. Improving federated learning personalization via model agnostic meta learning. arXiv:1909.12488, 2019.
    [17] Fallah A, Mokhtari A, Ozdaglar A. Personalized federated learning with theoretical guarantees: A model-agnostic meta-learning approach. Proceedings of the 34th International Conference on Neural Information Processing Systems. Vancouver: Curran Associates Inc., 2020. 300.
    [18] He CY, Annavaram M, Avestimehr S. Group knowledge transfer: Federated learning of large CNNs at the edge. Proceedings of the 34th International Conference on Neural Information Processing Systems. Vancouver: Curran Associates Inc., 2020. 1180.
    [19] Jeong E, Oh S, Kim H, et al. Communication-efficient on-device machine learning: Federated distillation and augmentation under non-IID private data. arXiv:1811.11479, 2018.
    [20] Jin H, Bai DS, Yao DZ, et al. Personalized edge intelligence via federated self-knowledge distillation. IEEE Transactions on Parallel and Distributed Systems, 2023, 34(2): 567–580.
    [21] Lee G, Shin Y, Jeong M, et al. Preservation of the global knowledge by not-true self knowledge distillation in federated learning. arXiv:2106.03097, 2021.
    [22] Krizhevsky A, Hinton G. Learning multiple layers of features from tiny images. https://www.researchgate.net/publication/265748773_Learning_Multiple_Layers_of_Features_from_Tiny_Images. [2024-10-12].
    [23] Arivazhagan MG, Aggarwal V, Singh AK, et al. Federated learning with personalization layers. arXiv:1912.00818, 2019.
    [24] Dinh CT, Tran N, Nguyen J. Personalized federated learning with Moreau envelopes. Proceedings of the 34th International Conference on Neural Information Processing Systems. Vancouver: Curran Associates Inc., 2020. 1796.
    [25] Zhang M, Sapra K, Fidler S, et al. Personalized federated learning with first order model optimization. Proceedings of the 9th International Conference on Learning Representations. OpenReview.net, 2021.
    Related
    Cited by
    Comments
    Comments
    分享到微博
    Submit
Get Citation

闵和祥,朱子奇.基于解耦自蒸馏的个性化联邦学习算法.计算机系统应用,,():1-8

Copy
Share
Article Metrics
  • Abstract:37
  • PDF: 144
  • HTML: 0
  • Cited by: 0
History
  • Received:October 22,2024
  • Revised:November 07,2024
  • Online: February 28,2025
Article QR Code
You are the first990937Visitors
Copyright: Institute of Software, Chinese Academy of Sciences Beijing ICP No. 05046678-3
Address:4# South Fourth Street, Zhongguancun,Haidian, Beijing,Postal Code:100190
Phone:010-62661041 Fax: Email:csa (a) iscas.ac.cn
Technical Support:Beijing Qinyun Technology Development Co., Ltd.

Beijing Public Network Security No. 11040202500063