gfd-ssl:-generative-federated-knowledge-distillation-based-semi-supervised-learning-|-international-journal-of-…-–-springer

GFD-SSL: generative federated knowledge distillation-based semi-supervised learning | International Journal of … – Springer

Abstract

Federated semi-supervised learning (Fed-SSL) algorithms have been developed to address the challenges of decentralized data access, data confidentiality, and costly data labeling in distributed environments. Most existing Fed-SSL algorithms are based on the federated averaging approach, which utilizes an equivalent model on all machines and replaces local models during the learning process. However, these algorithms suffer from significant communication overhead when transferring parameters of local models. In contrast, knowledge distillation-based Fed-SSL algorithms reduce communication costs by only transferring the output of local models on shared data between machines. However, these algorithms assume that all local data on the machines are labeled, and that there exists a large set of shared unlabeled data for training. These assumptions are not always feasible in real-world applications. In this paper, a knowledge distillation-based Fed-SSL algorithm has been presented, which does not make any assumptions about how the data is distributed among machines. Additionally, it artificially generates shared data required for the learning process. The learning process of the presented approach employs a semi-supervised GAN on local machines and has two stages. In the first stage, each machine trains its local model independently. In the second stage, each machine generates some artificial data in each step and propagates it to other machines. Each machine trains its discriminator with these data and the average output of all machines on these data. The effectiveness of this algorithm has been examined in terms of accuracy and the amount of communication among machines by using different data sets with different distributions. The evaluations reveal that, on average, the presented algorithm is 15% more accurate than state-of-the-art methods, especially in the case of non-IID data. In addition, in most cases, it yields better results than existing studies in terms of the amount of data communication among machines.

Access this article

Log in via an institution

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Data availability

The datasets supporting the results of this article are publicly available online and are referenced within the manuscript.

References

  1. Farooq U, Tariq N, Asim M, Baker T, Al-Shamma’a A (2022) Machine learning and the internet of things security: solutions and open challenges. J Parallel Distrib Comput 162:89–104. https://doi.org/10.1016/j.jpdc.2022.01.015

    Article  Google Scholar 

  2. Yang Q et al (2019) Federated machine learning: concept and applications. ACM Trans Intell Syst Technol (TIST) 10(2):1–19

    Article  Google Scholar 

  3. Zhang C et al (2021) A survey on federated learning. Knowl Based Syst 216:10

    Article  Google Scholar 

  4. Karimireddy SP et al (2020) Scaffold: stochastic controlled averaging for federated learning. In: International conference on machine learning. PMLR

  5. Sattler F et al (2019) Robust and communication-efficient federated learning from non-IID data. IEEE Trans Neural Netw Learn Syst 31(9):3400–3413

    Article  Google Scholar 

  6. Jeong W et al (2020) Federated semi-supervised learning with inter-client consistency. arXiv e-prints. arXiv:2006.12097

  7. Itahara S, Nishio T, Koda Y, Morikura M, Yamamoto K (2021) Distillation-based semi-supervised federated learning for communication-efficient collaborative training with non-IID private data. IEEE Trans Mob Comput 22(1):191–205

    Article  Google Scholar 

  8. Goodfellow I et al (2020) Generative adversarial networks. Commun ACM 63(11):139–144

    Article  MathSciNet  Google Scholar 

  9. Hinton G Vinyals O Dean J (2015) Distilling the knowledge in a neural network. Comput Sci 14(7):38–39

    Google Scholar 

  10. Berthelot D, Carlini N, Goodfellow I, Papernot N, Oliver A, Raffel CA (2019) Mixmatch: a holistic approach to semi-supervised learning. Adv Neural Inf Process Syst 32:1–11

  11. Diao E, Ding J, Tarokh V (2021) SemiFL: communication efficient semi-supervised federated learning with unlabeled clients. arXiv preprint arXiv:2106.01432

  12. Arazo E et al (2020) Pseudo-labeling and confirmation bias in deep semi-supervised learning. In: 2020 international joint conference on neural networks (IJCNN). IEEE

  13. Ding G et al (2019) Feature affinity-based pseudo labeling for semi-supervised person re-identification. IEEE Trans Multimed 21(11):2891–2902

    Article  Google Scholar 

  14. Jin Y et al (2020) Towards utilizing unlabeled data in federated learning: a survey and prospective. arXiv preprint arXiv:2002.11545

  15. Albaseer A et al (2020) Exploiting unlabeled data in smart cities using federated edge learning. In: 2020 international wireless communications and mobile computing (IWCMC). IEEE

  16. Lin H et al (2021) SemiFed: semi-supervised federated learning with consistency and pseudo-labeling. arXiv preprint arXiv:2108.09412

  17. Zhang Z et al (2021) Improving semi-supervised federated learning by reducing the gradient diversity of models. In: 2021 IEEE international conference on big data (big data), Orlando, FL, USA, pp. 1214–1225, https://doi.org/10.1109/BigData52589.2021.9671693.

  18. Liu Y et al (2020) RC-SSFL: towards robust and communication-efficient semi-supervised federated learning system. arXiv preprint arXiv:2012.04432

  19. Zhao Y et al (2020) Semi-supervised federated learning for activity recognition. arXiv preprint arXiv:2011.00851

  20. Zhong Z, Wang J, Bao W et al (2023) Semi-HFL: semi-supervised federated learning for heterogeneous devices. Complex Intell Syst 9:1995–2017. https://doi.org/10.1007/s40747-022-00894-4

    Article  Google Scholar 

  21. Guha N, Talwalkar A, Smith V (2019) One-shot federated learning. arXiv preprint arXiv:1902.11175

  22. Konečný J et al (2016) Federated learning: strategies for improving communication efficiency. arXiv preprint arXiv:1610.05492

  23. Wang B et al (2020) Graphfl: a federated learning framework for semi-supervised node classification on graphs. arXiv preprint arXiv:2012.04187

  24. Presotto R, Civitarese G, Bettini C (2022) Semi-supervised and personalized federated activity recognition based on active learning and label propagation. Pers Ubiquit Comput 26(5):1281–1298

    Article  Google Scholar 

  25. Park J, Yoon D, Yeo S, Oh S (2022) AMBLE: Adjusting mini-batch and local epoch for federated learning with heterogeneous devices. J Parallel Distrib Comput 170:13–23. https://doi.org/10.1016/j.jpdc.2022.07.009

    Article  Google Scholar 

  26. Shang X, Huang G, Lu Y, Lou J, Han B, Cheung YM, Wang H (2023) Federated semi-supervised learning with annotation heterogeneity. arXiv preprint arXiv:2303.02445

  27. Jeong E et al (2018) Communication-efficient on-device machine learning: federated distillation and augmentation under non-iid private data. arXiv preprint arXiv:1811.11479

  28. Li D Wang J (2019) FedMD: Heterogenous federated learning via model distillation. arXiv:1910.03581

  29. Lin T, Kong L, Stich SU, Jaggi M (2020) Ensemble distillation for robust model fusion in federated learning. Adv Neural Inf Process Syst 33:2351–2363

    Google Scholar 

  30. Gong X et al (2021) Ensemble attention distillation for privacy-preserving federated learning. In: Proceedings of the IEEE/CVF international conference on computer vision

  31. Rasmus A, Berglund M, Honkala M, Valpola H, Raiko T (2015) Semi-supervised learning with ladder networks. Adv Neural Inf Process Syst 28:1–9

  32. Sohn K et al (2020) Fixmatch: simplifying semi-supervised learning with consistency and confidence. Adv Neural Inf Process Syst 33:596–608

    Google Scholar 

  33. Li M, Li Q, Wang Y (2023) Class balanced adaptive pseudo labeling for federated semi-supervised learning. 2023 IEEE/CVF conference on computer vision and pattern recognition (CVPR). Vancouver, BC, Canada, pp 16292–16301. https://doi.org/10.1109/CVPR52729.2023.01563

  34. Kramer MA (1991) Nonlinear principal component analysis using autoassociative neural networks. AIChE J 37(2):233–243

    Article  Google Scholar 

  35. Kingma DP, Welling M (2019) An introduction to variational autoencoders. Found Trends® Mach Learn 12(4):307–392

    Article  Google Scholar 

  36. Creswell A et al (2018) Generative adversarial networks: an overview. IEEE Signal Process Mag 35(1):53–65

    Article  Google Scholar 

  37. Fan C, Liu P (2020) Federated generative adversarial learning. In: Chinese conference on pattern recognition and computer vision (PRCV). Springer

  38. Rasouli M, Sun T, Rajagopal R (2020) Fedgan: federated generative adversarial networks for distributed data. arXiv preprint arXiv:2006.07228

  39. Hardy C, Le Merrer E, Sericola B (2019) Md-gan: multi-discriminator generative adversarial networks for distributed datasets. In: 2019 IEEE international parallel and distributed processing symposium (IPDPS). IEEE

  40. Salimans T et al (2016) Improved techniques for training gans. Adv Neural Inf Process Syst 29:2234–2242

    Google Scholar 

  41. Chen M, Liao W, Zha H, Zhao T (2022) Distribution approximation and statistical estimation guarantees of generative adversarial networks. arXiv [cs.LG]

  42. Xiao H, Rasul K, Vollgraf R (2017) Fashion-MNIST: a novel image dataset for benchmarking machine learning algorithms. https://github.com/zalandoresearch/fashion-mnist. Accessed 16 Sep 2023

  43. Lewis D Reuters 21578 data set. http://kdd.ics.uci.edu/databases/reuters21578/reuters21578.html. Accessed 16 Sep 2023

  44. Maas AL, Daly RE, Pham PT, Huang D, Ng AY, Potts C (2011) Learning word vectors for sentiment analysis. In: Proceedings of the 49th annual meeting of the association for computational linguistics: Hum Language Tech pp 142–150, Portland, Oregon, USA.

  45. Chowdhury MEH et al (2020) Can AI help in screening viral and COVID-19 pneumonia? IEEE Access 8:132665–132676. https://doi.org/10.1109/ACCESS.2020.3010287

    Article  Google Scholar 

  46. Rahman T et al (2021) Exploring the effect of image enhancement techniques on COVID-19 detection using chest X-ray images. Comput Biol Med 132:104319

    Article  Google Scholar 

Download references

Author information

Authors and Affiliations

  1. Faculty of Computer Engineering, University of Isfahan, Isfahan, Iran

    Ali Karami, Reza Ramezani & Ahmad Baraani Dastjerdi

Contributions

All authors contributed to the conception, design, and execution of the research. They collectively conducted experiments, analyzed the data, and jointly wrote the manuscript. All authors reviewed and approved the final version of the manuscript.

Corresponding author

Correspondence to Reza Ramezani.

Ethics declarations

Conflict of interest

The authors declare no competing interests.

Additional information

Publisher’s Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Karami, A., Ramezani, R. & Baraani Dastjerdi, A. GFD-SSL: generative federated knowledge distillation-based semi-supervised learning. Int. J. Mach. Learn. & Cyber. (2024). https://doi.org/10.1007/s13042-024-02256-7

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1007/s13042-024-02256-7

Keywords