Abstract
Federated semi-supervised learning (Fed-SSL) algorithms have been developed to address the challenges of decentralized data access, data confidentiality, and costly data labeling in distributed environments. Most existing Fed-SSL algorithms are based on the federated averaging approach, which utilizes an equivalent model on all machines and replaces local models during the learning process. However, these algorithms suffer from significant communication overhead when transferring parameters of local models. In contrast, knowledge distillation-based Fed-SSL algorithms reduce communication costs by only transferring the output of local models on shared data between machines. However, these algorithms assume that all local data on the machines are labeled, and that there exists a large set of shared unlabeled data for training. These assumptions are not always feasible in real-world applications. In this paper, a knowledge distillation-based Fed-SSL algorithm has been presented, which does not make any assumptions about how the data is distributed among machines. Additionally, it artificially generates shared data required for the learning process. The learning process of the presented approach employs a semi-supervised GAN on local machines and has two stages. In the first stage, each machine trains its local model independently. In the second stage, each machine generates some artificial data in each step and propagates it to other machines. Each machine trains its discriminator with these data and the average output of all machines on these data. The effectiveness of this algorithm has been examined in terms of accuracy and the amount of communication among machines by using different data sets with different distributions. The evaluations reveal that, on average, the presented algorithm is 15% more accurate than state-of-the-art methods, especially in the case of non-IID data. In addition, in most cases, it yields better results than existing studies in terms of the amount of data communication among machines.
Access this article
Price excludes VAT (USA)
Tax calculation will be finalised during checkout.
Instant access to the full article PDF.
Data availability
The datasets supporting the results of this article are publicly available online and are referenced within the manuscript.
References
-
Farooq U, Tariq N, Asim M, Baker T, Al-Shamma’a A (2022) Machine learning and the internet of things security: solutions and open challenges. J Parallel Distrib Comput 162:89–104. https://doi.org/10.1016/j.jpdc.2022.01.015
-
Yang Q et al (2019) Federated machine learning: concept and applications. ACM Trans Intell Syst Technol (TIST) 10(2):1–19
-
Zhang C et al (2021) A survey on federated learning. Knowl Based Syst 216:10
-
Karimireddy SP et al (2020) Scaffold: stochastic controlled averaging for federated learning. In: International conference on machine learning. PMLR
-
Sattler F et al (2019) Robust and communication-efficient federated learning from non-IID data. IEEE Trans Neural Netw Learn Syst 31(9):3400–3413
-
Jeong W et al (2020) Federated semi-supervised learning with inter-client consistency. arXiv e-prints. arXiv:2006.12097
-
Itahara S, Nishio T, Koda Y, Morikura M, Yamamoto K (2021) Distillation-based semi-supervised federated learning for communication-efficient collaborative training with non-IID private data. IEEE Trans Mob Comput 22(1):191–205
-
Goodfellow I et al (2020) Generative adversarial networks. Commun ACM 63(11):139–144
-
Hinton G Vinyals O Dean J (2015) Distilling the knowledge in a neural network. Comput Sci 14(7):38–39
-
Berthelot D, Carlini N, Goodfellow I, Papernot N, Oliver A, Raffel CA (2019) Mixmatch: a holistic approach to semi-supervised learning. Adv Neural Inf Process Syst 32:1–11
-
Diao E, Ding J, Tarokh V (2021) SemiFL: communication efficient semi-supervised federated learning with unlabeled clients. arXiv preprint arXiv:2106.01432
-
Arazo E et al (2020) Pseudo-labeling and confirmation bias in deep semi-supervised learning. In: 2020 international joint conference on neural networks (IJCNN). IEEE
-
Ding G et al (2019) Feature affinity-based pseudo labeling for semi-supervised person re-identification. IEEE Trans Multimed 21(11):2891–2902
-
Jin Y et al (2020) Towards utilizing unlabeled data in federated learning: a survey and prospective. arXiv preprint arXiv:2002.11545
-
Albaseer A et al (2020) Exploiting unlabeled data in smart cities using federated edge learning. In: 2020 international wireless communications and mobile computing (IWCMC). IEEE
-
Lin H et al (2021) SemiFed: semi-supervised federated learning with consistency and pseudo-labeling. arXiv preprint arXiv:2108.09412
-
Zhang Z et al (2021) Improving semi-supervised federated learning by reducing the gradient diversity of models. In: 2021 IEEE international conference on big data (big data), Orlando, FL, USA, pp. 1214–1225, https://doi.org/10.1109/BigData52589.2021.9671693.
-
Liu Y et al (2020) RC-SSFL: towards robust and communication-efficient semi-supervised federated learning system. arXiv preprint arXiv:2012.04432
-
Zhao Y et al (2020) Semi-supervised federated learning for activity recognition. arXiv preprint arXiv:2011.00851
-
Zhong Z, Wang J, Bao W et al (2023) Semi-HFL: semi-supervised federated learning for heterogeneous devices. Complex Intell Syst 9:1995–2017. https://doi.org/10.1007/s40747-022-00894-4
-
Guha N, Talwalkar A, Smith V (2019) One-shot federated learning. arXiv preprint arXiv:1902.11175
-
Konečný J et al (2016) Federated learning: strategies for improving communication efficiency. arXiv preprint arXiv:1610.05492
-
Wang B et al (2020) Graphfl: a federated learning framework for semi-supervised node classification on graphs. arXiv preprint arXiv:2012.04187
-
Presotto R, Civitarese G, Bettini C (2022) Semi-supervised and personalized federated activity recognition based on active learning and label propagation. Pers Ubiquit Comput 26(5):1281–1298
-
Park J, Yoon D, Yeo S, Oh S (2022) AMBLE: Adjusting mini-batch and local epoch for federated learning with heterogeneous devices. J Parallel Distrib Comput 170:13–23. https://doi.org/10.1016/j.jpdc.2022.07.009
-
Shang X, Huang G, Lu Y, Lou J, Han B, Cheung YM, Wang H (2023) Federated semi-supervised learning with annotation heterogeneity. arXiv preprint arXiv:2303.02445
-
Jeong E et al (2018) Communication-efficient on-device machine learning: federated distillation and augmentation under non-iid private data. arXiv preprint arXiv:1811.11479
-
Li D Wang J (2019) FedMD: Heterogenous federated learning via model distillation. arXiv:1910.03581
-
Lin T, Kong L, Stich SU, Jaggi M (2020) Ensemble distillation for robust model fusion in federated learning. Adv Neural Inf Process Syst 33:2351–2363
-
Gong X et al (2021) Ensemble attention distillation for privacy-preserving federated learning. In: Proceedings of the IEEE/CVF international conference on computer vision
-
Rasmus A, Berglund M, Honkala M, Valpola H, Raiko T (2015) Semi-supervised learning with ladder networks. Adv Neural Inf Process Syst 28:1–9
-
Sohn K et al (2020) Fixmatch: simplifying semi-supervised learning with consistency and confidence. Adv Neural Inf Process Syst 33:596–608
-
Li M, Li Q, Wang Y (2023) Class balanced adaptive pseudo labeling for federated semi-supervised learning. 2023 IEEE/CVF conference on computer vision and pattern recognition (CVPR). Vancouver, BC, Canada, pp 16292–16301. https://doi.org/10.1109/CVPR52729.2023.01563
-
Kramer MA (1991) Nonlinear principal component analysis using autoassociative neural networks. AIChE J 37(2):233–243
-
Kingma DP, Welling M (2019) An introduction to variational autoencoders. Found Trends® Mach Learn 12(4):307–392
-
Creswell A et al (2018) Generative adversarial networks: an overview. IEEE Signal Process Mag 35(1):53–65
-
Fan C, Liu P (2020) Federated generative adversarial learning. In: Chinese conference on pattern recognition and computer vision (PRCV). Springer
-
Rasouli M, Sun T, Rajagopal R (2020) Fedgan: federated generative adversarial networks for distributed data. arXiv preprint arXiv:2006.07228
-
Hardy C, Le Merrer E, Sericola B (2019) Md-gan: multi-discriminator generative adversarial networks for distributed datasets. In: 2019 IEEE international parallel and distributed processing symposium (IPDPS). IEEE
-
Salimans T et al (2016) Improved techniques for training gans. Adv Neural Inf Process Syst 29:2234–2242
-
Chen M, Liao W, Zha H, Zhao T (2022) Distribution approximation and statistical estimation guarantees of generative adversarial networks. arXiv [cs.LG]
-
Xiao H, Rasul K, Vollgraf R (2017) Fashion-MNIST: a novel image dataset for benchmarking machine learning algorithms. https://github.com/zalandoresearch/fashion-mnist. Accessed 16 Sep 2023
-
Lewis D Reuters 21578 data set. http://kdd.ics.uci.edu/databases/reuters21578/reuters21578.html. Accessed 16 Sep 2023
-
Maas AL, Daly RE, Pham PT, Huang D, Ng AY, Potts C (2011) Learning word vectors for sentiment analysis. In: Proceedings of the 49th annual meeting of the association for computational linguistics: Hum Language Tech pp 142–150, Portland, Oregon, USA.
-
Chowdhury MEH et al (2020) Can AI help in screening viral and COVID-19 pneumonia? IEEE Access 8:132665–132676. https://doi.org/10.1109/ACCESS.2020.3010287
-
Rahman T et al (2021) Exploring the effect of image enhancement techniques on COVID-19 detection using chest X-ray images. Comput Biol Med 132:104319
Ethics declarations
Conflict of interest
The authors declare no competing interests.
Additional information
Publisher’s Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
About this article
Cite this article
Karami, A., Ramezani, R. & Baraani Dastjerdi, A. GFD-SSL: generative federated knowledge distillation-based semi-supervised learning. Int. J. Mach. Learn. & Cyber. (2024). https://doi.org/10.1007/s13042-024-02256-7
-
Received:
-
Accepted:
-
Published:
-
DOI: https://doi.org/10.1007/s13042-024-02256-7