Submitted:
08 January 2025
Posted:
09 January 2025
You are already at the latest version
Abstract
Federated Learning (FL) has emerged as a transformative approach to distributed machine learning, enabling the collaborative training of models across decentralized and private datasets. Unlike traditional centralized learning paradigms, FL ensures data privacy by keeping raw data localized on client devices while leveraging aggregated updates to build global models. This survey explores the critical aspects of efficient federated learning, including communication reduction, robustness to system and data heterogeneity, and scalability in real-world applications. We discuss key techniques such as model compression, asynchronous updates, personalized learning, and robust aggregation to address challenges posed by resource-constrained devices, non-IID data distributions, and adversarial environments. Applications of FL across diverse domains, including healthcare, finance, smart cities, and autonomous systems, highlight its potential to transform industries while preserving privacy and compliance with regulatory frameworks. The survey also identifies open challenges in scalability, privacy guarantees, fairness, and ethical considerations, providing future research directions to address these gaps. As FL continues to evolve, it holds the promise of enabling privacy-preserving, collaborative intelligence on a global scale, fostering innovation while addressing critical societal and technical challenges.
Keywords:
1. Introduction
2. Background and Fundamentals of Federated Learning
2.1. Federated Learning Overview
- Model Initialization: The central server initializes a global model and shares it with participating clients.
- Local Training: Each client trains the model locally on its private data, generating updated model parameters [22].
- Model Aggregation: Clients send their locally trained model updates to the server, which aggregates them to update the global model.
- Model Dissemination: The updated global model is redistributed to clients for the next training round.
2.2. Challenges in Federated Learning
2.2.1. Communication Overhead
2.2.2. System Heterogeneity
2.2.3. Data Heterogeneity
2.2.4. Privacy and Security
2.3. Types of Federated Learning
- Horizontal Federated Learning (HFL): Clients have data with similar feature spaces but different samples [33]. HFL is common in scenarios where clients operate in the same domain, such as hospitals sharing patient records.
- Vertical Federated Learning (VFL): Clients have data with different feature spaces but overlapping samples. VFL is relevant when organizations collaborate on common users with complementary datasets [34].
- Federated Transfer Learning (FTL): Combines the principles of FL and transfer learning to enable collaboration between clients with limited overlap in both features and samples [35].
2.4. Metrics for Evaluating Federated Learning
- Model Performance: Accuracy, precision, recall, or other metrics used to assess the quality of the global model.
- Communication Efficiency: Measured in terms of the number of communication rounds or the amount of data exchanged [36].
- System Efficiency: Computational cost, resource utilization, and scalability across heterogeneous clients [37].
- Fairness: The extent to which the global model benefits all clients equitably, particularly in the presence of data heterogeneity.
- Privacy and Security: Robustness against attacks and the level of privacy guarantees provided.
3. Techniques for Improving Communication Efficiency
3.1. Model Compression Techniques
- Gradient Quantization: Gradients are quantized to a lower number of bits, reducing the size of the transmitted data [40]. For instance, techniques like QSGD (Quantized Stochastic Gradient Descent) apply fixed-point representations to gradients.
3.2. Federated Averaging and Update Frequency
3.3. Asynchronous Communication
3.4. Communication-Efficient Aggregation
- Weighted Aggregation: Aggregating updates based on client contributions, such as the size of local datasets or the quality of local models, ensures efficient use of communication resources.
- Error Feedback Mechanisms: Techniques like federated dropout and residual feedback allow clients to focus on transmitting updates that contribute most to model improvement, while others are approximated or compressed.
3.5. Advanced Techniques and Hybrid Methods
- Split Learning with FL: Splitting the model across clients and the server reduces the size of updates, as only specific layers or features are communicated [52].
- Distillation-Based FL: Knowledge distillation is used to share distilled model representations instead of full model updates, significantly reducing communication costs.
- Adaptive Communication Schedules: Dynamically adjusting the frequency of communication based on training progress, model convergence, or client conditions reduces redundant communications.
3.6. Challenges and Trade-offs
- Accuracy vs. Communication Trade-off: Over-aggressive compression or sparsification may degrade model performance, requiring careful tuning of compression parameters.
- Computation vs. Communication Trade-off: Increasing local computations to reduce communication rounds may overwhelm resource-constrained devices.
- Staleness and Convergence: Asynchronous and sparse updates may lead to stale gradients or slower convergence, requiring robust aggregation mechanisms to mitigate these effects.
4. Managing System and Data Heterogeneity
4.1. Techniques for Managing System Heterogeneity
4.1.1. Adaptive Client Participation
4.1.2. Dynamic Model Architectures
4.1.3. Federated Dropout
4.1.4. Robust Aggregation in Heterogeneous Systems
4.2. Techniques for Addressing Data Heterogeneity
4.2.1. Personalized Federated Learning
- Fine-tuning the global model on local data.
- Training personalized layers while keeping shared layers common across clients [58].
- Using meta-learning techniques, such as Model-Agnostic Meta-Learning (MAML), to optimize for client-specific learning tasks.
4.2.2. Clustered Federated Learning
4.2.3. Regularization-Based Techniques
4.2.4. Data Sharing and Synthetic Data Generation
4.3. Fairness and Equity in Federated Learning
- Fair Aggregation: Adjusting the aggregation process to prevent dominant clients from overshadowing underrepresented clients.
- Equitable Resource Allocation: Ensuring that clients with limited resources are not excluded from participation or penalized during training.
- Performance Parity: Designing global models that provide equitable benefits across diverse clients, even in the presence of significant heterogeneity [61].
4.4. Challenges and Open Problems
- Scalability: Managing heterogeneity becomes increasingly complex as the number of clients grows [62].
- Trade-offs: Balancing efficiency, accuracy, and fairness often involves trade-offs that depend on the specific FL application [63].
- Dynamic Environments: Clients’ data distributions and system conditions may evolve over time, requiring adaptive mechanisms [64].
- Evaluation Metrics: Establishing robust and universally applicable metrics for evaluating the effectiveness of heterogeneity management techniques [65].
5. Applications of Federated Learning
5.1. Healthcare and Medical Research
5.1.1. Disease Diagnosis and Prognosis
5.1.2. Drug Discovery and Genomics
5.2. Finance and Banking
5.2.1. Fraud Detection
5.2.2. Credit Scoring and Risk Assessment
5.3. Smart Cities and IoT Networks
5.3.1. Traffic Management
5.3.2. Energy Management
5.4. Natural Language Processing and Recommendation Systems
5.4.1. Predictive Text and Language Models
5.4.2. Personalized Recommendations
5.5. Autonomous Systems and Robotics
5.5.1. Autonomous Vehicles
5.5.2. Robotic Swarms
5.6. Challenges in Real-World Applications
- Scalability: Managing large-scale FL deployments with thousands or millions of clients.
- Privacy and Security: Addressing threats such as model inversion attacks or poisoning attacks in sensitive applications.
- Heterogeneous Data: Adapting to diverse data distributions and quality across clients in practical scenarios.
- Regulatory Compliance: Navigating complex regulatory frameworks and ensuring FL implementations align with legal requirements.
6. Open Challenges and Future Directions
6.1. Scalability and Communication Efficiency
- Development of ultra-efficient communication protocols that minimize data exchange without sacrificing model performance [90].
- Exploration of hierarchical FL architectures to manage large-scale deployments by leveraging intermediate aggregators or edge servers.
- Dynamic client participation schemes that prioritize high-value clients while maintaining fairness and diversity [91].
6.2. Robustness to Adversarial Attacks
- Designing robust aggregation algorithms that detect and mitigate the influence of malicious clients [94].
- Incorporating blockchain-based frameworks to enhance trust and accountability in FL systems [95].
- Developing adversarial training techniques tailored to distributed and heterogeneous environments.
6.3. Privacy Enhancements
- Advancing lightweight privacy-preserving methods that ensure strong privacy guarantees with minimal impact on system efficiency [97].
- Integrating federated learning with privacy-enhancing technologies, such as homomorphic encryption and trusted execution environments.
- Establishing formal frameworks to quantify privacy risks and trade-offs in FL systems.
6.4. Handling System and Data Heterogeneity
- Developing adaptive aggregation methods that account for client diversity and data heterogeneity [98].
- Designing personalized FL frameworks that balance global model accuracy with individual client performance [99].
- Leveraging federated transfer learning to handle scenarios with minimal overlap in data distributions across clients [100].
6.5. Evaluation and Benchmarking
- Creating comprehensive FL benchmarks that account for diverse application scenarios, client heterogeneity, and privacy constraints [102].
- Establishing metrics that measure trade-offs among accuracy, efficiency, fairness, and privacy [103].
- Conducting large-scale, real-world FL experiments to validate theoretical advancements and identify practical bottlenecks.
6.6. Cross-Disciplinary Collaboration and Ethical Considerations
- Developing frameworks for fairness-aware FL that ensure equitable model performance across diverse populations [104].
- Investigating explainable federated learning to enhance the interpretability of models in sensitive domains such as healthcare and finance.
- Establishing guidelines and standards for the ethical deployment of FL systems, considering societal impacts and regulatory compliance.
6.7. Emerging Applications and Novel Paradigms
- Exploring FL in emerging fields like autonomous systems, metaverse applications, and decentralized AI ecosystems.
- Developing hybrid paradigms that integrate FL with other machine learning approaches, such as transfer learning and self-supervised learning [105].
- Investigating cross-silo FL deployments that involve collaboration among organizations with varying trust levels and regulatory constraints [106].
7. Conclusion
References
- Ryabinin, M.; Gorbunov, E.; Plokhotnyuk, V.; Pekhimenko, G. Moshpit sgd: Communication-efficient decentralized training on heterogeneous unreliable devices. Advances in Neural Information Processing Systems 2021, 34, 18195–18211. [Google Scholar]
- Ozkara, K.; Singh, N.; Data, D.; Diggavi, S. QuPeD: Quantized Personalization via Distillation with Applications to Federated Learning. Advances in Neural Information Processing Systems 2021, 34, 3622–3634. [Google Scholar]
- Yu, P.; Liu, Y. Federated object detection: Optimizing object detection model with federated learning. In Proceedings of the Proceedings of the 3rd International Conference on Vision, Image and Signal Processing, 2019, pp.
- Lu, Y.; Huang, X.; Zhang, K.; Maharjan, S.; Zhang, Y. Low-latency federated learning and blockchain for edge association in digital twin empowered 6G networks. IEEE Transactions on Industrial Informatics 2020, 17, 5098–5107. [Google Scholar] [CrossRef]
- Mashhadi, M.B.; Shlezinger, N.; Eldar, Y.C.; Gündüz, D. Fedrec: Federated learning of universal receivers over fading channels. In Proceedings of the 2021 IEEE Statistical Signal Processing Workshop (SSP). IEEE; 2021; pp. 576–580. [Google Scholar]
- Li, T.; Sanjabi, M.; Beirami, A.; Smith, V. Fair resource allocation in federated learning. arXiv preprint, arXiv:1905.10497 2019.
- Shi, S.; Wang, Q.; Zhao, K.; Tang, Z.; Wang, Y.; Huang, X.; Chu, X. A Distributed Synchronous SGD Algorithm with Global Top-k Sparsification for Low Bandwidth Networks. In Proceedings of the 2019 IEEE 39th International Conference on Distributed Computing Systems (ICDCS); 2019; pp. 2238–2247. [Google Scholar]
- Sui, D.; Chen, Y.; Zhao, J.; Jia, Y.; Xie, Y.; Sun, W. Feded: Federated learning via ensemble distillation for medical relation extraction. In Proceedings of the Proceedings of the 2020 conference on empirical methods in natural language processing (EMNLP), 2020, pp. 2118–2128.
- Cha, H.; Park, J.; Kim, H.; Bennis, M.; Kim, S.L. Proxy experience replay: Federated distillation for distributed reinforcement learning. IEEE Intelligent Systems 2020, 35, 94–101. [Google Scholar] [CrossRef]
- He, Y.; Zenk, M.; Fritz, M. CosSGD: Nonlinear Quantization for Communication-efficient Federated Learning. CORR 2020, abs/2012.08241.
- Shi, S.; Wang, Q.; Chu, X.; Li, B.; Qin, Y.; Liu, R.; Zhao, X. Communication-Efficient Distributed Deep Learning with Merged Gradient Sparsification on GPUs. In Proceedings of the IEEE INFOCOM 2020 - IEEE Conference on Computer Communications; 2020; pp. 406–415. [Google Scholar]
- Salehi, M.; Hossain, E. Federated learning in unreliable and resource-constrained cellular wireless networks. IEEE Transactions on Communications 2021, 69, 5136–5151. [Google Scholar] [CrossRef]
- Xu, G.; Li, H.; Liu, S.; Yang, K.; Lin, X. Verifynet: Secure and verifiable federated learning. IEEE Transactions on Information Forensics and Security 2019, 15, 911–926. [Google Scholar] [CrossRef]
- Xu, H.; Ho, C.Y.; Abdelmoniem, A.M.; Dutta, A.; Bergou, E.H.; Karatsenidis, K.; Canini, M.; Kalnis, P. Compressed communication for distributed deep learning: Survey and quantitative evaluation. Technical report, 2020.
- Sun, J.; Chen, T.; Giannakis, G.B.; Yang, Q.; Yang, Z. Lazily aggregated quantized gradient innovation for communication-efficient federated learning. IEEE Transactions on Pattern Analysis and Machine Intelligence 2020. [Google Scholar] [CrossRef] [PubMed]
- Haddadpour, F.; Kamani, M.M.; Mokhtari, A.; Mahdavi, M. Federated learning with compression: Unified analysis and sharp guarantees. In Proceedings of the International Conference on Artificial Intelligence and Statistics. PMLR; 2021; pp. 2350–2358. [Google Scholar]
- Li, S.; Qi, Q.; Wang, J.; Sun, H.; Li, Y.; Yu, F.R. GGS: General Gradient Sparsification for Federated Learning in Edge Computing. In Proceedings of the ICC 2020 - 2020 IEEE International Conference on Communications (ICC); 2020; pp. 1–7. [Google Scholar]
- Alistarh, D.; Grubic, D.; Li, J.; Tomioka, R.; Vojnovic, M. QSGD: Communication-efficient SGD via gradient quantization and encoding. Advances in neural information processing systems 2017, 30. [Google Scholar]
- Amiri, M.M.; Kulkarni, S.R.; Poor, H.V. Federated learning with downlink device selection. In Proceedings of the 2021 IEEE 22nd International Workshop on Signal Processing Advances in Wireless Communications (SPAWC). IEEE; 2021; pp. 306–310. [Google Scholar]
- White paper for federated learning in mobile communication networks; China Mobile Communications Research Institute, 2021.
- Sattler, F.; Korjakow, T.; Rischke, R.; Samek, W. Fedaux: Leveraging unlabeled auxiliary data in federated learning. IEEE Transactions on Neural Networks and Learning Systems 2021. [Google Scholar] [CrossRef] [PubMed]
- Zhao, Z.; Xia, J.; Fan, L.; Lei, X.; Karagiannidis, G.K.; Nallanathan, A. System optimization of federated learning networks with a constrained latency. IEEE Transactions on Vehicular Technology 2021, 71, 1095–1100. [Google Scholar] [CrossRef]
- Sun, L.; Lyu, L. Federated model distillation with noise-free differential privacy. arXiv preprint, arXiv:2009.05537 2020.
- McMahan, B.; Moore, E.; Ramage, D.; Hampson, S.; y Arcas, B.A. Communication-efficient learning of deep networks from decentralized data. In Proceedings of the Artificial intelligence and statistics. PMLR; 2017; pp. 1273–1282. [Google Scholar]
- Das, R.; Acharya, A.; Hashemi, A.; Sanghavi, S.; Dhillon, I.S.; Topcu, U. Faster non-convex federated learning via global and local momentum. arXiv preprint, arXiv:2012.04061 2020.
- Zhang, L.; Shen, L.; Ding, L.; Tao, D.; Duan, L.Y. Fine-tuning global model via data-free knowledge distillation for non-iid federated learning. In Proceedings of the Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2022, pp.10174–10183.
- Chen, S.; Shen, C.; Zhang, L.; Tang, Y. Dynamic aggregation for heterogeneous quantization in federated learning. IEEE Transactions on Wireless Communications 2021, 20, 6804–6819. [Google Scholar] [CrossRef]
- Liu, Y.; James, J.; Kang, J.; Niyato, D.; Zhang, S. Privacy-preserving traffic flow prediction: A federated learning approach. IEEE Internet of Things Journal 2020, 7, 7751–7763. [Google Scholar] [CrossRef]
- Yang, H.; Qiu, P.; Liu, J.; Yener, A. Over-the-Air Federated Learning with Joint Adaptive Computation and Power Control. arXiv preprint, arXiv:2205.05867 2022.
- Xia, S.; Zhu, J.; Yang, Y.; Zhou, Y.; Shi, Y.; Chen, W. Fast convergence algorithm for analog federated learning. In Proceedings of the ICC 2021-IEEE International Conference on Communications. IEEE; 2021; pp. 1–6. [Google Scholar]
- Kairouz, P.; McMahan, H.B.; Avent, B.; Bellet, A.; Bennis, M.; Bhagoji, A.N.; Bonawitz, K.; Charles, Z.; Cormode, G.; Cummings, R.; et al. Advances and open problems in federated learning. Foundations and Trends® in Machine Learning 2021, 14, 1–210. [Google Scholar] [CrossRef]
- Sturluson, S.P.; Trew, S.; Muñoz-González, L.; Grama, M.; Passerat-Palmbach, J.; Rueckert, D.; Alansary, A. FedRAD: Federated Robust Adaptive Distillation. arXiv preprint, arXiv:2112.01405 2021.
- Chen, M.; Shlezinger, N.; Poor, H.V.; Eldar, Y.C.; Cui, S. Communication-efficient federated learning. Proceedings of the National Academy of Sciences 2021, 118, e2024789118. [Google Scholar] [CrossRef]
- Wall, M.E.; Rechtsteiner, A.; Rocha, L.M. Singular value decomposition and principal component analysis. In A practical approach to microarray data analysis; Springer, 2003; pp. 91–109.
- Zhu, Z.; Hong, J.; Zhou, J. Data-free knowledge distillation for heterogeneous federated learning. In Proceedings of the International Conference on Machine Learning. PMLR; 2021; pp. 12878–12889. [Google Scholar]
- Lin, Y.; Han, S.; Mao, H.; Wang, Y.; Dally, W.J. Deep gradient compression: Reducing the communication bandwidth for distributed training. arXiv preprint, arXiv:1712.01887 2017.
- Mao, Y.; Zhao, Z.; Yang, M.; Liang, L.; Liu, Y.; Ding, W.; Lan, T.; Zhang, X.P. SAFARI: Sparsity enabled Federated Learning with Limited and Unreliable Communications. arXiv preprint, arXiv:2204.02321 2022.
- Sattler, F.; Marban, A.; Rischke, R.; Samek, W. Cfd: Communication-efficient federated distillation via soft-label quantization and delta coding. IEEE Transactions on Network Science and Engineering 2021. [Google Scholar] [CrossRef]
- Aji, A.F.; Heafield, K. Sparse communication for distributed gradient descent. arXiv preprint, arXiv:1704.05021 2017.
- Cho, Y.J.; Manoel, A.; Joshi, G.; Sim, R.; Dimitriadis, D. Heterogeneous Ensemble Knowledge Transfer for Training Large Models in Federated Learning. arXiv preprint, arXiv:2204.12703 2022.
- Azam, S.S.; Hosseinalipour, S.; Qiu, Q.; Brinton, C. Recycling Model Updates in Federated Learning: Are Gradient Subspaces Low-Rank? In Proceedings of the International Conference on Learning Representations, 2021.
- Wang, H.; Sievert, S.; Liu, S.; Charles, Z.; Papailiopoulos, D.; Wright, S. Atomo: Communication-efficient learning via atomic sparsification. Advances in Neural Information Processing Systems 2018, 31. [Google Scholar]
- Wei, X.; Shen, C. Federated learning over noisy channels. In Proceedings of the ICC 2021-IEEE International Conference on Communications. IEEE; 2021; pp. 1–6. [Google Scholar]
- Shi, S.; Tang, Z.; Wang, Q.; Zhao, K.; Chu, X. Layer-wise adaptive gradient sparsification for distributed deep learning with convergence guarantees. arXiv preprint, arXiv:1911.08727 2019.
- Zniyed, Y.; Nguyen, T.P.; et al. Efficient tensor decomposition-based filter pruning. Neural Networks 2024, 178, 106393. [Google Scholar]
- Philippenko, C.; Dieuleveut, A. Bidirectional compression in heterogeneous settings for distributed or federated learning with partial participation: tight convergence guarantees. arXiv preprint, arXiv:2006.14591 2020.
- Wu, W.; He, L.; Lin, W.; Mao, R.; Maple, C.; Jarvis, S. SAFA: A semi-asynchronous protocol for fast federated learning with low overhead. IEEE Transactions on Computers 2020, 70, 655–668. [Google Scholar] [CrossRef]
- Muhammad, K.; Ullah, A.; Lloret, J.; Del Ser, J.; de Albuquerque, V.H.C. Deep learning for safe autonomous driving: Current challenges and future directions. IEEE Transactions on Intelligent Transportation Systems 2020, 22, 4316–4336. [Google Scholar] [CrossRef]
- Zamir, R.; Feder, M. On universal quantization by randomized uniform/lattice quantizers. IEEE Transactions on Information Theory 1992, 38, 428–436. [Google Scholar] [CrossRef]
- Fan, X.; Wang, Y.; Huo, Y.; Tian, Z. Communication-efficient federated learning through 1-bit compressive sensing and analog aggregation. In Proceedings of the 2021 IEEE International Conference on Communications Workshops (ICC Workshops). IEEE; 2021; pp. 1–6. [Google Scholar]
- Mothukuri, V.; Parizi, R.M.; Pouriyeh, S.; Huang, Y.; Dehghantanha, A.; Srivastava, G. A survey on security and privacy of federated learning. Future Generation Computer Systems 2021, 115, 619–640. [Google Scholar] [CrossRef]
- Jia, X.; Song, S.; He, W.; Wang, Y.; Rong, H.; Zhou, F.; Xie, L.; Guo, Z.; Yang, Y.; Yu, L.; et al. Highly scalable deep learning training system with mixed-precision: Training imagenet in four minutes. arXiv preprint, arXiv:1807.11205 2018.
- Yu, C.; Tang, H.; Renggli, C.; Kassing, S.; Singla, A.; Alistarh, D.; Zhang, C.; Liu, J. Distributed learning over unreliable networks. In Proceedings of the International Conference on Machine Learning. PMLR; 2019; pp. 7202–7212. [Google Scholar]
- Jhunjhunwala, D.; Gadhikar, A.; Joshi, G.; Eldar, Y.C. Adaptive quantization of model updates for communication-efficient federated learning. In Proceedings of the ICASSP 2021-2021 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). IEEE; 2021; pp. 3110–3114. [Google Scholar]
- Cha, H.; Park, J.; Kim, H.; Kim, S.L.; Bennis, M. Federated reinforcement distillation with proxy experience memory. arXiv preprint, arXiv:1907.06536 2019.
- Eghlidi, N.F.; Jaggi, M. Sparse communication for training deep networks. arXiv preprint, arXiv:2009.09271 2020.
- Bottou, L. Large-scale machine learning with stochastic gradient descent. Proceedings of COMPSTAT’2010; Springer; 2010; pp. 177–186. [Google Scholar]
- Li, C.; Li, G.; Varshney, P.K. Communication-efficient federated learning based on compressed sensing. IEEE Internet of Things Journal 2021, 8, 15531–15541. [Google Scholar] [CrossRef]
- Kairouz, P.; Oh, S.; Viswanath, P. Extremal mechanisms for local differential privacy. Advances in neural information processing systems 2014, 27. [Google Scholar]
- Al-Qizwini, M.; Barjasteh, I.; Al-Qassab, H.; Radha, H. Deep learning algorithm for autonomous driving using googlenet. In Proceedings of the 2017 IEEE Intelligent Vehicles Symposium (IV). IEEE; 2017; pp. 89–96. [Google Scholar]
- Rieke, N.; Hancox, J.; Li, W.; Milletari, F.; Roth, H.R.; Albarqouni, S.; Bakas, S.; Galtier, M.N.; Landman, B.A.; Maier-Hein, K.; et al. The future of digital health with federated learning. NPJ digital medicine 2020, 3, 1–7. [Google Scholar] [CrossRef] [PubMed]
- Xiao, Y.; Shi, G.; Krunz, M. Towards ubiquitous AI in 6G with federated learning. arXiv preprint, arXiv:2004.13563 2020.
- Jeong, E.; Oh, S.; Kim, H.; Park, J.; Bennis, M.; Kim, S.L. Communication-efficient on-device machine learning: Federated distillation and augmentation under non-iid private data. arXiv preprint, arXiv:1811.11479 2018.
- Konečnỳ, J.; McMahan, H.B.; Yu, F.X.; Richtárik, P.; Suresh, A.T.; Bacon, D. Federated learning: Strategies for improving communication efficiency. arXiv preprint, arXiv:1610.05492 2016.
- Sahoo, A.K.; Pradhan, C.; Barik, R.K.; Dubey, H. DeepReco: deep learning based health recommender system using collaborative filtering. Computation 2019, 7, 25. [Google Scholar] [CrossRef]
- Arandjelovic, O.; Shakhnarovich, G.; Fisher, J.; Cipolla, R.; Darrell, T. Face recognition with image sets using manifold density divergence. In Proceedings of the 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR’05). IEEE, Vol. 1; 2005; pp. 581–588. [Google Scholar]
- Lin, Z.; Liu, H.; Zhang, Y.J.A. Relay-assisted cooperative federated learning. IEEE Transactions on Wireless Communications 2022. [Google Scholar] [CrossRef]
- Ahn, J.H.; Simeone, O.; Kang, J. Wireless federated distillation for distributed edge learning with heterogeneous data. In Proceedings of the 2019 IEEE 30th Annual International Symposium on Personal, Indoor and Mobile Radio Communications (PIMRC). IEEE; 2019; pp. 1–6. [Google Scholar]
- Zhang, X.; Hong, M.; Dhople, S.; Yin, W.; Liu, Y. Fedpd: A federated learning framework with optimal rates and adaptivity to non-iid data. arXiv preprint, arXiv:2005.11418 2020.
- Jiang, Z.; Wang, W.; Li, B.; Yang, Q. Towards Efficient Synchronous Federated Training: A Survey on System Optimization Strategies. IEEE Transactions on Big Data 2022. [Google Scholar] [CrossRef]
- Feynman, R.; Vernon Jr., F. The theory of a general quantum system interacting with a linear dissipative system. Annals of Physics 1963, 24, 118–173. [Google Scholar] [CrossRef]
- Mahmoudi, A.; Ghadikolaei, H.S.; Júnior, J.M.B.D.S.; Fischione, C. FedCau: A Proactive Stop Policy for Communication and Computation Efficient Federated Learning. arXiv preprint, arXiv:2204.07773 2022.
- Zeng, Q.; Du, Y.; Huang, K.; Leung, K.K. Energy-efficient radio resource allocation for federated edge learning. In Proceedings of the 2020 IEEE International Conference on Communications Workshops (ICC Workshops). IEEE; 2020; pp. 1–6. [Google Scholar]
- Ahn, J.H.; Simeone, O.; Kang, J. Cooperative learning via federated distillation over fading channels. In Proceedings of the ICASSP 2020-2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). IEEE; 2020; pp. 8856–8860. [Google Scholar]
- Oh, S.; Park, J.; Jeong, E.; Kim, H.; Bennis, M.; Kim, S.L. Mix2FLD: Downlink federated learning after uplink federated distillation with two-way mixup. IEEE Communications Letters 2020, 24, 2211–2215. [Google Scholar] [CrossRef]
- Jiang, D.; Shan, C.; Zhang, Z. Federated learning algorithm based on knowledge distillation. In Proceedings of the 2020 International Conference on Artificial Intelligence and Computer Engineering (ICAICE). IEEE; 2020; pp. 163–167. [Google Scholar]
- Zhang, L.; Wu, D.; Yuan, X. FedZKT: Zero-Shot Knowledge Transfer towards Resource-Constrained Federated Learning with Heterogeneous On-Device Models. In Proceedings of the 2022 IEEE 42nd International Conference on Distributed Computing Systems (ICDCS). IEEE; 2022; pp. 928–938. [Google Scholar]
- Vargaftik, S.; Basat, R.B.; Portnoy, A.; Mendelson, G.; Ben-Itzhak, Y.; Mitzenmacher, M. DRIVE: one-bit distributed mean estimation. Advances in Neural Information Processing Systems 2021, 34, 362–377. [Google Scholar]
- Nishio, T.; Yonetani, R. Client selection for federated learning with heterogeneous resources in mobile edge. In Proceedings of the ICC 2019-2019 IEEE international conference on communications (ICC). IEEE; 2019; pp. 1–7. [Google Scholar]
- Lian, X.; Zhang, C.; Zhang, H.; Hsieh, C.J.; Zhang, W.; Liu, J. Can decentralized algorithms outperform centralized algorithms? a case study for decentralized parallel stochastic gradient descent. Advances in Neural Information Processing Systems 2017, 30. [Google Scholar]
- Chen, M.; Poor, H.V.; Saad, W.; Cui, S. Convergence time optimization for federated learning over wireless networks. IEEE Transactions on Wireless Communications 2020, 20, 2457–2471. [Google Scholar] [CrossRef]
- Wu, C.; Zhu, S.; Mitra, P. Federated Unlearning with Knowledge Distillation. arXiv preprint, arXiv:2201.09441 2022.
- Grigorescu, S.; Trasnea, B.; Cocias, T.; Macesanu, G. A survey of deep learning techniques for autonomous driving. Journal of Field Robotics 2020, 37, 362–386. [Google Scholar] [CrossRef]
- Chen, M.; Mao, B.; Ma, T. FedSA: A staleness-aware asynchronous Federated Learning algorithm with non-IID data. Future Generation Computer Systems 2021, 120, 1–12. [Google Scholar] [CrossRef]
- Chang, H.; Shejwalkar, V.; Shokri, R.; Houmansadr, A. Cronus: Robust and heterogeneous collaborative learning with black-box knowledge transfer. arXiv preprint, arXiv:1912.11279 2019.
- Nori, M.K.; Yun, S.; Kim, I.M. Fast federated learning by balancing communication trade-offs. IEEE Transactions on Communications 2021, 69, 5168–5182. [Google Scholar] [CrossRef]
- Itahara, S.; Nishio, T.; Koda, Y.; Morikura, M.; Yamamoto, K. Distillation-Based Semi-Supervised Federated Learning for Communication-Efficient Collaborative Training with Non-IID Private Data. IEEE Transactions on Mobile Computing 2021. [Google Scholar] [CrossRef]
- Sun, C.; Jiang, T.; Zonouz, S.; Pompili, D. Fed2KD: Heterogeneous Federated Learning for Pandemic Risk Assessment via Two-Way Knowledge Distillation. In Proceedings of the 2022 17th Wireless On-Demand Network Systems and Services Conference (WONS). IEEE; 2022; pp. 1–8. [Google Scholar]
- Zniyed, Y.; Nguyen, T.P.; et al. Enhanced network compression through tensor decompositions and pruning. IEEE Transactions on Neural Networks and Learning Systems 2024. [Google Scholar]
- Gray, R.M.; Neuhoff, D.L. Quantization. IEEE transactions on information theory 1998, 44, 2325–2383. [Google Scholar] [CrossRef]
- Han, P.; Wang, S.; Leung, K.K. Adaptive Gradient Sparsification for Efficient Federated Learning: An Online Learning Approach. In Proceedings of the 2020 IEEE 40th International Conference on Distributed Computing Systems (ICDCS); 2020; pp. 300–310. [Google Scholar]
- Zaw, C.W.; Pandey, S.R.; Kim, K.; Hong, C.S. Energy-aware resource management for federated learning in multi-access edge computing systems. IEEE Access 2021, 9, 34938–34950. [Google Scholar] [CrossRef]
- Tramèr, F.; Kurakin, A.; Papernot, N.; Goodfellow, I.; Boneh, D.; McDaniel, P. Ensemble adversarial training: Attacks and defenses. arXiv preprint, arXiv:1705.07204 2017.
- Malekijoo, A.; Fadaeieslam, M.J.; Malekijou, H.; Homayounfar, M.; Alizadeh-Shabdiz, F.; Rawassizadeh, R. Fedzip: A compression framework for communication-efficient federated learning. arXiv preprint, arXiv:2102.01593 2021.
- Suresh, A.T.; Felix, X.Y.; Kumar, S.; McMahan, H.B. Distributed mean estimation with limited communication. In Proceedings of the International conference on machine learning. PMLR; 2017; pp. 3329–3337. [Google Scholar]
- Xu, H.; Kostopoulou, K.; Dutta, A.; Li, X.; Ntoulas, A.; Kalnis, P. DeepReduce: A Sparse-tensor Communication Framework for Federated Deep Learning. In Proceedings of the Advances in Neural Information Processing Systems; Ranzato, M.; Beygelzimer, A.; Dauphin, Y.; Liang, P.; Vaughan, J.W., Eds. Curran Associates, Inc., Vol. 34; 2021; pp. 21150–21163. [Google Scholar]
- Imteaj, A.; Amini, M.H. Fedar: Activity and resource-aware federated learning model for distributed mobile robots. In Proceedings of the 2020 19th IEEE International Conference on Machine Learning and Applications (ICMLA). IEEE; 2020; pp. 1153–1160. [Google Scholar]
- Reisizadeh, A.; Mokhtari, A.; Hassani, H.; Jadbabaie, A.; Pedarsani, R. Fedpaq: A communication-efficient federated learning method with periodic averaging and quantization. In Proceedings of the International Conference on Artificial Intelligence and Statistics. PMLR; 2020; pp. 2021–2031. [Google Scholar]
- Boyd, S.; Parikh, N.; Chu, E.; Peleato, B.; Eckstein, J.; et al. Distributed optimization and statistical learning via the alternating direction method of multipliers. Foundations and Trends® in Machine learning 2011, 3, 1–122. [Google Scholar] [CrossRef]
- Hard, A.; Rao, K.; Mathews, R.; Ramaswamy, S.; Beaufays, F.; Augenstein, S.; Eichner, H.; Kiddon, C.; Ramage, D. Federated learning for mobile keyboard prediction. arXiv preprint, arXiv:1811.03604 2018.
- Donoho, D.L. Compressed sensing. IEEE Transactions on information theory 2006, 52, 1289–1306. [Google Scholar] [CrossRef]
- Study on enablers for network automation for the 5G System (5GS); 3GPP TR 23.700-91, 2020.
- Zhu, Z.; Hong, J.; Drew, S.; Zhou, J. Resilient and Communication Efficient Learning for Heterogeneous Federated Systems. In Proceedings of the International Conference on Machine Learning. PMLR; 2022; pp. 27504–27526. [Google Scholar]
- Zhang, N.; Tao, M. Gradient statistics aware power control for over-the-air federated learning in fading channels. In Proceedings of the 2020 IEEE International Conference on Communications Workshops (ICC Workshops). IEEE; 2020; pp. 1–6. [Google Scholar]
- Yang, Z.; Chen, M.; Saad, W.; Hong, C.S.; Shikh-Bahaei, M.; Poor, H.V.; Cui, S. Delay minimization for federated learning over wireless communication networks. arXiv preprint, arXiv:2007.03462 2020.
- Wu, W.; He, L.; Lin, W.; Mao, R. Accelerating federated learning over reliability-agnostic clients in mobile edge computing systems. IEEE Transactions on Parallel and Distributed Systems 2020, 32, 1539–1551. [Google Scholar] [CrossRef]
- Wu, C.; Wu, F.; Liu, R.; Lyu, L.; Huang, Y.; Xie, X. Fedkd: Communication efficient federated learning via knowledge distillation. arXiv preprint, arXiv:2108.13323 2021.
- Sergeev, A.; Del Balso, M. Horovod: fast and easy distributed deep learning in TensorFlow. arXiv preprint, arXiv:1802.05799 2018.
- Bernstein, J.; Wang, Y.X.; Azizzadenesheli, K.; Anandkumar, A. signSGD: Compressed optimisation for non-convex problems. In Proceedings of the International Conference on Machine Learning. PMLR; 2018; pp. 560–569. [Google Scholar]
- Lin, T.; Kong, L.; Stich, S.U.; Jaggi, M. Ensemble distillation for robust model fusion in federated learning. Advances in Neural Information Processing Systems 2020, 33, 2351–2363. [Google Scholar]
- Yang, Z.; Chen, M.; Saad, W.; Hong, C.S.; Shikh-Bahaei, M. Energy efficient federated learning over wireless communication networks. IEEE Transactions on Wireless Communications 2020, 20, 1935–1949. [Google Scholar] [CrossRef]
- Stich, S.U.; Cordonnier, J.B.; Jaggi, M. Sparsified SGD with memory. Advances in Neural Information Processing Systems 2018, 31. [Google Scholar]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).
