Standard federated learning often operates under simplifying assumptions, such as clients having identically distributed data and similar computational resources. Real-world deployments frequently violate these assumptions. Client datasets typically exhibit statistical heterogeneity (Non-IID data), meaning the local data distribution Pk(x,y) for a client k can differ substantially from the global distribution P(x,y). Concurrently, variations in client hardware, network speed, and availability introduce systems heterogeneity. Both types can hinder convergence, degrade model accuracy, and lead to fairness issues.
This chapter provides techniques to address these common difficulties. We will investigate methods designed to mitigate the negative effects of Non-IID data and system variability. Recognizing that heterogeneity implies a single global model may not serve all clients optimally, we will also study personalization in federated learning. You will learn about strategies including:
4.1 Sources of Heterogeneity: Statistical and System
4.2 Techniques for Handling Non-IID Data Distributions
4.3 Clustered Federated Learning Approaches
4.4 Meta-Learning for Federated Personalization
4.5 Multi-Task Learning in Federated Settings
4.6 Model Pruning and Adaptation for Device Constraints
4.7 Practice: Simulating Non-IID Data and Mitigation
© 2025 ApX Machine Learning