Redirecting to original paper in 30 seconds...

Click below to go immediately or wait for automatic redirect

arxiv_ai 85% Match Research Paper Federated Learning researchers,ML engineers,Data scientists,AI ethicists 1 week ago

Local Performance vs. Out-of-Distribution Generalization: An Empirical Analysis of Personalized Federated Learning in Heterogeneous Data Environments

large-language-models β€Ί evaluation
πŸ“„ Abstract

Abstract: In the context of Federated Learning with heterogeneous data environments, local models tend to converge to their own local model optima during local training steps, deviating from the overall data distributions. Aggregation of these local updates, e.g., with FedAvg, often does not align with the global model optimum (client drift), resulting in an update that is suboptimal for most clients. Personalized Federated Learning approaches address this challenge by exclusively focusing on the average local performances of clients' models on their own data distribution. Generalization to out-of-distribution samples, which is a substantial benefit of FedAvg and represents a significant component of robustness, appears to be inadequately incorporated into the assessment and evaluation processes. This study involves a thorough evaluation of Federated Learning approaches, encompassing both their local performance and their generalization capabilities. Therefore, we examine different stages within a single communication round to enable a more nuanced understanding of the considered metrics. Furthermore, we propose and incorporate a modified approach of FedAvg, designated as Federated Learning with Individualized Updates (FLIU), extending the algorithm by a straightforward individualization step with an adaptive personalization factor. We evaluate and compare the approaches empirically using MNIST and CIFAR-10 under various distributional conditions, including benchmark IID and pathological non-IID, as well as additional novel test environments with Dirichlet distribution specifically developed to stress the algorithms on complex data heterogeneity.
Authors (3)
Mortesa Hussaini
Jan Theiß
Anthony Stein
Submitted
October 28, 2025
arXiv Category
cs.LG
arXiv PDF

Key Contributions

Provides an empirical analysis of Personalized Federated Learning (PFL) in heterogeneous data environments, comparing local performance against out-of-distribution generalization. It highlights that PFL often inadequately incorporates generalization, a key benefit of methods like FedAvg, and calls for a more balanced assessment.

Business Value

Helps in developing more robust and effective federated learning systems that perform well not only on local data but also generalize to unseen data, crucial for applications where data is distributed and diverse.