Redirecting to original paper in 30 seconds...

Click below to go immediately or wait for automatic redirect

arxiv_ml 95% Match Research Paper AI Researchers,Machine Learning Engineers,Privacy Experts,Developers working with decentralized data 3 weeks ago

FedMMKT:Co-Enhancing a Server Text-to-Image Model and Client Task Models in Multi-Modal Federated Learning

large-language-models › multimodal-llms
📄 Abstract

Abstract: Text-to-Image (T2I) models have demonstrated their versatility in a wide range of applications. However, adaptation of T2I models to specialized tasks is often limited by the availability of task-specific data due to privacy concerns. On the other hand, harnessing the power of rich multimodal data from modern mobile systems and IoT infrastructures presents a great opportunity. This paper introduces Federated Multi-modal Knowledge Transfer (FedMMKT), a novel framework that enables co-enhancement of a server T2I model and client task-specific models using decentralized multimodal data without compromising data privacy.
Authors (7)
Ningxin He
Yang Liu
Wei Sun
Xiaozhou Ye
Ye Ouyang
Tiegang Gao
+1 more
Submitted
October 14, 2025
arXiv Category
cs.LG
arXiv PDF

Key Contributions

Introduces Federated Multi-modal Knowledge Transfer (FedMMKT), a novel framework for co-enhancing a server Text-to-Image (T2I) model and client task-specific models using decentralized multimodal data without compromising privacy. It enables adaptation of T2I models to specialized tasks where data is scarce or private.

Business Value

Allows businesses to leverage sensitive, decentralized user data (e.g., from mobile devices) to create personalized and specialized AI services without centralizing raw data, enhancing privacy and utility.