Distributed Collaborative Learning with Representative Knowledge Sharing
Keren Li
Speaker
University of Alabama at Birmingham
Monday, Aug 4: 10:35 AM - 10:55 AM
Topic-Contributed Paper Session
Music City Center
Distributed Collaborative Learning (DCL) addresses critical challenges in privacy-aware machine learning by enabling indirect knowledge transfer across nodes with heterogeneous feature distributions. Unlike conventional federated learning approaches, DCL assumes non-IID data and prediction task distributions that span beyond local training data, requiring selective collaboration to achieve generalization. In this work, we propose a novel Collaborative Transfer Learning (CTL) framework that utilizes representative datasets and adaptive distillation weights to facilitate efficient and privacy-preserving collaboration. By quantifying node similarity via Distributed Energy Coefficients, approximated from Taylor-expanded energy distance, CTL dynamically selects optimal collaborators and refines local models through knowledge distillation on shared representative datasets. These representatives, locally constructed synthetic datasets that encode conditional information, serve as a common ground for knowledge exchange and model comparison. We highlight how Representative Learning enables quantification of model heterogeneity, facilitates transfer under non-IID task distributions, and supports scalable generalization. Simulations demonstrate the benefit of adaptive collaboration, with CTL achieving superior trade-offs between personalization and global coordination. We also discuss a taxonomy of data heterogeneity types, including newly defined model and representation divergences, and illustrate their relevance to node alignment and collaborative efficiency.
Collaborative Transfer Learning
Knowledge distillation
Contrastive Learning
Federated Learning
Representative Learning
You have unsaved changes.