Goto

Collaborating Authors

 gkr



A Proofs

Neural Information Processing Systems

Proof of Theorem 2. The Kantorovich Rubinstein (KR) distance is defined as follows. The optimal partial transport distance is defined as follows. We prove the contraposition of Theorem 2 for the KR distance. We solve the BHCP problem using this algorithm.


dba31bb5c75992690f20c2d3b370ec7c-AuthorFeedback.pdf

Neural Information Processing Systems

We thank all reviewers for their positive feedback. Figalli's distance, our algorithm computes it efficiently by setting λ We compare the distributions of crime locations. Our method is the first UOT method that can handle million-scale datasets. The cost is the ground distance between the centers of regions of the quadtree. GKR is reduced to the standard OT, thus metric.


Fusion of Global and Local Knowledge for Personalized Federated Learning

Huang, Tiansheng, Shen, Li, Sun, Yan, Lin, Weiwei, Tao, Dacheng

arXiv.org Artificial Intelligence

Personalized federated learning, as a variant of federated learning, trains customized models for clients using their heterogeneously distributed data. However, it is still inconclusive about how to design personalized models with better representation of shared global knowledge and personalized pattern. To bridge the gap, we in this paper explore personalized models with low-rank and sparse decomposition. Specifically, we employ proper regularization to extract a low-rank global knowledge representation (GKR), so as to distill global knowledge into a compact representation. Subsequently, we employ a sparse component over the obtained GKR to fuse the personalized pattern into the global knowledge. As a solution, we propose a two-stage proximal-based algorithm named \textbf{Fed}erated learning with mixed \textbf{S}parse and \textbf{L}ow-\textbf{R}ank representation (FedSLR) to efficiently search for the mixed models. Theoretically, under proper assumptions, we show that the GKR trained by FedSLR can at least sub-linearly converge to a stationary point of the regularized problem, and that the sparse component being fused can converge to its stationary point under proper settings. Extensive experiments also demonstrate the superior empirical performance of FedSLR. Moreover, FedSLR reduces the number of parameters, and lowers the down-link communication complexity, which are all desirable for federated learning algorithms. Source code is available in \url{https://github.com/huangtiansheng/fedslr}.