Supplementary of PRIOR: Personalized Prior for Reactivating the Information Overlooked in Federated Learning A Glossary, Some Basic Knowledge and Details about Implementations
–Neural Information Processing Systems
A.1 Glossary The main notations in this paper are shown in Table 5. E[ D (X, y)] (15) Eq. (16) is the definition of Bregman divergence: D Hence, the selected function g should be convex. Besides, the non-maximum entropy rule approach is also worth considering, but we focus on maximum entropy prior in this section. In this section, inspired by MAML, we briefly introduce a meta-step-based implementation method. The mean of the SX-family prior in Eq. (8) is used in regular term, which can be In this work, we use the approximation Bayesian methods. The local training process based on regular terms differs from Bayesian learning based on sampling, i.e., each time a model needs to be obtained by sampling the model There are three parts in Eq. (13) we need to deal with, and the first-order methods are as shown below: In recent years, PFL has found use not only in predictive tasks like mobile device input methods but also in areas where privacy is paramount, such as healthcare and finance.
Neural Information Processing Systems
Oct-8-2025, 18:14:01 GMT