A Little Is Enough: Circumventing Defenses For Distributed Learning
Baruch, Gilad, Baruch, Moran, Goldberg, Yoav
–Neural Information Processing Systems
Distributed learning is central for large-scale training of deep-learning models. However, it is exposed to a security threat in which Byzantine participants can interrupt or control the learning process. Previous attack models assume that the rogue participants (a) are omniscient (know the data of all other participants), and (b) introduce large changes to the parameters. Accordingly, most defense mechanisms make a similar assumption and attempt to use statistically robust methods to identify and discard values whose reported gradients are far from the population mean. We observe that if the empirical variance between the gradients of workers is high enough, an attacker could take advantage of this and launch a non-omniscient attack that operates within the population variance.
Neural Information Processing Systems
Mar-19-2020, 00:03:42 GMT