FedCD: Improving Performance in non-IID Federated Learning
Kopparapu, Kavya, Lin, Eric, Zhao, Jessica
Federated learning has been widely applied to enable decentralized devices, which each have their own local data, to learn a shared model. However, learning from real-world data can be challenging, as it is rarely identically and independently distributed (IID) across edge devices (a key assumption for current high-performing and low-bandwidth algorithms). We present a novel approach, FedCD, which clones and deletes models to dynamically group devices with similar data. Experiments on the CIFAR-10 dataset show that FedCD achieves higher accuracy and faster convergence compared to a FedAvg baseline on non-IID data while incurring minimal computation, communication, and storage overheads.
Jul-27-2020
- Country:
- North America > United States (0.30)
- Genre:
- Research Report (1.00)
- Industry:
- Health & Medicine (0.46)
- Technology: