Communication Efficient Distributed Learning for Kernelized Contextual Bandits
Li, Chuanhao, Wang, Huazheng, Wang, Mengdi, Wang, Hongning
–arXiv.org Artificial Intelligence
We tackle the communication efficiency challenge of learning kernelized contextual bandits in a distributed setting. Despite the recent advances in communication-efficient distributed bandit learning, existing solutions are restricted to simple models like multi-armed bandits and linear bandits, which hamper their practical utility. In this paper, instead of assuming the existence of a linear reward mapping from the features to the expected rewards, we consider non-linear reward mappings, by letting agents collaboratively search in a reproducing kernel Hilbert space (RKHS). This introduces significant challenges in communication efficiency as distributed kernel learning requires the transfer of raw data, leading to a communication cost that grows linearly w.r.t. time horizon $T$. We addresses this issue by equipping all agents to communicate via a common Nystr\"{o}m embedding that gets updated adaptively as more data points are collected. We rigorously proved that our algorithm can attain sub-linear rate in both regret and communication cost.
arXiv.org Artificial Intelligence
Oct-12-2022
- Country:
- North America > United States
- New Jersey > Mercer County
- Princeton (0.04)
- Oregon (0.04)
- Virginia > Albemarle County
- Charlottesville (0.14)
- New Jersey > Mercer County
- North America > United States
- Genre:
- Research Report (0.81)
- Technology:
- Information Technology
- Artificial Intelligence
- Machine Learning (1.00)
- Representation & Reasoning > Agents (1.00)
- Communications (1.00)
- Data Science > Data Mining (1.00)
- Artificial Intelligence
- Information Technology