Reinforcement Learning in Time-Varying Systems: an Empirical Study
Hamadanian, Pouya, Schwarzkopf, Malte, Sen, Siddartha, Alizadeh, Mohammad
–arXiv.org Artificial Intelligence
Recent research has turned to Reinforcement Learning (RL) to solve challenging decision problems, as an alternative to hand-tuned heuristics. RL can learn good policies without the need for modeling the environment's dynamics. Despite this promise, RL remains an impractical solution for many real-world systems problems. A particularly challenging case occurs when the environment changes over time, i.e. it exhibits non-stationarity. In this work, we characterize the challenges introduced by non-stationarity and develop a framework for addressing them to train RL agents in live systems. Such agents must explore and learn new environments, without hurting the system's performance, and remember them over time. To this end, our framework (1) identifies different environments encountered by the live system, (2) explores and trains a separate expert policy for each environment, and (3) employs safeguards to protect the system's performance. We apply our framework to two systems problems: straggler mitigation and adaptive video streaming, and evaluate it against a variety of alternative approaches using real-world and synthetic data. We show that each component of our framework is necessary to cope with non-stationarity.
arXiv.org Artificial Intelligence
Jan-14-2022
- Country:
- Asia > Middle East
- Jordan (0.04)
- North America > United States
- Massachusetts > Middlesex County
- Cambridge (0.28)
- New York
- Bronx County > New York City (0.04)
- Kings County > New York City (0.04)
- New York County > New York City (0.14)
- Queens County > New York City (0.04)
- Richmond County > New York City (0.04)
- Rhode Island > Providence County
- Providence (0.04)
- Massachusetts > Middlesex County
- Asia > Middle East
- Genre:
- Research Report > New Finding (0.67)
- Industry:
- Education (0.94)