point mdp
- North America > United States > Massachusetts > Middlesex County > Cambridge (0.05)
- Asia > Middle East > Jordan (0.04)
- North America > United States > Utah > Salt Lake County > Salt Lake City (0.04)
- Europe > Germany > Bavaria > Upper Bavaria > Ingolstadt (0.04)
- Research Report (0.68)
- Instructional Material (0.46)
- Health & Medicine (0.68)
- Transportation > Infrastructure & Services (0.50)
- Transportation > Ground > Road (0.50)
- North America > United States > Utah > Salt Lake County > Salt Lake City (0.05)
- North America > United States > New York (0.04)
- North America > United States > Massachusetts > Middlesex County > Cambridge (0.05)
- Asia > Middle East > Jordan (0.04)
- North America > United States > Utah > Salt Lake County > Salt Lake City (0.04)
- Europe > Germany > Bavaria > Upper Bavaria > Ingolstadt (0.04)
- Health & Medicine (0.68)
- Transportation > Infrastructure & Services (0.50)
- Transportation > Ground > Road (0.50)
The Impact of Task Underspecification in Evaluating Deep Reinforcement Learning
Jayawardana, Vindula, Tang, Catherine, Li, Sirui, Suo, Dajiang, Wu, Cathy
Evaluations of Deep Reinforcement Learning (DRL) methods are an integral part of scientific progress of the field. Beyond designing DRL methods for general intelligence, designing task-specific methods is becoming increasingly prominent for real-world applications. In these settings, the standard evaluation practice involves using a few instances of Markov Decision Processes (MDPs) to represent the task. However, many tasks induce a large family of MDPs owing to variations in the underlying environment, particularly in real-world contexts. For example, in traffic signal control, variations may stem from intersection geometries and traffic flow levels. The select MDP instances may thus inadvertently cause overfitting, lacking the statistical power to draw conclusions about the method's true performance across the family. In this article, we augment DRL evaluations to consider parameterized families of MDPs. We show that in comparison to evaluating DRL methods on select MDP instances, evaluating the MDP family often yields a substantially different relative ranking of methods, casting doubt on what methods should be considered state-of-the-art. We validate this phenomenon in standard control benchmarks and the real-world application of traffic signal control. At the same time, we show that accurately evaluating on an MDP family is nontrivial. Overall, this work identifies new challenges for empirical rigor in reinforcement learning, especially as the outcomes of DRL trickle into downstream decision-making.
- North America > United States > Massachusetts > Middlesex County > Cambridge (0.04)
- North America > United States > Utah > Salt Lake County > Salt Lake City (0.04)
- Asia > Middle East > Jordan (0.04)
- (2 more...)
- Transportation > Infrastructure & Services (1.00)
- Transportation > Ground > Road (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Reinforcement Learning (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning (0.46)
- Information Technology > Artificial Intelligence > Machine Learning > Learning Graphical Models > Undirected Networks > Markov Models (0.34)