PID-Inspired Inductive Biases for Deep Reinforcement Learning in Partially Observable Control Tasks
–Neural Information Processing Systems
Deep reinforcement learning (RL) has shown immense potential for learning to control systems through data alone. However, one challenge deep RL faces is that the full state of the system is often not observable. When this is the case, the policy needs to leverage the history of observations to infer the current state. At the same time, differences between the training and testing environments makes it critical for the policy not to overfit to the sequence of observations it sees at training time. As such, there is an important balancing act between having the history encoder be flexible enough to extract relevant information, yet be robust to changes in the environment.
Neural Information Processing Systems
Dec-26-2025, 15:40:53 GMT
- Technology: