Approximate Policy Iteration with a Policy Language Bias
–Neural Information Processing Systems
We explore approximate policy iteration, replacing the usual cost- function learning step with a learning step in policy space. We give policy-language biases that enable solution of very large relational Markov decision processes (MDPs) that no previous technique can solve. In particular, we induce high-quality domain-specific planners for clas- sical planning domains (both deterministic and stochastic variants) by solving such domains as extremely large MDPs.
Neural Information Processing Systems
Apr-6-2023, 16:04:04 GMT
- Technology: