Offline Model-based Adaptable Policy Learning Xiong-Hui Chen 1, Y ang Y u
–Neural Information Processing Systems
In reinforcement learning, a promising direction to avoid online trial-and-error costs is learning from an offline dataset. Current offline reinforcement learning methods commonly learn in the policy space constrained to in-support regions by the offline dataset, in order to ensure the robustness of the outcome policies.
Neural Information Processing Systems
Aug-14-2025, 06:58:14 GMT
- Country:
- Africa > Ethiopia
- Addis Ababa > Addis Ababa (0.04)
- Asia
- China > Jiangsu Province
- Nanjing (0.04)
- Myanmar > Tanintharyi Region
- Dawei (0.04)
- China > Jiangsu Province
- Europe
- Portugal (0.04)
- Sweden (0.04)
- United Kingdom > England
- Cambridgeshire > Cambridge (0.04)
- Greater London > London (0.04)
- North America
- Canada > British Columbia
- Vancouver (0.04)
- United States
- California > Los Angeles County
- Long Beach (0.04)
- Colorado > Denver County
- Denver (0.04)
- Massachusetts > Middlesex County
- Cambridge (0.04)
- California > Los Angeles County
- Canada > British Columbia
- Oceania > Australia
- Queensland > Brisbane (0.04)
- South America > Argentina
- Pampas > Buenos Aires F.D. > Buenos Aires (0.04)
- Africa > Ethiopia
- Genre:
- Research Report (0.68)
- Industry:
- Information Technology (0.46)
- Marketing (0.46)
- Technology: