Learning to Act and Observe in Partially Observable Domains
Bolander, Thomas, Gierasimczuk, Nina, Liberman, Andrés Occhipinti
–arXiv.org Artificial Intelligence
We consider a learning agent in a partially observable environment, with which the agent has never interacted before, and about which it learns both what it can observe and how its actions affect the environment. The agent can learn about this domain from experience gathered by taking actions in the domain and observing their results. We present learning algorithms capable of learning as much as possible (in a well-defined sense) both about what is directly observable and about what actions do in the domain, given the learner's observational constraints. We differentiate the level of domain knowledge attained by each algorithm, and characterize the type of observations required to reach it. The algorithms use dynamic epistemic logic (DEL) to represent the learned domain information symbolically. Our work continues that of Bolander and Gierasimczuk (2015), which developed DEL-based learning algorithms based to learn domain information in fully observable domains.
arXiv.org Artificial Intelligence
Sep-13-2021
- Country:
- Europe
- Latvia > Riga Municipality
- Riga (0.04)
- Netherlands > North Holland
- Amsterdam (0.04)
- Spain > Catalonia
- Tarragona Province > Tarragona (0.04)
- United Kingdom > England
- Cambridgeshire > Cambridge (0.14)
- Latvia > Riga Municipality
- North America > United States
- California > San Mateo County
- Menlo Park (0.04)
- Illinois > Cook County
- Chicago (0.04)
- Massachusetts > Middlesex County
- Cambridge (0.04)
- New York > Nassau County
- Mineola (0.04)
- California > San Mateo County
- South America > Brazil
- São Paulo (0.04)
- Europe
- Genre:
- Research Report (0.81)
- Technology: