M-Walk: Learning to Walk over Graphs using Monte Carlo Tree Search
Yelong Shen, Jianshu Chen, Po-Sen Huang, Yuqing Guo, Jianfeng Gao
–Neural Information Processing Systems
Learning to walk over a graph towards a target node for a given query and a source node is an important problem in applications such as knowledge base completion (KBC). It can be formulated as a reinforcement learning (RL) problem with a known state transition model. To overcome the challenge of sparse rewards, we develop a graph-walking agent called M-Walk, which consists of a deep recurrent neural network (RNN) and Monte Carlo Tree Search (MCTS). The RNN encodes the state (i.e., history of the walked path) and maps it separately to a policy and Q-values. In order to effectively train the agent from sparse rewards, we combine MCTS with the neural policy to generate trajectories yielding more positive rewards.
Neural Information Processing Systems
Mar-26-2025, 23:47:29 GMT
- Country:
- North America > United States (1.00)
- Industry:
- Leisure & Entertainment > Sports (0.46)
- Technology: