Reviews: Deep Generative Models with Learnable Knowledge Constraints
–Neural Information Processing Systems
Summary: The paper proposes a way to incorporate constraints into the learning of generative models through posterior regularization. In doing so, the paper draws connections between posterior regularization and policy optimization. One of the key contributions of this paper is that the constraints are modeled as extrinsic rewards and learned through inverse reinforcement learning. The paper studies an interesting and very practical problem and the contributions are substantial. The writing could definitely be made clearer for Sections 3 and 4, where the overloaded notation is often hard to follow. I have the following questions: 1.
Neural Information Processing Systems
Oct-8-2024, 05:56:06 GMT
- Technology: