The Power of the Noisy Channel: Unsupervised End-to-End Task-Oriented Dialogue with LLMs
King, Brendan, Flanigan, Jeffrey
–arXiv.org Artificial Intelligence
Training task-oriented dialogue systems typically requires turn-level annotations for interacting with their APIs: e.g. a dialogue state and the system actions taken at each step. These annotations can be costly to produce, error-prone, and require both domain and annotation expertise. With advances in LLMs, we hypothesize unlabelled data and a schema definition are sufficient for building a working task-oriented dialogue system, completely unsupervised. Using only (1) a well-defined API schema (2) a set of unlabelled dialogues between a user and agent, we develop a novel approach for inferring turn-level annotations as latent variables using a noisy channel model. We iteratively improve these pseudo-labels with expectation-maximization (EM), and use the inferred labels to train an end-to-end dialogue agent. Evaluating our approach on the MultiWOZ benchmark, our method more than doubles the dialogue success rate of a strong GPT-3.5 baseline.
arXiv.org Artificial Intelligence
Apr-23-2024
- Country:
- Europe
- North America
- Canada (0.14)
- United States > California (0.14)
- Genre:
- Overview > Innovation (0.34)
- Research Report > Promising Solution (0.34)
- Technology: