Shim, Sungrok
CoSQL: A Conversational Text-to-SQL Challenge Towards Cross-Domain Natural Language Interfaces to Databases
Yu, Tao, Zhang, Rui, Er, He Yang, Li, Suyi, Xue, Eric, Pang, Bo, Lin, Xi Victoria, Tan, Yi Chern, Shi, Tianze, Li, Zihan, Jiang, Youxuan, Yasunaga, Michihiro, Shim, Sungrok, Chen, Tao, Fabbri, Alexander, Li, Zifan, Chen, Luyao, Zhang, Yuwen, Dixit, Shreya, Zhang, Vincent, Xiong, Caiming, Socher, Richard, Lasecki, Walter S, Radev, Dragomir
It consists of 30k turns plus 10k annotated SQL queries, obtained from a Wizard-of-Oz (WOZ) collection of 3k dialogues querying 200 complex DBs spanning 138 domains. Each dialogue simulates a real-world DB query scenario with a crowd worker as a user exploring the DB and a SQL expert retrieving answers with SQL, clarifying ambiguous questions, or otherwise informing of unanswerable questions. When user questions are answerable by SQL, the expert describes the SQL and execution results to the user, hence maintaining a natural interaction flow. CoSQL introduces new challenges compared to existing task-oriented dialogue datasets: (1) the dialogue states are grounded in SQL, a domain-independent executable representation, instead of domain-specific slot-value pairs, and (2) because testing is done on unseen databases, success requires generalizing to new domains. CoSQL includes three tasks: SQL-grounded dialogue state tracking, response generation from query results, and user dialogue act prediction. We evaluate a set of strong baselines for each task and show that CoSQL presents significant challenges for future research. The dataset, baselines, and leaderboard will be released at https:// yale-lily.github.io/cosql .
SParC: Cross-Domain Semantic Parsing in Context
Yu, Tao, Zhang, Rui, Yasunaga, Michihiro, Tan, Yi Chern, Lin, Xi Victoria, Li, Suyi, Er, Heyang, Li, Irene, Pang, Bo, Chen, Tao, Ji, Emily, Dixit, Shreya, Proctor, David, Shim, Sungrok, Kraft, Jonathan, Zhang, Vincent, Xiong, Caiming, Socher, Richard, Radev, Dragomir
We present SParC, a dataset for cross-domainSemanticParsing inContext that consists of 4,298 coherent question sequences (12k+ individual questions annotated with SQL queries). It is obtained from controlled user interactions with 200 complex databases over 138 domains. We provide an in-depth analysis of SParC and show that it introduces new challenges compared to existing datasets. SParC demonstrates complex contextual dependencies, (2) has greater semantic diversity, and (3) requires generalization to unseen domains due to its cross-domain nature and the unseen databases at test time. We experiment with two state-of-the-art text-to-SQL models adapted to the context-dependent, cross-domain setup. The best model obtains an exact match accuracy of 20.2% over all questions and less than10% over all interaction sequences, indicating that the cross-domain setting and the con-textual phenomena of the dataset present significant challenges for future research. The dataset, baselines, and leaderboard are released at https://yale-lily.github.io/sparc.