Palmer, Martha
Speech Is Not Enough: Interpreting Nonverbal Indicators of Common Knowledge and Engagement
Palmer, Derek, Zhu, Yifan, Lai, Kenneth, VanderHoeven, Hannah, Bradford, Mariah, Khebour, Ibrahim, Mabrey, Carlos, Fitzgerald, Jack, Krishnaswamy, Nikhil, Palmer, Martha, Pustejovsky, James
Our goal is to develop an AI Partner that can provide support for group problem solving and social dynamics. In multi-party working group environments, multimodal analytics is crucial for identifying non-verbal interactions of group members. In conjunction with their verbal participation, this creates an holistic understanding of collaboration and engagement that provides necessary context for the AI Partner. In this demo, we illustrate our present capabilities at detecting and tracking nonverbal behavior in student task-oriented interactions in the classroom, and the implications for tracking common ground and engagement.
SCI 3.0: A Web-based Schema Curation Interface for Graphical Event Representations
Suchocki, Reece, Martin, Mary, Palmer, Martha, Brown, Susan
To understand the complexity of global events, one must navigate a web of interwoven sub-events, identifying those most impactful elements within the larger, abstract macro-event framework at play. This concept can be extended to the field of natural language processing (NLP) through the creation of structured event schemas which can serve as representations of these abstract events. Central to our approach is the Schema Curation Interface 3.0 (SCI 3.0), a web application that facilitates real-time editing of event schema properties within a generated graph e.g., adding, removing, or editing sub-events, entities, and relations directly through an interface.
Adapting Abstract Meaning Representation Parsing to the Clinical Narrative -- the SPRING THYME parser
Cai, Jon Z., Wright-Bettner, Kristin, Palmer, Martha, Savova, Guergana K., Martin, James H.
This paper is dedicated to the design and evaluation of the first AMR parser tailored for clinical notes. Our objective was to facilitate the precise transformation of the clinical notes into structured AMR expressions, thereby enhancing the interpretability and usability of clinical text data at scale. Leveraging the colon cancer dataset from the Temporal Histories of Your Medical Events (THYME) corpus, we adapted a state-of-the-art AMR parser utilizing continuous training. Our approach incorporates data augmentation techniques to enhance the accuracy of AMR structure predictions. Notably, through this learning strategy, our parser achieved an impressive F1 score of 88% on the THYME corpus's colon cancer dataset. Moreover, our research delved into the efficacy of data required for domain adaptation within the realm of clinical notes, presenting domain adaptation data requirements for AMR parsing. This exploration not only underscores the parser's robust performance but also highlights its potential in facilitating a deeper understanding of clinical narratives through structured semantic representations.
Linear Cross-document Event Coreference Resolution with X-AMR
Ahmed, Shafiuddin Rehan, Baker, George Arthur, Judge, Evi, Regan, Michael, Wright-Bettner, Kristin, Palmer, Martha, Martin, James H.
Event Coreference Resolution (ECR) as a pairwise mention classification task is expensive both for automated systems and manual annotations. The task's quadratic difficulty is exacerbated when using Large Language Models (LLMs), making prompt engineering for ECR prohibitively costly. In this work, we propose a graphical representation of events, X-AMR, anchored around individual mentions using a \textbf{cross}-document version of \textbf{A}bstract \textbf{M}eaning \textbf{R}epresentation. We then linearize the ECR with a novel multi-hop coreference algorithm over the event graphs. The event graphs simplify ECR, making it a) LLM cost-effective, b) compositional and interpretable, and c) easily annotated. For a fair assessment, we first enrich an existing ECR benchmark dataset with these event graphs using an annotator-friendly tool we introduce. Then, we employ GPT-4, the newest LLM by OpenAI, for these annotations. Finally, using the ECR algorithm, we assess GPT-4 against humans and analyze its limitations. Through this research, we aim to advance the state-of-the-art for efficient ECR and shed light on the potential shortcomings of current LLMs at this task. Code and annotations: \url{https://github.com/ahmeshaf/gpt_coref}
X-AMR Annotation Tool
Ahmed, Shafiuddin Rehan, Cai, Jon Z., Palmer, Martha, Martin, James H.
To illustrate the challenge of coreference across Semantic representations of events play a pivotal documents, consider the following example: Two role in natural language processing (NLP) tasks, facilitating news articles discuss a corporate acquisition. In the understanding and extraction of meaningful one article, the event is described as "Company A's information from text. Among the various purchase of Company B on July 1st, 2008" while approaches to represent events, Semantic Role Labeling in another article, it is referred to as "In 7/08 Company (SRL; Palmer et al. (2005)) and Abstract B was acquired by Company A." Establishing Meaning Representation (AMR; Banarescu et al. the coreference relationship between these two descriptions (2013)) have gained significant attention. In this is non-trivial, yet crucial for creating a paper, we delve into the realm of semantic event comprehensive representation of the acquisition representations, with a particular focus on a method event.
RESIN-EDITOR: A Schema-guided Hierarchical Event Graph Visualizer and Editor
Nguyen, Khanh Duy, Zhang, Zixuan, Suchocki, Reece, Li, Sha, Palmer, Martha, Brown, Susan, Han, Jiawei, Ji, Heng
In this paper, we present RESIN-EDITOR, an interactive event graph visualizer and editor designed for analyzing complex events. Our RESIN-EDITOR system allows users to render and freely edit hierarchical event graphs extracted from multimedia and multi-document news clusters with guidance from human-curated event schemas. RESIN-EDITOR's unique features include hierarchical graph visualization, comprehensive source tracing, and interactive user editing, which is more powerful and versatile than existing Information Extraction (IE) visualization tools. In our evaluation of RESIN-EDITOR, we demonstrate ways in which our tool is effective in understanding complex events and enhancing system performance. The source code, a video demonstration, and a live website for RESIN-EDITOR have been made publicly available.
CAMRA: Copilot for AMR Annotation
Cai, Jon Z., Ahmed, Shafiuddin Rehan, Bonn, Julia, Wright-Bettner, Kristin, Palmer, Martha, Martin, James H.
In this paper, we introduce CAMRA (Copilot for AMR Annotatations), a cutting-edge web-based tool designed for constructing Abstract Meaning Representation (AMR) from natural language text. CAMRA offers a novel approach to deep lexical semantics annotation such as AMR, treating AMR annotation akin to coding in programming languages. Leveraging the familiarity of programming paradigms, CAMRA encompasses all essential features of existing AMR editors, including example lookup, while going a step further by integrating Propbank roleset lookup as an autocomplete feature within the tool. Notably, CAMRA incorporates AMR parser models as coding co-pilots, greatly enhancing the efficiency and accuracy of AMR annotators. To demonstrate the tool's capabilities, we provide a live demo accessible at: https://camra.colorado.edu
GLEN: General-Purpose Event Detection for Thousands of Types
Zhan, Qiusi, Li, Sha, Conger, Kathryn, Palmer, Martha, Ji, Heng, Han, Jiawei
The progress of event extraction research has been hindered by the absence of wide-coverage, large-scale datasets. To make event extraction systems more accessible, we build a general-purpose event detection dataset GLEN, which covers 205K event mentions with 3,465 different types, making it more than 20x larger in ontology than today's largest event dataset. GLEN is created by utilizing the DWD Overlay, which provides a mapping between Wikidata Qnodes and PropBank rolesets. This enables us to use the abundant existing annotation for PropBank as distant supervision. In addition, we also propose a new multi-stage event detection model CEDAR specifically designed to handle the large ontology size in GLEN. We show that our model exhibits superior performance compared to a range of baselines including InstructGPT. Finally, we perform error analysis and show that label noise is still the largest challenge for improving performance for this new dataset. Our dataset, code, and models are released at \url{https://github.com/ZQS1943/GLEN}.}
Learning Semantic Role Labeling from Compatible Label Sequences
Li, Tao, Kazeminejad, Ghazaleh, Brown, Susan W., Palmer, Martha, Srikumar, Vivek
Semantic role labeling (SRL) has multiple disjoint label sets, e.g., VerbNet and PropBank. Creating these datasets is challenging, therefore a natural question is how to use each one to help the other. Prior work has shown that cross-task interaction helps, but only explored multitask learning so far. A common issue with multi-task setup is that argument sequences are still separately decoded, running the risk of generating structurally inconsistent label sequences (as per lexicons like Semlink). In this paper, we eliminate such issue with a framework that jointly models VerbNet and PropBank labels as one sequence. In this setup, we show that enforcing Semlink constraints during decoding constantly improves the overall F1. With special input constructions, our joint model infers VerbNet arguments from given PropBank arguments with over 99 F1. For learning, we propose a constrained marginal model that learns with knowledge defined in Semlink to further benefit from the large amounts of PropBank-only data. On the joint benchmark based on CoNLL05, our models achieve state-of-the-art F1's, outperforming the prior best in-domain model by 3.5 (VerbNet) and 0.8 (PropBank). For out-of-domain generalization, our models surpass the prior best by 3.4 (VerbNet) and 0.2 (PropBank).
Human-in-the-Loop Schema Induction
Zhang, Tianyi, Tham, Isaac, Hou, Zhaoyi, Ren, Jiaxuan, Zhou, Liyang, Xu, Hainiu, Zhang, Li, Martin, Lara J., Dror, Rotem, Li, Sha, Ji, Heng, Palmer, Martha, Brown, Susan, Suchocki, Reece, Callison-Burch, Chris
Schema induction builds a graph representation explaining how events unfold in a scenario. Existing approaches have been based on information retrieval (IR) and information extraction(IE), often with limited human curation. We demonstrate a human-in-the-loop schema induction system powered by GPT-3. We first describe the different modules of our system, including prompting to generate schematic elements, manual edit of those elements, and conversion of those into a schema graph. By qualitatively comparing our system to previous ones, we show that our system not only transfers to new domains more easily than previous approaches, but also reduces efforts of human curation thanks to our interactive interface.