Goto

Collaborating Authors

 clevrer-human


Supplementary Material for CLEVRER-Humans: Describing Physical and Causal Events the Human Way Jiayuan Mao MIT Xuelin Y ang

Neural Information Processing Systems

We bear all responsibility in case of violation of rights. The rest of this supplementary document is organized as the following. Next, in Section C, we describe the user interface for dataset collection. On average, we can obtain 29.4 descriptions per video, highlighting the advantage of our First, CLEVRER-Humans contains dense annotations of causal relations between physical events. The outer circle represents the general event families. We have lemmatized all verbs to remove the tense.



CLEVRER-Humans: Describing Physical and Causal Events the Human Way

Neural Information Processing Systems

Building machines that can reason about physical events and their causal relationships is crucial for flexible interaction with the physical world. However, most existing physical and causal reasoning benchmarks are exclusively based on synthetically generated events and synthetic natural language descriptions of the causal relationships. This design brings up two issues. First, there is a lack of diversity in both event types and natural language descriptions; second, causal relationships based on manually-defined heuristics are different from human judgments. To address both shortcomings, we present the CLEVRER-Humans benchmark, a video reasoning dataset for causal judgment of physical events with human labels. We employ two techniques to improve data collection efficiency: first, a novel iterative event cloze task to elicit a new representation of events in videos, which we term Causal Event Graphs (CEGs); second, a data augmentation technique based on neural language generative models. We convert the collected CEGs into questions and answers to be consistent with prior work. Finally, we study a collection of baseline approaches for CLEVRER-Humans question-answering, highlighting great challenges set forth by our benchmark.


Supplementary Material for CLEVRER-Humans: Describing Physical and Causal Events the Human Way Jiayuan Mao MIT Xuelin Y ang

Neural Information Processing Systems

We bear all responsibility in case of violation of rights. The rest of this supplementary document is organized as the following. Next, in Section C, we describe the user interface for dataset collection. On average, we can obtain 29.4 descriptions per video, highlighting the advantage of our First, CLEVRER-Humans contains dense annotations of causal relations between physical events. The outer circle represents the general event families. We have lemmatized all verbs to remove the tense.



CLEVRER-Humans: Describing Physical and Causal Events the Human Way

Neural Information Processing Systems

Building machines that can reason about physical events and their causal relationships is crucial for flexible interaction with the physical world. However, most existing physical and causal reasoning benchmarks are exclusively based on synthetically generated events and synthetic natural language descriptions of the causal relationships. This design brings up two issues. First, there is a lack of diversity in both event types and natural language descriptions; second, causal relationships based on manually-defined heuristics are different from human judgments. To address both shortcomings, we present the CLEVRER-Humans benchmark, a video reasoning dataset for causal judgment of physical events with human labels.


CLEVRER-Humans: Describing Physical and Causal Events the Human Way

Mao, Jiayuan, Yang, Xuelin, Zhang, Xikun, Goodman, Noah D., Wu, Jiajun

arXiv.org Machine Learning

Building machines that can reason about physical events and their causal relationships is crucial for flexible interaction with the physical world. However, most existing physical and causal reasoning benchmarks are exclusively based on synthetically generated events and synthetic natural language descriptions of causal relationships. This design brings up two issues. First, there is a lack of diversity in both event types and natural language descriptions; second, causal relationships based on manually-defined heuristics are different from human judgments. To address both shortcomings, we present the CLEVRER-Humans benchmark, a video reasoning dataset for causal judgment of physical events with human labels. We employ two techniques to improve data collection efficiency: first, a novel iterative event cloze task to elicit a new representation of events in videos, which we term Causal Event Graphs (CEGs); second, a data augmentation technique based on neural language generative models. We convert the collected CEGs into questions and answers to be consistent with prior work. Finally, we study a collection of baseline approaches for CLEVRER-Humans question-answering, highlighting the great challenges set forth by our benchmark.