Chen, Xinlei
Embodied Visual Recognition
Yang, Jianwei, Ren, Zhile, Xu, Mingze, Chen, Xinlei, Crandall, David, Parikh, Devi, Batra, Dhruv
Passive visual systems typically fail to recognize objects in the amodal setting where they are heavily occluded. In contrast, humans and other embodied agents have the ability to move in the environment, and actively control the viewing angle to better understand object shapes and semantics. In this work, we introduce the task of Embodied Visual Recognition (EVR): An agent is instantiated in a 3D environment close to an occluded target object, and is free to move in the environment to perform object classification, amodal object localization, and amodal object segmentation. To address this, we develop a new model called Embodied Mask R-CNN, for agents to learn to move strategically to improve their visual recognition abilities. We conduct experiments using the House3D environment. Experimental results show that: 1) agents with embodiment (movement) achieve better visual recognition performance than passive ones; 2) in order to improve visual recognition abilities, agents can learn strategical moving paths that are different from shortest paths.
CoDraw: Collaborative Drawing as a Testbed for Grounded Goal-driven Communication
Kim, Jin-Hwa, Kitaev, Nikita, Chen, Xinlei, Rohrbach, Marcus, Tian, Yuandong, Batra, Dhruv, Parikh, Devi
In this work, we propose a goal-driven collaborative task that contains language, vision, and action in a virtual environment as its core components. Specifically, we develop a Collaborative image-Drawing game between two agents, called CoDraw. Our game is grounded in a virtual world that contains movable clip art objects. The game involves two players: a Teller and a Drawer. The Teller sees an abstract scene containing multiple clip art pieces in a semantically meaningful configuration, while the Drawer tries to reconstruct the scene on an empty canvas using available clip art pieces. The two players communicate via two-way communication using natural language. We collect the CoDraw dataset of ~10K dialogs consisting of ~138K messages exchanged between human agents. We define protocols and metrics to evaluate the effectiveness of learned agents on this testbed, highlighting the need for a novel crosstalk condition which pairs agents trained independently on disjoint subsets of the training data for evaluation. We present models for our task, including simple but effective nearest-neighbor techniques and neural network approaches trained using a combination of imitation learning and goal-driven training. All models are benchmarked using both fully automated evaluation and by playing the game with live human agents.
nocaps: novel object captioning at scale
Agrawal, Harsh, Desai, Karan, Chen, Xinlei, Jain, Rishabh, Batra, Dhruv, Parikh, Devi, Lee, Stefan, Anderson, Peter
Image captioning models have achieved impressive results on datasets containing limited visual concepts and large amounts of paired image-caption training data. However, if these models are to ever function in the wild, a much larger variety of visual concepts must be learned, ideally from less supervision. To encourage the development of image captioning models that can learn visual concepts from alternative data sources, such as object detection datasets, we present the first large-scale benchmark for this task. Dubbed 'nocaps', for novel object captioning at scale, our benchmark consists of 166,100 human-generated captions describing 15,100 images from the Open Images validation and test sets. The associated training data consists of COCO image-caption pairs, plus Open Images image-level labels and object bounding boxes. Since Open Images contains many more classes than COCO, more than 500 object classes seen in test images have no training captions (hence, nocaps). We evaluate several existing approaches to novel object captioning on our challenging benchmark. In automatic evaluations these approaches show modest improvements over a strong baseline trained only on image-caption data. However, even when using ground-truth object detections, the results are significantly weaker than our human baseline - indicating substantial room for improvement.
Never-Ending Learning
Mitchell, Tom M. (Carnegie Mellon University) | Cohen, William (Carnegie Mellon University) | Hruschka, Estevam (University of Sao Carlos) | Talukdar, Partha (Indian Institute of Science) | Betteridge, Justin (Carnegie Mellon University) | Carlson, Andrew (Google) | Mishra, Bhavana Dalvi (Carnegien Mellon University) | Gardner, Matthew (Carnegie Mellon University) | Kisiel, Bryan (Carnegie Mellon University) | Krishnamurthy, Jayant (Carnegie Mellon University) | Lao, Ni (Google) | Mazaitis, Kathryn (Carnegie Mellon University) | Mohamed, Thahir (Carnegie Mellon University) | Nakashole, Ndapa (Carnegie Mellon University) | Platanios, Emmanouil Antonios (Ohioe State University) | Ritter, Alan (Carnegie Mellon University) | Samadi, Mehdi (Duolingo) | Settles, Burr (Carnegie Mellon University) | Wang, Richard (Carnegie Mellon University) | Wijaya, Derry (Carnegie Mellon University) | Gupta, Abhinav (Carnegie Mellon University) | Chen, Xinlei (Alpine Data Lab) | Saparov, Abulhair (Pittsburgh Supercomputer Center) | Greaves, Malcolm | Welling, Joel
Whereas people learn many different types of knowledge from diverse experiences over many years, most current machine learning systems acquire just a single function or data model from just a single data set. We propose a never-ending learning paradigm for machine learning, to better reflect the more ambitious and encompassing type of learning performed by humans. As a case study, we describe the Never-Ending Language Learner (NELL), which achieves some of the desired properties of a never-ending learner, and we discuss lessons learned. NELL has been learning to read the web 24 hours/day since January 2010, and so far has acquired a knowledge base with over 80 million confidence-weighted beliefs (e.g., servedWith(tea, biscuits) ). NELL has also learned millions of features and parameters that enable it to read these beliefs from the web. Additionally, it has learned to reason over these beliefs to infer new beliefs, and is able to extend its ontology by synthesizing new relational predicates. NELL can be tracked online at http://rtw.ml.cmu.edu, and followed on Twitter at @CMUNELL.
Large Scale Spectral Clustering with Landmark-Based Representation
Chen, Xinlei (Zhejiang University) | Cai, Deng (Zhejiang University)
Spectral clustering is one of the most popular clustering approaches. Despite its good performance, it is limited in its applicability to large-scale problems due to its high computational complexity. Recently, many approaches have been proposed to accelerate the spectral clustering. Unfortunately, these methods usually sacrifice quite a lot information of the original data, thus result in a degradation of performance. In this paper, we propose a novel approach, called Landmark-based Spectral Clustering (LSC), for large scale clustering problems. Specifically, we select $p\ (\ll n)$ representative data points as the landmarks and represent the original data points as the linear combinations of these landmarks. The spectral embedding of the data can then be efficiently computed with the landmark-based representation. The proposed algorithm scales linearly with the problem size. Extensive experiments show the effectiveness and efficiency of our approach comparing to the state-of-the-art methods.