Contextual knowledge is essential in answering questions given speciﬁc observations. While recent approaches to building commonsense knowledge basesvia text mining and/or crowdsourcing are successful,contextual knowledge is largely missing. To addressthis gap, this paper presents SocialExplain, a novel approach to acquiring contextual commonsense knowledge from explanations of social content. The acquisition process is broken into two cognitively simple tasks:to identify contextual clues from the given social content, and to explain the content with the clues. An experiment was conducted to show that multiple piecesof contextual commonsense knowledge can be identi-ﬁed from a small number of tweets. Online users veriﬁed that 92.45% of the acquired sentences are good,and 95.92% are new sentences compared with existingcrowd-sourced commonsense knowledge bases.
The central challenge in commonsense knowledge representation research is to develop content theories that achieve a high degree of both competency and coverage. We describe a new methodology for constructing formal theories in commonsense knowledge domains that complements traditional knowledge representation approaches by first addressing issues of coverage. These concepts are sorted into a manageable number of coherent domains, one of which is the representational area of commonsense human memory. These representational areas are then analyzed using more traditional knowledge representation techniques, as demonstrated in this article by our treatment of commonsense human memory.
Reasoning with commonsense knowledge plays an important role in various NLU tasks. Often the commonsense knowledge is needed to be extracted separately. In this paper we present our work of automatically extracting a certain type of commonsense knowledge. The knowledge resembles the kind that humans have about the events and the entities that participate in those events. One example of such knowledge is that "IF A bullying B causes T rescued Z THEN (possibly) Z = B ''. We call this knowledge an event-based conditional commonsense. Our approach involves semantic parsing of natural language sentences by using the Knowledge Parser (K-Parser) and extracting the knowledge, if found. We extracted about 19000 instances of such knowledge from the Open American National Corpus.
In the era of Big Data, knowledge integration is key for tasks such as social media aggregation, opinion mining, and cyber-issue detection. The integration of different kinds of knowledge coming from multiple sources, however, is often a problematic issue as it either requires a lot of manual effort in defining aggregation rules or suffers from noise generated by automatic integration techniques. In this work, we propose a method based on conceptual primitives for efficiently integrating pieces of knowledge coming from different common and common-sense resources, which we test in the field of concept-level sentiment analysis.
This paper presents a semantically grounded method for extracting commonsense knowledge. First, commonsense rules are identified, e.g., one cannot see imaginary objects. Second, those rules are combined with a basic semantic representation in order to infer commonsense knowledge facts, e.g. one cannot see a flying carpet. Further combinations of semantic relations with inferred commonsense facts are proposed and analyzed. Results show that this novel method is able to extract thousands of commonsense facts with little human interaction and high accuracy.