Not enough data to create a plot.
Try a different view from the menu above.
Simperl, Elena
Neural Wikipedian: Generating Textual Summaries from Knowledge Base Triples
Vougiouklis, Pavlos, Elsahar, Hady, Kaffee, Lucie-Aimée, Gravier, Christoph, Laforest, Frederique, Hare, Jonathon, Simperl, Elena
Most people do not interact with Semantic Web data directly. Unless they have the expertise to understand the underlying technology, they need textual or visual interfaces to help them make sense of it. We explore the problem of generating natural language summaries for Semantic Web data. This is non-trivial, especially in an open-domain context. To address this problem, we explore the use of neural networks. Our system encodes the information from a set of triples into a vector of fixed dimensionality and generates a textual summary by conditioning the output on the encoded vector. We train and evaluate our models on two corpora of loosely aligned Wikipedia snippets and DBpedia and Wikidata triples with promising results.
Towards Knowledge-Driven Annotation
Mrabet, Yassine (CRP Henri Tudor) | Gardent, Claire ( CNRS/LORIA ) | Foulonneau, Muriel (CRP Henri Tudor) | Simperl, Elena (University of Southampton) | Ras, Eric (CRP Henri Tudor)
While the Web of data is attracting increasing interest and rapidly growing in size, the major support of information on the surface Web are still multimedia documents. Semantic annotation of texts is one of the main processes that are intended to facilitate meaning-based information exchange between computational agents. However, such annotation faces several challenges such as the heterogeneity of natural language expressions, the heterogeneity of documents structure and context dependencies. While a broad range of annotation approaches rely mainly or partly on the target textual context to disambiguate the extracted entities, in this paper we present an approach that relies mainly on formalized-knowledge expressed in RDF datasets to categorize and disambiguate noun phrases. In the proposed method, we represent the reference knowledge bases as co-occurrence matrices and the disambiguation problem as a 0-1 Integer Linear Programming (ILP) problem. The proposed approach is unsupervised and can be ported to any RDF knowledge base. The system implementing this approach, called KODA, shows very promising results w.r.t. state-of-the-art annotation tools in cross-domain experimentations.
Crowdsourcing Tasks in Open Query Answering
Simperl, Elena (Karslruhe Institute of Technology) | Norton, Barry (Ontotext AD) | Vrandecic, Denny (Karlsruhe Institute of Technology)
Open query answering is the idea of answering queries that are not given using the vocabulary of the queried knowledge base but instead the vocabulary of the inquirer. Many aspects of open query answering can be tackled through the combination of human effort with algorithmic techniques. In this paper we explore its applicability to crowdsourcing, using a framework in which human and computational intelligence can co-exist by augmenting existing Linked Data and Linked Service technology with crowdsourcing functionality. We analyze how the task can be decomposed and translated into Mechanical Turk projects in order to achieve this vision.