Goto

Collaborating Authors

 Chen, Huajun


Long-tail Relation Extraction via Knowledge Graph Embeddings and Graph Convolution Networks

arXiv.org Artificial Intelligence

We propose a distance supervised relation extraction approach for long-tailed, imbalanced data which is prevalent in real-world settings. Here, the challenge is to learn accurate "few-shot" models for classes existing at the tail of the class distribution, for which little data is available. Inspired by the rich semantic correlations between classes at the long tail and those at the head, we take advantage of the knowledge from data-rich classes at the head of the distribution to boost the performance of the data-poor classes at the tail. First, we propose to leverage implicit relational knowledge among class labels from knowledge graph embeddings and learn explicit relational knowledge using graph convolution networks. Second, we integrate that relational knowledge into relation extraction model by coarse-to-fine knowledge-aware attention mechanism. We demonstrate our results for a large-scale benchmark dataset which show that our approach significantly outperforms other baselines, especially for long-tail relations.


Human-centric Transfer Learning Explanation via Knowledge Graph [Extended Abstract]

arXiv.org Machine Learning

Transfer learning which aims at utilizing knowledge learned from one problem (source domain) to solve another different but related problem (target domain) has attracted wide research attentions. However, the current transfer learning methods are mostly uninterpretable, especially to people without ML expertise. In this extended abstract, we brief introduce two knowledge graph (KG) based frameworks towards human understandable transfer learning explanation. The first one explains the transferability of features learned by Convolutional Neural Network (CNN) from one domain to another through pre-training and fine-tuning, while the second justifies the model of a target domain predicted by models from multiple source domains in zero-shot learning (ZSL). Both methods utilize KG and its reasoning capability to provide rich and human understandable explanations to the transfer procedure.


Attention-Based Capsule Networks with Dynamic Routing for Relation Extraction

arXiv.org Artificial Intelligence

A capsule is a group of neurons, whose activity vector represents the instantiation parameters of a specific type of entity. In this paper, we explore the capsule networks used for relation extraction in a multi-instance multi-label learning framework and propose a novel neural approach based on capsule networks with attention mechanisms. We evaluate our method with different benchmarks, and it is demonstrated that our method improves the precision of the predicted relations. Particularly, we show that capsule networks improve multiple entity pairs relation extraction.


Knowledge-based Transfer Learning Explanation

arXiv.org Artificial Intelligence

Machine learning explanation can significantly boost machine learning's application in decision making, but the usability of current methods is limited in human-centric explanation, especially for transfer learning, an important machine learning branch that aims at utilizing knowledge from one learning domain (i.e., a pair of dataset and prediction task) to enhance prediction model training in another learning domain. In this paper, we propose an ontology-based approach for human-centric explanation of transfer learning. Three kinds of knowledge-based explanatory evidence, with different granularities, including general factors, particular narrators and core contexts are first proposed and then inferred with both local ontologies and external knowledge bases. The evaluation with US flight data and DBpedia has presented their confidence and availability in explaining the transferability of feature representation in flight departure delay forecasting.


AAAI-07 Workshop Reports

AI Magazine

The AAAI-07 workshop program was held Sunday and Monday, July 22-23, in Vancouver, British Columbia, Canada. The program included the following thirteen workshops: (1) Acquiring Planning Knowledge via Demonstration; (2) Configuration; (3) Evaluating Architectures for Intelligence; (4) Evaluation Methods for Machine Learning; (5) Explanation-Aware Computing; (6) Human Implications of Human-Robot Interaction; (7) Intelligent Techniques for Web Personalization; (8) Plan, Activity, and Intent Recognition; (9) Preference Handling for Artificial Intelligence; (10) Semantic e-Science; (11) Spatial and Temporal Reasoning; (12) Trading Agent Design and Analysis; and (13) Information Integration on the Web.


AAAI-07 Workshop Reports

AI Magazine

The AAAI-07 workshop program was held Sunday and Monday, July 22-23, in Vancouver, British Columbia, Canada. The program included the following thirteen workshops: (1) Acquiring Planning Knowledge via Demonstration; (2) Configuration; (3) Evaluating Architectures for Intelligence; (4) Evaluation Methods for Machine Learning; (5) Explanation-Aware Computing; (6) Human Implications of Human-Robot Interaction; (7) Intelligent Techniques for Web Personalization; (8) Plan, Activity, and Intent Recognition; (9) Preference Handling for Artificial Intelligence; (10) Semantic e-Science; (11) Spatial and Temporal Reasoning; (12) Trading Agent Design and Analysis; and (13) Information Integration on the Web.