Collaborating Authors


AAAI Conferences

Recent progress in machine learning has lead to great advancements in robot intelligence and human-robot interaction (HRI). It is reported that robots can deeply understand visual scene information and describe the scenes in natural language using object recognition and natural language processing methods. Image-based question and answering (Q&A) systems can be used for enhancing HRI. However, despite these successful results, several key issues still remain to be discussed and improved. In particular, it is essential for an agent to act in a dynamic, uncertain, and asynchronous envi-ronment for achieving human-level robot intelligence. In this paper, we propose a prototype system for a video Q&A robot "Pororobot". The system uses the state-of-the-art machine learning methods such as a deep concept hierarchy model. In our scenario, a robot and a child plays a video Q&A game together under real world environments. Here we demonstrate preliminary results of the proposed system and discuss some directions as future works.

Which machine learning is the best? - Rebellion Research


Which machine learning is the best? The five top machine learning methods I prefer are dimension reduction, regression, ensembling methods, natural language processing and reinforcement learning. Dimension reduction is important mainly because of its use in feature engineering and data pre-processing, where proper dimension reduction can greatly save time and space while keeping the most important information of the dataset. A good dataset that makes sense is just as important as having a powerful model, just as the saying goes – "rubbish in, rubbish out." Regression, being usually the most intuitive method to come right to mind, is both easy to interpret and fast to run.

Beyond Offline Mapping: Learning Cross Lingual Word Embeddings through Context Anchoring Artificial Intelligence

Recent research on cross-lingual word embeddings has been dominated by unsupervised mapping approaches that align monolingual embeddings. Such methods critically rely on those embeddings having a similar structure, but it was recently shown that the separate training in different languages causes departures from this assumption. In this paper, we propose an alternative approach that does not have this limitation, while requiring a weak seed dictionary (e.g., a list of identical words) as the only form of supervision. Rather than aligning two fixed embedding spaces, our method works by fixing the target language embeddings, and learning a new set of embeddings for the source language that are aligned with them. To that end, we use an extension of skip-gram that leverages translated context words as anchor points, and incorporates self-learning and iterative restarts to reduce the dependency on the initial dictionary. Our approach outperforms conventional mapping methods on bilingual lexicon induction, and obtains competitive results in the downstream XNLI task.

Prompt-Learning for Fine-Grained Entity Typing Artificial Intelligence

As an effective approach to tune pre-trained language models (PLMs) for specific tasks, prompt-learning has recently attracted much attention from researchers. By using \textit{cloze}-style language prompts to stimulate the versatile knowledge of PLMs, prompt-learning can achieve promising results on a series of NLP tasks, such as natural language inference, sentiment classification, and knowledge probing. In this work, we investigate the application of prompt-learning on fine-grained entity typing in fully supervised, few-shot and zero-shot scenarios. We first develop a simple and effective prompt-learning pipeline by constructing entity-oriented verbalizers and templates and conducting masked language modeling. Further, to tackle the zero-shot regime, we propose a self-supervised strategy that carries out distribution-level optimization in prompt-learning to automatically summarize the information of entity types. Extensive experiments on three fine-grained entity typing benchmarks (with up to 86 classes) under fully supervised, few-shot and zero-shot settings show that prompt-learning methods significantly outperform fine-tuning baselines, especially when the training data is insufficient.

Skip-Gram Model


Natural Language Processing is the popular field of Artificial Intelligence. We go to process human language as text or speech to make computers alike humans in this process. Humans have a big amount of data written in a much careless format. That is a problem for any machine to find meaning from raw text. We essential to transforming this data into a vector format to make a machine learn from the raw text.