National Taiwan University
A Low-Cost Ethics Shaping Approach for Designing Reinforcement Learning Agents
Wu, Yueh-Hua (National Taiwan University) | Lin, Shou-De (National Taiwan University)
One idea to achieve incorrect but also unethical decisions. Reinforcement learning such goal is to collect enough ethical behavior data of human (Sutton and Barto 1998) is designed to tackle intricate acting toward the given goal, and then apply the inverse real-world problems in rather short time (Strehl et al. 2006; reinforcement learning (IRL) (Amin and Singh 2016; Evans, Brafman and Tennenholtz 2002) with a performance bound Stuhlmüller, and Goodman 2016; Ng, Russell, and others (Strehl, Li, and Littman 2009); however, it relies heavily on 2000; Sezener 2015) technique to learn an ethical agent that the quality of the reward functions provided as the inputs.
Order-Free RNN With Visual Attention for Multi-Label Classification
Chen, Shang-Fu (National Taiwan University) | Chen, Yi-Chen (National Taiwan University) | Yeh, Chih-Kuan (Carnegie Mellon University) | Wang, Yu-Chiang Frank (National Taiwan University)
While a number of research works (Zhang and Zhou 2006; Nam et al. 2014; Gong et al. 2013; Wei et al. 2014; We propose a recurrent neural network (RNN) based model Wang et al. 2016) start to advance the CNN architecture for image multi-label classification. Our model uniquely integrates for multi-label classification, CNN-RNN (Wang et al. and learning of visual attention and Long Short 2016) embeds image and semantic structures by projecting Term Memory (LSTM) layers, which jointly learns the labels both features into a joint embedding space. By further of interest and their co-occurrences, while the associated utilizing the component of Long Short Term Memory image regions are visually attended. Different from existing (LSTM) (Hochreiter and Schmidhuber 1997), a recurrent approaches utilize either model in their network architectures, neural network (RNN) structure is introduced to memorize training of our model does not require predefined long-term label dependency. As a result, CNN-RNN exhibits label orders. Moreover, a robust inference process is introduced promising multi-label classification performance with crosslabel so that prediction errors would not propagate and thus correlation implicitly preserved.
A Deep Model With Local Surrogate Loss for General Cost-Sensitive Multi-Label Learning
Hsieh, Cheng-Yu (National Taiwan University) | Lin, Yi-An (National Taiwan University) | Lin, Hsuan-Tien (National Taiwan University)
Multi-label learning is an important machine learning problem with a wide range of applications. The variety of criteria for satisfying different application needs calls for cost-sensitive algorithms, which can adapt to different criteria easily. Nevertheless, because of the sophisticated nature of the criteria for multi-label learning, cost-sensitive algorithms for general criteria are hard to design, and current cost-sensitive algorithms can at most deal with some special types of criteria. In this work, we propose a novel cost-sensitive multi-label learning model for any general criteria. Our key idea within the model is to iteratively estimate a surrogate loss that approximates the sophisticated criterion of interest near some local neighborhood, and use the estimate to decide a descent direction for optimization. The key idea is then coupled with deep learning to form our proposed model. Experimental results validate that our proposed model is superior to existing cost-sensitive algorithms and existing deep learning models across different criteria.
Scheduling in Visual Fog Computing: NP-Completeness and Practical Efficient Solutions
Chu, Hong-Min (National Taiwan University) | Yang, Shao-Wen (Intel) | Pillai, Padmanabhan (Intel) | Chen, Yen-Kuang (Intel)
The visual fog paradigm envisions tens of thousands of heterogeneous, camera-enabled edge devices distributed across the Internet, providing live sensing for a myriad of different visual processing applications. The scale, computational demands, and bandwidth needed for visual computing pipelines necessitates offloading intelligently to distributed computing infrastructure, including the cloud, Internet gateway devices, and the edge devices themselves. This paper focuses on the visual fog scheduling problem of assigning the visual computing tasks to various devices to optimize network utilization. We first prove this problem is NP-complete, and then formulate a practical, efficient solution. We demonstrate sub-minute computation time to optimally schedule 20,000 tasks across over 7,000 devices, and just 7-minute execution time to place 60,000 tasks across 20,000 devices, showing our approach is ready to meet the scale challenges introduced by visual fog.
Supporting ESL Writing by Prompting Crowdsourced Structural Feedback
Huang, Yi-Ching (National Taiwan University) | Huang, Jiunn-Chia (National Taiwan University) | Wang, Hao-Chuan (National Tsing Hua University) | Hsu, Jane Yung-jen (National Taiwan University)
Writing is challenging, especially for non-native speakers. To support English as a Second Language (ESL) writing, we propose StructFeed, which allows native speakers to annotate topic sentence and relevant keywords in texts and generate writing hints based on the principle of paragraph unity. First, we compared our crowd-based method with three naive machine learning (ML) methods and got the best performance on the identification of topic sentence and irrelevant sentence in the article. Next, we evaluated the StructFeed system with two feedback-generation mechanisms including feedback generated by one expert and by one crowd worker. The results showed that people who received feedback by StructFeed got the highest improvement after revision.
Learning Deep Latent Space for Multi-Label Classification
Yeh, Chih-Kuan (Academic Sinica) | Wu, Wei-Chieh (National Taiwan University) | Ko, Wei-Jen (National Taiwan University) | Wang, Yu-Chiang Frank (Academic Sinica)
Multi-label classification is a practical yet challenging task in machine learning related fields, since it requires the prediction of more than one label category for each input instance. We propose a novel deep neural networks (DNN) based model, Canonical Correlated AutoEncoder (C2AE), for solving this task. Aiming at better relating feature and label domain data for improved classification, we uniquely perform joint feature and label embedding by deriving a deep latent space, followed by the introduction of label-correlation sensitive loss function for recovering the predicted label outputs. Our C2AE is achieved by integrating the DNN architectures of canonical correlation analysis and autoencoder, which allows end-to-end learning and prediction with the ability to exploit label dependency. Moreover, our C2AE can be easily extended to address the learning problem with missing labels. Our experiments on multiple datasets with different scales confirm the effectiveness and robustness of our proposed method, which is shown to perform favorably against state-of-the-art methods for multi-label classification.
A Unified Algorithm for One-Cass Structured Matrix Factorization with Side Information
Yu, Hsiang-Fu (University of Texas at Austin) | Huang, Hsin-Yuan (National Taiwan University) | Dhillon, Inderjit (University of Texas at Austin) | Lin, Chih-Jen (National Taiwan University)
In many applications such as recommender systems and multi-label learning the task is to complete a partially observed binary matrix. Such PU learning (positive-unlabeled) problems can be solved by one-class matrix factorization (MF). In practice side information such as user or item features in recommender systems are often available besides the observed positive user-item connections. In this work we consider a generalization of one-class MF so that two types of side information are incorporated and a general convex loss function can be used. The resulting optimization problem is very challenging, but we derive an efficient and effective alternating minimization procedure. Experiments on large-scale multi-label learning and one-class recommender systems demonstrate the effectiveness of our proposed approach.
Multi-Modal Learning over User-Contributed Content from Cross-Domain Social Media
Lee, Wen-Yu (National Taiwan University)
The goal of the research is to discover and summarize data from the emerging social media into information of interests. Specifically, leveraging user-contributed data from cross-domain social media, the idea is to perform multi-modal learning for a given photo, aiming to present people’s description or comments, geographical information, and events of interest, closely related to the photo. These information then can be used for various purposes, such as being a real-time guide for the tourists to improve the quality of tourism. As a result, this research investigates modern challenges of image annotation, image retrieval, and cross-media mining, followed by presenting promising ways to conquer the challenges.
Domain-Constraint Transfer Coding for Imbalanced Unsupervised Domain Adaptation
Tsai, Yao-Hung Hubert (Academia Sinica) | Hou, Cheng-An (Carnegie Mellon University) | Chen, Wei-Yu (National Taiwan University) | Yeh, Yi-Ren (National Kaohsiung Normal University) | Wang, Yu-Chiang Frank (Academia Sinica)
Unsupervised domain adaptation (UDA) deals with the task that labeled training and unlabeled test data collected from source and target domains, respectively. In this paper, we particularly address the practical and challenging scenario of imbalanced cross-domain data. That is, we do not assume the label numbers across domains to be the same, and we also allow the data in each domain to be collected from multiple datasets/sub-domains. To solve the above task of imbalanced domain adaptation, we propose a novel algorithm of Domain-constraint Transfer Coding (DcTC). Our DcTC is able to exploit latent subdomains within and across data domains, and learns a common feature space for joint adaptation and classification purposes. Without assuming balanced cross-domain data as most existing UDA approaches do, we show that our method performs favorably against state-of-the-art methods on multiple cross-domain visual classification tasks.
Chinese Relation Extraction by Multiple Instance Learning
Chen, Yu-Ju (National Taiwan University) | Hsu, Jane Yung-jen (National Taiwan University)
Relation extraction, which learns semantic relations of concept pairs from text, is an approach for mining commonsense knowledge. This paper investigates an approach for relation extraction, which helps expand a commonsense knowledge base with little labor work. We proposed a framework that learns new pairs from Chinese corpora by adopting concept pairs in Chinese commonsense knowledge base as seeds. Multiple instance learning is utilized as the learning algorithm for predicting relation for unseen pairs. The performance of our system could be improved by learning multiple iterations. The results in each iteration are manually evaluated and processed to next iteration as seeds. Our experiments extracted new pairs for relations “AtLocation”, “CapableOf”, and “HasProperty”. This study showed that new pairs could be extracted from text without huge humans work.