Institute for Infocomm Research
CoupleNet: Paying Attention to Couples with Coupled Attention for Relationship Recommendation
Tay, Yi (Nanyang Technological University) | Tuan, Luu Anh (Institute for Infocomm Research) | Hui, Siu Cheung (Nanyang Technological University)
Dating and romantic relationships not only play a huge role in our personal lives but also collectively influence and shape society. Today, many romantic partnerships originate from the Internet, signifying the importance of technology and the web in modern dating. In this paper, we present a text-based computational approach for estimating the relationship compatibility of two users on social media. Unlike many previous works that propose reciprocal recommender systems for online dating websites, we devise a distant supervision heuristic to obtain real world couples from social platforms such as Twitter. Our approach, the CoupleNet is an end-to-end deep learning basedestimator that analyzes the social profiles of two users and subsequently performs a similarity match between the users. Intuitively, our approach performs both user profiling and match-making within a unified end-to-end framework. CoupleNet utilizes hierarchical recurrent neural models for learning representations of user profiles and subsequently coupled attention mechanisms to fuse information aggregated from two users.To the best of our knowledge, our approach is the first data-driven deep learning approach for our novel relationship recommendation problem. We benchmarkour CoupleNet against several machine learning and deep learning baselines. Experimental results show that our approach outperformsall approaches significantly in terms of precision. Qualitative analysis shows that our model is capable of also producing explainable results to users.
Deep Activity Recognition Models with Triaxial Accelerometers
Alsheikh, Mohammad Abu (Nanyang Technological University) | Selim, Ahmed (Trinity College Dublin) | Niyato, Dusit (Nanyang Technological University) | Doyle, Linda (Trinity College Dublin) | Lin, Shaowei (Institute for Infocomm Research) | Tan, Hwee-Pink (Singapore Management University)
Despite the widespread installation of accelerometers in almost all mobile phones and wearable devices, activity recognition using accelerometers is still immature due to the poor recognition accuracy of existing recognition methods and the scarcity of labeled training data. We consider the problem of human activity recognition using triaxial accelerometers and deep learning paradigms. This paper shows that deep activity recognition models (a) provide better recognition accuracy of human activities, (b) avoid the expensive design of handcrafted features in existing systems, and (c) utilize the massive unlabeled acceleration samples for unsupervised feature extraction. We show substantial recognition improvement on real world datasets over state-of-the-art methods of human activity recognition using triaxial accelerometers.
Collective Biobjective Optimization Algorithm for Parallel Test Paper Generation
Nguyen, Minh Luan (Institute for Infocomm Research) | Hui, Siu Cheung (Nanyang Technological University) | Fong, Alvis C. M. (University of Glasgow)
Parallel Test Paper Generation ( k -TPG) is a biobjective distributed resource allocation problem, which aims to generate multiple similarly optimal test papers automatically according to multiple user-specified criteria.Generating high-quality parallel test papers is challenging due to its NP-hardness in maximizing the collective objective functions.In this paper, we propose a Collective Biobjective Optimization (CBO) algorithm for solving k -TPG. CBO is a multi-step greedy-based approximation algorithm, which exploits the submodular property for biobjective optimization of k -TPG.Experiment results have shown that CBO has drastically outperformed the current techniques in terms of paper quality and runtime efficiency.
Modeling Users' Dynamic Preference for Personalized Recommendation
Liu, Xin (Institute for Infocomm Research)
Modeling the evolution of users' preference over time is essential for personalized recommendation. Traditional time-aware models like (1) time-window or recency based approaches ignore or deemphasize much potentially useful information, and (2) time-aware collaborative filtering (CF) approaches largely rely on the information of other users, thus failing to precisely and comprehensively profile individual users for personalization. In this paper, for implicit feedback data, we propose a personalized recommendation model to capture users' dynamic preference using Gaussian process. We first apply topic modeling to represent a user's temporal preference in an interaction as a topic distribution. By aggregating such topic distributions of the user's past interactions, we build her profile, where we treat each topic's values at different interactions as a time series. Gaussian process is then applied to predict the user's preference in the next interactions for top-N recommendation. Experiments conducted over two real datasets demonstrate that our approach outperforms the state-of-the-art recommendation models by at least 42.46% and 66.14% in terms of precision and Mean Reciprocal Rank respectively.
An Adaptive Gradient Method for Online AUC Maximization
Ding, Yi (Nanyang Technological University) | Zhao, Peilin (Institute for Infocomm Research) | Hoi, Steven C. H. (Singapore Management University) | Ong, Yew-Soon (Nanyang Technological University)
Learning for maximizing AUC performance is an important research problem in machine learning. Unlike traditional batch learning methods for maximizing AUC which often suffer from poor scalability, recent years have witnessed some emerging studies that attempt to maximize AUC by single-pass online learning approaches. Despite their encouraging results reported, the existing online AUC maximization algorithms often adopt simple stochastic gradient descent approaches, which fail to exploit the geometry knowledge of the data observed in the online learning process, and thus could suffer from relatively slow convergence. To overcome the limitation of the existing studies, in this paper, we propose a novel algorithm of Adaptive Online AUC Maximization (AdaOAM), by applying an adaptive gradient method for exploiting the knowledge of historical gradients to perform more informative online learning. The new adaptive updating strategy by AdaOAM is less sensitive to parameter settings due to its natural effect of tuning the learning rate. In addition, the time complexity of the new algorithm remains the same as the previous non-adaptive algorithms. To demonstrate the effectiveness of the proposed algorithm, we analyze its theoretical bound, and further evaluate its empirical performance on both public benchmark datasets and anomaly detection datasets. The encouraging empirical results clearly show the effectiveness and efficiency of the proposed algorithm.
Extracting Verb Expressions Implying Negative Opinions
Li, Huayi (University of Illinois at Chicago) | Mukherjee, Arjun (University of Houston) | Si, Jianfeng (Institute for Infocomm Research) | Liu, Bing (University of Illinois at Chicago)
Identifying aspect-based opinions has been studied extensively in recent years. However, existing work primarily focused on adjective, adverb, and noun expressions. Clearly, verb expressions can imply opinions too. We found that in many domains verb expressions can be even more important to applications because they often describe major issues of products or services. These issues enable brands and businesses to directly improve their products or services. To the best of our knowledge, this problem has not received much attention in the literature. In this paper, we make an attempt to solve this problem. Our proposed method first extracts verb expressions from reviews and then employs Markov Networks to model rich linguistic features and long distance relationships to identify negative issue expressions. Since our training data is obtained from titles of reviews whose labels are automatically inferred from review ratings, our approach is applicable to any domain without manual involvement. Experimental results using real-life review datasets show that our approach outperforms strong baselines.
Robust Subspace Clustering via Thresholding Ridge Regression
Peng, Xi (Institute for Infocomm Research) | Yi, Zhang (Sichuan University) | Tang, Huajin (Institute for Infocomm Research)
Given a data set from a union of multiple linear subspaces, a robust subspace clustering algorithm fits each group of data points with a low-dimensional subspace and then clusters these data even though they are grossly corrupted or sampled from the union of dependent subspaces. Under the framework of spectral clustering, recent works using sparse representation, low rank representation and their extensions achieve robust clustering results by formulating the errors (e.g., corruptions) into their objective functions so that the errors can be removed from the inputs. However, these approaches have suffered from the limitation that the structure of the errors should be known as the prior knowledge. In this paper, we present a new method of robust subspace clustering by eliminating the effect of the errors from the projection space (representation) rather than from the input space. We firstly prove that ell_1-, ell_2-, and ell_infty-norm-based linear projection spaces share the property of intra-subspace projection dominance, i.e., the coefficients over intra-subspace data points are larger than those over inter-subspace data points. Based on this property, we propose a robust and efficient subspace clustering algorithm, called Thresholding Ridge Regression (TRR). TRR calculates the ell2-norm-based coefficients of a given data set and performs a hard thresholding operator; and then the coefficients are used to build a similarity graph for clustering. Experimental studies show that TRR outperforms the state-of-the-art methods with respect to clustering quality, robustness, and time-saving.
How Long Will It Take? Accurate Prediction of Ontology Reasoning Performance
Kang, Yong-Bin (Monash University) | Pan, Jeff Z. (University of Aberdeen) | Krishnaswamy, Shonali (Institute for Infocomm Research) | Sawangphol, Wudhichart (Monash University) | Li, Yuan-Fang (Monash University)
For expressive ontology languages such as OWL 2 DL, classification is a computationally expensive taskโ2NEXPTIME-complete in the worst case. Hence, it is highly desirable to be able to accurately estimate classification time, especially for large and complex ontologies. Recently, machine learning techniques have been successfully applied to predicting the reasoning hardness category for a given (ontology, reasoner) pair. In this paper, we further develop predictive models to estimate actual classification time using regression techniques, with ontology metrics as features. Our large-scale experiments on 6 state-of-the-art OWL 2 DL reasoners and more than 450 significantly diverse ontologies demonstrate that the prediction models achieve high accuracy, good generalizability and statistical significance. Such prediction models have a wide range of applications. We demonstrate how they can be used to efficiently and accurately identify performance hotspots in a large and complex ontology, an otherwise very time-consuming and resource-intensive task.
Hybrid Heterogeneous Transfer Learning through Deep Learning
Zhou, Joey Tianyi (Nanyang Technological University) | Pan, Sinno Jialin (Institute for Infocomm Research) | Tsang, Ivor W. (University of Technology, Sydney) | Yan, Yan (University of Queensland)
Most previous heterogeneous transfer learning methods learn a cross-domain feature mapping between heterogeneous feature spaces based on a few cross-domain instance-correspondences, and these corresponding instances are assumed to be representative in the source and target domains respectively. However, in many real-world scenarios, this assumption may not hold. As a result, the constructed feature mapping may not be precisely due to the bias issue of the correspondences in the target or (and) source domain(s). In this case, a classifier trained on the labeled transformed-source-domain data may not be useful for the target domain. In this paper, we present a new transfer learning framework called Hybrid Heterogeneous Transfer Learning (HHTL), which allows the corresponding instances across domains to be biased in either the source or target domain. Specifically, we propose a deep learning approach to learn a feature mapping between cross-domain heterogeneous features as well as a better feature representation for mapped data to reduce the bias issue caused by the cross-domain correspondences. Extensive experiments on several multilingual sentiment classification tasks verify the effectiveness of our proposed approach compared with some baseline methods.
Source Free Transfer Learning for Text Classification
Lu, Zhongqi (Hong Kong University of Science and Technology) | Zhu, Yin (Hong Kong University of Science and Technology) | Pan, Sinno Jialin (Institute for Infocomm Research) | Xiang, Evan Wei (Baidu Inc.) | Wang, Yujing (Microsoft Research Asia, Beijing) | Yang, Qiang (Hong Kong University of Science and Technology)
Transfer learning uses relevant auxiliary data to help the learning task in a target domain where labeled data is usually insufficient to train an accurate model. Given appropriate auxiliary data, researchers have proposed many transfer learning models. How to find such auxiliary data, however, is of little research so far. In this paper, we focus on the problem of auxiliary data retrieval, and propose a transfer learning framework that effectively selects helpful auxiliary data from an open knowledge space (e.g. the World Wide Web). Because there is no need of manually selecting auxiliary data for different target domain tasks, we call our framework Source Free Transfer Learning (SFTL). For each target domain task, SFTL framework iteratively queries for the helpful auxiliary data based on the learned model and then updates the model using the retrieved auxiliary data. We highlight the automatic constructions of queries and the robustness of the SFTL framework. Our experiments on 20NewsGroup dataset and a Google search snippets dataset suggest that the framework is capable of achieving comparable performance to those state-of-the-art methods with dedicated selections of auxiliary data.