Goto

Collaborating Authors

 autorec


Multi-Metric AutoRec for High Dimensional and Sparse User Behavior Data Prediction

Liang, Cheng, Huang, Teng, He, Yi, Deng, Song, Wu, Di, Luo, Xin

arXiv.org Artificial Intelligence

User behavior data produced during interaction with massive items in the significant data era are generally heterogeneous and sparse, leaving the recommender system (RS) a large diversity of underlying patterns to excavate. Deep neural network-based models have reached the state-of-the-art benchmark of the RS owing to their fitting capabilities. However, prior works mainly focus on designing an intricate architecture with fixed loss function and regulation. These single-metric models provide limited performance when facing heterogeneous and sparse user behavior data. Motivated by this finding, we propose a multi-metric AutoRec (MMA) based on the representative AutoRec. The idea of the proposed MMA is mainly two-fold: 1) apply different $L_p$-norm on loss function and regularization to form different variant models in different metric spaces, and 2) aggregate these variant models. Thus, the proposed MMA enjoys the multi-metric orientation from a set of dispersed metric spaces, achieving a comprehensive representation of user data. Theoretical studies proved that the proposed MMA could attain performance improvement. The extensive experiment on five real-world datasets proves that MMA can outperform seven other state-of-the-art models in predicting unobserved user behavior data.


AutoRec: An Automated Recommender System

Wang, Ting-Hsiang, Song, Qingquan, Han, Xiaotian, Liu, Zirui, Jin, Haifeng, Hu, Xia

arXiv.org Machine Learning

For example, NCF [8] takes user-item implicit feedback data as inputs for the rating prediction task; and DeepFM [6] leverages both numerical and categorical data for the CTR prediction task. However, high degree of specialization comes at the expense of model adaptability and tuning complexity. As recommendation tasks evolve over time and additional types of data are collected, the originally apt model can either become obsolete or require tremendous tuning efforts. So far, several pipelines for recommender systems, e.g., OpenRec [16] and SMORe [4], tried to address the adaptability issue via providing modular base blocks that can be selected according to the context of recommendation. Nevertheless, both determining the blocks to use and tuning the model parameters are not straightforward when facing new data and changing tasks. In order to bridge the gap, we present AutoRec, which aims to provide an end-to-end solution to automate model selection and hyperparameter tuning. While many AutoML libraries, such as Auto-Sklearn [5] and TPOT [12] have shown promising results in general-purpose machine learning tasks (e.g., regression and hyperparameter tuning) and


Sparsity Normalization: Stabilizing the Expected Outputs of Deep Networks

Yi, Joonyoung, Lee, Juhyuk, Hwang, Sung Ju, Yang, Eunho

arXiv.org Machine Learning

The learning of deep models, in which a numerous of parameters are superimposed, is known to be a fairly sensitive process and should be carefully done through a combination of several techniques that can help to stabilize it. We introduce an additional challenge that has never been explicitly studied: the heterogeneity of sparsity at the instance level due to missing values or the innate nature of the input distribution. We confirm experimentally on the widely used benchmark datasets that this variable sparsity problem makes the output statistics of neurons unstable and makes the learning process more difficult by saturating non-linearities. We also provide the analysis of this phenomenon, and based on our analysis, we present a simple technique to prevent this issue, referred to as Sparsity Normalization (SN). Finally, we show that the performance can be significantly improved with SN on certain popular benchmark datasets, or that similar performance can be achieved with lower capacity. Especially focusing on the collaborative filtering problem where the variable sparsity issue has been completely ignored, we achieve new state-of-the-art results on Movielens 100k and 1M datasets, by simply applying Sparsity Normalization (SN).


DARec: Deep Domain Adaptation for Cross-Domain Recommendation via Transferring Rating Patterns

Yuan, Feng, Yao, Lina, Benatallah, Boualem

arXiv.org Machine Learning

Cross-domain recommendation has long been one of the major topics in recommender systems. Recently, various deep models have been proposed to transfer the learned knowledge across domains, but most of them focus on extracting abstract transferable features from auxilliary contents, e.g., images and review texts, and the patterns in the rating matrix itself is rarely touched. In this work, inspired by the concept of domain adaptation, we proposed a deep domain adaptation model (DARec) that is capable of extracting and transferring patterns from rating matrices {\em only} without relying on any auxillary information. We empirically demonstrate on public datasets that our method achieves the best performance among several state-of-the-art alternative cross-domain recommendation models.


Neural Ideal Point Estimation Network

Song, Kyungwoo (Korea Advanced Institute of Science and Technology) | Lee, Wonsung (Korea Advanced Institute of Science and Technology) | Moon, Il-Chul (Korea Advanced Institute of Science and Technology)

AAAI Conferences

Understanding politics is challenging because the politics take the influence from everything. Even we limit ourselves to the political context in the legislative processes; we need a better understanding of latent factors, such as legislators, bills, their ideal points, and their relations. From the modeling perspective, this is difficult 1) because these observations lie in a high dimension that requires learning on low dimensional representations, and 2) because these observations require complex probabilistic modeling with latent variables to reflect the causalities. This paper presents a new model to reflect and understand this political setting, NIPEN, including factors mentioned above in the legislation. We propose two versions of NIPEN: one is a hybrid model of deep learning and probabilistic graphical model, and the other model is a neural tensor model. Our result indicates that NIPEN successfully learns the manifold of the legislative bill's text, and NIPEN utilizes the learned low-dimensional latent variables to increase the prediction performance of legislators' votings. Additionally, by virtue of being a domain-rich probabilistic model, NIPEN shows the hidden strength of the legislators' trust network and their various characteristics on casting votes.