Goto

Collaborating Authors

 matrix factorization method


Nonnegative matrix factorization and the principle of the common cause

Khalafyan, E., Allahverdyan, A. E., Hovhannisyan, A.

arXiv.org Machine Learning

--Nonnegative matrix factorization (NMF) is a known unsupervised data-reduction method. The principle of the common cause (PCC) is a basic methodological approach in probabilistic causality, which seeks an independent mixture model for the joint probability of two dependent random variables. It turns out that these two concepts are closely related. This relationship is explored reciprocally for several datasets of gray-scale images, which are conveniently mapped into probability models. On one hand, PCC provides a predictability tool that leads to a robust estimation of the effective rank of NMF . Unlike other estimates (e.g., those based on the Bayesian Information Criteria), our estimate of the rank is stable against weak noise. We show that NMF implemented around this rank produces features (basis images) that are also stable against noise and against seeds of local optimization, thereby effectively resolving the NMF nonidentifiability problem. On the other hand, NMF provides an interesting possibility of implementing PCC in an approximate way, where larger and positively correlated joint probabilities tend to be explained better via the independent mixture model. We work out a clustering method, where data points with the same common cause are grouped into the same cluster . We also show how NMF can be employed for data denoising. Nonnegative matrix factorization (NMF) was proposed and developed in data science [1]-[3].


Interpretable Topic Extraction and Word Embedding Learning using row-stochastic DEDICOM

Hillebrand, Lars, Biesner, David, Bauckhage, Christian, Sifa, Rafet

arXiv.org Artificial Intelligence

The DEDICOM algorithm provides a uniquely interpretable matrix factorization method for symmetric and asymmetric square matrices. We employ a new row-stochastic variation of DEDICOM on the pointwise mutual information matrices of text corpora to identify latent topic clusters within the vocabulary and simultaneously learn interpretable word embeddings. We introduce a method to efficiently train a constrained DEDICOM algorithm and a qualitative evaluation of its topic modeling and word embedding performance.


Design of an basis-projected layer for sparse datasets in deep learning training using gc-ms spectra as a case study

Chang, Yu Tang, Chen, Shih Fang

arXiv.org Artificial Intelligence

Deep learning (DL) models encompass millions or even billions of parameters and learn complex patterns from big data. However, not all data are initially stored in a suitable formation to effectively train a DL model, e.g., gas chromatography-mass spectrometry (GC-MS) spectra and DNA sequence. These datasets commonly contain many zero values, and the sparse data formation causes difficulties in optimizing DL models. A DL module called the basis-projected layer (BPL) was proposed to mitigate the issue by transforming the sparse data into a dense representation. The transformed data is expected to facilitate the gradient calculation and finetuned process in a DL training process. The dataset, example of a sparse dataset, contained 362 specialty coffee odorant spectra detected from GC-MS. The BPL layer was placed at the beginning of the DL model. The tunable parameters in the layer were learnable projected axes that were the bases of a new representation space. The layer rotated these bases when its parameters were updated. When the number of the bases was the same as the original dimension, the increasing percentage of the F1 scores was 8.56%. Furthermore, when the number was set as 768 (the original dimension was 490), the increasing percentage of the F1 score was 11.49%. The layer not only maintained the model performance and even constructed a better representation space in analyzing sparse datasets.


MACHINE LEARNING AS AN APPLICATION OF LINEAR ALGEBRA

#artificialintelligence

In this blog, you will discover why machine learning practitioners should study linear algebra to improve their skills and capabilities as practitioners. After reading this blog, you will understand how can linear algebra be applied in machine learning. Linear algebra is the study of vector spaces, lines and planes, and mappings that are used for linear transforms. It was initially formalized in the 1800s to find the unknowns in linear equations systems, and hence it is relatively a young field of study. Linear Algebra is an essential field of mathematics that can also be called the mathematics of data.


GitHub - ethen8181/machine-learning: machine learning tutorials (mainly in Python3)

#artificialintelligence

This is a continuously updated repository that documents personal journey on learning data science, machine learning related topics. The content aims to strike a good balance between mathematical notations, educational implementation from scratch using Python's scientific stack including numpy, numba, scipy, pandas, matplotlib, pyspark etc. and open-source library usage such as scikit-learn, fasttext, huggingface, onnx, xgboost, lightgbm, pytorch, keras, tensorflow, gensim, h2o, ortools, ray tune etc. Notes related to advertising domain. Information Retrieval, some examples are demonstrated using ElasticSearch. End to end project including data preprocessing, model building. Includes: Quick review of necessary statistic concepts.


Generalizable Cross-Graph Embedding for GNN-based Congestion Prediction

Ghose, Amur, Zhang, Vincent, Zhang, Yingxue, Li, Dong, Liu, Wulong, Coates, Mark

arXiv.org Artificial Intelligence

Presently with technology node scaling, an accurate prediction model at early design stages can significantly reduce the design cycle. Especially during logic synthesis, predicting cell congestion due to improper logic combination can reduce the burden of subsequent physical implementations. There have been attempts using Graph Neural Network (GNN) techniques to tackle congestion prediction during the logic synthesis stage. However, they require informative cell features to achieve reasonable performance since the core idea of GNNs is built on the message passing framework, which would be impractical at the early logic synthesis stage. To address this limitation, we propose a framework that can directly learn embeddings for the given netlist to enhance the quality of our node features. Popular random-walk based embedding methods such as Node2vec, LINE, and DeepWalk suffer from the issue of cross-graph alignment and poor generalization to unseen netlist graphs, yielding inferior performance and costing significant runtime. In our framework, we introduce a superior alternative to obtain node embeddings that can generalize across netlist graphs using matrix factorization methods. We propose an efficient mini-batch training method at the sub-graph level that can guarantee parallel training and satisfy the memory restriction for large-scale netlists. We present results utilizing open-source EDA tools such as DREAMPLACE and OPENROAD frameworks on a variety of openly available circuits. By combining the learned embedding on top of the netlist with the GNNs, our method improves prediction performance, generalizes to new circuit lines, and is efficient in training, potentially saving over $90 \%$ of runtime.


Multi-view Clustering with Deep Matrix Factorization and Global Graph Refinement

#artificialintelligence

Multi-view clustering is an important yet challenging task in machine learning and data mining community. One popular strategy for multi-view clustering is matrix factorization which could explore useful feature representations at lower-dimensional space and therefore alleviate dimension curse. However, there are two major drawbacks in the existing work: i) most matrix factorization methods are limited to shadow depth, which leads to the inability to fully discover the rich hidden information of original data. Few deep matrix factorization methods provide a basis for the selection of the new representation's dimensions of different layers. To tackle the above issues, we propose a novel Multi-View Clustering method with Deep semi-NMF and Global Graph Refinement (MVC-DMF-GGR) in this paper. Firstly, we capture new representation matrices for each view by hierarchical decomposition, then learn a common graph by approximating a combination of graphs which are reconstructed from these new representations to refine the new representations in return.


Learn NLP the Stanford Way -- Lesson 2

#artificialintelligence

In the previous post, we introduced NLP. To find out word meanings with the Python programming language, we used the NLTK package and worked our way into word embeddings using the gensim package and Word2vec. Since we only touched the Word2Vec technique from a 10,000-feet overview, we are now going to dive deeper into the training method to create a Word2vec model. The Word2vec (Mikolov et al. 2013)[1][2] is not a singular technique or algorithm. It's actually a family of neural network architectures and optimization techniques that can produce good results learning embeddings for large datasets.


Introduction To Recommender Systems- 2: Deep Neural Network Based Recommendation Systems

#artificialintelligence

It is my second article on the Recommendation systems. In my previous article, I have talked about content-based and collaborative filtering systems. I will encourage you to go through the article if you have any confusion. In this article, we are going to see how Deep Learning is used in Recommender systems. We will go through the recommender system's candidate generation architecture of Youtube.


10 Examples of Linear Algebra in Machine Learning - Machine Learning Mastery

@machinelearnbot

Linear algebra is a sub-field of mathematics concerned with vectors, matrices, and linear transforms. It is a key foundation to the field of machine learning, from notations used to describe the operation of algorithms to the implementation of algorithms in code. Although linear algebra is integral to the field of machine learning, the tight relationship is often left unexplained or explained using abstract concepts such as vector spaces or specific matrix operations. In this post, you will discover 10 common examples of machine learning that you may be familiar with that use, require and are really best understood using linear algebra. In this post, we will review 10 obvious and concrete examples of linear algebra in machine learning.