Goto

Collaborating Authors

 Pados, Dimitris A.


Outlier Processing Via L1-Principal Subspaces

AAAI Conferences

With the advent of big data, there is a growing demand for smart algorithms that can extract relevant information from high-dimensional large data sets, potentially corrupted by faulty measurements (outliers). In this context, we present a novel line of research that utilizes the robust nature of L1-norm subspaces for data dimensionality reduction and outlier processing. Specifically, (i) we use the euclidean-distance between original and L1-norm-subspace projected samples as a metric to assign weight to each sample point, (ii) perform (K=2)-means clustering over the one-dimensional weights discarding samples corresponding to the outlier cluster, and (iii) compute the robust L1-norm principal subspaces over the reduced โ€œcleanโ€ data set for further applications. Numerical studies included in this paper from the fields of (i) data dimesnionality reduction, (ii) direction-of-arrival estimation, (iii) image fusion, and (iv) video foreground extarction demonstrate the efficacy of the proposed outlier processing algorithm in designing robust low-dimensional subspaces from faulty high-dimensional data.


Efficient L1-Norm Principal-Component Analysis via Bit Flipping

arXiv.org Machine Learning

It was shown recently that the $K$ L1-norm principal components (L1-PCs) of a real-valued data matrix $\mathbf X \in \mathbb R^{D \times N}$ ($N$ data samples of $D$ dimensions) can be exactly calculated with cost $\mathcal{O}(2^{NK})$ or, when advantageous, $\mathcal{O}(N^{dK - K + 1})$ where $d=\mathrm{rank}(\mathbf X)$, $K


Some Options for L1-Subspace Signal Processing

arXiv.org Machine Learning

We describe ways to define and calculate $L_1$-norm signal subspaces which are less sensitive to outlying data than $L_2$-calculated subspaces. We focus on the computation of the $L_1$ maximum-projection principal component of a data matrix containing N signal samples of dimension D and conclude that the general problem is formally NP-hard in asymptotically large N, D. We prove, however, that the case of engineering interest of fixed dimension D and asymptotically large sample support N is not and we present an optimal algorithm of complexity $O(N^D)$. We generalize to multiple $L_1$-max-projection components and present an explicit optimal $L_1$ subspace calculation algorithm in the form of matrix nuclear-norm evaluations. We conclude with illustrations of $L_1$-subspace signal processing in the fields of data dimensionality reduction and direction-of-arrival estimation.