Tempczyk, Piotr
A Wiener process perspective on local intrinsic dimension estimation methods
Tempczyk, Piotr, Garncarek, Łukasz, Filipiak, Dominik, Kurpisz, Adam
Local intrinsic dimension (LID) estimation methods have received a lot of attention in recent years thanks to the progress in deep neural networks and generative modeling. In opposition to old non-parametric methods, new methods use generative models to approximate diffused dataset density and scale the methods to high-dimensional datasets like images. In this paper, we investigate the recent state-of-the-art parametric LID estimation methods from the perspective of the Wiener process. We explore how these methods behave when their assumptions are not met. We give an extended mathematical description of those methods and their error as a function of the probability density of the data.
Polite Teacher: Semi-Supervised Instance Segmentation with Mutual Learning and Pseudo-Label Thresholding
Filipiak, Dominik, Zapała, Andrzej, Tempczyk, Piotr, Fensel, Anna, Cygan, Marek
The proposed architecture relies on the Teacher-Student mutual learning framework. To filter out noisy pseudo-labels, we use confidence thresholding for bounding boxes and mask scoring for masks. The approach has been tested with CenterMask, a single-stage anchor-free detector. To the best of our knowledge, this is one of the first works tackling the problem of semi-supervised instance segmentation and the first one devoted to an anchor-free detector. The advent of deep learning transformed computer vision pipelines both in academia and industry. However, progress is often hindered, since deep learning models are expensive to train for several reasons. Leaving the hardware and computational expenses aside, the vast share of costs often comes from providing the right amount of samples to learn from. For a number of supervised problems in computer vision, it is relatively easy to obtain data. However, labelling them is often the real source of expenses. Semi-supervised learning methods are tailored to deal with the situation in which there are enough data samples, but access to the labels is severely limited.
LIDL: Local Intrinsic Dimension Estimation Using Approximate Likelihood
Tempczyk, Piotr, Michaluk, Rafał, Garncarek, Łukasz, Spurek, Przemysław, Tabor, Jacek, Goliński, Adam
Most of the existing methods for estimating the local intrinsic dimension of a data distribution do not scale well to high-dimensional data. Many of them rely on a non-parametric nearest neighbors approach which suffers from the curse of dimensionality. We attempt to address that challenge by proposing a novel approach to the problem: Local Intrinsic Dimension estimation using approximate Likelihood (LIDL). Our method relies on an arbitrary density estimation method as its subroutine and hence tries to sidestep the dimensionality challenge by making use of the recent progress in parametric neural methods for likelihood estimation. We carefully investigate the empirical properties of the proposed method, compare them with our theoretical predictions, and show that LIDL yields competitive results on the standard benchmarks for this problem and that it scales to thousands of dimensions. What is more, we anticipate this approach to improve further with the continuing advances in the density estimation literature.
n-CPS: Generalising Cross Pseudo Supervision to n networks for Semi-Supervised Semantic Segmentation
Filipiak, Dominik, Tempczyk, Piotr, Cygan, Marek
We present n-CPS - a generalisation of the recent state-of-the-art cross pseudo supervision (CPS) approach for the task of semi-supervised semantic segmentation. In n-CPS, there are n simultaneously trained subnetworks that learn from each other through one-hot encoding perturbation and consistency regularisation. We also show that ensembling techniques applied to subnetworks outputs can significantly improve the performance. To the best of our knowledge, n-CPS paired with CutMix outperforms CPS and sets the new state-of-the-art for Pascal VOC 2012 with (1/16, 1/8, 1/4, and 1/2 supervised regimes) and Cityscapes (1/16 supervised).