Goto

Collaborating Authors

 fixmatch



FlexMatch: Boosting Semi-Supervised Learning with Curriculum Pseudo Labeling

Neural Information Processing Systems

The recently proposed FixMatch achieved state-of-the-art results on most semi-supervised learning (SSL) benchmarks. However, like other modern SSL algorithms, FixMatch uses a pre-defined constant threshold for all classes to select unlabeled data that contribute to the training, thus failing to consider different learning status and learning difficulties of different classes. To address this issue, we propose Curriculum Pseudo Labeling (CPL), a curriculum learning approach to leverage unlabeled data according to the model's learning status. The core of CPL is to flexibly adjust thresholds for different classes at each time step to let pass informative unlabeled data and their pseudo labels. CPL does not introduce additional parameters or computations (forward or backward propagation).


FixMatch: Simplifying Semi-Supervised Learning with Consistency and Confidence

Neural Information Processing Systems

Semi-supervised learning (SSL) provides an effective means of leveraging unlabeled data to improve a model's performance. This domain has seen fast progress recently, at the cost of requiring more complex methods. In this paper we propose FixMatch, an algorithm that is a significant simplification of existing SSL methods. FixMatch first generates pseudo-labels using the model's predictions on weakly-augmented unlabeled images. For a given image, the pseudo-label is only retained if the model produces a high-confidence prediction. The model is then trained to predict the pseudo-label when fed a strongly-augmented version of the same image. Despite its simplicity, we show that FixMatch achieves state-of-the-art performance across a variety of standard semi-supervised learning benchmarks, including 94.93% accuracy on CIFAR-10 with 250 labels and 88.61% accuracy with 40 - just 4 labels per class. We carry out an extensive ablation study to tease apart the experimental factors that are most important to FixMatch's success.



CalibrateMix: Guided-Mixup Calibration of Image Semi-Supervised Models

Rahman, Mehrab Mustafy, Mohan, Jayanth, Sosea, Tiberiu, Caragea, Cornelia

arXiv.org Artificial Intelligence

Semi-supervised learning (SSL) has demonstrated high performance in image classification tasks by effectively utilizing both labeled and unlabeled data. However, existing SSL methods often suffer from poor calibration, with models yielding overconfident predictions that misrepresent actual prediction likelihoods. Recently, neural networks trained with {\tt mixup} that linearly interpolates random examples from the training set have shown better calibration in supervised settings. However, calibration of neural models remains under-explored in semi-supervised settings. Although effective in supervised model calibration, random mixup of pseudolabels in SSL presents challenges due to the overconfidence and unreliability of pseudolabels. In this work, we introduce CalibrateMix, a targeted mixup-based approach that aims to improve the calibration of SSL models while maintaining or even improving their classification accuracy. Our method leverages training dynamics of labeled and unlabeled samples to identify ``easy-to-learn'' and ``hard-to-learn'' samples, which in turn are utilized in a targeted mixup of easy and hard samples. Experimental results across several benchmark image datasets show that our method achieves lower expected calibration error (ECE) and superior accuracy compared to existing SSL approaches.



A Experimental Details

Neural Information Processing Systems

Unless otherwise noted, the same value is used for all experiments. In CIFAR10 and CIFAR100, 50 samples per each known class are used. We utilize the training split of ImageNet-30 for training and test one for evaluation. Testing was done for 3,000 test samples. Here, we show the ablation of SOCR for a model training with FixMatch loss.



Cost-Sensitive Self-Training for Optimizing Non-Decomposable Metrics Anonymous Author(s) Affiliation

Neural Information Processing Systems

Coverage etc. which can be expressed as a linear form of confusion matrix. We leave this as an open direction for further work. Also in this work we considered datasets where unlabeled data distribution doesn't significantly differ In this section, we show that minimization of weighted consistency regularizer Eq. In this section, we provide some examples for assumptions introduced in Sec. 3 and a proof of Theorem 5. We provide proof of these examples in Sec. In this subsection, we provide a proof of Theorem 5 assuming Lemma 10.