Goto

Collaborating Authors

 label propagation


Graph-based Semi-Supervised Learning via Maximum Discrimination

Katz, Nadav, Jaffe, Ariel

arXiv.org Machine Learning

Semi-supervised learning (SSL) addresses the critical challenge of training accurate models when labeled data is scarce but unlabeled data is abundant. Graph-based SSL (GSSL) has emerged as a popular framework that captures data structure through graph representations. Classic graph SSL methods, such as Label Propagation and Label Spreading, aim to compute low-dimensional representations where points with the same labels are close in representation space. Although often effective, these methods can be suboptimal on data with complex label distributions. In our work, we develop AUC-spec, a graph approach that computes a low-dimensional representation that maximizes class separation. We compute this representation by optimizing the Area Under the ROC Curve (AUC) as estimated via the labeled points. We provide a detailed analysis of our approach under a product-of-manifold model, and show that the required number of labeled points for AUC-spec is polynomial in the model parameters. Empirically, we show that AUC-spec balances class separation with graph smoothness. It demonstrates competitive results on synthetic and real-world datasets while maintaining computational efficiency comparable to the field's classic and state-of-the-art methods.



Video Diffusion Models Excel at Tracking Similar-Looking Objects Without Supervision

Zhang, Chenshuang, Zhang, Kang, Chung, Joon Son, Kweon, In So, Kim, Junmo, Mao, Chengzhi

arXiv.org Artificial Intelligence

Distinguishing visually similar objects by their motion remains a critical challenge in computer vision. Although supervised trackers show promise, contemporary self-supervised trackers struggle when visual cues become ambiguous, limiting their scalability and generalization without extensive labeled data. We find that pre-trained video diffusion models inherently learn motion representations suitable for tracking without task-specific training. This ability arises because their denoising process isolates motion in early, high-noise stages, distinct from later appearance refinement. Capitalizing on this discovery, our self-supervised tracker significantly improves performance in distinguishing visually similar objects, an underexplored failure point for existing methods. Our method achieves up to a 6-point improvement over recent self-supervised approaches on established benchmarks and our newly introduced tests focused on tracking visually similar items. Visualizations confirm that these diffusion-derived motion representations enable robust tracking of even identical objects across challenging viewpoint changes and deformations.


Few Shot Semi-Supervised Learning for Abnormal Stop Detection from Sparse GPS Trajectories

Sabir, Muhammad Ayub, Pang, Junbiao, Wu, Jiaqi, Ashraf, Fatima

arXiv.org Artificial Intelligence

Abnormal stop detection (ASD) in intercity coach transportation is critical for ensuring passenger safety, operational reliability, and regulatory compliance. However, two key challenges hinder ASD effectiveness: sparse GPS trajectories, which obscure short or unauthorized stops, and limited labeled data, which restricts supervised learning. Existing methods often assume dense sampling or regular movement patterns, limiting their applicability. To address data sparsity, we propose a Sparsity-Aware Segmentation (SAS) method that adaptively defines segment boundaries based on local spatial-temporal density. Building upon these segments, we introduce three domain-specific indicators to capture abnormal stop behaviors. To further mitigate the impact of sparsity, we develop Locally Temporal-Indicator Guided Adjustment (LTIGA), which smooths these indicators via local similarity graphs. To overcome label scarcity, we construct a spatial-temporal graph where each segment is a node with LTIGA-refined features. We apply label propagation to expand weak supervision across the graph, followed by a GCN to learn relational patterns. A final self-training module incorporates high-confidence pseudo-labels to iteratively improve predictions. Experiments on real-world coach data show an AUC of 0.854 and AP of 0.866 using only 10 labeled instances, outperforming prior methods. The code and dataset are publicly available at \href{https://github.com/pangjunbiao/Abnormal-Stop-Detection-SSL.git}


Transfer Learning in a Transductive Setting

Marcus Rohrbach, Sandra Ebert, Bernt Schiele

Neural Information Processing Systems

Category models for objects or activities typically rely on supervised learning requiring sufficiently large training sets. Transferring knowledge from known categories to novel classes with no or only a few labels is far less researched even though it is a common scenario. In this work, we extend transfer learning with semi-supervised learning to exploit unlabeled instances of (novel) categories with no or only a few labeled instances. Our proposed approach Propagated Semantic Transfer combines three techniques. First, we transfer information from known to novel categories by incorporating external knowledge, such as linguistic or expert-specified information, e.g., by a mid-level layer of semantic attributes.


3a835d3215755c435ef4fe9965a3f2a0-Reviews.html

Neural Information Processing Systems

First provide a summary of the paper, and then address the following criteria: Quality, clarity, originality and significance. This paper addresses the problem of assigning weights to edges in a graph for label propagation. The assumption is that the graph structure (and training labels) are provided. The proposed method is simple to implement and appears to perform well against standard approaches. While many additional experiments could be run, I feel the evaluation is adequate.


First of all, we wish to sincerely thank the anonymous reviewers for their time and efforts in reviewing our NeurIPS

Neural Information Processing Systems

In the revised version, we will make this clearer in the "Related Work" section. Figure 2 illustrates how the classification model (i.e. ( t 1) The parameter σ corresponds to the width of Gaussian kernel, which is fixed to be 1 in this paper (pp.3, footnote 1).


We would like to thank all reviewers for your helpful suggestions

Neural Information Processing Systems

We would like to thank all reviewers for your helpful suggestions. In what follows we answer your questions to help clarify the notations and experiments. "motivation not clear": G-SSL and specifically label propagation used to be an important semi-supervised learning We will try to include more real-world examples in the revised version. ": following the standard notations, we use The feature matrix is composed of labeled part and unlabeled part. "Datasets not explained": we will add more descriptions of datasets in the next version.



Manifold-based Similarity Adaptation for Label Propagation

Neural Information Processing Systems

Label propagation is one of the state-of-the-art methods for semi-supervised learning, which estimates labels by propagating label information through a graph. Label propagation assumes that data points (nodes) connected in a graph should have similar labels. Consequently, the label estimation heavily depends on edge weights in a graph which represent similarity of each node pair. We propose a method for a graph to capture the manifold structure of input features using edge weights parameterized by a similarity function. In this approach, edge weights represent both similarity and local reconstruction weight simultaneously, both being reasonable for label propagation. For further justification, we provide analytical considerations including an interpretation as a cross-validation of a propagation model in the feature space, and an error analysis based on a low dimensional manifold model. Experimental results demonstrated the effectiveness of our approach both in synthetic and real datasets.