Semi-Supervised Sparse Gaussian Classification: Provable Benefits of Unlabeled Data
–Neural Information Processing Systems
The premise of semi-supervised learning (SSL) is that combining labeled and unlabeled data yields significantly more accurate models. Despite empirical successes, the theoretical understanding of SSL is still far from complete. In this work, we study SSL for high dimensional sparse Gaussian classification. To construct an accurate classifier a key task is feature selection, detecting the few variables that separate the two classes. For this SSL setting, we analyze information theoretic lower bounds for accurate feature selection as well as computational lower bounds, assuming the low-degree likelihood hardness conjecture.
Neural Information Processing Systems
Mar-19-2025, 01:55:07 GMT