PAC Generalization Bounds for Co-training
Dasgupta, Sanjoy, Littman, Michael L., McAllester, David A.
–Neural Information Processing Systems
The rule-based bootstrapping introduced by Y arowsky, and its co-training variant by Blum and Mitchell, have met with considerable empirical success. Earlier work on the theory of co-training has been only loosely related to empirically useful co-training algorithms. Here we give a new P ACstyle bound on generalization error which justifies both the use of confidences -- partial rules and partial labeling of the unlabeled data -- and the use of an agreement-based objective function as suggested by Collins and Singer. Our bounds apply to the multiclass case, i.e., where instances are to be assigned one of