Conditioning and Processing: Techniques to Improve Information-Theoretic Generalization Bounds
–Neural Information Processing Systems
Obtaining generalization bounds for learning algorithms is one of the main subjects studied in theoretical machine learning. In recent years, information-theoretic bounds on generalization have gained the attention of researchers. This approach provides an insight into learning algorithms by considering the mutual information between the model and the training set. In this paper, a probabilistic graphical representation of this approach is adopted and two general techniques to improve the bounds are introduced, namely conditioning and processing. In conditioning, a random variable in the graph is considered as given, while in processing a random variable is substituted with one of its children.
Neural Information Processing Systems
Oct-11-2024, 06:20:08 GMT
- Technology: