Goto

Collaborating Authors

 interpretability


A Appendix

Neural Information Processing Systems

A.1 Prototype-based Graph Information Bottleneck - Eq. 4 From Eq. 3, the GIB objective is: min We perform ablation studies to examine the effectiveness of our model (i.e., PGIB and PGIB In Figure 7, the " with all " setting represents our final model that includes all the components. We conduct experiments on graph classification using different readout functions for PGIB. We illustrate the reasoning process on two datasets, i.e., MUT AG and BA2Motif, in Figure 8. PGIB Then, PGIB computes the "points contributed" to predicting each class by multiplying the similarity We have conducted additional qualitative analysis. It is crucial that the prototypes not only contain key structural information from the input graph but also ensure a certain level of diversity since each class is represented by multiple prototypes. Its goal is to make the masked subgraph's prediction as close as possible to the original graph, which helps to detect substructures significant




Feature Learning for Interpretable, Performant Decision Trees

Neural Information Processing Systems

Points were sampled uniformly in the bands denoted by dashed lines. We posit that these barriers are due, at least in part, to the sensitivity of decision trees to transformations of the input resulting from greedy construction and simple decision rules. Of these, key limitation is the latter; even if we replace greedy construction with a perfect tree learner, simple distributions can nonetheless require an arbitrarily large axis-aligned tree to fit.