Reviews: Characterizing Bias in Classifiers using Generative Models

Neural Information Processing Systems 

The paper proposes a method to study certain types of biases in the data-generating mode, which could, for example, translate to discrimination and unfairness in the classification setting. The reviewers agree with the importance and relevance of the proposed framework. Personally, I found the whole narrative a bit surprising, or unusual, since there is not one unique problem of "bias", but multiple types of biases, which are largely acknowledged in the causal inference literature. In particular, if I understood the paper correctly, the authors are really discussing the mismatch between the proportion of units sampled to the study versus of the underlying population relative to certain features, which in the sciences is called (sampling) selection bias. In order to avoid readers to get confused, I would try to be more specific in the title and add a short discussion articulating the specific type of bias considered in the proposed work.