Review for NeurIPS paper: Multi-Stage Influence Function

Neural Information Processing Systems 

My main concern is that the authors did not compare the proposed method with any latent baselines. For example, we can also use to the uncertainty (the loss value or the entropy of the predicted distribution) of the model as an indicator to identify those problematic examples in the pre-training data. As the proposed method in this paper does not show very impressive results in the experiments (the Pearson's correlation is only 0.4 0.6 in Figure 1), it may not outperform this simple baseline. "At the pretraining stage, we train the models with examples from two classes ("bird" vs. "frog") for CIFAR-10 and four classes (0, 1, 2, and 3) for MNIST". The transfer tasks in these settings may be too easy.