Unsupervised or Indirectly Supervised Learning
Uncoupled Regression from Pairwise Comparison Data Junya Honda
Uncoupled regression is the problem to learn a model from unlabeled data and the set of target values while the correspondence between them is unknown. Such a situation arises in predicting anonymized targets that involve sensitive information, e.g., one's annual income. Since existing methods for uncoupled regression often require strong assumptions on the true target function, and thus, their range of applications is limited, we introduce a novel framework that does not require such assumptions in this paper. Our key idea is to utilize pairwise comparison data, which consists of pairs of unlabeled data that we know which one has a larger target value. Such pairwise comparison data is easy to collect, as typically discussed in the learning-to-rank scenario, and does not break the anonymity of data. We propose two practical methods for uncoupled regression from pairwise comparison data and show that the learned regression model converges to the optimal model with the optimal parametric convergence rate when the target variable distributes uniformly. Moreover, we empirically show that for linear models the proposed methods are comparable to ordinary supervised regression with labeled data.
Reviews: MelGAN: Generative Adversarial Networks for Conditional Waveform Synthesis
Quality: This paper suffers from a few critical issues. Clarity: The experiment setting ups can be described with more details. Sec 3.2 and 3.4 is missing important information such as the datasets used for conducting the experiments. Significance: Although the quality of the proposed model remains unclear because of the previously mentioned critical issues, it's a significant work because it's the first GAN-based model for spectrogram-to-waveform conversion which seems to be working at some degree. It's significantly over-claimed: 1) claiming state-of-the-art for spectrogram-to-waveform conversion (line 6) with MOS 3.09 is surprising; many previous works are at a much higher level (e.g.
Reviews: MelGAN: Generative Adversarial Networks for Conditional Waveform Synthesis
The paper describes a successful approach for non-autoregressive spectrogram inversion based on Generative Adversarial Networks. The reviewers noted that even though the results are not at the level of state-of-the-art, the paper addresses a difficult and timely problem, with a convincing experimental validation and ablation study. The rebuttal addressed the main concerns of the reviewers; the authors should nonetheless make sure to address other concerns in the camera-ready version.
Reviews: A Unified Framework for Data Poisoning Attack to Graph-based Semi-supervised Learning
As the topic of the this work is very interesting, I'm not frustrated by the presentation, which has a large space for improvement. First, the motivation of the work is not so clear. Why is that an important concern to be able to poison a G-SSL model? The three references [1-3] provided seem at least a decade ago. Are they widely applied in modern security/privacy sensitive applications?
Reviews: A Unified Framework for Data Poisoning Attack to Graph-based Semi-supervised Learning
I recommend acceptance of this article subject to some corrections. The novelty of considering data poisoning in GSSL can be interesting for the NeurIPS community. The results presented here are founded by theoretical results and experiments are conducted. It opens new lines of research and can trigger new results. Because the paper has several weak points, we insist on the importance to handle the following comments. However, the overall structure of the presentation is acceptable.
Review for NeurIPS paper: Unsupervised Data Augmentation for Consistency Training
Additional Feedback: The main comment I have regarding the paper is that the authors do not provide adequate justification as to why the advanced data augmentation work compared to the simple ones and when to apply them. This same intuition can be applied for other semi-supervised methods like nearest neighbor and label propagation. These methods will assign the same labels to unlabeled data examples within its component in a graph. This is intuitive but does not explain why the noise from the advanced data augmentation methods are better for semi-supervised learning or provide guarantees for when they work. I acknowledge that I read the rebuttal and thank the authors for providing explanations to the questions and concerns I had.
Review for NeurIPS paper: Graph Contrastive Learning with Augmentations
Summary and Contributions: This paper proposes a contrastive learning algorithm to learn graph representations in an unsupervised manner. It is an extension of SimCLR [1] applied to learn graph representations that can be used for different graph classification tasks, either in semi-supervised learning, unsupervised learning or transfer learning scenarios. To do so, the authors propose several graph augmentation techniques that are needed for the contrastive learning algorithm, and analyse its effects on different types of datasets. The four different types of data augmentation techniques explored in the paper are: node dropping, edge perturbation, attribute masking and subgraph. In their empirical study, the authors explore the effect of these data augmentation techniques in different kinds of graph structure data like social networks and biochemical molecules, showing that different techniques work better on each domain, depending on the nature of the structure represented by the graph. This pre-training technique shows promising results across different datasets and tasks.
Reviews: MarginGAN: Adversarial Training in Semi-Supervised Learning
The main contribution of this paper is in setting up a 3 player game for semi-supervised learning where the generator tries to maximize the margin of the examples it generates in competition with a classifier the traditional GAN approach of fooling a discriminator. This idea is novel to my knowledge. One small reservation I have with this method is that as the quality of the GAN and generated images increases the margin maximization for the classifier for generated examples becomes counter productive (as acknowledged by the authors) which requires careful early stopping. But this is standard practice with GANs and it should not be held against this paper. The paper is generally of high quality and significance but these could be improved by a broader treatment of related works.
MarginGAN: Adversarial Training in Semi-Supervised Learning
A Margin Generative Adversarial Network (MarginGAN) is proposed for semisupervised learning problems. Like Triple-GAN, the proposed MarginGAN consists of three components--a generator, a discriminator and a classifier, among which two forms of adversarial training arise. The discriminator is trained as usual to distinguish real examples from fake examples produced by the generator. The new feature is that the classifier attempts to increase the margin of real examples and to decrease the margin of fake examples. On the contrary, the purpose of the generator is yielding realistic and large-margin examples in order to fool the discriminator and the classifier simultaneously. Pseudo labels are used for generated and unlabeled examples in training. Our method is motivated by the success of large-margin classifiers and the recent viewpoint that good semi-supervised learning requires a "bad" GAN. Experiments on benchmark datasets testify that MarginGAN is orthogonal to several state-of-the-art methods, offering improved error rates and shorter training time as well.
Reviews: MarginGAN: Adversarial Training in Semi-Supervised Learning
The paper formulates semi-supervised learning as a 3 player game among a generator, a classifier, and a discriminator. The generator and discriminator compete to train realistic examples, as in usual GANs, and the key new idea is that the classifier tries to maximize the margin of real examples and minimize the margin of fake examples. The method both improves predictive performance and greatly reduces training time. The reviewers agree that it is a significant contribution.