Goto

Collaborating Authors

 multiple discriminator


MCL-GAN: GenerativeAdversarialNetworks withMultipleSpecializedDiscriminators

Neural Information Processing Systems

There are several GAN literature that adopts multiple discriminators [9-14]. Among them, GMAN [10] is closely related to our approach in the sense that it utilizes an ensemble predictionofdiscriminators.



MCL-GAN: Generative Adversarial Networks with Multiple Specialized Discriminators

Neural Information Processing Systems

We propose a framework of generative adversarial networks with multiple discriminators, which collaborate to represent a real dataset more effectively. Our approach facilitates learning a generator consistent with the underlying data distribution based on real images and thus mitigates the chronic mode collapse problem. From the inspiration of multiple choice learning, we guide each discriminator to have expertise in a subset of the entire data and allow the generator to find reasonable correspondences between the latent and real data spaces automatically without extra supervision for training examples. Despite the use of multiple discriminators, the backbone networks are shared across the discriminators and the increase in training cost is marginal. We demonstrate the effectiveness of our algorithm using multiple evaluation metrics in the standard datasets for diverse tasks.


Figure 1 Additional analysis

Neural Information Processing Systems

We answer each question below. Our strong empirical results backs our design choice. As noted in the main paper (see section 3.3 Framework Design), we learn target-to-source alignment Thus, the '+trad' must be replaced with /check. The clear improvement demonstrates its efficacy. The followings are the results: [Ours 32.0 / ADVENT 29.1 / Adaptseg R1,R4: end-to-end training The end-to-end training causes the model to diverge.



Collective Learning Mechanism based Optimal Transport Generative Adversarial Network for Non-parallel Voice Conversion

Dhar, Sandipan, Akhter, Md. Tousin, Jana, Nanda Dulal, Das, Swagatam

arXiv.org Artificial Intelligence

Collective Learning Mechanism based Optimal Transport Generative Adversarial Network for Non-parallel V oice Conversion Sandipan Dhar 1, Md. Email: sandipandhartsk03@gmail.com, tousin@cse.iitb.ac.in, ndjana.cse@nitdgp.ac.in, swagatam.das@isical.ac.in Abstract --After demonstrating significant success in image synthesis, Generative Adversarial Network (GAN) models have likewise made significant progress in the field of speech synthesis, leveraging their capacity to adapt the precise distribution of target data through adversarial learning processes. Notably, in the realm of State-Of-The-Art (SOT A) GAN-based V oice Conversion (VC) models, there exists a substantial disparity in naturalness between real and GAN-generated speech samples. Furthermore, while many GAN models currently operate on a single generator discriminator learning approach, optimizing target data distribution is more effectively achievable through a single generator multi-discriminator learning scheme. Hence, this study introduces a novel GAN model named Collective Learning Mechanism-based Optimal Transport GAN (CLOT - GAN) model, incorporating multiple discriminators, including the Deep Convolutional Neural Network (DCNN) model, Vision Transformer (ViT), and conformer . The objective of integrating various discriminators lies in their ability to comprehend the formant distribution of mel-spectrograms, facilitated by a collective learning mechanism. Simultaneously, the inclusion of Optimal Transport (OT) loss aims to precisely bridge the gap between the source and target data distribution, employing the principles of OT theory.


MCL-GAN: Generative Adversarial Networks with Multiple Specialized Discriminators

Neural Information Processing Systems

We propose a framework of generative adversarial networks with multiple discriminators, which collaborate to represent a real dataset more effectively. Our approach facilitates learning a generator consistent with the underlying data distribution based on real images and thus mitigates the chronic mode collapse problem. From the inspiration of multiple choice learning, we guide each discriminator to have expertise in a subset of the entire data and allow the generator to find reasonable correspondences between the latent and real data spaces automatically without extra supervision for training examples. Despite the use of multiple discriminators, the backbone networks are shared across the discriminators and the increase in training cost is marginal. We demonstrate the effectiveness of our algorithm using multiple evaluation metrics in the standard datasets for diverse tasks.


Adversarial Permutation Invariant Training for Universal Sound Separation

Postolache, Emilian, Pons, Jordi, Pascual, Santiago, Serrà, Joan

arXiv.org Artificial Intelligence

Universal sound separation consists of separating mixes with arbitrary sounds of different types, and permutation invariant training (PIT) is used to train source agnostic models that do so. In this work, we complement PIT with adversarial losses but find it challenging with the standard formulation used in speech source separation. We overcome this challenge with a novel I-replacement context-based adversarial loss, and by training with multiple discriminators. Our experiments show that by simply improving the loss (keeping the same model and dataset) we obtain a non-negligible improvement of 1.4 dB SI-SNRi in the reverberant FUSS dataset. We also find adversarial PIT to be effective at reducing spectral holes, ubiquitous in mask-based separation models, which highlights the potential relevance of adversarial losses for source separation.


Multi-Adversarial Variational Autoencoder Networks

Imran, Abdullah-Al-Zubaer, Terzopoulos, Demetri

arXiv.org Machine Learning

The unsupervised training of GANs and VAEs has enabled them to generate realistic images mimicking real-world distributions and perform image-based unsupervised clustering or semi-supervised classification. Combining the power of these two generative models, we introduce Multi-Adversarial Variational autoEncoder Networks (MAVENs), a novel network architecture that incorporates an ensemble of discriminators in a VAE-GAN network, with simultaneous adversarial learning and variational inference. We apply MAVENs to the generation of synthetic images and propose a new distribution measure to quantify the quality of the generated images. Our experimental results using datasets from the computer vision and medical imaging domains---Street View House Numbers, CIFAR-10, and Chest X-Ray datasets---demonstrate competitive performance against state-of-the-art semi-supervised models both in image generation and classification tasks.


Domain Partitioning Network

Csaba, Botos, Boukhayma, Adnane, Kulharia, Viveka, Horváth, András, Torr, Philip H. S.

arXiv.org Machine Learning

Standard adversarial training involves two agents, namely a generator and a discriminator, playing a mini-max game. However, even if the players converge to an equilibrium, the generator may only recover a part of the target data distribution, in a situation commonly referred to as mode collapse. In this work, we present the Domain Partitioning Network (DoPaNet), a new approach to deal with mode collapse in generative adversarial learning. We employ multiple discriminators, each encouraging the generator to cover a different part of the target distribution. To ensure these parts do not overlap and collapse into the same mode, we add a classifier as a third agent in the game. The classifier decides which discriminator the generator is trained against for each sample. Through experiments on toy examples and real images, we show the merits of DoPaNet in covering the real distribution and its superiority with respect to the competing methods. Besides, we also show that we can control the modes from which samples are generated using DoPaNet.