Machine Learning
2) Showed how CANs and the SL follow naturally from introducing constraints into the discriminator of GANs; 3)
We thank all the reviewers for their careful reviews. None of these follow trivially from Xu et al. In molecule generation, validity is encouraged using the reward-based approach of MolGANs, not the SL. As a result diversity increases (91% to 98.98%) and so does uniqueness (2.4 to 2.5). We will clarify this in the discussion.
A Bayesian Theory of Conformity in Collective Decision Making
Koosha Khalvati, Saghar Mirbagheri, Seongmin A. Park, Jean-Claude Dreher, Rajesh PN Rao
In collective decision making, members of a group need to coordinate their actions in order to achieve a desirable outcome. When there is no direct communication between group members, one must decide based on inferring others' intentions from their actions. The inference of others' intentions is called "theory of mind" and can involve different levels of reasoning, from a single inference of a hidden variable to considering others partially or fully optimal and reasoning about their actions conditioned on one's own actions (levels of "theory of mind"). In this paper, we present a new Bayesian theory of collective decision making based on a simple yet most commonly observed behavior: conformity. We show that such a Bayesian framework allows one to achieve any level of theory of mind in collective decision making. The viability of our framework is demonstrated on two different experiments, a consensus task with 120 subjects and a volunteer's dilemma task with 29 subjects, each with multiple conditions.
Black-Box Optimization with Local Generative Surrogates Supplementary Material A Surrogates Implementation Details
A.1 GAN Implementation For the training of the GANs we have used conditional generative network, with three hidden layers of size 100 and conditional discriminative network with two hidden layers of size 100. For all the hidden layers except the last one we have used tanh activation. For the last hidden layer leaky_relu was used. The conditioning is performed via concatenating the input noise z with input parameters ฯ. The learning rate and batch size is set to 0.0008 and 512 correspondingly.
Black-Box Optimization with Local Generative Surrogates Department of Physics National Research University Imperial College London Higher School of Economics United Kingdom
We propose a novel method for gradient-based optimization of black-box simulators using differentiable local surrogate models. In fields such as physics and engineering, many processes are modeled with non-differentiable simulators with intractable likelihoods. Optimization of these forward models is particularly challenging, especially when the simulator is stochastic. To address such cases, we introduce the use of deep generative models to iteratively approximate the simulator in local neighborhoods of the parameter space. We demonstrate that these local surrogates can be used to approximate the gradient of the simulator, and thus enable gradient-based optimization of simulator parameters. In cases where the dependence of the simulator on the parameter space is constrained to a low dimensional submanifold, we observe that our method attains minima faster than baseline methods, including Bayesian optimization, numerical optimization, and approaches using score function gradient estimators.
a878dbebc902328b41dbf02aa87abb58-AuthorFeedback.pdf
We thank the reviewers for their thoughtful comments! We agree that we might have been under-selling scalability of our method (R3) w.r.t. We are glad the text is clearly written (R2, R4) with a broad literature review (R1, R2). We note this on lines [146-149]. Similarly, for visual clarity, error bands were not included for all baselines in Figure 1.
HOI-Swap: Swapping Objects in Videos with Hand-Object Interaction Awareness Mi Luo 1 Changan Chen 1 Kristen Grauman
We study the problem of precisely swapping objects in videos, with a focus on those interacted with by hands, given one user-provided reference object image. Despite the great advancements that diffusion models have made in video editing recently, these models often fall short in handling the intricacies of hand-object interactions (HOI), failing to produce realistic edits--especially when object swapping results in object shape or functionality changes. To bridge this gap, we present HOI-Swap, a novel diffusion-based video editing framework trained in a self-supervised manner. Designed in two stages, the first stage focuses on object swapping in a single frame with HOI awareness; the model learns to adjust the interaction patterns, such as the hand grasp, based on changes in the object's properties. The second stage extends the single-frame edit across the entire sequence; we achieve controllable motion alignment with the original video by: (1) warping a new sequence from the stage-I edited frame based on sampled motion points and (2) conditioning video generation on the warped sequence. Comprehensive qualitative and quantitative evaluations demonstrate that HOI-Swap significantly outperforms existing methods, delivering high-quality video edits with realistic HOIs.
Efficient Communication in Multi-Agent Reinforcement Learning via Variance Based Control
Sai Qian Zhang, Qi Zhang, Jieyu Lin
Multi-agent reinforcement learning (MARL) has recently received considerable attention due to its applicability to a wide range of real-world applications. However, achieving efficient communication among agents has always been an overarching problem in MARL. In this work, we propose Variance Based Control (VBC), a simple yet efficient technique to improve communication efficiency in MARL. By limiting the variance of the exchanged messages between agents during the training phase, the noisy component in the messages can be eliminated effectively, while the useful part can be preserved and utilized by the agents for better performance. Our evaluation using multiple MARL benchmarks indicates that our method achieves 2 10 lower in communication overhead than state-of-the-art MARL algorithms, while allowing agents to achieve better overall performance.