Goto

Collaborating Authors

 style parameter


Complex Instruction Following with Diverse Style Policies in Football Games

Sun, Chenglu, Shen, Shuo, Hu, Haonan, Zhou, Wei, Chen, Chen

arXiv.org Artificial Intelligence

Despite advancements in language-controlled reinforcement learning (LC-RL) for basic domains and straightforward commands (e.g., object manipulation and navigation), effectively extending LC-RL to comprehend and execute high-level or abstract instructions in complex, multi-agent environments, such as football games, remains a significant challenge. To address this gap, we introduce Language-Controlled Diverse Style Policies (LCDSP), a novel LC-RL paradigm specifically designed for complex scenarios. LCDSP comprises two key components: a Diverse Style Training (DST) method and a Style Interpreter (SI). The DST method efficiently trains a single policy capable of exhibiting a wide range of diverse behaviors by modulating agent actions through style parameters (SP). The SI is designed to accurately and rapidly translate high-level language instructions into these corresponding SP . Through extensive experiments in a complex 5v5 football environment, we demonstrate that LCDSP effectively comprehends abstract tactical instructions and accurately executes the desired diverse behavioral styles, showcasing its potential for complex, real-world applications.


Appendix of Memory with No Forgetting

Neural Information Processing Systems

Figure 7: (a) The generator architecture adopted in this paper. It inherits the architecture from GP-GAN which is the same as the green/frozen part of our GAN memory (see Figure 1(a)). Given a target task/data ( e.g., Flowers, Cathedrals, or Cats), all the parameters are trainable and fine-tuned to fit the target data. FC/Conv layer, one need to apply it to all layers in real implementation. "None" means no modulation from target data is applied, we only use the modulation from the Only" means that we replace the Only" means that we replace the Only" means that we replace the b "All" means all style parameters from target data "All" is obtained via a similar way to that of Figure 2(a); " means using the style parameters from target data without "None" and "All" are obtained via a similar way to that of Figure 2(a); "FC" is obtained by applying a newly designed style parameters which copies the style parameters " is obtained by copying the designed style parameters under the "FC" setting and replacing " is obtained by copying the designed style parameters under the " "Our" is our GAN memory; "NoNorm" is a modified version of our GAN memory which removes the normalization on "NoBias" is a modified version of our GAN memory which removes the bias term Here we discuss the detailed techniques for interpolation among different generative processes with our GAN memory and show more examples in Figure 8, 9, 10, and 11.



Appendix of Memory with No Forgetting

Neural Information Processing Systems

Figure 7: (a) The generator architecture adopted in this paper. It inherits the architecture from GP-GAN which is the same as the green/frozen part of our GAN memory (see Figure 1(a)). Given a target task/data ( e.g., Flowers, Cathedrals, or Cats), all the parameters are trainable and fine-tuned to fit the target data. FC/Conv layer, one need to apply it to all layers in real implementation. "None" means no modulation from target data is applied, we only use the modulation from the Only" means that we replace the Only" means that we replace the Only" means that we replace the b "All" means all style parameters from target data "All" is obtained via a similar way to that of Figure 2(a); " means using the style parameters from target data without "None" and "All" are obtained via a similar way to that of Figure 2(a); "FC" is obtained by applying a newly designed style parameters which copies the style parameters " is obtained by copying the designed style parameters under the "FC" setting and replacing " is obtained by copying the designed style parameters under the " "Our" is our GAN memory; "NoNorm" is a modified version of our GAN memory which removes the normalization on "NoBias" is a modified version of our GAN memory which removes the bias term Here we discuss the detailed techniques for interpolation among different generative processes with our GAN memory and show more examples in Figure 8, 9, 10, and 11.



OpenVoice: Versatile Instant Voice Cloning

Qin, Zengyi, Zhao, Wenliang, Yu, Xumin, Sun, Xin

arXiv.org Artificial Intelligence

We introduce OpenVoice, a versatile voice cloning approach that requires only a short audio clip from the reference speaker to replicate their voice and generate speech in multiple languages. OpenVoice represents a significant advancement in addressing the following open challenges in the field: 1) Flexible Voice Style Control. OpenVoice enables granular control over voice styles, including emotion, accent, rhythm, pauses, and intonation, in addition to replicating the tone color of the reference speaker. The voice styles are not directly copied from and constrained by the style of the reference speaker. Previous approaches lacked the ability to flexibly manipulate voice styles after cloning. 2) Zero-Shot Cross-Lingual Voice Cloning. OpenVoice achieves zero-shot cross-lingual voice cloning for languages not included in the massive-speaker training set. Unlike previous approaches, which typically require extensive massive-speaker multi-lingual (MSML) dataset for all languages, OpenVoice can clone voices into a new language without any massive-speaker training data for that language. OpenVoice is also computationally efficient, costing tens of times less than commercially available APIs that offer even inferior performance. To foster further research in the field, we have made the source code and trained model publicly accessible. We also provide qualitative results in our demo website. Prior to its public release, our internal version of OpenVoice was used tens of millions of times by users worldwide between May and October 2023, serving as the backend of MyShell.


GANzilla: User-Driven Direction Discovery in Generative Adversarial Networks

Evirgen, Noyan, Chen, Xiang 'Anthony'

arXiv.org Artificial Intelligence

Generative Adversarial Network (GAN) is widely adopted in numerous application areas, such as data preprocessing, image editing, and creativity support. However, GAN's 'black box' nature prevents non-expert users from controlling what data a model generates, spawning a plethora of prior work that focused on algorithm-driven approaches to extract editing directions to control GAN. Complementarily, we propose a GANzilla: a user-driven tool that empowers a user with the classic scatter/gather technique to iteratively discover directions to meet their editing goals. In a study with 12 participants, GANzilla users were able to discover directions that (i) edited images to match provided examples (closed-ended tasks) and that (ii) met a high-level goal, e.g., making the face happier, while showing diversity across individuals (open-ended tasks).


Generating, With Style: The Mechanics Behind NVIDIA's Highly Realistic GAN Images

#artificialintelligence

Although the z vector is just sampled randomly, our ultimate goal is to create a mapping between the distribution of images and our reference distribution Z, such that each vector in z corresponds to a plausibly real image. As a result, despite being meaningless at first, each particular z ends up corresponding to and encoding properties of the image that it will produce. In their simplest form, transposed convolutions work by learning a filter matrix (for example, 3x3), and multiplying that by the value at each pixel to expand its information outward spatially. Each of the single "pixels" in a 4x4 representation influences the values in a 3x3 patch of output; these patches overlap and sum to create the final "blown out" representation. The visual above, while good for building simplified intuition, is a little misleading, since it makes it look like the values of the enlarged patch have to be spun out of a single piece of information from a single pixel.