artstation
Prompt Pirates Need a Map: Stealing Seeds helps Stealing Prompts
Mächtle, Felix, Shetty, Ashwath, Sander, Jonas, Loose, Nils, Pirk, Sören, Eisenbarth, Thomas
Diffusion models have significantly advanced text-to-image generation, enabling the creation of highly realistic images conditioned on textual prompts and seeds. Given the considerable intellectual and economic value embedded in such prompts, prompt theft poses a critical security and privacy concern. In this paper, we investigate prompt-stealing attacks targeting diffusion models. We reveal that numerical optimization-based prompt recovery methods are fundamentally limited as they do not account for the initial random noise used during image generation. We identify and exploit a noise-generation vulnerability (CWE-339), prevalent in major image-generation frameworks, originating from PyTorch's restriction of seed values to a range of $2^{32}$ when generating the initial random noise on CPUs. Through a large-scale empirical analysis conducted on images shared via the popular platform CivitAI, we demonstrate that approximately 95% of these images' seed values can be effectively brute-forced in 140 minutes per seed using our seed-recovery tool, SeedSnitch. Leveraging the recovered seed, we propose PromptPirate, a genetic algorithm-based optimization method explicitly designed for prompt stealing. PromptPirate surpasses state-of-the-art methods, i.e., PromptStealer, P2HP, and CLIP-Interrogator, achieving an 8-11% improvement in LPIPS similarity. Furthermore, we introduce straightforward and effective countermeasures that render seed stealing, and thus optimization-based prompt stealing, ineffective. We have disclosed our findings responsibly and initiated coordinated mitigation efforts with the developers to address this critical vulnerability.
- North America > United States > Washington > King County > Seattle (0.04)
- North America > United States > Utah > Salt Lake County > Salt Lake City (0.04)
- North America > United States > Maryland > Baltimore (0.04)
- (8 more...)
- Media (1.00)
- Information Technology > Security & Privacy (1.00)
Dynamic Prompt Optimizing for Text-to-Image Generation
Mo, Wenyi, Zhang, Tianyu, Bai, Yalong, Su, Bing, Wen, Ji-Rong, Yang, Qing
Text-to-image generative models, specifically those based on diffusion models like Imagen and Stable Diffusion, have made substantial advancements. Recently, there has been a surge of interest in the delicate refinement of text prompts. Users assign weights or alter the injection time steps of certain words in the text prompts to improve the quality of generated images. However, the success of fine-control prompts depends on the accuracy of the text prompts and the careful selection of weights and time steps, which requires significant manual intervention. To address this, we introduce the \textbf{P}rompt \textbf{A}uto-\textbf{E}diting (PAE) method. Besides refining the original prompts for image generation, we further employ an online reinforcement learning strategy to explore the weights and injection time steps of each word, leading to the dynamic fine-control prompts. The reward function during training encourages the model to consider aesthetic score, semantic consistency, and user preferences. Experimental results demonstrate that our proposed method effectively improves the original prompts, generating visually more appealing images while maintaining semantic alignment. Code is available at https://github.com/Mowenyii/PAE.
- North America > United States > Minnesota > Hennepin County > Minneapolis (0.14)
- North America > Canada > Ontario > Toronto (0.04)
- Asia > Middle East > Republic of Türkiye > Batman Province > Batman (0.04)
- (7 more...)
Finetuning Text-to-Image Diffusion Models for Fairness
Shen, Xudong, Du, Chao, Pang, Tianyu, Lin, Min, Wong, Yongkang, Kankanhalli, Mohan
The rapid adoption of text-to-image diffusion models in society underscores an urgent need to address their biases. Without interventions, these biases could propagate a skewed worldview and restrict opportunities for minority groups. In this work, we frame fairness as a distributional alignment problem. Our solution consists of two main technical contributions: (1) a distributional alignment loss that steers specific characteristics of the generated images towards a user-defined target distribution, and (2) adjusted direct finetuning of diffusion model's sampling process (adjusted DFT), which leverages an adjusted gradient to directly optimize losses defined on the generated images. Empirically, our method markedly reduces gender, racial, and their intersectional biases for occupational prompts. Gender bias is significantly reduced even when finetuning just five soft tokens. Crucially, our method supports diverse perspectives of fairness beyond absolute equality, which is demonstrated by controlling age to a $75\%$ young and $25\%$ old distribution while simultaneously debiasing gender and race. Finally, our method is scalable: it can debias multiple concepts at once by simply including these prompts in the finetuning data. We share code and various fair diffusion model adaptors at https://sail-sg.github.io/finetune-fair-diffusion/.
- Asia > Singapore (0.04)
- North America > United States > New York (0.04)
- North America > United States > Maryland (0.04)
- (4 more...)
- Media (1.00)
- Leisure & Entertainment > Sports > Martial Arts (1.00)
- Health & Medicine > Therapeutic Area (0.93)
- Education > Educational Setting (0.93)
No Longer Trending on Artstation: Prompt Analysis of Generative AI Art
McCormack, Jon, Llano, Maria Teresa, Krol, Stephen James, Rajcic, Nina
Image generation using generative AI is rapidly becoming a major new source of visual media, with billions of AI generated images created using diffusion models such as Stable Diffusion and Midjourney over the last few years. In this paper we collect and analyse over 3 million prompts and the images they generate. Using natural language processing, topic analysis and visualisation methods we aim to understand collectively how people are using text prompts, the impact of these systems on artists, and more broadly on the visual cultures they promote. Our study shows that prompting focuses largely on surface aesthetics, reinforcing cultural norms, popular conventional representations and imagery. We also find that many users focus on popular topics (such as making colouring books, fantasy art, or Christmas cards), suggesting that the dominant use for the systems analysed is recreational rather than artistic.
- Oceania > Australia (0.04)
- North America > United States > Texas > Dallas County > Dallas (0.04)
- North America > United States > New York > New York County > New York City (0.04)
- (7 more...)
Studying Artist Sentiments around AI-generated Artwork
Ali, Safinah, Breazeal, Cynthia
Art created using generated Artificial Intelligence has taken the world by storm and generated excitement for many digital creators and technologists. However, the reception and reaction from artists have been mixed. Concerns about plagiarizing their artworks and styles for datasets and uncertainty around the future of digital art sparked movements in artist communities shunning the use of AI for generating art and protecting artists' rights. Collaborating with these tools for novel creative use cases also sparked hope from some creators. Artists are an integral stakeholder in the rapidly evolving digital creativity industry and understanding their concerns and hopes inform responsible development and use of creativity support tools. In this work, we study artists' sentiments about AI-generated art. We interviewed 7 artists and analyzed public posts from artists on social media platforms Reddit, Twitter and Artstation. We report artists' main concerns and hopes around AI-generated artwork, informing a way forward for inclusive development of these tools.
- North America > United States > Massachusetts (0.04)
- North America > United States > Virginia (0.04)
- North America > United States > Pennsylvania (0.04)
- (4 more...)
- Research Report (0.83)
- Personal > Interview (0.46)
BeautifulPrompt: Towards Automatic Prompt Engineering for Text-to-Image Synthesis
Cao, Tingfeng, Wang, Chengyu, Liu, Bingyan, Wu, Ziheng, Zhu, Jinhui, Huang, Jun
Recently, diffusion-based deep generative models (e.g., Stable Diffusion) have shown impressive results in text-to-image synthesis. However, current text-to-image models often require multiple passes of prompt engineering by humans in order to produce satisfactory results for real-world applications. We propose BeautifulPrompt, a deep generative model to produce high-quality prompts from very simple raw descriptions, which enables diffusion-based models to generate more beautiful images. In our work, we first fine-tuned the BeautifulPrompt model over low-quality and high-quality collecting prompt pairs. Then, to ensure that our generated prompts can generate more beautiful images, we further propose a Reinforcement Learning with Visual AI Feedback technique to fine-tune our model to maximize the reward values of the generated prompts, where the reward values are calculated based on the PickScore and the Aesthetic Scores. Our results demonstrate that learning from visual AI feedback promises the potential to improve the quality of generated prompts and images significantly. We further showcase the integration of BeautifulPrompt to a cloud-native AI platform to provide better text-to-image generation service in the cloud.
Prompt Stealing Attacks Against Text-to-Image Generation Models
Shen, Xinyue, Qu, Yiting, Backes, Michael, Zhang, Yang
Text-to-Image generation models have revolutionized the artwork design process and enabled anyone to create high-quality images by entering text descriptions called prompts. Creating a high-quality prompt that consists of a subject and several modifiers can be time-consuming and costly. In consequence, a trend of trading high-quality prompts on specialized marketplaces has emerged. In this paper, we propose a novel attack, namely prompt stealing attack, which aims to steal prompts from generated images by text-to-image generation models. Successful prompt stealing attacks direct violate the intellectual property and privacy of prompt engineers and also jeopardize the business model of prompt trading marketplaces. We first perform a large-scale analysis on a dataset collected by ourselves and show that a successful prompt stealing attack should consider a prompt's subject as well as its modifiers. We then propose the first learning-based prompt stealing attack, PromptStealer, and demonstrate its superiority over two baseline methods quantitatively and qualitatively. We also make some initial attempts to defend PromptStealer. In general, our study uncovers a new attack surface in the ecosystem created by the popular text-to-image generation models. We hope our results can help to mitigate the threat. To facilitate research in this field, we will share our dataset and code with the community.
- Europe > Italy > Calabria > Catanzaro Province > Catanzaro (0.04)
- Asia > Middle East > Jordan (0.04)
Mixture of Diffusers for scene composition and high resolution image generation
Diffusion methods have been proven to be very effective to generate images while conditioning on a text prompt. However, and although the quality of the generated images is unprecedented, these methods seem to struggle when trying to generate specific image compositions. In this paper we present Mixture of Diffusers, an algorithm that builds over existing diffusion models to provide a more detailed control over composition. By harmonizing several diffusion processes acting on different regions of a canvas, it allows generating larger images, where the location of each object and style is controlled by a separate diffusion process.
The Best Illustrations From ArtStation User's AI Protests
For the past week, ArtStation--the world's most popular portfolio site for professional (and amateur!) artists working in the entertainment business--has been rocked by protests from its users, after owners Epic Games refused to offer adequate protections against the growing threat of AI-generated imagery. For the first few days of that protest, most users simply pasted a clean, bold image by Alexander Nanitchkov, using repetition in numbers to have the site's front page looking like this: As the days have marched on, though, and ArtStation and Epic refuse to offer more suitable protections for the very artworks their site is designed for, artists have moved on and have decided to come up with pieces that are a bit more elaborate, and personal. I thought I'd highlight some of my favourites in this post. You'll find links to their passionate, creative and deeply human portfolios of each artist responsible in the names under each image. For Xbox Series X, Xbox Series S, Xbox One, & Windows devices The new Xbox Wireless Controller feels natural with its lightweight design, textured triggers, and hybrid D-pad.
Artists stage mass protest against AI-generated artwork on ArtStation
On Tuesday, members of the online community ArtStation began widely protesting AI-generated artwork by placing "No AI Art" images in their portfolios. By Wednesday, the protest images dominated ArtStation's trending page. The artists seek to criticize the presence of AI-generated work on ArtStation and to potentially disrupt future AI models trained using artwork found on the site. Early rumblings of the protest began on December 5 when Bulgarian artist Alexander Nanitchkov tweeted, "Current AI'art' is created on the backs of hundreds of thousands of artists and photographers who made billions of images and spend time, love and dedication to have their work soullessly stolen and used by selfish people for profit without the slightest concept of ethics." Nanitchkov also posted a stark logo featuring the letters "AI" in white uppercase behind the circular strike-through symbol.