Goto

Collaborating Authors

 glaze


Interpreting Structured Perturbations in Image Protection Methods for Diffusion Models

Martin, Michael R., Chan, Garrick, Ma, Kwan-Liu

arXiv.org Artificial Intelligence

Recent image protection mechanisms such as Glaze and Nightshade introduce imperceptible, adversarially designed perturbations intended to disrupt downstream text-to-image generative models. While their empirical effectiveness has been demonstrated, the internal structure, detectability, and representational behavior of these perturbations remain poorly understood. In this study, we demonstrated a systematic explainable AI analysis of image protection perturbations using a unified framework that integrates white-box feature-space inspection and black-box signal-level probing. Through latent-space clustering, feature-channel activation analysis, occlusion-based spatial sensitivity mapping, and frequency-domain spectral characterization, we revealed that modern protection mechanisms operate as structured, low-entropy perturbations that remain tightly coupled to underlying image content across representational, spatial, and spectral domains in all evaluated cases. We showed that protected images preserve content-driven feature organization with protection-specific substructure rather than inducing global representational drift. Detectability is governed by interacting effects of perturbation entropy, spatial deployment, and frequency alignment as revealed through combined synthetic and spectral analyses, with sequential protection amplifying detectable structure rather than suppressing it. Frequency-domain analysis further demonstrated that Glaze and Nightshade redistribute energy along dominant image-aligned frequency axes rather than introducing spectrally diffuse noise. These results suggested that contemporary image protection operates through structured feature-level deformation rather than semantic dislocation, providing mechanistic insight into why protection signals remain visually subtle yet consistently detectable. This work advances the interpretability of adversarial image protection and informs the design of future defenses and detection strategies for generative AI systems.




This tool strips away anti-AI protections from digital art

MIT Technology Review

To be clear, the researchers behind LightShed aren't trying to steal artists' work. They just don't want people to get a false sense of security. "You will not be sure if companies have methods to delete these poisons but will never tell you," says Hanna Foerster, a PhD student at the University of Cambridge and the lead author of a paper on the work. And if they do, it may be too late to fix the problem. AI models work, in part, by implicitly creating boundaries between what they perceive as different categories of images.


Four ways to protect your art from AI

MIT Technology Review

Artists and writers have launched several lawsuits against AI companies, arguing that their work has been scraped into databases for training AI models without consent or compensation. Tech companies have responded that anything on the public internet falls under fair use. But it will be years until we have a legal resolution to the problem. Unfortunately, there is little you can do if your work has been scraped into a data set and used in a model that is already out there. You can, however, take steps to prevent your work from being used in the future.


GLEAN: Generative Learning for Eliminating Adversarial Noise

Kim, Justin Lyu, Woo, Kyoungwan

arXiv.org Artificial Intelligence

In the age of powerful diffusion models such as DALL-E and Stable Diffusion, many in the digital art community have suffered style mimicry attacks due to fine-tuning these models on their works. The ability to mimic an artist's style via text-to-image diffusion models raises serious ethical issues, especially without explicit consent. Glaze, a tool that applies various ranges of perturbations to digital art, has shown significant success in preventing style mimicry attacks, at the cost of artifacts ranging from imperceptible noise to severe quality degradation. The release of Glaze has sparked further discussions regarding the effectiveness of similar protection methods. In this paper, we propose GLEAN- applying I2I generative networks to strip perturbations from Glazed images, evaluating the performance of style mimicry attacks before and after GLEAN on the results of Glaze. GLEAN aims to support and enhance Glaze by highlighting its limitations and encouraging further development.


2024 Innovator of the Year: Shawn Shan builds tools to help artists fight back against exploitative AI

MIT Technology Review

Now artists are fighting back. And some of the most powerful tools they have were built by Shawn Shan, 26, a PhD student in computer science at the University of Chicago (and MIT Technology Review's 2024 Innovator of the Year). Shan got his start in AI security and privacy as an undergraduate there and participated in a project that built Fawkes, a tool to protect faces from facial recognition technology. But it was conversations with artists who had been hurt by the generative AI boom that propelled him into the middle of one of the biggest fights in the field. Soon after learning about the impact on artists, Shan and his advisors Ben Zhao (who made our Innovators Under 35 list in 2006) and Heather Zheng (who was on the 2005 list) decided to build a tool to help. They gathered input from more than a thousand artists to learn what they needed and how they would use any protective technology.


Why artists are becoming less scared of AI

MIT Technology Review

Researchers from Google DeepMind asked 20 professional comedians to use popular AI language models to write jokes and comedy performances. The comedians said that the tools were useful in helping them produce an initial "vomit draft" that they could iterate on, and helped them structure their routines. But the AI was not able to produce anything that was original, stimulating, or, crucially, funny. My colleague Rhiannon Williams has the full story. As Tuhin Chakrabarty, a computer science researcher at Columbia University who specializes in AI and creativity, told Rhiannon, humor often relies on being surprising and incongruous.


Worried About AI Killing Art? This App Offers a Refuge--If Its Founder Can Keep the Lights On

WIRED

"I was about to go to bed and then realized we had this interview," Jingna Zhang tells me. It's 9 am in Seattle, where she's currently living. The photographer and art director has been pulling all-nighters trying to keep up with demand for her social platform for artists, Cara, which recently exploded in popularity in response to widespread opposition to Meta's policies around art and artificial intelligence. More users has led to an onslaught of complications, including a hefty 96,000 bill from the social network's cloud storage provider, as well as service outages. Cara began as a side project, but its newfound prominence means that Zhang is now an accidental startup founder, keeping the hours to match.


Organic or Diffused: Can We Distinguish Human Art from AI-generated Images?

Ha, Anna Yoo Jeong, Passananti, Josephine, Bhaskar, Ronik, Shan, Shawn, Southen, Reid, Zheng, Haitao, Zhao, Ben Y.

arXiv.org Artificial Intelligence

The advent of generative AI images has completely disrupted the art world. Distinguishing AI generated images from human art is a challenging problem whose impact is growing over time. A failure to address this problem allows bad actors to defraud individuals paying a premium for human art and companies whose stated policies forbid AI imagery. It is also critical for content owners to establish copyright, and for model trainers interested in curating training data in order to avoid potential model collapse. There are several different approaches to distinguishing human art from AI images, including classifiers trained by supervised learning, research tools targeting diffusion models, and identification by professional artists using their knowledge of artistic techniques. In this paper, we seek to understand how well these approaches can perform against today's modern generative models in both benign and adversarial settings. We curate real human art across 7 styles, generate matching images from 5 generative models, and apply 8 detectors (5 automated detectors and 3 different human groups including 180 crowdworkers, 4000+ professional artists, and 13 expert artists experienced at detecting AI). Both Hive and expert artists do very well, but make mistakes in different ways (Hive is weaker against adversarial perturbations while Expert artists produce higher false positives). We believe these weaknesses will remain as models continue to evolve, and use our data to demonstrate why a combined team of human and automated detectors provides the best combination of accuracy and robustness.