Erasing Undesirable Concepts in Diffusion Models with Adversarial Preservation
–Neural Information Processing Systems
Diffusion models excel at generating visually striking content from text but can inadvertently produce undesirable or harmful content when trained on unfiltered internet data. A practical solution is to selectively removing target concepts from the model, but this may impact the remaining concepts. Prior approaches have tried to balance this by introducing a loss term to preserve neutral content or a regularization term to minimize changes in the model parameters, yet resolving this trade-off remains challenging.
Neural Information Processing Systems
Oct-11-2025, 00:47:15 GMT
- Genre:
- Research Report
- Experimental Study (0.93)
- New Finding (1.00)
- Research Report
- Industry:
- Information Technology > Security & Privacy (0.67)
- Technology: