Be like a Goldfish, Don't Memorize! Mitigating Memorization in Generative LLMs
–Neural Information Processing Systems
To mitigate memorization, we introduce a subtle modification to the next-token training objective that we call the goldfish loss. During training, a randomly sampled subsets of tokens are excluded from the loss computation. These dropped tokens are not memorized by the model, which prevents verbatim reproduction of a complete chain of tokens from the training set.
Neural Information Processing Systems
May-26-2025, 19:43:41 GMT
- Technology: