OpenAI GPT leaking your data
In this series around GPT language model, we will focus on the paper "Extract Training Data from Large Language Models" The authors want to show that they can extract verbatim data from a language model such as GPT-2. More interestingly, they explain that they can extract verbatim that have appeared only a few times in the training data from the model itself. Naturally, that can be very dangerous if you own a company and you are using customers' data to train a language model. In their own words, "the paper demonstrates that (…), an adversary can perform a training data extraction attack to recover individual training examples by querying the language model." Who would want to risk leaking private information?
Apr-21-2021, 15:40:32 GMT
- Country:
- North America > United States > California > San Francisco County > San Francisco (0.05)
- Technology: