Fight Back Against Jailbreaking via Prompt Adversarial Tuning Yichuan Mo1Y uji Wang 2 Zeming Wei 3 Yisen Wang 1,4 1

Neural Information Processing Systems 

While Large Language Models (LLMs) have achieved tremendous success in various applications, they are also susceptible to jailbreaking attacks.

Similar Docs  Excel Report  more

TitleSimilaritySource
None found