'Many-shot jailbreak': lab reveals how AI safety features can be easily bypassed