Anti-adversarial Learning: Desensitizing Prompts for Large Language Models