How to Protect Models against Adversarial Unlearning?