FairPair: A Robust Evaluation of Biases in Language Models through Paired Perturbations