The Mismeasure of Man and Models: Evaluating Allocational Harms in Large Language Models