HInter: Exposing Hidden Intersectional Bias in Large Language Models