Bridging the Evaluation Gap: Leveraging Large Language Models for Topic Model Evaluation