DivLogicEval: A Framework for Benchmarking Logical Reasoning Evaluation in Large Language Models