DialogBench: Evaluating LLMs as Human-like Dialogue Systems