LLMCheckup: Conversational Examination of Large Language Models via Interpretability Tools