Can LLMs Explain Themselves Counterfactually?