Interpretability is in the Mind of the Beholder: A Causal Framework for Human-interpretable Representation Learning