Distilling Inductive Bias: Knowledge Distillation Beyond Model Compression