Learning with Less: Knowledge Distillation from Large Language Models via Unlabeled Data