LLM-GROP: Visually Grounded Robot Task and Motion Planning with Large Language Models