Enhancing Cognitive Robotics with Commonsense through LLM-Generated Preconditions and Subgoals

Bachner, Ohad, Gamliel, Bar

arXiv.org Artificial Intelligence 

Autonomous robots are increasingly deployed in dynamic and unstructured environments, where they must plan and execute complex tasks under uncertainty. Classical planning approaches, typically modeled in PDDL and solved with heuristic search, provide a principled foundation for task planning (Edelkamp and Schr odl, 2011; Geffner and Bonet, 2013). However, these methods rely on explicit domain models that enumerate preconditions and effects of actions. In practice, such models often omit implicit commonsense knowledge, for example, that a container must be upright before pouring, or that water must be boiled before making tea. The absence of such knowledge can lead to plans that are logically correct but physically invalid. Cognitive robotics research seeks to bridge symbolic reasoning with robot perception and control (Ghallab et al., 2004). While significant progress has been made in integrating planning with motion control and execution, robots still lack the ability to autonomously infer commonsense constraints that humans consider obvious. Large Language Models (LLMs), trained on massive corpora of human knowledge, present a promising avenue for addressing this gap. LLMs can generate likely preconditions, subgoals, and contextual constraints from natural language task descriptions, potentially enriching classical planning models. 1