Training Code Language Models with Comprehensive Semantics Reasoning
–Neural Information Processing Systems
Code Large Language Models (Code LLMs) have excelled at tasks like code completion but often miss deeper semantics such as execution effects and dynamic states. This paper aims to bridge the gap between Code LLMs' reliance on static text data and the need for semantic understanding for complex tasks like debugging and program repair. We introduce a novel strategy, monologue reasoning, to train Code LLMs to reason comprehensive semantics, encompassing high-level functional descriptions, local execution effects of individual statements, and overall input/output behavior, thereby linking static code text with dynamic execution states.
Neural Information Processing Systems
May-29-2025, 22:53:58 GMT
- Country:
- North America > United States (0.28)
- Genre:
- Research Report > Experimental Study (0.93)
- Industry:
- Education (0.92)
- Information Technology (0.92)
- Technology: