Can LLMs Implicitly Learn Numeric Parameter Constraints in Data Science APIs?
–Neural Information Processing Systems
Data science (DS) programs, typically built on popular DS libraries (such as Py-Torch and NumPy) with thousands of APIs, serve as the cornerstone for various mission-critical domains such as financial systems, autonomous driving software, and coding assistants. Recently, large language models (LLMs) have been widely applied to generate DS programs across diverse scenarios, such as assisting users for DS programming or detecting critical vulnerabilities in DS frameworks. Such applications have all operated under the assumption, that LLMs can implicitly model the numerical parameter constraints in DS library APIs and produce valid code. However, this assumption has not been rigorously studied in the literature. In this paper, we empirically investigate the proficiency of LLMs to handle these implicit numerical constraints when generating DS programs.
Neural Information Processing Systems
May-29-2025, 18:24:07 GMT
- Country:
- Asia > Middle East
- UAE (0.14)
- North America > United States
- California (0.14)
- Illinois (0.14)
- Asia > Middle East
- Genre:
- Research Report
- Experimental Study (0.93)
- New Finding (1.00)
- Research Report
- Industry:
- Information Technology > Security & Privacy (0.46)
- Technology: