Evaluating the Limitations of Local LLMs in Solving Complex Programming Challenges
Matotek, Kadin, Cassel, Heather, Amiruzzaman, Md, Ngo, Linh B.
–arXiv.org Artificial Intelligence
This study examines the performance of today's open-source, locally hosted large-language models (LLMs) in handling complex competitive programming tasks with extended problem descriptions and contexts. Building on the original Framework for AI-driven Code Generation Evaluation (FACE), the authors retrofit the pipeline to work entirely offline through the Ollama runtime, collapsing FACE's sprawling per-problem directory tree into a handful of consolidated JSON files, and adding robust checkpointing so multi-day runs can resume after failures. The enhanced framework generates, submits, and records solutions for the full Kattis corpus of 3,589 problems across eight code-oriented models ranging from 6.7-9 billion parameters. The submission results show that the overall pass@1 accuracy is modest for the local models, with the best models performing at approximately half the acceptance rate of the proprietary models, Gemini 1.5 and ChatGPT-4. These findings expose a persistent gap between private, cost-controlled LLM deployments and state-of-the-art proprietary services, yet also highlight the rapid progress of open models and the practical benefits of an evaluation workflow that organizations can replicate on in-house hardware.
arXiv.org Artificial Intelligence
Sep-22-2025
- Country:
- North America > United States
- New Mexico > Bernalillo County
- Albuquerque (0.04)
- New York > New York County
- New York City (0.04)
- Pennsylvania
- Chester County > West Chester (0.04)
- Delaware County > Chester (0.04)
- New Mexico > Bernalillo County
- North America > United States
- Genre:
- Research Report > New Finding (1.00)
- Industry:
- Information Technology (1.00)
- Technology: