Human-Level Reasoning: A Comparative Study of Large Language Models on Logical and Abstract Reasoning
–arXiv.org Artificial Intelligence
Evaluating reasoning ability in Large Language Models (LLMs) is important for advancing artificial intelligence, as it transcends mere linguistic task performance. It involves understanding whether these models truly understand information, perform inferences, and are able to draw conclusions in a logical and valid way. This study compare logical and abstract reasoning skills of several LLMs - including GPT, Claude, DeepSeek, Gemini, Grok, Llama, Mistral, Perplexity, and Sabi a - using a set of eight custom-designed reasoning questions. The LLM results are benchmarked against human performance on the same tasks, revealing significant differences and indicating areas where LLMs struggle with deduction.
arXiv.org Artificial Intelligence
Oct-29-2025
- Country:
- North America > United States (0.04)
- South America > Brazil
- Santa Catarina (0.04)
- Genre:
- Research Report (0.82)
- Industry:
- Education (0.46)
- Technology: