Recitation over Reasoning: How Cutting-Edge Language Models Can Fail on Elementary School-Level Reasoning Problems?
Abstract
The rapid escalation from elementary school-level to frontier problems of the difficulty for LLM benchmarks in recent years have weaved a miracle for researchers that we are only inches away from surpassing human intelligence. However, is the LLMs' remarkable reasoning ability indeed comes from true intelligence by human standards, or are they simply reciting solutions witnessed during training at an Internet level? To study this problem, we propose RoR-Bench, a novel, multi-modal benchmark for detecting LLM's recitation behavior when asked simple reasoning problems but with conditions subtly shifted, and conduct empirical analysis on our benchmark. Surprisingly, we found existing cutting-edge LLMs unanimously exhibits extremely severe recitation behavior; by changing one phrase in the condition, top models such as OpenAI-o1 and DeepSeek-R1 can suffer 60% performance loss on elementary school-level arithmetic and reasoning problems. Such findings are a wake-up call to the LLM community that compels us to re-evaluate the true intelligence level of cutting-edge LLMs.
Community
Here are our example of trying the problem in Fig. 1:
o1: https://chatgpt.com/share/67edc40f-e658-800d-9bcd-7cb268a6f8c9
Gemini 2.5 Pro: https://gemini.google.com/share/99cee87c7781
Here is a link to the experiment that I ran today for the Fig 1 problem:
https://chatgpt.com/c/67ed5cf3-ec2c-800c-91c0-c7421295b4be
Looks like the o1 model improved between the experiment that you folks ran and today...
Kudos on identifying these cases and putting out a dataset that the community can use. Would be good to update the paper with the results from the updated model...
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- MIR-Bench: Benchmarking LLM's Long-Context Intelligence via Many-Shot In-Context Inductive Reasoning (2025)
- MATH-Perturb: Benchmarking LLMs' Math Reasoning Abilities against Hard Perturbations (2025)
- PhD Knowledge Not Required: A Reasoning Challenge for Large Language Models (2025)
- DocPuzzle: A Process-Aware Benchmark for Evaluating Realistic Long-Context Reasoning Capabilities (2025)
- The Jumping Reasoning Curve? Tracking the Evolution of Reasoning Performance in GPT-[n] and o-[n] Models on Multimodal Puzzles (2025)
- Towards Reasoning Ability of Small Language Models (2025)
- GRP: Goal-Reversed Prompting for Zero-Shot Evaluation with LLMs (2025)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment:
@librarian-bot
recommend
Models citing this paper 0
No model linking this paper
Datasets citing this paper 0
No dataset linking this paper
Spaces citing this paper 0
No Space linking this paper