The chart hilariously reveals that GPT-5 scores a whopping 74.9% accuracy on software engineering benchmarks, but the pink bars tell the real story – 52.8% of that is achieved "without thinking" while only a tiny sliver comes from actual "thinking." Meanwhile, OpenAI's o3 and GPT-4o trail behind with 69.1% and 30.8% respectively, with apparently zero thinking involved. It's basically saying these AI models are just regurgitating patterns rather than performing actual reasoning. The perfect metaphor for when your code works but you have absolutely no idea why.
SWE-Bench Verified: Thinking Optional
6 months ago
384,506 views
1 shares
ai-memes, machine-learning-memes, gpt-memes, benchmarks-memes, software-engineering-memes | ProgrammerHumor.io
More Like This
Artificial Intelligence Vs Natural Stupidity
10 months ago
309.3K views
0 shares
Ancient IBM Wisdom That The Bosses Just Straight Up Promptly Forgot
3 months ago
317.2K views
0 shares
The Future Of Communication
9 months ago
233.3K views
0 shares
Current State Of Microsoft
1 day ago
180.1K views
0 shares
Too Afraid To Ask About LLM Benchmarks
9 months ago
248.8K views
0 shares
The AI Bust: Expectations vs. Painful Reality
9 months ago
243.4K views
1 shares
Loading more content...
AI
AWS
Agile
Algorithms
Android
Apple
Bash
C++
Csharp