The chart hilariously reveals that GPT-5 scores a whopping 74.9% accuracy on software engineering benchmarks, but the pink bars tell the real story – 52.8% of that is achieved "without thinking" while only a tiny sliver comes from actual "thinking." Meanwhile, OpenAI's o3 and GPT-4o trail behind with 69.1% and 30.8% respectively, with apparently zero thinking involved. It's basically saying these AI models are just regurgitating patterns rather than performing actual reasoning. The perfect metaphor for when your code works but you have absolutely no idea why.
SWE-Bench Verified: Thinking Optional
7 months ago
428,706 views
1 shares
ai-memes, machine-learning-memes, gpt-memes, benchmarks-memes, software-engineering-memes | ProgrammerHumor.io
More Like This
No Tear Was Dropped
2 months ago
340.7K views
0 shares
How To Get 2 Billion Valuation
4 months ago
252.4K views
0 shares
AI Really Does Replace Juniors
8 months ago
365.4K views
0 shares
North Korean Software Engineers Were Sweating Yesterday
15 days ago
338.8K views
2 shares
Guess I'll Wait It Out...
2 months ago
279.7K views
0 shares
Developer Tools (affiliate)
The Developer's Path To Enlightenment
1 year ago
320.2K views
1 shares
Loading more content...
AI
AWS
Agile
Algorithms
Android
Apple
Bash
C++
Csharp