The chart hilariously reveals that GPT-5 scores a whopping 74.9% accuracy on software engineering benchmarks, but the pink bars tell the real story – 52.8% of that is achieved "without thinking" while only a tiny sliver comes from actual "thinking." Meanwhile, OpenAI's o3 and GPT-4o trail behind with 69.1% and 30.8% respectively, with apparently zero thinking involved. It's basically saying these AI models are just regurgitating patterns rather than performing actual reasoning. The perfect metaphor for when your code works but you have absolutely no idea why.
SWE-Bench Verified: Thinking Optional
3 months ago
300,730 views
1 shares
ai-memes, machine-learning-memes, gpt-memes, benchmarks-memes, software-engineering-memes | ProgrammerHumor.io
More Like This
The Future Is Here: Just Not The One We Need
7 months ago
211.2K views
0 shares
Better Prompting: The Modern Programmer's Paradox
1 month ago
194.4K views
0 shares
I Just Want A Normal Browser
24 days ago
181.9K views
1 shares
We Are Fine
4 months ago
207.2K views
0 shares
The Real Heroes Of Programming
4 months ago
259.6K views
0 shares
Programming Styles War
10 days ago
58.6K views
0 shares
Loading more content...
AI
AWS
Agile
Algorithms
Android
Apple
Bash
C++
Csharp