The chart hilariously reveals that GPT-5 scores a whopping 74.9% accuracy on software engineering benchmarks, but the pink bars tell the real story – 52.8% of that is achieved "without thinking" while only a tiny sliver comes from actual "thinking." Meanwhile, OpenAI's o3 and GPT-4o trail behind with 69.1% and 30.8% respectively, with apparently zero thinking involved. It's basically saying these AI models are just regurgitating patterns rather than performing actual reasoning. The perfect metaphor for when your code works but you have absolutely no idea why.
SWE-Bench Verified: Thinking Optional
4 months ago
336,107 views
1 shares
ai-memes, machine-learning-memes, gpt-memes, benchmarks-memes, software-engineering-memes | ProgrammerHumor.io
More Like This
Thanks Copilot For The Div Inception
9 months ago
342.9K views
0 shares
Ninety-Five Percent AI Generated
8 months ago
240.4K views
0 shares
When Copilot Goes Off The Rails
9 months ago
246.5K views
1 shares
I'm Not Crazy, I'm Training A Model
6 months ago
295.4K views
0 shares
They're Just Like Us: AI Learns The Art Of Procrastination
1 month ago
204.5K views
1 shares
Loading more content...
AI
AWS
Agile
Algorithms
Android
Apple
Bash
C++
Csharp