The chart hilariously reveals that GPT-5 scores a whopping 74.9% accuracy on software engineering benchmarks, but the pink bars tell the real story – 52.8% of that is achieved "without thinking" while only a tiny sliver comes from actual "thinking." Meanwhile, OpenAI's o3 and GPT-4o trail behind with 69.1% and 30.8% respectively, with apparently zero thinking involved. It's basically saying these AI models are just regurgitating patterns rather than performing actual reasoning. The perfect metaphor for when your code works but you have absolutely no idea why.
SWE-Bench Verified: Thinking Optional
8 months ago
452,064 views
1 shares
ai-memes, machine-learning-memes, gpt-memes, benchmarks-memes, software-engineering-memes | ProgrammerHumor.io
More Like This
Tower Of Hanoi: Humbling Humans And AI Alike
10 months ago
318.1K views
0 shares
Only One Prompt Away
7 months ago
269.0K views
1 shares
The Evolution Of Tech Rage: From Windows Search To AI Assistants
11 months ago
324.4K views
0 shares
Why Am I Doing This
2 months ago
370.2K views
0 shares
I'M Not Calling It By Its „Real" Name Anymore, Sry Slopdella
2 months ago
259.9K views
0 shares
Loading more content...
AI
AWS
Agile
Algorithms
Android
Apple
Bash
C++
Csharp