The chart hilariously reveals that GPT-5 scores a whopping 74.9% accuracy on software engineering benchmarks, but the pink bars tell the real story – 52.8% of that is achieved "without thinking" while only a tiny sliver comes from actual "thinking." Meanwhile, OpenAI's o3 and GPT-4o trail behind with 69.1% and 30.8% respectively, with apparently zero thinking involved. It's basically saying these AI models are just regurgitating patterns rather than performing actual reasoning. The perfect metaphor for when your code works but you have absolutely no idea why.
SWE-Bench Verified: Thinking Optional
2 months ago
277,049 views
1 shares
ai-memes, machine-learning-memes, gpt-memes, benchmarks-memes, software-engineering-memes | ProgrammerHumor.io
More Like This
Are You Sure Buddy
4 months ago
153.1K views
0 shares
Be Sure To Register Your Place At Vibe Con
5 months ago
174.8K views
4 shares
Who's Gonna Tell Him
5 months ago
199.4K views
0 shares
Elegant Code Misused Is Bad Code
7 months ago
185.1K views
0 shares
Intel Powers Students' Wallets Into Oblivion
6 months ago
204.6K views
0 shares
Developer Tools (affiliate)
The Only Right Way To Implement AI Reasoning
2 months ago
182.8K views
0 shares
Loading more content...
AI
AWS
Agile
Algorithms
Android
Apple
Bash
C++
Csharp