The chart hilariously reveals that GPT-5 scores a whopping 74.9% accuracy on software engineering benchmarks, but the pink bars tell the real story – 52.8% of that is achieved "without thinking" while only a tiny sliver comes from actual "thinking." Meanwhile, OpenAI's o3 and GPT-4o trail behind with 69.1% and 30.8% respectively, with apparently zero thinking involved. It's basically saying these AI models are just regurgitating patterns rather than performing actual reasoning. The perfect metaphor for when your code works but you have absolutely no idea why.
SWE-Bench Verified: Thinking Optional
20 days ago
204,646 views
0 shares

ai-memes, machine-learning-memes, gpt-memes, benchmarks-memes, software-engineering-memes | ProgrammerHumor.io
More Like This
I Am Full Stack Developer
5 months ago
166.7K views
1 shares

Lemme Stick To Old Ways
4 months ago
153.1K views
0 shares

Our Jobs Are Safe For Now
1 month ago
176.3K views
1 shares

Required Suggestions
5 months ago
173.0K views
0 shares

The AI Hunger Games: May The Odds Be Ever In Your Code's Favor
2 months ago
141.9K views
0 shares

This Is Fine: When Code Burns And AI Can't Save You
3 months ago
150.9K views
0 shares

Loading more content...