The chart hilariously reveals that GPT-5 scores a whopping 74.9% accuracy on software engineering benchmarks, but the pink bars tell the real story – 52.8% of that is achieved "without thinking" while only a tiny sliver comes from actual "thinking." Meanwhile, OpenAI's o3 and GPT-4o trail behind with 69.1% and 30.8% respectively, with apparently zero thinking involved. It's basically saying these AI models are just regurgitating patterns rather than performing actual reasoning. The perfect metaphor for when your code works but you have absolutely no idea why.
SWE-Bench Verified: Thinking Optional
2 months ago
255,119 views
1 shares

ai-memes, machine-learning-memes, gpt-memes, benchmarks-memes, software-engineering-memes | ProgrammerHumor.io
More Like This
Rubber Duck Debugging With Extra Steps
2 months ago
190.8K views
2 shares

ChatGPT Vs Programmers: First Time?
5 months ago
176.3K views
0 shares

When AI Models Train On Your NPM Packages
4 months ago
248.7K views
0 shares

The Leather-to-Suit Price Hike Indicator
4 months ago
159.5K views
0 shares

Future Of Software Development If GPT Bros Win
6 months ago
167.3K views
1 shares

Tech Gadgets (affiliate)
Gambling vs. Vibe Coding: Same Addiction, Different Casino
1 month ago
240.2K views
5 shares

Loading more content...