The chart hilariously reveals that GPT-5 scores a whopping 74.9% accuracy on software engineering benchmarks, but the pink bars tell the real story – 52.8% of that is achieved "without thinking" while only a tiny sliver comes from actual "thinking." Meanwhile, OpenAI's o3 and GPT-4o trail behind with 69.1% and 30.8% respectively, with apparently zero thinking involved. It's basically saying these AI models are just regurgitating patterns rather than performing actual reasoning. The perfect metaphor for when your code works but you have absolutely no idea why.
SWE-Bench Verified: Thinking Optional
3 hours ago
30,056 views
0 shares

ai-memes, machine-learning-memes, gpt-memes, benchmarks-memes, software-engineering-memes | ProgrammerHumor.io
More Like This
Reality Is Often Disappointing
3 months ago
124.8K views
0 shares

Co Pilot Go Brrrr
3 months ago
196.8K views
0 shares

Fix The Error
4 months ago
167.6K views
0 shares

The Algorithmic Sacrifice
2 months ago
159.2K views
0 shares

The Perfect Timing Algorithm
2 months ago
95.9K views
0 shares

Faster, But At What Cost?
4 months ago
148.8K views
0 shares

Loading more content...