GPT5 Is Horrible

8 druskacik 2 8/8/2025, 8:25:34 AM old.reddit.com ↗

Comments (2)

A_D_E_P_T · 2h ago
For my use cases it doesn't appear to be any better than o3/o3-pro.

Also, GPT-5 pro is a lot slower than o3-pro. My two most recent queries taking 17 and 18 minutes, whereas o3-pro would probably take 4-5.

Surprisingly, at generic writing-prose tasks (e.g. compose a poem in the style of X), GPT-5 is still noticeably inferior to DeepSeek and Kimi.

Honestly, I'm tempted to cancel my $200/month subscription.

This probably means that we're in the "slow AI / stagnation" timeline, so at least we're not going to get paperclipped by 2027.

NitpickLawyer · 1h ago
> For my use cases it doesn't appear to be any better than o3/o3-pro.

It doesn't have to be better, it has to be "as close to those as possible", while being cost efficient to run and serve at scale.

> This probably means that we're in the "slow AI / stagnation" timeline

I'd say we're more in the "ok, we got the capabilities in huge models, now we need to make them smaller, faster and scalable" timeline. If they capture ~80-90% of the capabilities of their strongest models while reducing costs a lot (they've gone from 40-60$/Mtok to 10$/Mtok) then they're starting to approach a break-even point and slowly make money off of serving tokens.

There's also a move towards having specialised models (code w/ claude, long context w/ gemini, etc) and oAI seem to have gone in this direction as well. They've said for a long time that gpt5 will be a "systems" update and not necessarily a core model update. That is, they have a routing model that takes a query and routes it towards the best model for the task. Once devs figure out how to use this to their advantage, the "vibes" will improve.