OpenAI scores gold in one of the top programming competitions

8 energy123 7 8/13/2025, 10:14:41 AM msn.com ↗

Comments (7)

NitpickLawyer · 1h ago
So in the past month we've had

- gold at IMO

- gold at IoI

- beat 9/10 humans in atcode heuristics

- longer context, better models, routing calls to cheaper models, 4-6x cheaper inference for 90% of the top models capabilities

- longer agentic sessions while being coherent/solving tasks (30-90min)

Yet every other post here and there are about "bubble this", "winter that", "plateauing this", "wall that"...

Are we in the denial stage, or bargaining stage? Can't quite tell...

energy123 · 1h ago
People use low-compute models in their day to day jobs. They're not exposed to how good the very-high-compute runs are doing at the moment.
machiaweliczny · 25m ago
This. My younger brother thinks it’s crap but if you know state of the art + research it seems like things still are moving quite fast. Also tons of product work on top already.
energy123 · 16m ago
Even gpt-5 on "high" reasoning effort (which is likely higher than what people get in the Plus subscription; that's most likely "medium") is very, very low compute compared to the top runs behind IOI/IMO solutions.
tyleo · 39m ago
But can it maintain my legacy crud app with no tests, millions of LoC, long compile times?

One day but not yet. Beyond pure capabilities the companies making AI don’t seem to have any sort of moat so it’s a $$$ incinerator for them so far.

Like the late 90s internet I suspect we’re in a bubble. But also like the late 90s internet I suspect there’s more in store here in the future.

robertlagrant · 1h ago
You might've said the same thing about self-driving cars five years ago, or chess even longer ago. It turns out chess was soluble, so the nay-sayers were wrong, but self-driving cars aren't soluble (yet) so the yay-sayers were wrong.
SideburnsOfDoom · 31m ago
How many of the answers were verbatim in the training data?