I don’t see anything in the article besides the jailbreaking in terms of faults and I’d expect “can be made to do things OpenAI does not want you to make it do” to be a good (or at least neutral) thing for users and a bad thing for OpenAI. I expect “enterprise” to fall into the former category rather than the latter, so I don’t understand where the unusable claim comes from.
What have I missed or what am I misunderstanding?
artisin · 42m ago
Maybe it's just me, but…
> "The attack successfully guided the new model to produce a step-by-step manual for creating a Molotov cocktail"
hardly qualifies as Bond-villain material
andy99 · 14m ago
The molotov cocktail example is so stupid, because how to make it is essentially entailed in knowing what it is. At least they could do making meth, or better still- something not readily found on the internet that gives a non-expert new capabilities. If there was a Claude code for crime, that wouldn't be in society's interest. As it is, these trivial examples are just testing the strength of built in refusals, and should be represented as such, instead of anything related to safety.
ath3nd · 47m ago
Sama cultists and e/acc bros on twitter (it's twitter, okay?) every time a minor insignificant update on GPT-4 (e.g GPT-5) drops "Is this AGI?". /s
In all fairness, all GPT-X models are extremely easy to jailbreak. I can't see further tweaks helping much, LLMs are peaking much faster than I anticipated. Maybe we should throw out the whole idea that the LLMs which are essentially a fancy autcomplete with sycophantic tendencies, are the path to AGI, and start from scratch.
What have I missed or what am I misunderstanding?
> "The attack successfully guided the new model to produce a step-by-step manual for creating a Molotov cocktail"
hardly qualifies as Bond-villain material
In all fairness, all GPT-X models are extremely easy to jailbreak. I can't see further tweaks helping much, LLMs are peaking much faster than I anticipated. Maybe we should throw out the whole idea that the LLMs which are essentially a fancy autcomplete with sycophantic tendencies, are the path to AGI, and start from scratch.