Does anyone think the current AI approach will hit a dead end?
22 rh121 51 9/8/2025, 2:44:08 AM
Billions of dollars spent, incredible hype that we will have AGI in several years. Does anyone think the current deep learning / neural net based AI approach will eventually hit a dead end and not be able to deliver its promises? If yes, why?
I realize this question is somewhat loosely defined. No doubt the current approach will continue to improve and yield results so it might not be easy to define "dead end".
In the spirit of things, I want to see whether some people think the current direction is wrong and won't get us to the final destination.
We'll get more incremental updates and nice features:
* more context size
* less hallucinations
* more prompt control (or the illusion of)
But we won't get AGI this way.
From the very beginning LLMs were shown to be incapable of synthesising new ideas. They don't sit there and think; they can only connect dots within a paradigm that we give them. You may give me examples of AI discovering new medicines and math proofs as a counter-argument but I see that as re-enforcing the above.
Paired with data and computional scaling issues, I just don't see it happening. They will remain a useful tool, but won't become AGI.
And whether they stay affordable is a question of time; all the big players are burning mountains of cash just to edge out the competition in terms of adoption.
Is there a level of adoption that can justify the current costs to run these things?
I'd argue they don't synthesize any ideas, even old ones. They skip that classic step to emit text, and the human reading that text generates their own idea and (unconsciously, incorrectly) assumes there must've an original idea that caused the text on the other side.
So perhaps it's more like: "LLMs aren't great at indirectly triggering humans into imagining useful novel ideas." (Especially when the user is trying to avoid having to think.)
Yeah, I know, it sounds like quibbling, but I believe it's necessary. This whole subject is an epistemic and anthropomorphic minefield. A lot of our habitual language connotations and metaphors can mislead.
Nobody has the tools to begin proving a negative [0] in either of those cases, and it's possible they'll eventually occur... But so what?
Just because it could happen someday does not mean it's happening now. Instead, we have decades of seeing humans excite themselves into perceiving semantics that aren't present [0], and nobody's provided a compelling reason to believe that this time things are finally different.
[0] https://en.wikipedia.org/wiki/Burden_of_proof_(philosophy)
[1] https://en.wikipedia.org/wiki/ELIZA_effect
If LLMs and statistics can't encode semantics, how can do chatbots perform long-form translations with appropriate contexts? How do codebreakers use statistics to break an adversary's communications?
Sometimes the statistics are semantic, like when "orange" and "arancia" the picture of that fruit all mean the same thing, but Orange the wireless carrier and orange the color are different. Those are connections/probabilities humans also learn via repeated exposure in different contexts.
I'm not arguing that LLMs are synthesizing new ideas (or old ones), but that they ARE capable of deriving semantic meaning from statistics. Rather than:
> language, based solely on statistical data, shorn of semantics
Isn't it more like:
> language, based solely on statistical data, with meanings emerging from clusters in the data
To a degree. The problem is that they don't actually understand the ideas in the training data. (Yeah, you can say we don't know how humans actually understand ideas. True, but not the point. However we understand ideas, LLMs don't do that.) And so they can only synthesize new ideas by rearranging words. This is much less than that human thinking. In particular, it seems that it could only generate ideas that are only new recombinations, not breakthrough ideas.
I don't think that follows: Manipulating a (lossy, imperfect) encoding [0] isn't the same as manipulating the thing it was intended to evoke.
If it is true, then... Well, it's not true in the same way anybody is excited about, because it means "synthesizing new ideas" is something we've been able to do for many decades and which you can easily script up right now at home [1].
[0] https://en.wikipedia.org/wiki/Encoding_(semiotics)
[1] https://benhoyt.com/writings/markov-chain/
>Starting from random play, and given no domain knowledge except the game rules, AlphaZero achieved within 24 hours a superhuman level of play in the games of chess and shogi [a similar Japanese board game] as well as Go, and convincingly defeated a world-champion program in each case https://www.theguardian.com/technology/2017/dec/07/alphazero...
And MuZero did similar, surpassing AlphaGo. So while LLMs may be bad at thinking and learning, other neural net arrangements may do well.
This is seen by what people term as hallucinations. AI seeks to please and will lie to you and invent things in order to please you. We can scale it up to give it more knowledge but ultimately those failures still creep in.
There will have to be a fundamentally new design for this to be overcome. What we have now is an incredible leap forward but it has already stalled on what it can deliver.
On top of that we're at what, year four of the AI "revolution"? And yet ChatGPT is still the only AI tool that is somewhat recognizable and used by the general public. Other AI-based tools are either serving a niche (Cursor, Windsurf), serve as toys (DALL-E, Veo) or are so small and insignificant that barely anyone is using them. If I go to any job board and browse the job offers no company seems to be naming any specific AI-powered tools that they want people to be proficient in. I don't think I've ever seen any company - big or small - either bragging that they've used generative AI as a significant driver in their project or claiming that thanks to implementing AI they've managed to cut x% costs or drive up their revenue by y%. Open source doesn't seem to have much going on either, I don't think there are examples of any projects that got a huge boost from generative AI.
Considering how much money was pumped into these solutions and how much buzz this topic has generated all over the internet in the past 4 years it seems at least bizarre that the actual adoption seems to be so insignificant. In many other areas of tech 4 years would be considered almost an eternity and yet this technology somehow gets a pass. This topic has puzzled me a for a while now but only in this year I've noticed other people pointing out these issues as well.
[1] https://www.researchgate.net/publication/381009719_Hydra_Enh...
Language is only one aspect of a conscious mind, there are others like the ones that handle executive function, spatial and logical thinking, reasoning, emotional regulation, and many others. LLMs only deal with the language part and that’s not nearly enough to build a true AGI— a conscious mind that lives inside computer that we can control.
Intelligence is an emergent property that comes as a result of all distinct functions of the brain (whether biological or artificial) being deeply intertwined.
Anywho, I think JN Research is onto something with the Adaptrons/Primite and hope you guys are able to take it as far as it’ll go.
The DeepDream stuff seemed quite dreamlike (https://en.wikipedia.org/wiki/DeepDream)
Kids learn to speak before they learn to think about what they're saying. A 2/3 year old can start regurgitating sentences and forming new ones which sound an awful lot like real speech, but it seems like it's often just the child trying to fit in, they don't really understand what they're saying.
I used to joke my kids talking was sometimes just like typing a word on my phone and then just hitting the next predictive word that shows up. Since then it's evolved in a way that seems similar to LLMs.
The actually process of thought seems slightly divorced from the ability to pattern match words, but the patter matching serves as a way to communicate it. I think we need a thinking machine to spit out vectors that the LLM can convert into language. So I don't think they are a dead end, I think they are just missing the other half of the puzzle.
I don’t know AI, but I’m of the few that’s grateful for what it is at the moment. I'm coding with the free mini model and it has saved me a ton of time and I’ve learned a lot.
But I could be totally wrong because im certainly not an expert in these fields.
The definition of AGI is very subjective. Clearly, current models are not as good as the top humans, but the progress in the last 20 years has been immense, and the ways in which these models fall short are becoming subtle and hard to articulate.
Yes transformers do not learn in the same way as humans do, but that's in part an intentional design decision, because humans have a big flaw: our memories can't be separated from our computation, they can't be uploaded, downloaded, modified or backed up. Do we really want AGI to have a bus factor? With transformers you can take the context that is fed to one model and feed it to another model, try doing that with a human!
In a sense, the transformer model with its context is an improvement on the architecture of the human brain. But we do need more tricks to make context as flexible as human long term memory.
What exactly is intelligence? Nobody really knows and understands yet where the “natural” comes from.
Hence all we do so far is nothing but a sophisticated cargo culting.
Case closed.
In my future I also saw lots and lots of cheap GPU chips and hardware, much gaming but fewer "developers" and a mostly flat-lined software economy for at least 8 years. Taiwan was still independent but was suffering from an economic recession.
The first big settlement for using stolen data has come (Anthropic). How you extricate the books archive and claimamants' works is unknown.
I believe that LLM's in verticals are being fed expert/cleaned data, but wasn't that always the case, i.e. knowledge bases? Much less data and power needed (less than ∞) Oh, and much less investment, IMO.
I'm bullish on AI's being generally capable by 2030. That date seems to be a 50/50 line for many in the field.
Personally, I fear the worst.
The bubble is bursting. Hope y'all are alright in the midst of it.
Imagine someone in 2006: "All these cynics and doomsayers about the US housing market are just envious and bitter that they didn't get in on the ground floor with their own investments."
Perhaps some were, but if we're going to start discounting opinions based on ulterior motives, there's a much bigger elephant in the room.
But all of your comments seem to be dismissive of other people's opinions.
Nice try, Claude!
All those LLMs benchmarks are terrible. LLMs gets better at it, but users don't perceive it. LLMs haven't improved that much the last year.
For AI in general, the future is bright. Now we have a lot of brain power and hardware available, more new research will pop up.
A LONG TIME AGO, Claude (your favorite LLM model?) Shannon has shown that entropy is a fundamental limit. There may be limitations we aren't aware of 'intelligence'.
Despite what experts say, Superintelligence or AGI might not even exist.
Is AGI knowing all the possible patterns in the universe? Nobody can even properly define it. But it is wrong, as not every intelligent thing isn't a pattern.
But are cars going to drive themselves by using similar inputs than a human? Yes, probably soon
Also many improvements to machinery, factories and productivity. They will shape the economy to a new format. No superintelligence or AGI needed. Just 'human'-level pattern recognition.
There are fundamental limitations with transformers that will not go away for as long as AI equates transformers.
The first one is the lack of understanding/control by humans. Orgs want guarantees that these systems won't behave unexpectedly while also wanting innovative and useful behaviour from them. Ultimately, most if not all neural nets are black boxes so understanding the reasoning for a specific action at a given time, let alone their behaviour in general is just not feasible due to their sheer complexity. We just don't understand why the behave the way they do in a scientific way anyway than we understand why a specific rabbit did a specific action at that particular moment in a way that can be used to make accurate predictions about when it will do that action again. Due to our lack of understanding, we just cannot control these things accurately. We either block some of their useful abilities to reduces the changes of undesired behaviour or you are you exposed to it. This trade-off is just a fundamental limitation of the fact that transformers are used nowadays are neural nets and as such have all the limitations that they have.
The second one is our inability to completely stop the hallucinations. From what my understanding, this is inherently tied to the very nature of how transformers based LLMs produce output. There is no understanding of the notion of truth or real world. It's just emulating patterns seeing in its training data, it just so happens that some of those don't correlate with real world facts (truth) even if they correlate with human grammar. In so far as there is no understanding of the notion of truth as separate from patterns in data, however implicit, hallucinations will continue. And there is no reason to believe that we will come up with a revolutionary way to train these systems in a way that they understand truth and not just grammar.
The third one is learning, models can't learn or remember as such, context learning is a trick to emulate learning but it's extremely inefficient and not scalable and models don't really have the ability to manipulate it the way humans or other animals can do. This is probably the most damning of them all as you cannot possible have a human level General Artificial that is unable to learn new skills on its own.
I would bet money on there not being significant progress before 2030. By significant progress I mean, the ability to do something that they could not do before at all regardless of the amount of training thrown at them given the same computing resources we have now.
That's starting to run dry, hence the predictions of progress stalling, but it doesn't factor in the option of using vast volumes of synthetic data. Some of the "thinking" models are already using generated problem sets, but this is just the tip of the iceberg.
There are so, so many ways in which synthetic data could be generated! Some random examples are:
- Introduce a typo into a working program or configuration file. Train the AI to recognise the typo based on the code and the error message.
- Bulk-generate homework or exam problems with the names, phrasing, quantities, etc... randomised.
- Train the AI not just on GitHub repos, but the full Git history of every branch, including errors generated by the compiler for commits that fail to build.
- Compile C/C++ code to various machine-code formats, assembly, LLVM IR, etc... and train the AI to reverse-engineer from the binary output to the source.
... and on and on.