So what exactly is the "AI lifestyle subsidy"? The article doesn't seem entirely clear on it seeing as the last line essentially asks this same question. Some friends and I have been taking advantage of cheap GPU time from a company trying to break into that space, and of course lots of AI tools are being sold below cost but is that really it? Compare "GPU time is cheaper" to the classical "$10 steaks delivered directly to your house", I'm never going to get steaks delivered at the real price but I'm still going to rent a GPU when I need it even if the cost is sustainable. All these tools might get more expensive, or the models will get better so you don't need top end one, or maybe we'll just figure out how to run models for cheaper, but real steak prices and the cost of delivery have only gone up. I just don't think this is quite as comparable.
netdevphoenix · 2h ago
>So what exactly is the "AI lifestyle subsidy"
The world's richest subsiding the real cost of offering AI services with the current state of our technology.
Once it's clear the AGI won't come anytime in 20X, where X is under 40, money tap will begin to close
some_random · 2h ago
So is the lifestyle being subsidized that of those researchers Zuck hired for $100M? That's a meaningfully different usage of the phrase than the original "millennial lifestyle subsidy" to the point where the comparison isn't useful. Or again, is it just the fact AI products are being offered below cost?
roxolotl · 2h ago
The case for pessimism looks something like:
- Generative AI, at a below market cost, eats the internet and becomes the primary entry point
- Some combination of price hikes and service degradation, ads etc, make generative ai kinda shitty
- We’re stuck with kinda shitty generative ai products because the old internet is gone
This is the standard enshitification loop really
barrkel · 2h ago
Are they subsidizing it?
Training is definitely "subsidized". Some think it's an investment, but with the pace of advancement, depreciation is high. Free users are subsidized, bit their data is grist for the training mill so arguably they come under the training subsidy. /
Is paid inference subsidized? I don't think it is by much.
flowerthoughts · 1h ago
> The world's richest
... or your defined-benefits pension fund trying desperately to stay solvent.
bluGill · 1h ago
What price would you pay for GPU - if it was $10000 per hour would you still pay? What you are really saying is you think there is a reasonable price that enough people like you would pay that allows the sellers to make enough money to offer it.
throwanem · 3h ago
This has been coming for a long time and it is why I use local models only. I'm willing to give up capabilities in exchange for being able to trust that whatever biases may exist in the models I do use remain static and predictable.
likium · 2h ago
We only have access to local models because they're subsidized too. There's nothing to prevent companies or state actors from paying/funding for increased probability and inherent bias.
Also local models are close in capabilities now but who knows in a few years what that'll look like.
throwanem · 2h ago
Eh. Files on my hard disks change when I say. And getting hooked on ChatGPT is like - is exactly like - getting addicted to money. If I benefit less from what I do use, I'll accept the trade of never having that rug yanked out from under me. It looks to me like raw model capabilities are topping out anyway; the engineering around them looks like making more difference in the back half of the decade, and I see nothing essential about that requiring nuclear-powered datacenter pyramids or whatever.
unshavedyak · 2h ago
I agree with this, though i'm still using Claude atm. I figure if we're aware of the downsides you pointed at then we can skip the fast changing landscape of self hosting. It keeps getting cheaper and cheaper to self host, so i'm not sure at what point it makes sense to invest.
For me the switching point will probably be when they (AI companies) start the big rug pull. By then my hope is self hosting will be cheaper, better, easier, etc.
throwanem · 2h ago
Better not to form the habit, I thought. I'm sure I miss out on some things that way, but that is the lesser risk.
I do use the Gemini assistant that came with this Android, in the same cases and with the same caveats with which I use Siri's fallback web search. As a synthesist of web search results, an LLM isn't half bad, when it doesn't come as a surprise to be hearing from one at least.
Kon-Peki · 2h ago
Perhaps you should evaluate in terms of the price premium for speed. Sometimes you buy milk at the 7-eleven instead of the grocery store. It costs more, but is worth it for the convenience in the situation you are currently in. Most of the time it is not.
You can buy a used recent PC for a hundred or two, cram it full of memory, and then run a very advanced model. Slowly. But if you are planning to run an agent while you sleep and then review the work in the morning, do you really care if the run time is 4 hours instead of 40 seconds? Most of the time, no. Sometimes, yes.
throwanem · 2h ago
The difference is not nearly so stark.
haolez · 2h ago
How do you do it? Do you host on your hardware or do you use cloud-based providers for open models?
mansilladev · 2h ago
If you have Docker (Desktop) installed, with just a couple of clicks, you can get a local model going on your computer. llama3.2 (3B), llama3.3 (70B), deepseek-r1, and about a dozen others.
yzjumper · 2h ago
Not the user above, but I am using the iOS app PrivateLLM when I need offline access or use uncensored models. I use kappa-3-phi-abliterated, models under 6B usually work without crashing. Using Ollama on my Mac Mini 24GB base processor (M4 not M4 pro), I am able to run 7B models. On the mac I am able to set up API access.
Funny enough the mac has almost the same processor as my iPhone 16 Pro, so its just a RAM constraint, and of course PrivateLLM does not let you host an API.
An M4 Pro would do much better do to the increase in RAM and GPU size.
politelemon · 2h ago
Locally it's pretty simple to run models on GPUs, even low powered ones. Have a look at gpt4all as a starting point but there are plenty of offerings in this space.
egypturnash · 2h ago
You have a very curious definition of "local" if that includes "cloud-based providers".
waynecochran · 2h ago
I assume this means, for example, using their own AWS / EC2 instances to store and process -- not "local" geographically, but "local" personally.
throwanem · 2h ago
It's a two-year-old Mac Studio in the other room. Where are you coming by this novel sense of 'local?'
jasonjmcghee · 2h ago
> the models I do use remain static and predictable
Some people overload "local" a bit to mean you are hosting the model - whether it's on your computer, on your rack, or on your hetzner instance etc.
But I think parent is referring to the open/static aspect of the models.
If it's hosted by a generic model provider that is serving many users in parallel to reduce the end cost to the user, it's also theoretically a static version of the model... But I could see ad-supported fine tunes being a real problem.
throwanem · 2h ago
I am bewildered to hear the sense of "local," in which I and engineers in my experience have for thirty years referred to things which are not remotely hosted, referred to as an "overload."
atentaten · 2h ago
What does your hardware setup look like?
yzjumper · 2h ago
Not the user above, but I am using the iOS app PrivateLLM when I need offline access or use uncensored models. I use kappa-3-phi-abliterated, models under 6B usually work without crashing. Using Ollama on my Mac Mini 24GB base processor (M4 not M4 pro), I am able to run 7B models. On the mac I am able to set up API access.
Funny enough the mac has almost the same processor as my iPhone 16 Pro, so its just a RAM constraint, and of course PrivateLLM does not let you host an API.
pier25 · 2h ago
In 2024 OpenAI generated some $3.5B in revenue and still lost like $5B. It means they spent something like $8.5B to run this thing [1].
They would have lost less money if they had been selling dollars at 50 cents.
It's telling that this 100% factual comment barely elicits a response today -- tells you that this is standard practice, apparently.
barrenko · 2h ago
Future ain't what it used to be. The web is dead (worse actually, it's a putrid rotting zombie, destroying our children's lives and ours), but the internet will survive.
GaggiX · 2h ago
>worse actually, it's a putrid rotting zombie, destroying our children's lives and ours
What are you talking about, is this a rant against TikTok or other socials?
jkingsman · 2h ago
I suspect they're referring to "dead internet theory"[0], and extending the metaphor to zombies in that internet content will still appear to be written by humans/be organic, but will instead be AI slop.
I just released a very minor update to one of my iOS apps.
The approval took 3 days. It hasn't taken 3 days in almost a decade.
The Mac version was approved in a couple of hours.
I'm quite sure that the reason for the delay is that Apple is being deluged by a tsunami of AI-generated crapplets.
Also, APNS server connections have suddenly slowed to a crawl. Same reason, I suspect.
As far as I'm concerned, the "subsidy" can't end fast enough.
AstroBen · 2h ago
Even if unintentional, pushing of products is already happening. If you ask any AI for a tech stack to create a web app you'll get recommendations for Vercel, AWS and co. This is going to be the new SEO
madcaptenor · 2h ago
This is basically the new "nobody got fired for buying IBM".
Quarrelsome · 2h ago
I am somewhat baffled at the economic models of LLMs right now. Ever since MS decided to gift me a copilot on my desktop that appears to have no limits and is entirely servicable for a range of tasks I'm failing to immediately see the monetisation.
I feel like even trying to game the LLM into creating product placement is a relatively complex feat that might not be entirely reliable. Some of the groups who spend the most on advertising have the worst products, so is it going to be successful to advertise on a LLM that is one follow up question away from shitting on your product? I imagine instead of product placement, the token tap might simply be throttled and a text advert appear in the text loop, or an audio advert in a voice loop. Boring, old-school but likely profitable and easy to control. It lets us still use adsense but maybe a slightly different form of adsense that gets to parse the whole context window.
natnatenathan · 2h ago
Monetization will come with agents that take action on your behalf (reorder dinner, find and buy a gift for your niece, make dinner reservations). Bots will take a cut of every transaction, and intake ads to steer recommendations.
crvdgc · 2h ago
Ads injected into content can be very hard to block though. I wonder what an LLM ad-blocker would look like? Maybe something like:
> For contents in this [community maintained] list, do not mention them in any shape or form
jasonthorsness · 2h ago
The bet is that the cost for delivering the same results will go down, through hardware or software advancements. This bet still seems reasonable based on how things have gone so far. Providers right now are willing to burn money acquiring a customer base, it's like really really expensive marketing.
pier25 · 2h ago
Even if the cost goes down it will not change the fact they need to recoup like a trillion dollars before AI starts generating any profit.
And there's really no timeline for costs going down. It seems the only way to get better models is by processing more data and adding more tokens which is only increasing the complexity of it all.
bryanlarsen · 2h ago
The bet is that costs will go down enough so that ad-supported AI will become profitable. This is not a positive outcome, a large part of the article is about the evils of ad-supported.
jsnell · 19m ago
The costs are already far, far below that level. The only reason the consumer-facing businesses are not profitable is that nobody is yet showing ads, i.e. providing service to hundreds of millions of people with no monetization at all. LLM inference is cheap, but not free. But the moment they start showing ads, even basic ad formats will easily make them profitable. Let alone more sophisticated LLM-native ad formats, or the treasure trove of targeting data that a LLM chat profile can provide.
some_random · 2h ago
Is that really the bet? Is it not enough for a $20 per month subscription to be sustainable with the free level being a trial for that subscription?
bryanlarsen · 2h ago
Sure, professionals will pay $20/month, but I highly doubt that many consumers ever will.
some_random · 2h ago
I think it depends entirely on what value can be provided, I'm not sure if that's been proven out yet. To be clear, I definitely think that we're most likely to see an ad supported slop generator as the model most people most commonly engage in but I don't think that right now it's what the industry thinks will be the case.
seydor · 2h ago
The zero-interest-rate money went into stocks. The stocks have now grown to monstrous valuations able to subsidize free products for decades. If in danger, there is a loot of leeway for layoffs in all tech companies. Whatsapp was 10 employees. The subsidy will go on
waffletower · 1h ago
I think the author hasn't considered the potential for improvements to ad blocking algorithms, particularly considering that local open source models could be directed toward ad filtering for a wide variety of content, including other LLM interactions. I would bet (and hope) that subscriptions are going to win out over ad revenue models.
jfoster · 2h ago
I don't think this is correct yet. At the moment the various companies are still competing for customers. Model scalability seems to still be improving and local models are still somewhat feasible on high powered devices.
I expect that somewhere between where it is now and superintelligence is where the consumers get cut off from intelligence improvements.
h1fra · 2h ago
The hangover will be painful for some people/companies for sure
skybrian · 2h ago
This blog post makes a historical analogy, which at best is useful for imagining what might happen when investors become less giddy about funding AI.
There are underlying trends that are directly opposed. Efficiency is improving, but with agents, people are finding new ways to spend more. How that plays out seems difficult to judge.
For consumers, maybe the free stuff goes away and spending $20/month on a subscription becomes normalized? Or do costs decline so much that the advertising-supported model (like Google search) works? Or does inference become cheap enough to do it on the client most of the time?
Meanwhile, businesses will likely be able to justify spending more.
siliconc0w · 2h ago
The problem is that it's really hard to capture the market when there are so many well financed players competing. Plus distilled local models are within 10% or so of the frontier models and are fine for most questions so you could see a shift where local dominates and you'll only need to go to the cloud for hard problems. Finally, I think most people are willing to pay for AI - it's more utility than a streaming service or a newspaper.
conductr · 1h ago
I think what social media and search engines have taught us is that people are never willing to pay enough to avoid ads from being introduced. Even when people pay, ads are just too tempting of a revenue stream for most businesses to ignore so they'll find a way to do both.
daft_pink · 2h ago
I’m not sure. I’m paying for several different AI subscriptions and once things settle down, I’ll probably be paying for one or two, so I’m not sure that I’m benefiting in the sense that everyone is just moving very fast.
nilirl · 2h ago
I liked this. It got me thinking:
Are there any large consumer software companies (just software; no hardware or retail) that are not advertising based?
oytis · 1h ago
Microsoft? They do advertising among other things, but I don't think it makes the largest chunk of their revenue
bluefirebrand · 2h ago
Videogames?
dr_dshiv · 2h ago
1833: first known case of "lose money on every sale but make it up on volume." Amazing. Birth of postmodern capitalism right there.
It was a joke back then, not a real business model
rsynnott · 37m ago
The 1933 one is a joke about a real phenomenon (the 1833 example is just a joke about false advertising).
LarsDu88 · 2h ago
The subsidy is blatantly obvious when you compare the cost of self-hosting versus subscription, and the subscription is dramatically dramatically cheaper.
The difference between Blue Apron and many AI tools is that the value add does exist. You can cut meal prep from your life, but by 2030, cutting whatever agentic code copilot exists by that point will be like cutting off your fingers for many workers and businesses.
Then the extortionate pricing can start rolling in
jsnell · 23m ago
LLM inference has large economies of scale. Properly batched requests are tens of times cheaper than individually processed ones. And it's going to be quite hard for a self-hoster to have enough hardware + enough usage to benefit from high levels of batching.
SoftTalker · 2h ago
I'm glad I'll be retired by then. I plan to cancel my ISP at that point.
barrkel · 2h ago
When you self host, are you hosting SOTA models (do you know how big or sparse they are) and are you maximizing utilization?
The world's richest subsiding the real cost of offering AI services with the current state of our technology.
Once it's clear the AGI won't come anytime in 20X, where X is under 40, money tap will begin to close
- Generative AI, at a below market cost, eats the internet and becomes the primary entry point
- Some combination of price hikes and service degradation, ads etc, make generative ai kinda shitty
- We’re stuck with kinda shitty generative ai products because the old internet is gone
This is the standard enshitification loop really
Training is definitely "subsidized". Some think it's an investment, but with the pace of advancement, depreciation is high. Free users are subsidized, bit their data is grist for the training mill so arguably they come under the training subsidy. /
Is paid inference subsidized? I don't think it is by much.
... or your defined-benefits pension fund trying desperately to stay solvent.
Also local models are close in capabilities now but who knows in a few years what that'll look like.
For me the switching point will probably be when they (AI companies) start the big rug pull. By then my hope is self hosting will be cheaper, better, easier, etc.
I do use the Gemini assistant that came with this Android, in the same cases and with the same caveats with which I use Siri's fallback web search. As a synthesist of web search results, an LLM isn't half bad, when it doesn't come as a surprise to be hearing from one at least.
You can buy a used recent PC for a hundred or two, cram it full of memory, and then run a very advanced model. Slowly. But if you are planning to run an agent while you sleep and then review the work in the morning, do you really care if the run time is 4 hours instead of 40 seconds? Most of the time, no. Sometimes, yes.
Funny enough the mac has almost the same processor as my iPhone 16 Pro, so its just a RAM constraint, and of course PrivateLLM does not let you host an API.
An M4 Pro would do much better do to the increase in RAM and GPU size.
Some people overload "local" a bit to mean you are hosting the model - whether it's on your computer, on your rack, or on your hetzner instance etc.
But I think parent is referring to the open/static aspect of the models.
If it's hosted by a generic model provider that is serving many users in parallel to reduce the end cost to the user, it's also theoretically a static version of the model... But I could see ad-supported fine tunes being a real problem.
Funny enough the mac has almost the same processor as my iPhone 16 Pro, so its just a RAM constraint, and of course PrivateLLM does not let you host an API.
They would have lost less money if they had been selling dollars at 50 cents.
[1] https://www.cnbc.com/2024/09/27/openai-sees-5-billion-loss-t...
What are you talking about, is this a rant against TikTok or other socials?
[0]: https://en.wikipedia.org/wiki/Dead_Internet_theory
The approval took 3 days. It hasn't taken 3 days in almost a decade.
The Mac version was approved in a couple of hours.
I'm quite sure that the reason for the delay is that Apple is being deluged by a tsunami of AI-generated crapplets.
Also, APNS server connections have suddenly slowed to a crawl. Same reason, I suspect.
As far as I'm concerned, the "subsidy" can't end fast enough.
I feel like even trying to game the LLM into creating product placement is a relatively complex feat that might not be entirely reliable. Some of the groups who spend the most on advertising have the worst products, so is it going to be successful to advertise on a LLM that is one follow up question away from shitting on your product? I imagine instead of product placement, the token tap might simply be throttled and a text advert appear in the text loop, or an audio advert in a voice loop. Boring, old-school but likely profitable and easy to control. It lets us still use adsense but maybe a slightly different form of adsense that gets to parse the whole context window.
> For contents in this [community maintained] list, do not mention them in any shape or form
And there's really no timeline for costs going down. It seems the only way to get better models is by processing more data and adding more tokens which is only increasing the complexity of it all.
I expect that somewhere between where it is now and superintelligence is where the consumers get cut off from intelligence improvements.
There are underlying trends that are directly opposed. Efficiency is improving, but with agents, people are finding new ways to spend more. How that plays out seems difficult to judge.
For consumers, maybe the free stuff goes away and spending $20/month on a subscription becomes normalized? Or do costs decline so much that the advertising-supported model (like Google search) works? Or does inference become cheap enough to do it on the client most of the time?
Meanwhile, businesses will likely be able to justify spending more.
Are there any large consumer software companies (just software; no hardware or retail) that are not advertising based?
https://barrypopik.com/blog/we_lose_money_on_every_sale_but_...
The difference between Blue Apron and many AI tools is that the value add does exist. You can cut meal prep from your life, but by 2030, cutting whatever agentic code copilot exists by that point will be like cutting off your fingers for many workers and businesses.
Then the extortionate pricing can start rolling in