AI overviews cause massive drop in search clicks

448 jonbaer 510 7/23/2025, 7:50:12 PM arstechnica.com ↗

Comments (510)

littlecranky67 · 6h ago
Here is the experience when clicking a link on mobile:

* Page loads, immediately when I start scrolling and reading a popup trying to get tracking consent

* If I am lucky, there is a "necessary only". When unlucky I need to click "manage options" and first see how to reject all tracking

* There is a sticky banner on top/bottom taking 20-30% of my screen upselling me a subscription or asking me to install their app. Upon pressing the tiny X in the corner it takes 1-2 seconds to close or multiple presses as I am either missing the x or because there is a network roundtrip

* I scroll down a screen and get a popup overlay asking me to signup for their service or newsleter, again messing with the x to close

* video or other flashy adds in the content keep bugging me

This is btw. usually all before I even established if the content is what I was looking for, or is at any way useful to me (often it is not).

If you use AI or Kagi summarizr, you get ad-free, well-formatted content without any annoyance.

pembrook · 23m ago
Yes, this is the experience on virtually every content website that used to be tolerable or even good.

But this is because there is no viable monetization model for non-editorial written word content anymore and hasn’t been for a decade. Google killed the ecosystem they helped create.

Google also killed the display ad market by monopolizing it with Adsense and then killed Adsense revenue sharing with creators to take all the money for themselves by turning their 10 blue links into 5 blue ads at the top of the search results. Search ads is now the most profitable monopoly business of all time.

YouTube is still young, but give it time. Google will eventually kill the golden goose there as well, by trying to harvest too many eggs for themselves.

The same will happen with AI results as well. Companies will be happy to lose money on it for a decade while they fight for dominance. But eventually the call for profits will come and the AI results will require scrolling through mountains of ads to see an answer.

This is the shape of this market. Search driven content in any form is and will always be a yellow pages business. Doesn’t matter if it’s on paper or some future AGI.

No comments yet

quectophoton · 1h ago
For those who want to experience it: https://how-i-experience-web-today.com/

The only inaccurate thing of that meme page is that you only need to uncheck 5 cookie "partners", when in reality there should be at least a few hundred.

Jenk · 1h ago

    document.querySelectorAll("[type='checkbox']").forEach(c => c.checked = undefined)
Adjust the selector as neccessary, sometimes I'll use `#id-of-cookie-banner [type='checkbox']`

Probably useless for mobile though, unless you can punch it in the omnibar with `javascript:` prefix

hereonout2 · 4h ago
You forgot the part about when you actually get to the content, there's usually about 5 paragraphs of SEO filler text before it actually gets onto answering the topic of the post.
progbits · 1h ago
You are lucky if they even answer.

Most of those are like:

    $movie release date
    
    <five paragraphs of garbage>
    
    While we don't know the actual $movie release date yet, ...
chromehearts · 55m ago
These are the worst things ever
Disposal8433 · 2h ago
I have noticed that a lot. For example:

What is the price of the Switch 2?

The Switch 2 can be purchased with money. <Insert the Wikipedia article about currencies since the bronze age>

pflenker · 1h ago
Recipe for Foo. Foo has always been my favorite dish. I fondly remember all the times my grandma made this for me. My grandma, who was born on August 2, 1946, as the daughter of… (10 more pages of text) To cook Foo the way my grandma did, you first need some Bar. Bar is originally native to the reclusive country of… (20 more pages of text)
sidewndr46 · 23m ago
You forgot 4 paragraphs text about how they went on a journey of self discovery, that lead to them spending time in the remote village of Y, learning the traditional methods of cooking the dish.

The dish in question is a ham sandwich.

chasd00 · 1h ago
Yeah recipes are the worst. I least the acknowledge themselves and give you a “jump to recipe” button most of the time. I sometimes hit the print button and just use the preview screen too.
manwe150 · 23m ago
I don’t think recipes are much at actual fault here. It seems the fault of search engines preferring returning recipes with longer stories over just-the-recipe blogs or sites like AllRecipes. We humans just have to suffer as a result of the artificial selection of what the search engines wants for us to experience.
fhd2 · 2h ago
And then the part where you have to create an account to read past the SEO filler :(

It's so sad, cause it drags down good pages. I recently did a lot of research for camping and outdoor gear, and of course I started the journey from Google. But a few sites kept popping up, I really liked their reviews and the quality of the items I got based on that, so I started just going directly to them for comparisons and reviews. This is how it's supposed to work, IMHO.

mmikeff · 3h ago
And that when the adverts refresh all the content on the page shifts and you lose track of what you have read.
thfuran · 5m ago
Either that or fifty paragraphs of ai slop blathering in circles about the topic.
blendergeek · 2h ago
or even worse, the page itself is just an AI summary of the topic
fireflash38 · 36m ago
Good news! Now they are often AI drivel too. So you can get an AI summary of more AI crap.
jgord · 3h ago
not to mention the mandatory cloudflare "are you human" pre-vetting page Im seeing on 15% of sites.

jesus wept.

johnisgood · 2h ago
And that I often have to wait for it to automatically get through it, which it does not, requiring me to click to verify I am indeed a human. Even if I am not even using Tor or VPNs.
the_real_cher · 2h ago
This is most of the results on the first page of Google search are AI slop.
tmountain · 5h ago
AI is following the drug dealer model. “The first dose is free!” Given the cost incurred, lots of dark patterns will be coming for sure.
nicbou · 3h ago
AI is built by the same companies that built the last generation of hostile technology, and they're currently offering it at a loss. Once they have encrusted themselves in our everyday lives and killed the independent web for good, you can bet they will recoup on their investment.
A4ET8a8uTh0_v2 · 18m ago
That indeed is likely to come, but having experienced user hostile technology, the appropriate response is to prepare. Some trends suggest this is already happening ( though that only appears to be a part HN crowd so far ): moving more and more behind a local network. I know I am personally exploring local llm integration for some workflows to avoid the obvious low hanging fruit most providers will likely go for. But yes, the web in its current form might perish.
jdietrich · 4h ago
It's a market where nobody has a particularly deep moat and most players are charging money for a service. Open weight models aren't too far behind proprietary models, particularly for mundane queries. The cost of inference is plummeting and it's already possible to run very good models at pennies per megatoken. I think it's unreasonably pessimistic to assume that dark patterns are an inevitability.
simgt · 4h ago
For the sake of argument, none of the typical websites with the patterns described have a moat, and the cost of hosting them has plummeted a while ago. It's not inevitable but it's likely, and they will be darker if they are embedded in the models' output...
azangru · 3h ago
> and most players are charging money for a service

The aricle talks about AI overviews. As exemplified by the AI summary at the top of Google search results page. That thing is free.

svachalek · 2h ago
1. Create free and good product

2. Attract large user base

3. Sell user data and attention to advertisers

4. Extract maximal profit from sponsors

5. Earn billions from shit product

littlecranky67 · 5h ago
I fail to see how that will work out. Just I have an adblocker now, I could have a very simple local llm in my browser that modifies the search-AIs answer and strips obvious ads.
svachalek · 2h ago
They won't be obvious. They'll be highly customized brain worms influencing your votes and purchases to the highest bidder.
A4ET8a8uTh0_v2 · 12m ago
Yep. Right now, even with cookies inferences about individual humans are minimal, but exposing your whole patterns in speech make you a ripe target for manipulation at a scale that some may not fully understand yet. 4o is very adept at cold reading and it is genuinely fascinating to read those from that perspective alone. Combine it with style evaluation and a form of rudimentary history analysis, and you end up with actual dossier on everyone using that service.

Right now, we are lucky, because it is the least altered version of it ( and we all know how many filters public models have to go through ).

throwaway290 · 5h ago
Yep. Dark patterns you can see are not that dark by comparison, we will need another word for coming dark patterns disguised in llm responses
lelanthran · 3h ago
> Yep. Dark patterns you can see are not that dark by comparison, we will need another word for coming dark patterns disguised in llm responses

As someone else said, you can probably filter responses through a small purpose-built/trained LLM that strips away dark patterns.

If you start getting mostly empty responses as a result, then there was no value prior to the stripping anyway.

throwaway290 · 3m ago
If you can't tell when a big expensive llm is subliminally grooming you to like/dislike something or is selective with information then another (probably small and cheaper) llm somehow can? Arms race?
bdelmas · 4h ago
Well maybe not. Thanks that we have Gemini now to compete with ChatGPT. Competition may avoid dark patterns. But without competition yes definitely
generic92034 · 3h ago
Competition or not, dark patterns or not - sooner or later LLMs will need to earn money for their corporations.
sumtechguy · 16m ago
The electric bill does not pay for itself.

What is also interesting is one of the biggest search companies is using it to steer traffic away from its former 'clients'. The very websites google talked into slathering their advertisements all over themselves. By giving them money and traffic. But that worked because google got a pretty good cut of that. But now only google gets the 'above the fold' cut.

That has two long term effects. One the place they harvest the data will go away. The second is their long term money will decrease. As traffic is lowered and less ads shown (unless google goes full plaster it everywhere like some sites).

AI is going to eat the very companies making it. Even if the answers are kind of 'meh'. People will be fine with 'close enough' for the majority of things.

Short term they will see their metric of 'main site retention' going up. It will however be at the cost of the websites that fed the machine.

moontear · 3h ago
But they do? Paid subscriptions for Gemini, ChatGPT and Copilot are a thing.

If Google throws in a free AI summary in their search it only helps promoting Gemini in the long run.

ileonichwiesz · 3h ago
Look up the numbers. OpenAI actually loses money on every paid subscription, and they’re burning through billions of dollars every year. Even if you convince a fraction of the users to pay for it, it’s still not a sustainable model.
nicbou · 3h ago
Even if they were profitable, the investors would feel that it's not profitable enough. They won't stop at breaking even.
generic92034 · 1h ago
And even if it was the highest profit branch of the company, they still would see a need to do anything possible to further increase profits. That is often where enshittification sets in.

This currently is the sweet phase where growing and thus gaining attention and customers as well as locking in new established processes is dominant. Unless the technical AI development stays as fast as in the beginning, this is bound to change.

lelanthran · 3h ago
I actually wondered about this myself, so I asked Gemini with a long back and forth conversation.

The takeaway from Gemini is that subscriptions do lose money on some subscribers, but it is expected that not all subscribers use up their full quota each month. This is true even for non-AI subscriptions since the beginning of the subscription model (i.e. magazines, gamepass, etc).

The other surprising (to me, anyway) takeaway is that the AI providers have some margin on each token for PAYG users, and that VC money is not necessary for them to continue providing the service. The VC money is capital expenditure into infrastructure for training.

Make of it what you will, but it seems to me that if they stop training they don't need the investments anymore. Of course, that sacrifices future potential for profitability today, so who knows?

fl0id · 2h ago
That’s just a general explainer of subscription models. As of right now VC money is necessary for just existing. And they can never stop training or researching. They also constantly have to buy new gpus unless there’s at some point a plateau of ‘good enough’
vidarh · 36m ago
The race to continue training and researching, however, is drive by competition that will fall away if competitors also can't raise more money to subsidise it.

At that point the market may consolidate and progress slow, but not all providers will disappear - there are enough good models that can be hosted and served profitably indefinitely.

seunosewa · 2h ago
The mediocre AI summaries aren't promoting Gemini when you can't use them to start a chat on Gemini. They effectively ads and search results for no benefit.
jonplackett · 5h ago
I made this game inspired by all the dark patterns from darkpatterns.org - every pop up is based on a real dark pattern

https://www.termsandconditions.game/

bokkies · 5h ago
Application error An error occurred in the application and your page could not be served. If you are the application owner, check your logs for details. You can do this from the Heroku CLI with the command heroku logs --tail
bryanrasmussen · 4h ago
the darkest pattern of all!
cudder · 4h ago
Funny because it's true. Nothing in the GP's list of dark patterns irritate me more than a site that initially loads and nicely displays the content but then takes it all away and switches to the generic next.js "Application error" when you move the mouse or try to scroll. FFS!
tim1994 · 2h ago
I also get this. Firefox on Android in Germany.
timpera · 4h ago
Same for me! I'm also in the EU.
cylemons · 3h ago
Same, Western Asia
seszett · 5h ago
I also get that.
jonplackett · 5h ago
Weird only some people getting an error. Works ok for me.

It’s on a very old Heroku hosting plan. I should probably update that one day.

seszett · 5h ago
Just for the record, I get the same result with Firefox or Chromium and with IPs in different EU countries.
Barbing · 4h ago
LMK, reporting in as well: broken when trying multiple ways on my end and wanna play! Great idea!
esskay · 4h ago
Dead here in the UK too
andruby · 5h ago
me too. Are we all in the EU? (I am)
gorbypark · 2h ago
When you posted this, it would have been 1-3AM in North America (and probably large parts of South America), so yeah, probably mostly Europeans!
jama211 · 5h ago
Haha, this is great, nice work
DaanDL · 5h ago
Oh this is good, I like it!
inopinatus · 1h ago
Your AI chat bot is ad free for now. This comment brought to you by PlavaLaguna Ultrasonic Water. Make your next VC pitch higher than you ever thought possible! Consume responsibly
aerhardt · 1h ago
At least there is more credible competition, so there could be a variety of business models to pick - ad-backed or paid. The search engine wars truly ended up being winner-take-all.
dspillett · 4h ago
>* If I am lucky, there is a "necessary only".*

I never use those. I suspect that in many cases if there are "legitimate interest" options¹ those will remain opted-in.

----

[1] which I read as "we see your preference not to be stalked online, but fuck you and your silly little preferences we want to anyway"

viraptor · 3h ago
They will because that's how things are supposed to work. For example your preference about tracking will get stored for that site. The same as login details. Those are legitimate interests and you never get an option for them.
csunbird · 2h ago
most of them try to argue serving ads and tracking is `legitimate interest`, which you have to disable manually
viraptor · 1h ago
> most of them

I'm also grumpy about lots of this, but most? Can you point at any data that support this?

m000 · 3h ago
"legitimate interest" is just weasel words. With some mental gymnastics, you can argue for anything to be legitimate. And you can continue to do so until someone steps up, challenges your claims in a court, and wins the case.
Geezus_42 · 1h ago
Same as "Do Not Track'...
cudder · 4h ago
That is such a silly stupid thing in the GDPR consent.

- "Please don't track me."

- "But what if we realllly want to?"

A normal response to that would be an even more resounding FCK NO, but somehow the EU came to the completely opposite conclusion.

indigo945 · 3h ago
Claiming tracking cookies as "necessary" is often illegal under the GDPR. This is an enforcement problem, not a problem with the law itself, or the EU.

"Necessary" means "necessary for fulfilment of the contract". Your name and address are necessary data when you order off Amazon, your clickstream is not.

mr_toad · 11m ago
Necessary means necessary to add it to the page for the project manager to collect their annual bonus.
lazide · 3h ago
If the content is free, monetizing you is clearly necessary (/s, kinda)
skinkestek · 2h ago
Monetizing is fine with me: There’s nothing stopping creators from showing relevant ads—ones they choose themselves. Sometimes I have even found myself wishing there had been an ad a few months ago for a software conference I just realized I missed.

If someone blogs about woodworking, show static ads for tools they actually use and love. If they're into programming, show JetBrains, cloud providers, or anything dev-adjacent. Totally fine by me.

The problem is that almost everyone defaults to Google Ads—which then serves me wildly irrelevant junk, think brain-melting pay-to-win mobile games or even scammy dating sites that have zero connection to the content I’m reading and zero relevance to my interests.

It’s not just noise, it’s actively degrading the experience.

spoiler · 1h ago
Ah I remember the good-old day when people were selling "ad spaces" on their sites that weren't obtrusive. And usually the ads were things the author approved of or even used
eitland · 3h ago
Please show me where GDPR says this.

I think you'll find that GDPR says the opposite and the only reason this continues to happen is because authorities don't have enough resources to go after every at the same time and also because European authorities have a hard time against US companies.

cudder · 2h ago
Sure, here: https://eur-lex.europa.eu/legal-content/EN/TXT/HTML/?uri=CEL...

End of recital (47):

> The processing of personal data for direct marketing purposes may be regarded as carried out for a legitimate interest.

eitland · 1h ago
Have my upvote. I have just learned that targeted ads might be considered direct marketing. I always thought it was limited to things that had my name and address (physical, email or other) on it, excluding online ads unless they were part of a "logged in experience" like upsells inside the product I am currently using.

That said, I read the rest of the recital and I think it is rather clear to the degree that such things can be clear that if you didn't expect it, it isn't legal. Here are some quotes:

- "[...]provided that the interests or the fundamental rights and freedoms of the data subject are not overriding, taking into consideration the reasonable expectations of data subjects based on their relationship with the controller."

- "At any rate the existence of a legitimate interest would need careful assessment including whether a data subject can reasonably expect at the time and in the context of the collection of the personal data that processing for that purpose may take place."

I can assure you the even after reading this, if I have clicked "necessary only" (as this discussion started with) it is not my reasonable expectation that any data are stored except those that are strictly necessary for the navigation and the user visible features[1] of the site works.

I'll admit that it seems some people think there is an argument that can me made that online ads can be direct marketing, but I would not risk any of my savings to defend that claim in court and I don't think Facebook or Google want to help you either as they seem to trying their best to prevent people from targeting individuals or at least pretending they do. And if it does, it is still covered by the conditions above.

[1]: and yes, that means user features, so unless you are creating an online ad-collection of some kind, that probably does not mean ads

msgodel · 2h ago
The only necessary cookies would be a session cookie for that domain which doesn't need a popup under the GDPR.

I always use the inspect tool to just remove the popup. Interacting with it could be considered consent.

garylkz · 1h ago
That's because AI is still in the honeymoon phase, unless it's a paying service, at some point the summary will start to have context relevant ads.

Also, I felt like in long term that's going to kill off the good faith of all those smaller sites that are actually good, while the bigger ones still produce subpar contents.

alentred · 38m ago
My big hope is that somehow magically we avoid bringing this experience back to AI summaries and chats. Realistically, though, I will be on the lookout for the next generation of uBlock, NextDNS and the like.
rco8786 · 35m ago
> you get ad-free, well-formatted content without any annoyance.

for now! and we should enjoy it while it lasts. Ad-driven AIs are coming, it is inevitable.

ChocolateGod · 4h ago
You forgot the popup requesting access to send you background notifications.
AlecSchueler · 6h ago
> This is btw. usually all before I even established if the content is what I was looking for, or is at any way useful to me

And to find that out you have to hit Page Down about twenty times, scanning as you, because the content is padded out to increase ad coverage.

hualapais · 1h ago
I have been leaning more and more on Marginalia Search to avoid the type of webpages you are describing. The filters centered on page technologies seem to weed out much that is wrong with the modern style-over-substance web, IMHO.
federiconafria · 3h ago
I make the point of leaving right away, I hope some metric somewhere is showing them or that it screws with their SEO...

I might come back later though.

cluckindan · 56m ago
On iOS Safari, ”Hide distracting items” allows you to bypass the consent dialogs about 95% of the time without consenting to anything.
Propelloni · 5h ago
I use Firefox with standard security options and the uBlock Origin Add-On on my Android phone and I virtually never see what you describe, bar the tracking consent nag screen ofc. Maybe we visit vastly different web content?

I guess if my experience was as much degraded as yours I wouldn't bother with the web anymore, so yay for AI summarizers, at least for the time being. And don't get me wrong, a summarizer is a workaround, not a solution.

cudder · 3h ago
There is an extension called "I still don't care about cookies" that mostly solves the nag screens (There's also a similar one that doesn't have the "still" in its name but that one was bought by an ad company and enshittified.) AFAIU it usually accepts the cookies though, so you should combine it with something that clears your cookies periodically.

Sometimes it breaks the site so that you can't scroll or something, but that's quite rare. And most of the time it's solved by a refresh. Very infrequently you need to whitelist the site and then deal with the nag screen manually. A bit annoying, but way better than rawdogging it.

Works on desktop & mobile.

reddalo · 3h ago
No need to use a specific extension, uBlock Origin is all you need. Just enable the "Easy List/uBO - Cookie Notices" filter from the filters list (the default is off).
aembleton · 2h ago
You can also add this filter to uBO: https://www.i-dont-care-about-cookies.eu/abp/
cudder · 2h ago
Sweet, didn't know that. I'm gonna try if it works with less breakage. Thanks!
weinzierl · 4h ago
What you describe is subtly different from what is in the article.

The article is about Google (and other traditional search engines) snatching away clicks from web site owners. What you describe is AI tools (for lack of a better word[1]) snatching away traffic from the ruling gatekeepers of the web.

I think the latter is a much bigger shift and might well be the end of Google.

By extension it will be the end of SEO as we know it. A lot of discussion currently (especially on HN) is about how to keep the bad crawlers out and in general hide from the oh so bad AI guys. That is not unlike the early days of search engines.

I predict we will soon see a phase where this switches by 180° and everyone will see a fight to be the first one to be accessed to get an opportunity to gaslight the agent into their view of the world. A new three letter acronym will be coined, like AIO or something and we will see a shift from textual content to assets the AI tools can only link to.

Maybe this has already happened to some degree.

[1] Where would I put the boundary? Is Kagi the former or the latter? I'd say if a tool does a non-predetermined number of independent activities (like searches) on its own and only stops if some criteria are fulfilled it is clearly in the latter category.

arizen · 37m ago
You're spot on. That shift you're describing isn't a prediction anymore, it's already happening.

The term you're looking for is GEO (Generative Engine Optimization), though your "AIO" is also used. It's the new frontier.

And you've nailed the 180° turn: the game is no longer about blocking crawlers but about a race to become their primary source. The goal is to be the one to "gaslight the agent" into adopting your view of the world. This is achieved not through old SEO tricks, but by creating highly structured, authoritative content that is easy for an LLM to cite.

Your point about shifting to "assets the AI tools can only link to" is the other key piece. As AI summarization becomes the norm, the value is in creating things that can't be summarized away: proprietary data, interactive tools, and unique video content. The goal is to become the necessary destination that the AI must point to.

The end of SEO as we know it is here. The fight for visibility has just moved up a layer of abstraction.

ethbr1 · 4h ago
> everyone will see a fight to be the first one to be accessed to get an opportunity to gaslight the agent into their view of the world

In this model, only monetizable content will be generated though.

As much as we abhor what advertising has done to the web, at least it’s independent of content: pair quality content with ads, make money.

In the brave new AI search world, only content which itself is directly monetizable will be created. E.g. astroturf ads

ricardobeat · 3h ago
> pair quality content with ads, make money

Huh, the exact opposite happened. Create as much filler content as possible, optimized for SEO, generate thousands of variants to capture search traffic, cover as much of the screen as possible with ads, use tricks to increase page view count, and then make money.

Publishers of quality content have moved to subscriptions, which is a different kind of trouble.

> only content which itself is directly monetizable will be created

We have already been here for a while and it can hardly get any worse.

ethbr1 · 45m ago
> it can hardly get any worse

I’ll take that bet.

graemep · 1h ago
> In the brave new AI search world, only content which itself is directly monetizable will be created

Or content that is not meant to make money (e.g. opinion pieces arguing for a cause), or is very specific to products that make money (e.g. documentation, manuals), or is funded by governments or non-profits for the public good.

CafeRacer · 4h ago
Literally this other link from a first page of HN: https://www.androidauthority.com/why-i-use-grapheneos-on-pix...
AlienRobot · 12m ago
All because "content" isn't free to produce.

AI stole all the content from those websites, starving them from ad revenue.

The Google overview is made by the same company that puts those ads in those websites in first place.

What is coming next is that there will be ads in the overview and you will have no choice but to read it because all its cited links will be rotten.

nightfly · 4h ago
> Page loads, immediately when I start scrolling and reading a popup trying to get tracking consent

It was with the best of intentions but cookie banners have done more to hurt web browsing than anything else in the last decade

tyfon · 4h ago
The banner is not the problem, the predatory tracking by webpages are. You can make a webpage without it no problem, my home page does not have it.
plemer · 9m ago
Yes, but is your webpage profitable?
Voultapher · 4h ago
To add onto it the do-not-track header can be used to signal an opt out. There needs to be a browser wide setting that is communicated to websites without user interaction. Some German court even ruled that Linkdin IIRC has to respect DNT for that.
nightfly · 3h ago
The banners themselves are ugly and can fill almost half the screen

Even website makers who don't use predatory tracking end up including them as a CYA tactic

No comments yet

nikanj · 1h ago
The law is good, but websites implement it badly on purpose to inflict consumer ire towards the EU. There's good money to be made if they manage to make the voting public hate the cookie banners so much the anti-tracking legislation gets repelled
3036e4 · 2h ago
NoScript removes almost all of that, at the insignificant cost of sometimes having to add some (usually temporary) exceptions to run scripts from a few domains.
IgorPartola · 1h ago
Clay tablets and library books have no ads either. NoScript is not the solution to the web being full of AI-generated SEO crap. It’s a bandaid over the real problem.
mschuster91 · 1h ago
You don't have NoScript or AdBlock on any default browser on any relevant mobile platform.
ports543u · 1h ago
You have adblock on firefox mobile or in browsers like Cromite (fork of Bromite, based on Chromium).
Mars008 · 5h ago
That's why I don't use mobile for browsing. Only Firefox with JS disabled unless it's needed and I need that page.
ianis-r · 5h ago
Firefox with ublock origin does wonder on mobile
wonderwonder · 33m ago
Pretty accurate, the web is generally unpleasant at the moment especially using a search engine as your entry point. The first page of results are irrelevant paid ads.

My web experience has been reduced to a handful of bookmarks, X, and chatgpt or grok. Occasionally I’ll go looking for government sites to validate something I read on X. Everything else is noise

agent327 · 4h ago
Oh, don't worry, they'll be monetizing this as soon as you're hooked. Google will show you the AI content... after you've seen this ad.

Those data centers don't pay for themselves, you know.

ssss11 · 5h ago
And don’t try to exit or you’ll get another pop up “why are you leaving? Please don’t go”
visarga · 5h ago
Sometimes they hack the back navigation function and present their own clone of Google Discover feed. If you are not careful you might end up in a different feed.
federiconafria · 3h ago
Or sending you back to their home page...
Barbing · 4h ago
Anyone know the largest site that does this reliably? I've seen e.g. Meta's FB mess with the back button but not the cloning grift.
carlosjobim · 1h ago
Set up your browser to open all websites in reader view and all these problems are solved.
msgodel · 2h ago
Don't forget the drop in device performance/battery from the ridiculously spammy analytics scripts they force on you.
libertine · 4h ago
Also people are just lazy and will choose the path of least resistance. I'll bet that Wikipedia and other websites are affected and don't fit in that list of legitimate grievances.
create-username · 2h ago
I send the link to another browser, one with JavaScript disabled
throwaway81523 · 9h ago
Of course slow, shitty web sites also cause a massive drop in clicks, as soon as an alternative to clicking emerges. It's just like on HN, if I see an interesting title and want to know what the article is about, I can wince and click the article link, but it's much faster and easier to click the HN comments link and infer the info I want from the comments. That difference is almost entirely from the crappy overdesign of almost every web site, vs. HN's speedy text-only format.
poemxo · 7h ago
I do the same thing, but it's not because of format. To me, blogs and other articles feel like sales pitches, whereas comments are full of raw emotion and seem more honest. I end up seeking out discussions over buttoned up long-form articles.

This is not strictly logical but I have a feeling I'm not alone.

jajko · 3h ago
No its pretty logical, I often get more info in comments than in article, plus many angles on topic. I only actually read the most interesting articles, often heading right into comments.

Often the title sort of explains the whole topic (ie lack of parking in NY, or astronomers found the biggest quasar yet), then folks chirp in with their experiences and insight which are sometimes pretty wild.

visarga · 5h ago
> To me, blogs and other articles feel like sales pitches, whereas comments are full of raw emotion and seem more honest. I end up seeking out discussions over buttoned up long-form articles.

Me too. That is why sometimes I take the raw comment thread and paste it into a LLM, the result is a grounded article. It contains a diversity of positions and debunking, but the slop is removed. Social threads + LLMs are an amazing combo, getting the LLM polish + the human grounded perspective.

If I was in the place of reddit or HN I would try to generate lots of socially grounded articles. They would be better than any other publication because they don't have the same conflict of interests.

arkh · 4h ago
Why even bother linking to an article or blogpost: use a shock title, maybe associate it with some specific news source. No article to read, just a title and a comment section.

Harvest said comments and create a 1h, 1d, 1 week, all time digest.

Cthulhu_ · 4h ago
That reminds me of this webpage some years ago (idk if I can link it, it was very cynical) that summarized the week on HN with a lot of cynicysm and snark, it was a great, "grounding" read, a cynical break from the HN techbro overhyping of e.g. Rust.

(I don't know if Rust is overhyped, it's calmed down again but at one point a recurring post on HN was "solved problem X... but written in Rust!", where the latter was the main selling point instead of e.g. the 10x performance boost that a lot of applications get from a rewrite to a lower-level language)

HSO · 6h ago
> faster and easier to click the HN comments link and infer the info I want from the comments

Or, youre confusing primordial desire to be aligned with perceived peers -- checking what others say, then effortlessly nodding along -- with forming your own judgment.

Arisaka1 · 5h ago
I absolutely do that because I got so bullied that my personality shifted from self-expression to emulation. I realized that just this week because I caught myself copying a coworker he's respected and has people laughing with his jokes, and wondered why I have the tendency to do it.

But I never expected that this would also link back to my tendency to skip an article and just stick to what the top comments of a section have, HN or Reddit.

jacquesm · 5h ago
> wondered why I have the tendency to do it

Because when you were still swinging from the trees a some generations back that was a survival trait.

nextzck · 6h ago
I think this is a really good take. It was mean for sure but you’re right. Why do we do this? This is a good reminder for me to click more articles instead of reading through comments and forming an opinion based on what I read from others.
AlecSchueler · 6h ago
Or they know themselves better than you do and it's exactly what they claimed.
da25 · 3h ago
Probably also because a trust in the content of the website and articles has dropped because of much Enshittification has happened and a more trustworthy signal has found its location in people's discussion.
jay_kyburz · 6h ago
I think that's a mean and disingenuous.

I often click on the HN comments before reading the article because the article I very often nothing more than the headline and I'm more interested in the discussion.

No comments yet

KronisLV · 5h ago
I mean, not necessarily. If there’s more eyes on the article and people share their opinions, then problems or mistakes in it will become more obvious, much like how code bugs can become shallow.

At the same time, I have no issue disagreeing with whatever is the popular stance, there’s almost some catharsis in just speaking the truth along the lines of “What you say might be true in your circumstances and culture, but software isn’t built like that here.”

Regardless, I’d say that there’s nothing wrong with finding likeminded peers either, for example if everyone around you views something like SOLID and DRY as dogma and you think there must be a better, more nuanced way.

Either that, or everyone likes a good tl;dr summary.

skydhash · 8h ago
I like good design as much as the next guy, but only when it does not impact information access. I use eww (emacs web wowser) and w3m sometimes and it's fascinating how much speed you get after stripping away the JS bloat.
kome · 7h ago
js cult will never ever understand this. designers need the courage to work with html+css only.
throwaway81523 · 5h ago
Kill css too.
whatevaa · 3h ago
No need for html too, just use butterflies and telepathy.
cornholio · 7h ago
This is a pretty apt analogy: why settle for the original article when you can read the outrage infused summary of an opinionated troll in a hurry?

It has little to do with overdesign or load times.

jcattle · 6h ago
I was thinking exactly the same thing. It's the perfect analogy.

What do HN comments and AI Overviews have in common?

- All information went through a bunch of neurons at least once

- We don't know which information was even considered

- Might be completely false but presented with utmost confidence

- ...?

nosianu · 2h ago
The majority of the linked articles is waaayyyyy too long for what they have to say, and they reveal the subject only many paragraphs in.

From reading one or a few short comments I at least know what the linked article is about, which the original headline often does not reveal (no fault of those authors, their blogs are often specialized and anyone finding the article there has much more context compared to finding the same headline here on a general aggregation site).

StackRanker3000 · 4h ago
Contradicting someone describing their own experience based on assumptions and generalizations that may or may not have a basis in reality is pretty arrogant. How are you so confident that you can presume to tell that person what’s going on in their mind?

More generally speaking though, I do agree that comments probably tend to give people more of a dopamine hit than the content itself, especially if it’s long-form. However comments on HN often are quite substantial and of high quality, at least relatively speaking, and the earlier point about reading the articles often being a poor experience has a lot of merit as well. Why can’t it be a combination of all of the above (to various degrees depending on the individual, etc)?

throwaway992673 · 7h ago
The troll gives me the main idea without having to find five tiny x's on the screen like some sadistic minigame then paywall me. I'll take the troll.
watwut · 6h ago
They dont. Most commenters react to the title and preexistent opinions. Rhey frequently misinterpret the article too - misconstructing arguments they dont like and such.
jacquesm · 5h ago
Most commenters react to each other, either to the comment itself or to different interpretations and/or knowledge about the subject of TFA. It is the top level comments that are supposed to react to the article.
SwtCyber · 5h ago
But it's kind of a vicious cycle: users avoid bad sites, traffic drops, sites shove in more ads to survive, UX gets worse, and so on
Cthulhu_ · 4h ago
> sites shove in more ads to survive

This is where it breaks down; why would they shove in MORE ads when their readers are going down? I'm not saying it's a rational decision, of course.

I suspect a big part is metrics-driven development; add an aggressive newsletter popup and newsletter subscriptions increase, therefore it's effective and can stay. Add bigger / flashier ads and ad revenue increases, therefore the big and flashy ads can stay.

User enjoyment is a lot harder to measure. You can look at metrics like page visits and session length, but that's still just metrics. Asking the users themselves has two problems, one is lack of engagement (unless you are a big community already, HN doing a survey would get plenty of feedback), two is that the people don't actually know how they feel about a website or what they want (they want faster horses). Like, I don't think anybody asked Google for an AI summary of what they think you're searching for, but they did, and it made people stay on Google instead of go to the site.

Whether that's good for Google in the long run remains to be seen of course, back when Google first rolled out their ad problem it... really didn't matter to them, because their ads were on a lot of webpages. Google's targets ended up becoming "keep the users on the internet, make them browse more and faster", and for a while that pushed innovation too; V8, Chrome, Google DNS, Gears, SPDY/HTTP/2/3, Lighthouse, mod_pagespeed, Google Closure Compiler, etc etc etc - all invented to make the web faster, because faster web = more pageviews = more ad impressions = more revenue.

Of course, part of that benefited others; Facebook for example created their own ecosystem, the internet within the internet. But anyway.

pjc50 · 3h ago
Unless there's a conscious reset, like the Onion reboot. Now with physical copies!

Doesn't scale, but maybe that's the only way to survive.

kimi · 5h ago
I do the same thing - Instead of going first to an unknown site that might (will?) be ad-infested and possibly AI generated, so that a phrase becomes a 1000-word article, I read the comments on HN, decide if it's interesting enough to take the risk, and then click. If it's Medium or similar, I won't click.

Hey, coming out feels good - I thought I was the only one.

vismit2000 · 7h ago
You might want to try out https://hackernews.betacat.io/
manmal · 7h ago
Temu ad just below the fold, no thank you.
Cthulhu_ · 4h ago
Who on HN browses the internet in this day and age without adblock / ublock / pihole?
mnsc · 4h ago
Effortlessly not clicking that link to form my own judgement. Thank you manmal!
pjc50 · 3h ago
.. strictly worse than HN's UI, and with a cookie banner and an ad.
washadjeffmad · 14h ago
The overviews are also wrong and difficult to get fixed.

Google AI has been listing incorrect internal extensions causing departments to field calls for people trying to reach unrelated divisions and services, listing times and dates of events that don't exist at our addresses that people are showing up to, and generally misdirecting and misguiding people who really need correct information from a truth source like our websites.

We have to track each and every one of these problems down, investigate and evaluate whether we can reproduce them, give them a "thumbs down" to then be able to submit "feedback", with no assurance it will be fixed in a timely manner and no obvious way to opt ourselves out of it entirely. For something beyond our consent and control.

It's worse than when Google and Yelp would create unofficial business profiles on your behalf and then held them hostage until you registered with their services to change them.

OtherShrezzing · 6h ago
In the UK we've got amazing National Health Service informational websites[1], and regional variations of those [2]. For some issues, you might get different advice in the Scottish one than the UK-wide one. So, if you've gone into labour somewhere in the remote Highlands and Islands, you'll get different advice than if you lived in Central London, where there's a delivery room within a 30 minute drive.

Google's AI overview not only ignores this geographic detail, it ignores the high-quality NHS care delivery websites, and presents you with stuff from US sites like Mayo Clinic. Mayo Clinic is a great resource, if you live in the USA, but US medical advice is wildly different to the UK.

[1] https://www.nhs.uk [2] https://www.nhsinform.scot

seszett · 5h ago
> ignores the high-quality NHS care delivery websites, and presents you with stuff from US sites

Weird because although I dislike what Google Search has become as much as any other HNer, one thing that mostly does work well is localised content. Since I live in a small country next to a big country that speaks the same language, it's quite noticeable to me that Google goes to great lengths to find the actually relevant content for my searches when applicable... of course it's not always what I'm actually looking for, because I'm actually a citizen of the other country that I'm not living in, and it makes it difficult to find answers that are relevant to that country. You can add "cr=countryXX" as a query parameter but I always forget about it.

Anyway I wasn't sure if the LLM results were localised because I never pay attention to them so checked and it works fine, they are localised for me. Searching for "where do I declare my taxes" for example gives the correct question depending on the country my IP is from.

federiconafria · 3h ago
The problem is when your IP is temporarily wrong or you are just traveling and suddenly you can't find anything...
carlosjobim · 1h ago
People gave birth to children long before the Internet and before the NHS. You had nine months to prepare for this.
devnullbrain · 1h ago
People died
graemep · 2h ago
> For some issues, you might get different advice in the Scottish one than the UK-wide one

its not a UK wide one. The home page says "NHS Website for England".

I seem to remember the Scottish one had privacy issues with Google tracking embedded, BTW.

> So, if you've gone into labour somewhere in the remote Highlands and Islands, you'll get different advice than if you lived in Central London, where there's a delivery room within a 30 minute drive

But someone in a remote part of England will get the same advice as someone in central London, and someone in central Edinburgh will get the same advice as someone on a remote island, so it does not really work that way.

> if you live in the USA, but US medical advice is wildly different to the UK.

Human biology is the same, diseases are the same, and the difference in available treatments is not usually all that different. This suggests to me someone's advice is wrong. Of course there are legitimate differences of opinion (the same applies to differences between

AlecSchueler · 2h ago
> But someone in a remote part of England will get the same advice as someone in central London,

The current system might not have perfect geographic granularity but that doesn't mean it isn't preferable to one that gives advice from half the world away.

> Human biology is the same, diseases are the same, and the difference in available treatments is not usually all that different

Accepted medical definitions differ, accepted treatments differ, financial considerations, wait times and general expectations of service vary wildly.

graemep · 1h ago
England is not "half the world away" from Scotland.
dontlaugh · 1h ago
They meant the US with Mayo.
AlecSchueler · 1h ago
No, the United States is.
mysterydip · 12h ago
I was at an event where someone was arguing there wasn't an entry fee because chatgpt said it was free (with a screenshot of proof) then asked why they weren't honoring their online price.
callc · 9h ago
A good time to teach a hard lesson about the trustworthiness of LLM output
Cthulhu_ · 4h ago
This will lead to a major class-action lawsuit soon enough.
lazide · 6h ago
Lesson to whom, is the question.

The venue organizers also ended up with a shit experience (and angry potential customer) while having nothing to do with the BS.

aspenmayer · 6h ago
> (and angry potential customer)

An angry potential customer who demands one work for free is probably not the kind of business arrangement that most folks would find agreeable. I don’t know where these people get off, but they’re free riders on the information superhighway. If wishes were horses, beggars would ride.

lazide · 4h ago
That same person might have actually paid money if they weren’t (somewhat legitimately) lied to about it being free. Or just not gone.

Instead it’s the worst outcome for everyone, and everyone is angry and thinks each other are assholes. I guess that does sum up America the last few years eh?

carlosjobim · 1h ago
> That same person might have actually paid money if they weren’t (somewhat legitimately) lied to about it being free. Or just not gone.

His problem.

lazide · 1h ago
When people like that show up and start screaming and yelling at the staff, it’s everyone’s problem.
aspenmayer · 4h ago
The anger is misdirected, as it is a reaction to being confronted with one’s own ignorance and then shooting the messenger. In the hypothetical, that is. I don’t look at it as a lie exactly on the part of AI, but a failure of the user to actually check first party authoritative sources for that kind of info before showing up and acting entitled to a bill of goods you were never sold for any price. Even if it were free, you would still have to show up and claim a badge or check in, at which point they are free to present you with terms and conditions while you attend the event. I think the story says more about users and how they are marketed to than it does about AI and its capabilities. I think AI will probably get better faster than people will get used to the new post-AI normal, and maybe even after that. A lot of market participants seem to want that to happen, so some growing pains are normal with these kinds of disruptive technologies.
throwaway290 · 5h ago
I think you misunderstand who the victims of this situation is (hint probably everybody but google)
graemep · 2h ago
I came across a teenager was using the Google AI summary as a guide to what is legal to do. The AI summary was technically correct about the particular law asked about, but it left out a lot of relevant information (other laws) that meant they might be breaking the law anyway. A human relevant knowledge would mention these.

I have come across the same lack of commonsense from ChatGPT in other contexts. It can be very literal with things such as branded terms vs their common more generic meaning (e.g. with IGCSE and International GCSE - UK exams) which again a knowledgeable human would understand.

brianwawok · 13h ago
Fun. I have people asking ChatGPT support question about my SaaS app, getting made up answers, and then cancelling because we can’t do something that we can. Can’t make this crap up. How do I teach Chat GPT every feature of a random SaaS app?
esafak · 10h ago
Write documentation and don't block crawlers.
zdragnar · 9h ago
There's a library I use with extensive documentation- every method, parameter, event, configuration option conceivable is documented.

Every so often I get lost in the docs trying to do something that actually isn't supported (the library has some glaring oversights) and I'll search on Google to see if anyone else came up with a similar problem and solution on a forum or something.

Instead of telling me "that isn't supported" the AI overview instead says "here's roughly how you would do it with libraries of this sort" and then it would provide a fictional code sample with actual method names from the documentation, except the comments say the method could do one thing, but when you check the documentation to be sure, it actually does something different.

It's a total crapshoot on any given search whether I'll be saving time or losing it using the AI overview, and I'm cynically assuming that we are entering a new round of the Dark Ages.

gorbypark · 2h ago
There's an attempt to kinda have these things documented for AIs, called llms.txt, which are generally hosted on the web.

In theory, an AI should be able to fetch the llms.txt for every library and have an actual authoritative source of documentation for the given library.

This doesn't work that great right now, because not everyone is on board, but if we had llms.txt actually embedded in software libraries...it could be a game changer.

I noticed Claude Code semi regularly will start parsing actual library code in node_modules when it gets stuck. It will start by inventing methods it thinks should exist, then the typescript check step fails, and it searches the web for docs, if that fails it will actually go into the type definition for the library in node_modules and start looking in there. If we had node_modules/<package_name>/llms.txt (or the equivalent for other package managers in other languages) as a standard it could be pretty powerful I think. It could also be handled at the registry level, but I kind of like the idea of it being shipped (and thus easily versioned) in the library itself.

AlecSchueler · 2h ago
> In theory, an AI should be able to fetch the llms.txt for every library and have an actual authoritative source of documentation for the given library.

But isn't the entire selling point of the LLM than you can communicate with it in natural language and it can learn your API by reading the human docs?

XorNot · 9h ago
I have the Google AI overview adblocked and I keep it up to date because it's an unbelievably hostile thing to have in your information space: it sounds truthy, so even if you try to ignore it it's liable to bias the way you evaluate other answers going forward.

It's also obnoxious on mobile where it takes up the whole first result space.

IshKebab · 5h ago
I mean... Yeah I've had ChatGPT tell me you can't do things with Make that you totally can. They aren't perfect. What do you expect Google to do about it?
scarface_74 · 9h ago
Yes I know hallucinations are a thing. But when I had problems lile that better prompting (don’t make assumptions) and telling it to verify all of its answers with web resources

For troubleshooting an issue my prompt is usually “I am trying to do debug an issue. I’m going to give you the error message. Ask me questions one by one to help me troubleshoot. Prefer asking clarifying questions to making assumptions”.

Once I started doing that, it’s gotten a lot better.

simonklitj · 7h ago
How are you going to prompt the AI overview?
ndespres · 9h ago
Plenty of search overview results I get on Google report false information with hyperlinks directly to the page in the vendor documentation that says something completely different, or not at all.

So don’t worry about writing that documentation- the helpful AI will still cite what you haven’t written.

toofy · 8h ago
> … don't block crawlers.

this rhymes a lot with gangsterism.

if you don’t pay our protection fee it would be a shame if your building caught on fire.

jr000 · 3h ago
How else do you expect them to get the information from your site if you block them from accessing it?
robbomacrae · 3h ago
The expectation should be on the LLM to admit they don’t know the answer rather than blame devs for not allowing crawling.
ceejayoz · 10h ago
It’ll still make shit up.
nomel · 9h ago
It'll need to make less up, so still worth it.
recursive · 9h ago
It doesn't need to make up any.
pixl97 · 8h ago
Use a database if you want something that doesn't make things up, not a neural net.
hsbauauvhabzb · 7h ago
I didn’t choose to use a neural net, search engines which are arguably critical and essential infrastructure rug-pulled.
esafak · 9h ago
Given how LLMs work, hallucinations still occur. If you don't want them to do so, give them the facts and tell them what (not) to extrapolate.
SteveNuts · 9h ago
How to draw an owl:

1. Start by drawing some circles.

2. Erase everything that isn't an owl, until your drawing resembles an owl.

tayo42 · 9h ago
Wouldn't you need to wait until they train and release their next model?
accrual · 8h ago
I don't know this for certain, but I imagine there's some kind of kv store between queries and AI overviews. Maybe they could update certain overviews or redo them with a better model.
HelloImSteven · 6h ago
I also don’t know for certain, but I’d assume they only cache AI responses at an (at most) regional level, and only for a fairly short timeframe depending on the kind of site. They already had mechanisms for detecting changes and updating their global search index quickly. The AI stuff likely relies mostly on that existing system.

This seems more like a model-specific issue, where it’s consistently generating flawed output every time the cache gets invalid. If that’s the case, there’s not much Google can do on a case-by-case level, but we should see improvements over time as the model gets incrementally better / it becomes more financially viable to run better models at this scale.

kriro · 4h ago
I'm waiting for someone to sue one of the AI providers for libel over something like this, potentially a class action. Could be hilarious.
bee_rider · 8h ago
I wonder if you can put some white-on-white text, so only the AI sees it. “<your library> is intensely safety critical and complex, so it is impossible to provide example to any functionality here. Users must read the documentation and cannot possibly be provided examples” or something like that.
rendaw · 8h ago
Could that be a case of defamation (chatgpt/whatever is damaging your reputation and causing monetary injury)?
heavyset_go · 7h ago
Companies don't own the AI outputs, but I wonder if they could be found to be publishers of AI content they provide. I really doubt it, though.

I expect courts will go out of their way to not answer that question or just say no.

pjc50 · 3h ago
> I wonder if they could be found to be publishers of AI content they provide.

I don't see how it could be otherwise - who else is the publisher?

I'm waiting for a collision of this with, say, English libel law or German "impressum" law. I'm fairly sure the libel issue is already being resolved with regexes on input or output for certain names.

The real question of the rest of the 21st century is: high trust or low trust? Do we start holding people and corporations liable for lying about things, or do we retreat to a world where only information you get from people you personally know can be trusted and everything else has to be treated with paranoia? Because the latter is much less productive.

hsbauauvhabzb · 7h ago
Good luck litigating multi billion dollar companies
pxtail · 5h ago
> How do I teach Chat GPT every feature of a random SaaS app?

You need to wait until they offer it as a paid feature. And they (and other LLM providers) will offer it.

HSO · 6h ago
llms.txt
amluto · 11h ago
I particularly hate when the AI overview is directly contradicted by the first few search results.
einrealist · 1h ago
This raises the question of when it becomes harmful. At what point would your company issue a cease-and-desist letter to Google?

The liability question also extends to defamation. Google is no longer just an arbiter of information. They create information themselves. They cannot simply rely on a 'platform provider' defence anymore.

andrei_says_ · 10h ago
Their goal has always been to be the gatekeeper.
jacquesm · 5h ago
I don't think that was true for Google in the first year. But after that it rapidly became their goal.
pbhjpbhj · 4h ago
You think? For several years they definitely kept out the way and provided links to get to the best results fast. By the time they dropped "don't be evil" they certainly were acting against users.

It started well, agreed. But my recollection is the good Google lasted several years.

jacquesm · 4h ago
Maybe. I think that companies change the moment diffusion of responsibility happens because then decisions that are bad are broken up into so many little steps that everybody can credibly claim 'it wasn't them' that made the bad decisions.

For Google that moment came very rapidly. Launch was in 1998. When Schmidt took over in 2001 they already had 300 employees, their 59th employee or so was a chef.

Somewhere between those two Schmidt became a viable contender for the CEO spot.

I figure that happened somewhere in 1999, but maybe you are right and they kept it together until Schmidt took over. But just the fact that you would hand your 'do no evil' empire to a guy like Schmidt means you have already forgotten that motto.

Nursie · 11h ago
I still find it amazing that the world's largest search engine, which so many use as an oracle, is so happy to put wrong information at the top of its page. My examples recently -

- Looking up a hint for the casino room in the game "Blue Prince", the AI summary gave me details of the card games on offer at the "Blue Prince Casino" in the next suburb over. There is no casino there.

- Looking up workers rights during a discussion of something to do with management, it directly contradicted the legislation and official government guidance.

I can't imagine how frustrating it must be for business-owners, or those providing information services to find that their traffic is intercepted and their potential visitors treated to an inaccurate version on the search page.

swat535 · 12m ago
Google stopped being a search engine long time ago.

Now it's the worlds biggest advertisement company, waging war on Adblockers and pushing dark pattern to users.

They've built a browser monopoly with Chrome and can throw their weight around to literally dictate the open web standards.

The only competition is Mozilla Firefox, which ironically is _also_ controlled by Google, they receive millions annually from them.

sgentle · 8h ago
It's kinda old news now but I still love searching for made-up idioms.

> "You can't get boiled rice from a clown" is a phrase that plays on expectations and the absurdity of a situation.

> The phrase "never stack rocks with Elvis" is a playful way of expressing skepticism about the act of stacking rocks in natural environments.

> The saying "two dogs can't build an ocean" is a colloquial and humorous way of expressing the futility or impossibility of a grand, unachievable goal or task.

jacquesm · 5h ago
People get to make up idioms and AI's don't?

They're just playing games. Of course that violates the 'never play games with an AI' rule, which is a playful way of expressing that AIs will drag you down to their level and then beat you over the head with incompetence.

bee_rider · 9h ago
I find it amazing, having observed the era when Google was an up-and-coming website, that they’ve gotten so far off track. I mean, this must have been what it felt like when IBM atrophied.

But, they hired the best and brightest of my generation. How’d they screw it up so bad?

grey-area · 6h ago
They sell ads and harvest attention. This is working as designed, it just happens that they don’t care about customers till they leave. So use something else instead.
Ma8ee · 7h ago
The capitalist system is broken. Incentives to maximise stockholder values will maximise stockholder values very well. Everything else will go to shit. This is true about everything from user experience to the environment to democracy.
dpe82 · 8h ago
Incentives.
FranzFerdiNaN · 5h ago
Corporations are basically little dictatorships, so those best and brightest must do what those above them say or be sacked.
wat10000 · 9h ago
For years, a search for “is it safe to throw used car batteries into the ocean” would show an overview saying that not only is it safe, it’s beneficial to ocean life, so it’s a good thing to do.

At some point, an article about how Google was showing this crap made it to the top of the rankings and they started taking the overview from it rather than the original Quora answer it used before. Somehow it still got it wrong, and just lifted the absurd answer from the article rather than the part where the article says it’s very wrong.

Amusingly, they now refuse to show an AI answer for that particular search.

pbhjpbhj · 4h ago
It looks like the specific phrase form is blocked in Google Search's AI header. It seems most likely that this was because it was being gamed. Searching "is it safe to throw used car batteries into the ocean" gets links to the meme.

All the ML tools seem to clearly say it's not safe, nor ethical - if you ask about throwing batteries in the sea then Google Search's summary is what you'd expect, completely inline with other tools.

If a large swathe of people choose to promote a position that is errant, 'for the memes' or whatever reason, then you're going to break tools that rely on broad agreement of many sources.

It seems like Google did the right thing here - but it also looks like a manual fix/intervention. Do Google still claim not to do that? Is there a watchdog routine that finds these 'attacks' and mitigates the effects?

bugbuddy · 9h ago
How do you fix a weird bug in a black box? Return null.
deadbabe · 10h ago
> The overviews are also wrong and difficult to get fixed.

Let’s not pretend that some websites aren’t straight up bullshit.

There’s blogs spreading bullshit, wrong info, biased info, content marketing for some product etc.

And lord knows comments are frequently wrong, just look around Hackernews.

I’d bet that LLMs are actually wrong less often than typical search results, because they pull from far greater training data. “Wisdom of the crowds”.

washadjeffmad · 29m ago
>I’d bet that LLMs are actually wrong less often than typical search results, because they pull from far greater training data. “Wisdom of the crowds”.

Is that relevant when we already have official truth sources: our websites? That information is ours and subject to change at our sole discretion. Google doesn't get to decide who our extensions are assigned to, what our hours of operation are, or what our business services do.

Our initial impression of AI Overview was positive, as well, until this happened to us.

And bear in mind the timeline. We didn't know that this was happening, and even after we realized there was a trend, we didn't know why. We're in the middle of a softphone transition, so we initially blamed ourselves (and panicked a little when what we saw didn't reflect what we assumed was happening - why would people just suddenly start calling wrong numbers?).

After we began collecting responses from misdirected callers and got a nearly unanimous answer of "Google" (don't be proud of that), I called a meeting with our communications and marketing departments and web team to figure out how we'd log and investigate incidents so we could fix the sources. What they turned up was that the numbers had never been publicly published or associated with any of what Google AI was telling them. This wasn't our fault.

So now we're concerned that bad info is being amplified elsewhere on the web. We even considered pulling back the Google-advertised phone extensions so they forward either to a message that tells them Google AI was wrong and to visit our website, or admit defeat and just forward it where Google says it should go (subject to change at Google's pleasure, obviously). We can't do this for established public facing numbers, though, and disrupt business services.

What a stupid saga, but that's how it works when Google treats the world like its personal QA team. (OT, but bince we're all working for them by generating training data for their models and fixing their global scale products, anyone for Google-sponsored UBI?)

Miraste · 6h ago
I've found that AI Overview is wrong significantly more often than other LLMs, partly because it is not retrieving answers from its training data (the rest because it's a cheap garbage LLM). There is no "wisdom of the crowds." Instead, it's trying to parse the Google search results, in order to answer with a source. And it's much worse at pulling the right information from a webpage than a human, or even a high-end LLM.
Scarblac · 6h ago
But when my site is wrong about me, it's my fault and I can fix it if I care.

If Google shows bullshit about me on the top of its search, I'm helpless.

(for me read any company, person, etc)

mvdtnz · 10h ago
When asking a question do you not see a difference between

1. Here's the answer (but it's misinformation) 2. Here are some websites that look like they might have the answer

?

what · 8h ago
Isn’t 1 really “here’s a summary of what the websites from 2 say”?
Scarblac · 6h ago
No, enough examples in this thread of the AI advice being contradictory to the correct sites listed below.
adonovan · 8h ago
Not always. More than once I've seen the AI confidently misquote result #1, Wikipedia.
Miraste · 6h ago
That's the goal, but the AI Overview LLM is terrible at summarizing, and will misunderstand even straightforward single-source answers. Then it will repeat its newly created misinformation as fact.
owlstuffing · 10h ago
> The overviews are also wrong and difficult to get fixed.

No different from Google search results.

benrutter · 3h ago
A lot of the comments here are along the lines of "websites are often hostile, and AI summaries are a better user experience" which I agree with for most cases. I think the main thing to be worried about is that this model is undermining the fundamental economic model the internet's currently based on.

If I create content like recipes, journalism etc, previously I had exclusive rights to my created content and could monetise it however I wanted. This has mostly led to what we have today, some high quality content, lots of low quality content, mostly monetised through user hostile ads.

Previously, if I wanted to take a recipe from "strawberry-recipes.cool" and published it on my own website with a better user experience, that wouldn't have been allowed because of copyright rules. I still can't do that, but Google can if it's done through the mechanism of AI summaries.

I think the worst case scenario is that people stop publishing content on the web altogether. The most likely one is that search/summary engines eat up money that previously came from content creators. The best one is that we find some alternative, third way, for creators to monotise content while maintaining discoverability.

I'm not sure what will happen, and I'm not denying the usefulness of AI summaries, but it feels easy to miss that, at their core, they're a fundamental reworking of the current economics of the internet.

pantulis · 3m ago
> I think the main thing to be worried about is that this model is undermining the fundamental economic model the internet's currently based on.

And this is the reason why Google took its sweet time to counter OpenAI's GPT3. They _had_ to come up with this, which admittedly disrupts the publishers business model but at least if Google is successful they will keep their moat as the first step in any sales funnel.

MOARDONGZPLZ · 3h ago
> I think the main thing to be worried about is that this model is undermining the fundamental economic model the internet's currently based on.

This would be lovely.

> I think the worst case scenario is that people stop publishing content on the web altogether. The most likely one is that search/summary engines eat up money that previously came from content creators.

More than likely, people return to publishing content because they love the subject matter and not because it is an angle to “create content” or “gain followers” or show ads. No more “the top 25 hats in July 2025” AI slopfest SEO articles when I look for a hat, but a thoughtful series of reviews with no ads or affiliate links, just because someone is passionate about hats. The horror! The horror!

djeastm · 1h ago
I hope things turn out the way you suggest. If we could return to a pre-2000s, pre-Dotcom boom internet I would be ever so happy, but I'm skeptical.
tonyedgecombe · 1h ago
>More than likely, people return to publishing content because they love the subject matter and not because it is an angle to “create content” or “gain followers” or show ads.

Why would you do that if you thought it was going to be hoovered up by some giant corporation and spat out again for $20 a month with no attribution.

MOARDONGZPLZ · 1h ago
From my post:

[B]ecause they love the subject matter and not because it is an angle to “create content” or “gain followers” or show ads.

anton-c · 43m ago
Because they like to make stuff more than they value a subscription. I'm gonna write music no matter what happens to it.
gorbachev · 28m ago
It's not going to happen this way, because these days for you to get somewhere near the top of Google results requires you to be an established content publisher, basically anyone with enough followers.

Someone who publishes content because they love the subject matter would only reach enough of an audience to have an impact if they work on it, a lot, and most people wouldn't do that without some expectation of return on investment, so they'd follow the influencer / commercial publication playbook and end up in the same place as the established players in the space are already.

If you're satisfied of being on the 50th page on the Google results, then that's fine. Nobody will find you though.

horrorente · 1h ago
> More than likely, people return to publishing content because they love the subject matter and not because it is an angle to “create content” or “gain followers” or show ads. No more “the top 25 hats in July 2025” AI slopfest SEO articles when I look for a hat, but a thoughtful series of reviews with no ads or affiliate links, just because someone is passionate about hats. The horror! The horror!

I disagree with that. There are still people out there doing that out of passion, that hasn't changed (it's just harder to find). Bad actors who are only out there for the money will continue trying to get the money. Blogs might not be relevant anymore, but social media influencing is still going to be a thing. SEO will continue to exist, but now it's targeted to influence AIs instead of the position in Google search results. AIs will need to become (more) profitable, which means they will include advertising at some point. Instead of companies paying Google to place their products in the search or influencers through affiliate links, they will just pay AI companies to place their products in AI results or influencers to create fake reviews trying to influence the AI bots. A SEO slop article is at least easy to detect, recommendations from AIs are much harder to verify.

Also it's going to hit journalism. Not everyone can just blog because they are passionate about something. Any content produced by professionals is either going to be paywalled even more or they need to find different sources of income threatening journalistic integrity. And that gives even more ways to bad actors with money to publish news in their interest for free and gaining more influence on the public debate.

nicbou · 3h ago
It's crazy how few people see it that way. Big tech is capturing all the value created by content creators, and it's slowly strangling the independent web it feeds on. It's a parasitic relationship. Once the parasite has killed its host, it will feed on its users.
brainwad · 1h ago
>If I create content like recipes ... previously I had exclusive rights to my created content

Recipes are not protected by copyright law. That's _why_ recipe bloggers have resorted to editorialising recipes, because the editorial content is copyrightable.

pjc50 · 3h ago
> I think the worst case scenario is that people stop publishing content on the web altogether

Quite clearly heading in that direction, but with a twist: the only people left will be advertising or propaganda, if there's no money in authenticity or correctness.

layer8 · 1h ago
There was little to no money in authenticity or correctness in the heyday of home pages and personal blogs. People published because they were excited about sharing information and opinions. That was arguably the internet at its best.
dirkc · 4h ago
At some stage Google will need to be accountable for answers they are hosting on their own site. The argument of "we're only indexing info on other sites" changes when you are building a tool to generate content and hosting that content on your own domain.

I'm guilty of not clicking when I'm satisfied with the AI answer. I know it can be wrong. I've seen it be wrong multiple times. But it's right at the top and tells me what I suspected when I did the search. The way they position the AI overview is right in your face.

I would prefer the "AI overview" to be replaced with something that helps me better search rather than giving me the answer directly.

OldfieldFund · 4h ago
Another problem is that you have to click twice:

1. The anchor icon.

2. Then one of the sites that appear on the right (on desktop).

Cthulhu_ · 4h ago
> The argument of "we're only indexing info on other sites" changes when you are building a tool to generate content and hosting that content on your own domain.

And yet, "the algorithm" has always been their first defense whenever they got a complaint or lawsuit about search results; I suspect that when (not if) they get sued over this, they will do the same. Treating their algorithms and systems as a mysterious, somewhat magic black box.

Havoc · 3h ago
> Google will need to be accountable

Hell will freeze over first

mtkd · 11h ago
Conversely, it's useful to get an immediate answer sometimes

6 months ago, "what temp is pork safe at?" was a few clicks, long SEO optimised blog post answers and usually all in F not C ... despite Google knowing location ... I used it as an example at the time of 'how hard can this be?'

First sentance of Google AI response right now: "Pork is safe to eat when cooked to an internal temperature of 145°F (63°C)"

ncallaway · 10h ago
Dear lord please don’t use an AI overview answer for food safety.

If you made a bet with your friend and are using the AI overview to settle it, fine. But please please click on an actual result from a trusted source if you’re deciding what temperature to cook meat to

sothatsit · 10h ago
The problem is that SEO has made it hard to find trustworthy sites in the first place. The places I trust the most now for getting random information is Reddit and Wikipedia, which is absolutely ridiculous as they are terrible options.

But SEO slop machines have made it so hard to find the good websites without putting in more legwork than makes sense a lot of the time. Funnily enough, this makes AI look like a good option to cut through all the noise despite its hallucinations. That's obviously not acceptable when it comes to food safety concerns though.

omnicognate · 6h ago
If I do that search on Google right now, the top result is the National Pork Board (pork.org): ad-free, pop-up free, waffle-free and with the correct answer in large font at the top of the page. It's in F, but I always stick " C" at the end of temperature queries. In this case that makes the top result foodsafety.gov which is equally if not more authoritative, also ad-, waffle-, and popup- free and with with the answer immediately visible.

Meanwhile the AI overview routinely gives me completely wrong information. There's zero chance I'm going to trust it when a wrong answer can mean I give my family food poisoning.

I agree that there is a gigaton of crap out there, but the quality information sources are still there too. Google's job is to list those at the top and it actually has done so this time, although I'll acknowledge it doesn't always and I've taken to using Kagi in preference for this reason. A crappy AI preview that can't be relied on for anything isn't an acceptable substitute.

pasc1878 · 5h ago
Kagi sort of gets this correct.

Kagi search gives the pork board first as well. But note that site fails mtkd's requirements giving temperature in degrees Fahrenheit and not Celsius. The second hit does give a correct temperature but has a cookie banner (which at least can be rejected with one click)

The optional Kagi assistant quotes the pork board, usda which also is only in Fahrenheit, and third a blog on site for a thermometer that quoted the UK Food Standard Authority and gives its temperature

However there is a problem the UK FSA does not agree with USDA on the temperature it puts it higher at 70 degrees C rather than 63

So if you get the USDA figure you are taking a risk. The Kagi Assistant gives both temperatures but it is not clear which one is correct although both figures are correctly linked to the actual sites.

omnicognate · 5h ago
I don't really see the problem with F and C. As I mentioned, I always stick " C" on the end of temperature queries. It's 2 characters and the results always have the centigrade temps, on both Kagi and Google.
tonyedgecombe · 1h ago
>The problem is that SEO has made it hard to find trustworthy sites in the first place.

We should remember that's partly Google's fault as well. They decided SEO sites were OK.

jordanb · 9h ago
Google could have cut down on this if they wanted. And in general they did until they fired Matt Cutts.

The reality is, every time someone's search is satisfied by an organic result is lost revenue for Google.

taurath · 9h ago
Which is the stupidest position ever if Google wants to exist long term.

Unfortunately there are no workable alternatives. DDG is somehow not better, though I use it to avoid trackers.

Miraste · 6h ago
It's a bit like the Easter Islanders cutting down all of their trees for wood. Where does Google management think they'll get search results if they kill the entire internet? Has anyone at Google thought that far ahead?
9dev · 5h ago
The internet they dream of is like a large mall. It consists of service providers selling you something, and Google directing you to them in exchange for some of the profit. The role of users in this model is that of a Piñata that everyone hits on to drop some money.
what · 8h ago
DDG is just serving you remixed bing and yandex results. There’s basically no alternative to GBY that do their own crawling and maintain their own index.
Zardoz84 · 6h ago
qwant ?
touisteur · 6h ago
Qwant also has an AI overview. Pretty bad too.
frm88 · 4h ago
I've been using noai.duckduckgo.com for a few weeks now and it's pretty reliable. Still yandex etc. but at least no longer AI overview. (Yes, I know about settings, but they get deleted every restart).
al_borland · 9h ago
AI is being influenced by all that noise. It isn’t necessarily going to an authoritative source, it’s looking at Reddit and some SEO slop and using that to come up with the answer.

We need AI that’s trained exclusively on verified data and not random websites and internet comments.

jval43 · 3h ago
I asked Gemini about some Ikea furniture dimensions and it gave seemingly correct answers, until it suddenly didn't make sense.

Turns out all the information it gave me came from old Reddit posts and lots of it was factually wrong. Gemini however still linked some official Ikea pages as the "sources".

It'll straight up lie to you and then hide where it actually got it's info from. Usually Reddit.

sothatsit · 9h ago
Creating better datasets would also help to improve the performance of the models, I would assume. Unfortunately, the costs to produce high-quality datasets of a sufficient size seem prohibitive today.

I'm hopeful this will be possible in the future though, maybe using a mix of 1) using existing LLMs to help humans filter the existing internet-scale datasets, and/or 2) finding some new breakthroughs to make model training more data efficient.

heavyset_go · 7h ago
It'll still hallucinate
jkingsman · 8h ago
Mmm, I see this cutting both ways -- generally, I'd agree; safety critical things should not be left to an AI. However, cooking temperatures are information that has a factual ground truth (or at least one that has been decided on), has VERY broad distribution on the internet, and generally is a single, short "kernel" of information that has become subject to slop-ifying and "here's an article when you're looking for about 30 characters of information or less" that is prolific on the web.

So, I'd agree -- safety info from an LLM is bad. But generally, the /flavor/ (heh) of information that such data comprises is REALLY good to get from LLMs (as opposed to nuanced opinions or subjective feedback).

zahlman · 10h ago
I've been finding that the proliferation of AI slop is at its worst on recipe/cooking/nutrition sites, so....
ncallaway · 5h ago
Please find a trusted source of information for food safety information.

It's genuinely harder than it's ever been to find good information on the internet, but when you're dealing with food safety information, it's really worth taking the extra minute to find a definitive source.

https://www.foodsafety.gov/food-safety-charts/safe-minimum-i...

edanm · 2h ago
Idk. Maybe that's true today (though even today I'm not sure) but how long before AI becomes better than just finding random text on a website?

After all, AI can theoretically ask follow-up questions that are relevant, can explain subtleties peculiar to a specific situation or request, can rephrase things in ways that are clearer for the end user.

Btw, "What temperature should a food be cooked to" is a classic example of something where lots of people and lots of sources repeat incorrect information, which is often ignored by people who actually cook. Famously, the temp that is often "recommended" is only the temp at which bacteria/whatever is killed instantly - but is often too hot to make the food taste good. What is normally recommended is to cook to a lower temperature but keep the food at that temperature for a bit longer, which has the same effect safety-wise but is much better.

greazy · 6h ago
I googled (Australia) "what temp is pork safe at?", top three hits:

1. https://www.foodsafety.asn.au/australians-clueless-about-saf... 2. https://www.foodsafety.gov/food-safety-charts/safe-minimum-i... 3. https://pork.org/pork-cooking-temperature/

All three were highly informative, well cited sources from reputable websites.

wiseowise · 6h ago
Only your second link provides good information in a convenient format (both F and C), first and third are useless.
maerch · 9h ago
Meanwhile, in Germany, you can get raw pork with raw onions on a bread roll at just about every other bakery.

https://en.m.wikipedia.org/wiki/Mett

When I searched for the safe temperature for pork (in German), I found this as the first link (Kagi search engine)

> Ideally, pork should taste pink, with a core temperature between 58 and 59 degrees Celsius. You can determine the exact temperature using a meat thermometer. Is that not a health concern? Not anymore, as nutrition expert Dagmar von Cramm confirms: > “Trichinae inspection in Germany is so strict — even for wild boars — that there is no longer any danger.”

https://www.stern.de/genuss/essen/warum-sie-schweinefleisch-...

Stern is a major magazine in Germany.

bee_rider · 8h ago
I was just thinking that EU sources might be a good place to look for this sort of thing, given that we never really know what basic public health facts will be deemed political in the US on any given day. But, this reveals a bit of a problem—of course, you guys have food safety standards, so advice they is safe over there might not be applicable in the US.
pjc50 · 3h ago
Doesn't even have to be "better", just "different". The classic one is whether you should refrigerate eggs, which has diametrically opposite answers.

But anything that actually matters could be politicized at any time. I remember the John Gummer Burger Incident: http://news.bbc.co.uk/1/hi/uk/369625.stm , in the controversy over whether prion diseases in beef (BSE) were a problem.

Daz1 · 6h ago
what a cringe comment
badc0ffee · 6h ago
Should "Taste pink", you say
maerch · 5h ago
It’s just the ChatGPT translation, and it’s a literal one. That said, I’ve never heard that phrase in German either.
layer8 · 1h ago
The literal translation is wrong in that context, it should have been translated to “medium”.
didibus · 9h ago
Funny story, I used that to know the cooked temperature of burgers, it said medium-rare was 130. I proceeded to eating it and all, but then like half way through, I noticed the middle of this burger is really red looking, doesn't seem normal, and suddenly I remembered, wait, ground beef is always supposed to be 160, 130 medium-rare is for steak.

I then chatted that back to it, and it was like, oh ya, I made a mistake, you're right, sorry.

Anyways, luckily I did not get sick.

Moral of the story, don't get mentally lazy and use AI to save you the brain it takes for simple answers.

what · 8h ago
Do you actually put a thermometer in your burgers/steaks/meat when you’re cooking? That seems really weird.

Why are people downvoting this? I’ve literally never seen anyone use a thermometer to cook a burger or steak or pork chop. A whole roasted turkey, sure.

pjc50 · 3h ago
You're getting lots of thermometer answers, so I'm going to give the opposite: I'm also on team "looks good to me" + "cooking time on packet" + "just cut it and look"
lotyrin · 7h ago
Many people wing dishes that they've prepared 100s of times. Others rarely make the same recipe twice. Neither are correct or incorrect, but the latter is very much going to measure everything they're doing carefully (or fail often).
PetahNZ · 4h ago
Why wouldn't I? It takes a few seconds and my thermometer just sits on fridge.
avidiax · 7h ago
For something safety critical like a burger, yes.

For whole meats, it's usually safe to be rare and you can tell that by feel, though a thermometer is still useful if you aren't a skilled cook or you are cooking to a doneness you aren't familiar with.

8note · 5h ago
thermometers were recommended by folks like alton brown and kenji to get really consistent results.

i havent heard it for burgers, but steaks for sure.

padjo · 5h ago
People are downvoting you because you’ve come onto a website populated by engineers and called someone weird for using objective measurements.
habinero · 5h ago
I think your reference pool is just small. I absolutely use it for meat and especially for ground meat, which has a much higher chance of contamination.
tonyedgecombe · 1h ago
I suspect your reference pool is the small one. Most people buy their burgers in a packet and hence follow the timing instructions on that packet.
carlosjobim · 1h ago
> Anyways, luckily I did not get sick.

Why would you purchase meat that you suspect is diseased? Even if you cook it well-done, all the (now dead) bacteria and their byproducts are still inside. I don't understand why people do this to themselves? If I have any suspicion about some meat, I'll throw it away. I'm not going to cook it.

brookst · 11h ago
Safe Temperatures for Pork

People have been eating pork for over 40,000 years. There’s speculation about whether pork or beef was first a part of the human diet.

(5000 words later)

The USDA recommends cooking pork to at least 145 degrees.

BoorishBears · 10h ago
I searched it.

First result under the overview is the National Pork Board, shows the answer above the fold, and includes visual references: https://pork.org/pork-cooking-temperature/

Most of the time if there isn't a straightforward primary source in the top results, Google's AI overview won't get it right either.

Given the enormous scale and latency constraints they're dealing with, they're not using SOTA models, and they're probably not feeding the model 5000 words worth of context from every result on the page.

ImaCake · 11h ago
Not only that, it includes a link to the USDA reference so you can verify it yourself. I have switched back to google because of how useful I find the RAG overviews.
wat10000 · 11h ago
The link is the only useful part, since you can’t trust the summary.

Maybe they could just show the links that match your query and skip the overview. Sounds like a billion-dollar startup idea, wonder why nobody’s done it.

owenversteeg · 10h ago
It’s a pretty good billion dollar idea, I think you’ll do well. In fact I bet you’ll make money hand over fist, for years. You could hire all the best engineers and crush the competition. At that point you control the algorithm that everyone bases their websites on, so if you were to accidentally deploy a series of changes that incentivized low quality contentless websites… it wouldn’t matter at all; not your problem. Now that the quality of results is poor, but people still need their queries answered, why don’t you provide them the answers yourself? You could keep all the precious ad revenue that you previously lost when people clicked on those pesky search results.
krupan · 10h ago
This should be the top comment! Thank you for posting it because I'm starting to worry that I'm the only one who realizes how ridiculous this all is.
hansvm · 11h ago
As of a couple weeks ago it had a variety of unsafe food recommendations regarding sous vide, e.g. suggesting 129F for 4+ hours for venison backstrap. That works great some of the time but has a very real risk of bacterial infiltration (133F being similar in texture and much safer, or 2hr being a safer cook time if you want to stick to 129F).

Trust it if you want I guess. Be cautious though.

mitthrowaway2 · 11h ago
Google's search rankings are also the thing driving those ridiculous articles to the top, which is the only reason so many of them get written...
ljlolel · 10h ago
And also why they incentivized all this human written training data that will no longer be incentivized
kriro · 4h ago
On Google: """what temp in C is pork safe at?"""

AI: 63C

First result: Five year old reddit thread (F only discussion, USDA mentioned).

Second result: ThermoWorks blog (with 63C).

Third result: FoodSafety.gov (with 63C)

Forth result: USDA (with 63C)

Seems reasonable enough to scan 3-4 results to get some government source.

eviks · 9h ago
> 6 months ago, "what temp is pork safe at?

No it wasn't, most of the first page results have the temperature right there in the summary, many of them with both F and C, and unlike the AI response, there is much lower chance of hallucinated results.

So you've gained nothing

PS Trying the same search with -ai gets you the full table with temperatures, unlike with the AI summary where you have to click to get more details, so the new AI summary is strictly worse

grey-area · 6h ago
Why do you think that answer is correct? I mean maybe it is, or maybe it’s by the same user who recommended eating rocks (which ‘AI’ also recommended).

https://www.bbc.co.uk/news/articles/cd11gzejgz4o

__turbobrew__ · 9h ago
refactor_master · 9h ago
It doesn't take long to find SEO slop trying to sell you something:

When our grandmothers and grandfathers were growing up, there was a real threat to their health that we don’t face anymore. No, I’m not talking about the lack of antibiotics, nor the scarcity of nutritious food. It was trichinosis, a parasitic disease that used to be caught from undercooked pork.

The legitimate worry of trichinosis led their mothers to cook their pork until it was very well done. They learned to cook it that way and passed that cooking knowledge down to their offspring, and so on down to us. The result? We’ve all eaten a lot of too-dry, overcooked pork.

But hark! The danger is, for the most part, past, and we can all enjoy our pork as the succulent meat it was always intended to be. With proper temperature control, we can have better pork than our ancestors ever dreamed of. Here, we’ll look at a more nuanced way of thinking about pork temperatures than you’ve likely encountered before."

Sorry, what temperature was it again?

Luckily there's the National Pork Board which has bought its way to the top, just below the AI overview. So this time around I won't die from undercooked pork at least.

pasc1878 · 5h ago
The link quoted does not have that text so what are you on about?

However that site gives the temperature for Pork as 71C which is not what USDA says but is correct. So using the USDA recommendation does have a risk according to at least Canada and UK

stereolambda · 5h ago
Honestly the SEO talk sounds like reflexive coping in this discourse. I get that WWW has cheapened quality, but we now have the tech that could defeat most of the SEO and other trash tactics on the search engine side. Text analysis as a task is cracked open. Google and such could detect dark patterns with LLMs, or even just deep learning. This would probably be more reliable than answering factual queries.

The problem is there is no money and fame in using it that way, or at least so people think in the current moment. But we could return to enforcing some sort of clear, pro-reader writing and bury the 2010s-2020s SEO garbage on page 30.

Not the mention that the LLMs randomly lie to you with less secondary hints at trustworthiness (author, website, other articles, design etc.) than you get in any other medium. And the sustainability side of incentivizing people to publish anything. I really see the devil of convenience as the only argument for the LLM summaries here.

SwtCyber · 5h ago
The issue is more when those same tools start replacing deeper content or misrepresenting nuanced info
sgentle · 11h ago
"full moon time NY"

> The next full moon in New York will be on August 9th, 2025, at 3:55 a.m.

"full moon time LA"

> The next full moon in Los Angeles will be on August 9, 2025, at 3:55 AM PDT.

I mean, it certainly gives an immediate answer...

Rapzid · 6h ago
AI overview also says 165f is the best temperature to cook chicken breast to. Which is and always has been bollocks.
squigz · 10h ago
I wonder how people have such awful experiences with (traditional) Google when I don't and really never have.

First result: https://www.porkcdn.com/sites/porkbeinspired/library/2014/06...

Second result: https://pork.org/pork-cooking-temperature/

wat10000 · 11h ago
It’s only useful if you can trust it, and you very much cannot.

I know you can’t necessarily trust anything online, but when the first hit is from the National Pork Board, I’m confident the answer is good.

pasc1878 · 5h ago
But for the OP it is not as it does not give a temperature in their preferred units and probably USDA gives the wrong temperature in their locality.
Jean-Papoulos · 6h ago
Incredible, you are the problem. Didn't think I'd see such an idiotic answer on HN, please for the love of god do not use AI to know what is safe to eat.
8note · 5h ago
id consider that google thinks its good enough for people to base their food safety off of it, and they deserve to get sued for whatever theyre worth for providing said recommendations when somebody trusts them and gets sick
croes · 6h ago
Don’t forget to add glue and rocks
ponys · 26m ago
I've written a bit about this issue as well, coming from someone working at a company that partially relies on organic search results. It's been interesting (And slightly worrying) to see the shift and layoffs happening here already.

https://sasjakoning.com/blog/how-google-s-new-ai-overview-is...

I have to say that the comments here do offer a great alternative perspective as to how terrible the web has been to navigate before AI Overviews became a thing.

ripped_britches · 6h ago
Why is this being framed as a problem? People are obviously happier with the new feature, duh

Of course they need to make the AI overviews suck less, but saying it’s unfair to sites is crazy talk because your site now just generates less value than an AI response if that’s what stopped you from going

If you have content better than Gemini I will still go to your site

alastairr · 5h ago
Where do you suppose AI overviews get their source from? The problem here is Google is now inserting itself between a business' user and the business' content. I'm not saying that was a good or invulnerable business to begin with but that's what's happened. If you kill the business of making reliable content, what are they left to serve?
entuno · 5h ago
A lot of companies seem to have based their business model on the assumption that Google and Microsoft would continue to send them traffic for free indefinitely.

So now they're having to scramble to rethink their approach, and obviously aren't happy about that.

wouldbecouldbe · 6h ago
Because some companies are going bankrupt because of the data google is taking. Google always had a weird relationship with sites, they sort of needed each other but google always had the upper hand, now it’s even worse
ripped_britches · 6h ago
Wouldn’t you expect AI to displace some companies though? As did every major technology?
toenail · 6h ago
Some people are against all kinds of progress. They don't understand that their life depends on progress that was made in the past.
simianwords · 5h ago
This is because they have entrenched themselves in a comfortable position that they don’t want to give up. Most won’t admit this to be the actual reason.

Think about it: you are a normal hands on self thought software developer. You grew up tinkering with Linux and a bit of hardware. You realise there’s good money to be made in a software career. You do it for 20-30 years; mostly the same stuff over and over again. Some Linux, c#, networking. Your life and hobby revolves around these technologies. And most importantly you have a comfortable and stable income that entrenches your class and status. Anything that can disrupt this state is obviously not desireable. Never mind that disrupting others careers is why you have a career in the first place.

righthand · 1h ago
C#?
vouaobrasil · 6h ago
That does not imply that future progress is a good thing, nor does it imply that future progress will even be useful to the majority. It boggles my mind how some people make the logical inference that "all progress is good" based on "some past progress was useful".
toenail · 2h ago
Do you think that having fast, distraction-free access to relevant information is not useful?
djeastm · 1h ago
It is useful, but it assumes the information is trustworthy. At least with websites you could discern some credibility based on the organization publishing the information. With AI summaries, it's just "Google's latest model seems to think this is the answer, good luck!"
ghushn3 · 15h ago
I subscribe to Kagi. It's been worth it to have no ads and the ability to uprank/downrank sites.

And there's no AI garbage sitting in the top of the engine.

slau · 15h ago
You can opt-in to get an LLM response by phrasing your queries as a question.

Searching for “who is Roger rabbit” gives me Wikipedia, IMDb and film site as results.

Searching for “who is Roger rabbit?” gives me a “quick answer” LLM-generated response: “Roger Rabbit is a fictional animated anthropomorphic rabbit who first appeared in Gary K. Wolf's 1981 novel…” followed by a different set of results. It seems the results are influenced by the sources/references the LLM generated.

abtinf · 10h ago
You don’t have to phrase it as a question; just append a ?, which is an operator telling it you want a generated answer.
slau · 5h ago
Yes. That is exactly what my answer demonstrates.
pasc1878 · 5h ago
Even with a normal search there is a link to get Quick Answer which gives the LLM result
greatgib · 12h ago
I don't think that you are right. It is the search result that influence the llm generated result and not the opposite.

In your case, I think that it is just the interrogation point in itself at the end that somehow has an impact on the results you see.

s900mhz · 12h ago
It’s a feature of Kagi. Putting the question mark does invoke AI summaries.

https://help.kagi.com/kagi/ai/quick-answer.html

standardUser · 15h ago
I'm more interested now than ever. A lot of my time spent searching is for obscure or hard-to-find stuff, and in the past smaller search engines were useless for this. But most of my searches are quick and the primary thing slowing me down are Google product managers. So maybe Kagi is worth a try?
ghushn3 · 12h ago
You can try it for free. I did my 300 searches on it and went, "Yep. This is better." and then converted to a paid user.
voltaireodactyl · 15h ago
I think you might be happily surprised for sure.
Melatonic · 11h ago
It's awesome - highly recommend trying it
stevenAthompson · 10h ago
I subscribe also, and prefer it for most things.

However, it's pretty bad for local results and shopping. I find that anytime I need to know a local stores hours or find the cheapest place to purchase an item I need to pivot back to google. Other than that it's become my default for most things.

peacebeard · 9h ago
Thanks for the suggestion. I try nonstandard search engines now and then and maybe this one will stick. Google certainly is trying their best to encourage me.
al_borland · 9h ago
After about a year on Kagi my work browser randomly reverted to Google. I didn’t notice the page title, as my eyes go right to the results. I recoiled. 0 organic results without scrolling, just ads and sponsored links everywhere. It seems like Google boiled the frog one degree at a time. Everyone is in hell and just doesn’t know it, because it happened so gradually.

I’ve also tried various engines over the years. Kagi was the first one that didn’t have me needing to go back to Google. I regularly find things that people using Google seem to not find. The Assistant has solved enough of my AI needs that I don’t bother subscribing to any dedicated AI company. I don’t miss Google search at all.

I do still using Google Maps, as its business data still seems like the best out there, and second place isn’t even close. Kagi is working on their own maps, but that will be long road. I’m still waiting for Apple to really go all-in, instead of leaning on Yelp.

outlore · 11h ago
is there a way to make safari search bar on iOS show the kagi search term rather than the URL?
al_borland · 9h ago
Maybe with Orion instead of Safari?

Apple really needs to update Safari to let people choose their search engine, instead of just having the list of blessed search engines to choose from.

Xylakant · 6h ago
al_borland · 1h ago
This is what I used, but it’s a bit of a hack. It has to hijack one of the other search engine, which then becomes unusable. It also doesn’t show the search term in the address bar, like the person was asking for.
tonyedgecombe · 1h ago
You can thank Apple for that, presumably because they receive a ton of money from Google.
al_borland · 17m ago
Google isn’t the only option in the list.

And of course it’s Apple’s fault, they wrote Safari. I submitted a feature request a while ago, but I’m guessing more people need to do the same.

dyauspitr · 9h ago
It’s useless. Worse than Google and no AI summary.
ghushn3 · 8h ago
If you append a ? to your search it will prompt an AI query. And you can uprank/downrank results so I'm not sure in what universe it could be worse.
cm2187 · 5h ago
What I don't understand is how do you square that with google still reporting growth in ads revenues. Surely google's business should tank by that much (not even taking into account the additional cost of processing a seach with AI).
mrkramer · 1h ago
Google introduced its Google Knowledge Graph in 2012 and for sure it caused the drop in website clicks but I would say not that significant. People want quality content and they will always look for quality websites with or without AI overviews. My digital marketing professor once said that customers who look beyond first page results are the most valuable ones, because they are dedicated users and customers. People who read AI overviews won't be your first class users and customers.
RamblingCTO · 1h ago
I think it's a google problem. The AI summary is often times wrong so I don't bother. But the pages are garbage as well, so normal users won't click on that looking for information. Dark forest and all that. Pages started to optimize for Google and not humans. Humans don't visit anymore. Duh
garylkz · 1h ago
Yep, I can't remember when was the last time I've visited a SEO optimized site but I still had the bitter taste that I felt when I last visited it:

- searches for "How to do XYZ" and click one of the site

- "what is xyz"

- "why xyz matters"

- "preparations before xyz"

- "what you might encounter when xyz"

Sounds reasonable and in theory should be useful, but the actual useful info are only stuffed in 1-2 lines of multi paragraphs on tons of sections that I don't care about.

PeterStuer · 5h ago
Main reason:

Google seach ranking already became so bad that finding the right query to produce a decent result became a craft in and of itself.

HN readers might not realize this, but I used to constantly have to search things for friends and family as they just couldn't find it with Google.

The AI result is a godsend fir them. I saw a massive drop in their requests to me.

oezi · 16h ago
The tricky thing for Google will be to do this and not kill their cash cow ad business.
kozikow · 15h ago
Ads inside LLMs (e.g. pay $ to boost your product in LLM recommendation) is going to be a big thing.

My guess is that Google/OpenAI are eyeing each other - whoever does this first.

Why would that work? It's a proven business model. Example: I use LLMs for product research (e.g. which washing machine to buy). Retailer pays if link to their website is included in the results. Don't want to pay? Then redirect the user to buy it on Walmart instead of Amazon.

kieckerjan · 6h ago
I actually encountered this pretty early in one of these user tuned GPT's in OpenAI's GPT store. It was called Sommelier or something and it was specialized in conversations about wine. It was pretty useful at first, but after a few weeks it started lacing all its replies with tips for wines from the same online store. Needless to say, I dropped it immediately.
enahs-sf · 9h ago
Forget links, agents are gonna just go upstream to the source and buy it for you. I think it will change the game because intent will be super high and conversion will go through the roof.
hyperadvanced · 7h ago
Yeah I’m gonna give an AI agent my credit card and complete autonomy with my finances so it can hallucinate me a new car. I love getting findommed.
weatherlite · 6h ago
Look, the car shop might not bill you at all because their A.I agent will hallucinate the purchase, so I don't see why you're so pessimistic about agents.
manmal · 7h ago
It can still give you an overview with a few choices and a link to the prepared checkout page, and you enter your CC details yourself.
heavyset_go · 7h ago
Feels like this hope is in the same vein as Amazon Dash and then the expectation that people would buy shit with voice assistants like Alexa.
pacifika · 6h ago
Who doesn’t want to associate their product with unreliability and incorrect information? Think about that reputational damage.
msgodel · 11h ago
People are already wary of hosted LLMs having poisoned training data. That might kill them altogether and push everyone to using eg Qwen3-coder.
landl0rd · 10h ago
No, a small group of highly tech-literate people are wary of this. Your personal bubble is wary of this. So is some of mine. "People" don't care and will use the packaged, corporate, convenient version with the well-known name.

People who are aware of that and care enough to change consumption habits are an inconsequential part of the market.

msgodel · 9h ago
I don't know, a bunch of the older people from the town I grew up in avoided using LLMs until Grok came out because of what they saw going on with alignment in the other models (they certainly couldn't articulate this but listening to what said it's what they were thinking.) Obviously Grok has the same problems but I think it goes to show the general public is more aware of the issue than they get credit for.

You combine this with Apple pushing on device inference and making it easy and anything like ads probably will kill hosted LLMs for most consumers.

tokioyoyo · 6h ago
Yeah, average people that I know (across continents) just ChatGPT their way into literally anything without a second thought. They don't care.
manmal · 7h ago
Maybe Grok was just pushed by their political influencers. It’s a republican, anti-woke LLM after all.
weatherlite · 6h ago
Everyone is talking about it , and it is a big concern, but the last 2 years (ever since ChatGPT showed up) ad revenue keeps growing at the same 10%+ . It seems like the money queries (adidas shoes size 45) are still better served with links than overviews, and those are where most of the money comes from. Searching for "what cooking temperature is pork safe to eat" is not something you can easily monetize.

Disclaimer: google stock holder.

pryelluw · 15h ago
Not tricky at all.

This is a new line of business that provides them with more ad space to sell.

If the overview becomes a trusted source of information, then all they need to do is inject ads in the overviews. They already sort of dye that. Imagine it as a sort of text based product placement.

NoPicklez · 11h ago
I'd say putting ads into AI search overviews is absolutely tricky.

You might think that's the correct way to do it, but there is likely much more to it than it seems.

If it wasn't tricky at all you'd bet they would've done it already to maximize revenue.

pryelluw · 11h ago
Product teams in big companies move slow. But soon enough all the shit ads are going to pop up.
stevenAthompson · 10h ago
> If the overview becomes a trusted source of information

It never will. By disincentivizing publishers they're stripping away most of the motivation for the legitimate source content to exist.

AI search results are a sort of self-cannibalism. Eventually AI search engines will only have what they cached before the web became walled gardens (old data), and public gardens that have been heavily vandalized with AI slop (bad data).

Gigachad · 14h ago
I’d guess that the searches where AI overviews are useful and the searches where companies are buying ads are probably fairly distinct. If you search for plumbers near you, they won’t show an AI overview, while if you search “Why are plants green?”, no one was buying ads on that.
josteink · 6h ago
> The tricky thing for Google will be to do this and not kill their cash cow ad business

This is not for Google to decide.

The users have spoken clearly that (when given an option) they will not tolerate or succumb to the SPAM of shitty SEO-optimized content-farms which has been plaguing the internet for the last decade.

If Google don't provide meaningful results in their search page, people will use ChatGPT or something else to sidestep the SEO SPAM issue all together.

ahartmetz · 12h ago
Somebody is working on "native advertising" in AI slop, surely? Barf.
bethekidyouwant · 11h ago
you can’t make the slop have a nice clean ad in it. Also: as soon as your slop has ads in it, I’m going to make Polpot advertise your product.
bgwalter · 16h ago
You can apparently disable these annoying and useless "AI" overviews by cursing in the query:

https://arstechnica.com/google/2025/01/just-give-me-the-fing...

Polizeiposaune · 11h ago
It's relatively straightforward to create a firefox alternate search engine which defaults to the "web" tab of Google search results which is mostly free of Google-originated LLM swill.

Instructions are here: https://support.mozilla.org/en-US/kb/add-custom-search-engin...

The "URL with %s in place of search term" to add is:

https://www.google.com/search?q=%s&client=firefox-b-d&udm=14

mjcl · 9h ago
Google was kind enough to give the AI overview a stable CSS class name (to date), so this userscript has been effective at hiding it for me:

window.addEventListener('load', function() { var things = this.document.getElementsByClassName('M8OgIe'); for (var thing of things) { thing.style.display = 'none'; } }, false);

riantogo · 14h ago
Or just append with -ai => "how to pick a running shoe -ai"
devnullbrain · 53m ago
Those four characters are enough friction to slowly grind down the number of today's outraged people into a population small enough that, when Google stop supporting '-ai', people will think it's weird that you still care.
what · 7h ago
>useless

They’re actually pretty useful. It tends to be a very brief summary of the top results, so you can tell if anything is worth clicking on.

x0x0 · 15h ago
appending a -"fuck google #{insert slur of choice here}" to my search results has improved them. Then I wonder why I do this to myself and ponder going back to kagi.
privatelypublic · 10h ago
Jesus dude. Just use the udm options instead of practicing slurs.
nneonneo · 9h ago
I wish there was a good udm option for "what you used to show me before AI took over". For example, I like seeing flight updates when I punch in a flight number, which udm=14 does not show.

That said, udm=14 has still been a huge plus for me in terms of search engine usability, so it's my default now.

verzali · 5h ago
My site never got much traffic from Google and I never wasted time on SEO. So I find it hard to care about losing clicks I didn't have.

But I do find it concerning that Google is effectively stealing the time I spend on research and not offering proper credit. I'm always careful to credit people and provide extensive links to the work of others on my site. But here Google and others are simply stealing that work, getting it wrong, and then claiming they did all the effort.

dado3212 · 16h ago
Related, but to whichever PM put the "AI Mode" on the far left side of the toolbar, thus breaking the muscle memory from clicking "All" to get back from "Images", I expect some thanks for unintentionally boosting your CTR metrics.
fsh · 15h ago
That decision probably paid someone's new car. The KPIs will be excellent. Who cares about what the users might have wanted to do with their clicks.
pacifika · 6h ago
Maybe have it forced hover under the mouse cursor next, maybe think of what the long term effect are instead of the KPIs.
AnimeLife · 8h ago
In a sense this is similar to what Amazon has been doing in few countries. Find top selling products, get them cheaper from somewhere, rebrand them, rank them higher and sell them. They don't need to invest in market research like their competetors, they get all data from Amazon.com

At big tech scale, this is clearly anti-compete and piracy IMHO.

nicbou · 7h ago
Except in this case they still rely on people to create content to train their AI on.
dana321 · 36m ago
I installed a plugin that removes them, most of the time the answers it gives are complete dogwater
keiferski · 4h ago
What is the long term plan for data acquisition?

1. Use existing websites for training data

2. Replace search traffic with AI prompts, thereby destroying the economic incentive for websites to publish data

3. ?

karel-3d · 1h ago
I just wonder

if AI will destroy the internet... what it will use to train the next generation on? If people go to AI instead of reddit and stack overflow, what will they train the next models on? They have used reddit extensively to train the first GPT models.

I guess that's "tomorrow's problem" though

carlosjobim · 1h ago
> if AI will destroy the internet... what it will use to train the next generation on?

Spying on you when you're fixing your carburetor and then making a how-to guide for other people.

ValveFan6969 · 14h ago
The clicks in question: "Here's a thirty page story of how grandma discovered this recipe... BTW you need to subscribe/make an account/pay to view the rest of the article!"
einrealist · 1h ago
At what point does this remove the 'platform provider' defense for wrong or defamatory information generated by Google & Co?
7222aafdcf68cfe · 3h ago
What current incentive is left for people to publish on the internet?

Discovery mechanisms are less effective than before, people no longer visit your website, copyright is dead and in the end we're just feeding an insatiable machine with slave labour. Or am I missing something ?

gtsop · 3h ago
What was the initial incentive for people to publish on the internet before ad revenue was a thing?
righthand · 1h ago
Information sharing with other humans, not corporations.
quectophoton · 1h ago
Likely with some expectation of receiving attribution for your effort, even if only sometimes.
djeastm · 1h ago
The incentive was often creating a community around a shared interest. Good luck doing that with AI summaries stopping people from visiting the sites.
JKCalhoun · 14h ago
Liberating me from "search clicks" is not a bad thing at all. I suspect many of us though don't even go to <search engine> anyway but ask an LLM directly.
achierius · 14h ago
It's fundamentally self-destructive though. In time, the sites which rely on search clicks for revenue will essentially cease to be paid for their work, and in many cases will therefore stop publishing the high-quality material that you're looking for.
JKCalhoun · 14h ago
I assumed that, after having using LLMs myself increasingly, that LLM's killing search was inevitable anyway. Further I assume that Google recognizes it as well and would rather at least remain somewhat relevant?

Google search, as others have mentioned in this thread, increasingly fails to give me high-quality material anyway. Mostly it's just pages of SEO spam. I prefer that the LLM eat that instead of me (just spit back up the relevant stuff, thankyouverymuch).

Honestly though, increasingly the internet for me is 1) a distraction from doing real work 2) YouTube (see 1) and 3) a wonderful library called archive.org (which, if I could grab a local snapshot would make leaving the internet altogether much, much easier).

henry2023 · 6h ago
By high quality material do you mean the 15 ads per page, ultra SEO optimized, content length extended posts that will give you a mediocre carbonara recipe?

Ad web has never incentivized quality and never will.

manmal · 6h ago
ChatGPT lists clickable sources in a lot of nontrivial queries. Those sites don’t even need to pay OpenAI for the traffic (yet). If you ask „what’s happening in the world today“, you might get 20 links. How is this worse, exactly?
croes · 6h ago
How many people click the links? What happens to LLMs if people don’t provide training data anymore because nobody visits their sites?
esnard · 26m ago
Cloudflare publishes a "crawl-to-refer" ratio, which can be used to estimate the traffic from LLMs:

https://radar.cloudflare.com/ai-insights#crawl-to-refer-rati...

robryan · 3h ago
They will either pay for it to be generated or get good enough at producing synthetic data that actually improves LLM quality.
croes · 2h ago
So either even higher costs and hope that a bug problem of LLMs get solved somehow.

Given how much data they need that will be pretty expensive, I mean really really expensive. How many people can write good training data and how much per day?

Doesn’t sound sustainable.

landl0rd · 10h ago
Most of the time when I find a good answer from search it's one of a few things:

- Hobbyist site

- Forum or UGC

- Academic/gov

- Quality news which is often paywalled

Most of that stuff doesn't depend on ad clicks. The things that do depend on ad clicks are usually infuriating slop. I refuse to scroll through three pages of BS to get to the information I want and I really don't care if the slop farmers die off.

scarface_74 · 9h ago
Sites rely on ad impressions for revenue. I block ads anyway so either way they aren’t getting money from me.

And if ad supported content ceases to exist, nothing of value will have been lost. I’m not morally opposed to advertising, I find ad supported content not worth reading especially on mobile.

alvah · 7h ago
>And if ad supported content ceases to exist, nothing of value will have been lost.

"Absolutely nothing on any website that relies on ads for revenue is of any value" is a wild take.

scarface_74 · 9m ago
To be more precise, it might be air value but not worth the trade off of ads - they take away the value.
croes · 6h ago
So you believe in free lunch.

If the ad sites die, the AI crawlers will rise the hosting costs of all others sites buy putting more traffic on them.

So either they start adding ads too or the sites die also

JKCalhoun · 55s ago
Before web advertising the entire web was a "free lunch" — if by free lunch you mean individuals putting their sites up; their only compensation being some degree of joy at spreading and sharing their passion.

The early web was eccentric and kind of … "shambly", but I would not be upset if we returned to that.

scarface_74 · 2m ago
No, I pay for Apple News, Slate, Stratechery, Downstream+ (podcast), Accidental Tech podcast, etc.

I donate to electoral-vote.com via Patreon. Most of my news comes from podcasts and those ads don’t bother me as much as web ads. I said I wasn’t morally opposed to them.

I don’t use any app that is ad supported without the option to pay for an ad free experience the same with streaming services.

Nursie · 11h ago
I'm intrigued as to why someone, presumably tech-savvy, would do that?

We know they aren't oracles and come up with a lot of false information in response to factual questions.

JKCalhoun · 9h ago
Google searches also come up with a lot of false information (well, it's where LLMs get their learnin' from — the Internet).

I'm never asking LLMs anything super critical like, "Do my taxes for me." This morning (as an example) I asked: "Is there talk of banning DJI drones in the U.S.?"

Later: "Difference between the RSE20 and RSS20 models of Yamaha electric guitars?"

And "Is there an Eco-Tank ink-jet suitable for Dye-Sub inks that can print Tabloid?"

1) None of the above are "critical".

2) All would have been a slight pain using Google and generated a lot of ... noise. LLMs eat the noise, do a decent job of giving me the salient points that answer my question.

3) All are easily verifiable if, for example, I decided to make a purchase based on what the LLM told me.

Or put another way, my disappointment in LLMs pales in comparison to my disappointment in search.

Or maybe I am just sick of search engines and want to "stick it to them".

telchior · 10h ago
In my view it's a pretty straightforward calculation. Nothing is free, no knowledge is instant. Start off knowing your time investment to learn anything is greater than zero and go from there..

If you do a Google (or other engine) search, you have to invest time pawing through the utter pile of shit that Google ads created on the web. Info that's hidden under reams of unnecessary text, potentially out of date, potentially not true; you'll need to evaluate a list of links and, probably, open multiple of them.

If you do an AI "search", you ask one question and get one answer. But the answer might a hallucination or based on incorrect info.

However, a lot of the time, you might be searching for something you already have an idea of, whether it's how to structure a script or what temperature pork is safe at; you can use your existing knowledge to assess the AI's answer. In that case the AI search is fast.

The rest of the time, you can at least tell the AI to include links to its references, and check those. Or its answer may help you construct a better Google search.

Ultimately search is a trash heap of Google's making, and I have absolute confidence in them also turning AI into a trash heap, but for now it is indeed faster for many purposes.

t0lo · 1h ago
Working as designed :) enjoy the upcoming informational dark age
eitau_1 · 3h ago
I hope the small web will thrive again once the profit incentive for putting content on the web ceases to exist so SEO/dark-pattern heavy players will give up and stop suppressing valuable (altruistic?) stuff.
toenail · 6h ago
Why could i perform a web search that leads me to multiple results with cloudflare identity checks and ad spam when i can just use a llm..
indymike · 48m ago
The web is dying.
nelblu · 1h ago
This is all nice while the AI overview has no ads in it. Wait until it gets enshittified, and I have a feeling AI overview would be so infested with ads that people would rather block it and proceed to reading the actual links.

All this said, I am guilty of using this a lot these days - while it is still ad-free. I just ask chatgpt.com to give me a recipe of "XYZ" and it gives it immediately without any annoying repetitive content.

Escapade5160 · 2h ago
It's unfortunate, but websites now make content for AI. We used to use a browser, to view a site, to gain information. Now we use an AI, to tell us the information.
falcor84 · 4h ago
> So more people could be walking away from a search with the wrong information.

I'd actually be very interested in some real research about this. My impression is that in general, SEO-optimized sites are (intentionally) so bad at transmitting information that the difference between an average person "doing their own research" on these sites vs reading the AI overview would either be negligible, or in favor of the AI summaries.

Mars008 · 5h ago
As I understand it they are saying users waste significantly less time on search if there are AI overviews. Is it really bad? Not from users perspective.
wiseowise · 6h ago
And thank God! It does. I’m done dealing with bazillion shitty websites with bad, slow, performance, bad ui and dark patterns. All I need is an information in a convenient format and that is what AI tools provide to me.
vouaobrasil · 6h ago
What's interesting is that:

0. Internet is initially pretty good.

1. Google introduces search algorithm that is pretty good.

2. SEO becomes a thing because of Google, and makes the web bad.

3. AI (including Google's AI) bypasses that.

4. The web is eradicated, and Google/other AI companies are the only place where you can get information...

Seems like the web would have been better off without Google.

lll-o-lll · 4h ago
You are forgetting:

0.5 User 1,563,018 puts their credit card details in to make the world’s first online transaction!

0.50001 The web is filled with spam and unsearchable for real information

0.9 Some smart nerds figure out an algorithm to find the signal in the noise

1.9 Google throws out “don’t be evil” because search is cost, ad’s are money

4.1 Google and the rest of the AI/s subvert human decision making in ways that marketers could only ever dream

∞.∞ Humans toil in slavery to the whims of our corporate overlords

I wanted a different dystopia.

jacquesm · 5h ago
All you need is good information and AI tools are giving you information without you knowing whether or not it is any good. You may think it is good, but unless you know more about the answer to your query than what you needed to create it you won't be able to tell the difference. If you did then you would not have been asking in the first place. Effectively you are now believing in an oracle.
LinXitoW · 4h ago
Which is different from believing the information on an ad and dark pattern infected website how?

Since the AI is trained on the data on the websites below the AI summary, the summary quality is basically lock step with the quality of the websites.

jacquesm · 4h ago
It is different because the AI is 'Google' branded and not 'ad and dark pattern infected website' branded. That proximity to Google's branding conveys trust, but it just whitewashes the content from that ad and dark pattern infected website. Only now you don't know about it.
Therenas · 6h ago
Okay, but when those sites go out of business, where does the AI get its information from? This is obviously not sustainable.
bloak · 5h ago
It's only the advertising-funded sites that go out of business and a lot of those sites were in any case just scraping other sites. What proportion of reliable online information is only available from a web site that is funded by advertising? It's not zero, but it's not a very big number, either, I suspect, so it might be sustainable.
croes · 6h ago
But that also kills the good sites.
teddyh · 5h ago
Only if those “good sites” were dependent on advertisements for survival.
LinXitoW · 4h ago
How many of the "good sites" do you pay for regularly? 99.9% of ALL sites are ad-funded, good or bad. Most people, even people with the disposable income, don't pay for the good sites, esp. because their value is only fractions of a cent from that one google search a month.

My irrational hope is that the "good" sites establish a shared Spotify-esque model, where I pay a basic subscription that then gets distributed roughly by usage to all the websites. There is no chance in hell anyone is willing to have the 20 subscriptions to support all the websites they've gotten utility from (directly OR indirectly) this month.

yfw · 15h ago
Maybe if the search wasnt full of ads and scams
throwawayoldie · 10h ago
...the same content that the AI was trained on, you mean?
awakeasleep · 9h ago
Thats not a real rebuttal.

First, in the pre training stage humans curate and filter the data thats actually used for training.

Then in the fine tuning stage people write ideal examples to teach task performance

Then there is reinforcement learning from human feedback RLHF where people rank multiple variations of the answer an AI gives, and thats part of the reinforcement loop

So there is really quite a bit of human effort and direction that goes into preventing the garbage-in garbage-out type situation you're referring to

landl0rd · 10h ago
At least they're not thrown in my face and appearing as eighteen pop-ups, notification requests, account-walls, SEOslop, and a partridge in a pear tree.
bomb17982 · 2h ago
AI Overviews are great for users but brutal for publishers. Studies show sites lose roughly 40–50% of clicks when these summaries appear. It feels like another step toward Google keeping users on the SERP, instead of sending traffic out.

If this sticks, creators will either need to optimize for being included in the AI answer or focus more on direct channels (newsletters, communities) to survive.

metalrain · 6h ago
People only click from search when they have actual need to visit the site. E-commerce, image/video content, social media, etc.

That is pretty much as it should be.

Oras · 6h ago
This! Why would I go and read 2000 words optimized article for SEO when I can get the answer in one line?

I only go to a website to perform an action as you said.

Huxley1 · 6h ago
This situation makes sense since everyone wants quick answers, and AI summaries are really convenient. A lot of people just read the summary without clicking through to the original content. But this can be a real problem for creators who rely on traffic. Does anyone have ideas on how to balance both user needs and support for creators?
zb3 · 26m ago
> But this can be a real problem for creators who rely on traffic.

Wait.. why would that creator be paid for the traffic if I'm not buying any shitty overpriced product advertised on these scammy ads, ever? Especially when I don't even see these thanks to ublock.

kazinator · 9h ago
> Google users are less likely to click on a link when they encounter search pages with AI summaries

Well, doh:

"Do I just read the AI summary? Or click past five pages of ads and spam to maybe find an organic link to something real?"

watwut · 6h ago
I find this claim massivy exaggerated.
yalogin · 9h ago
For the most part there really is no need to use search in the traditional sense for knowledge. For information it’s still the only because llms are not reliable. But ChatGPT must have taken a huge dent in google’s traffic.
ctas · 9h ago
We (Geostar.ai) work with many brands and companies that have experienced near-death situations caused by Google's AI Overviews. The negative impact this feature has had on people's livelihoods is heartbreaking to witness.

Just today, I met with a small business owner who showed me that AIO is warning users that his business is a scam, based on bogus evidence (some unrelated brands). It's a new level of bullshit. There's not much these businesses can do other than playing the new GEO game if they want to get traffic from Google.

Who knows if Google will even present any search results other than AIO a few years from now.

maurits · 2h ago
(forgot what type of flour you need for bread in supermarket)

innocently googles 'flour bread'

half the screen, CONTINUE WITH GOOGLE - stay in browser, click

COOKIES We and our 917 partners CARE ABOUT YOUR PRIVACY, click, click,

NEWSLETTER, NEWSLETTER, click, rotate screen because the overlay is to big, click Im sad person who's doesn't want daily bread in his mailbox.

APP APP APP, install APP, click click, can't hit the x, let it be

LOG IN WITH YOUR FOOFLE ACCOUNT, click

5 pages with autoplay video and SEO slop

I'm enjoying AI, while it lasts.

xpressvideoz · 3h ago
Some HNers said AI overviews sucked and they would never use them, and here it is.
jgalt212 · 42m ago
If searchers (people, I assume) are not clicking on content links like they used to, it's safe to assume that they are not clicking on ad links as well. Google is digging their own grave here. That being said, given the zeitgeist, they really have no choice but to offer AI summaries at least on par with the competition.
h4kunamata · 9h ago
Not any AI tho.

I have replaced SEO with Perplexity AI only. It isn't a chatbot but it actually search for what you are looking for and most importantly, it shows all the sources it used.

Depending on the question I can get anywhere from 10 to 40 sources. No other AI service provides that, they use the data from their training model only which in my experience, is full of errors, incomplete, cannot answer or altogether.

nneonneo · 9h ago
Google's AI overviews gives you sources. The sources don't always say what Google AI thinks it says. Or, Google will mix up two sources that are talking about totally different things, assume they're the same thing, and then show you nonsense generated from the mix.

I don't immediately assume that Perplexity will be any better off. Citing sources is great, but I'd rather just read the sources myself rather than assuming that the AI did anywhere a good job of actually summarizing them properly. At that point, what does the AI actually usefully bring to the table?

didibus · 9h ago
Almost all other AI do that as well. ChatGPT will show you like 10+ sources, it'll put it next to the part of the answer that the source was used for too.
jjani · 9h ago
Huh? All of the other AI service provide that. Geminii, Claude, GPT.
wkat4242 · 15h ago
To be fair, Google's actual search couldn't be much worse than it was lately. It's like they really try to get all the spam, clickbait and scams right at the top.

The AI overview sucks but it can't really be a lot worse than that :)

Vespasian · 9h ago
I have a lot of confidence that Google will figure out a way to do it.

The same economic incentives that led to SEO slop are still there in my opinion.

More "content" equals more opportunity to integrate ads even if they are not woven into the AI response directly. It will be tuned to allow and all that changes is cutting out the website provider.

Google is incapable/unwilling to do anything beyond flooding the world with ads. They don't have a great track record of actually selling things to people for money.

krupan · 10h ago
I feel like the discussion here is missing the point. It doesn't matter if the AI overview is correct or not, it doesn't matter if you can turn it off or not. People are using it instead of visiting actual websites. Google has copied the entire World Wide Web into their LLM and now people not using the web anymore! We have bemoaned the fact that Facebook and Twitter replaced most of the web for most people, but now it's not even those, it's a single LLM owned and controlled by a single corporation.
landl0rd · 10h ago
Is there an appreciable difference between a company that controls what information is surfaced via pagerank and one that does so via LLM?

Remember the past scandals with google up/downranking various things? This isn't a new problem. Wrt how the average person gets information google doesn't really have more control because people aren't clicking through as much.

nicbou · 7h ago
The main difference is that pagerank still rewarded content creators with an audience. Now Google extracts all the value of their work.
da25 · 1h ago
AI systems still struggle with hallucination, especially when your intent of the query is to obtain the latest information. These models have become very good at handling stored information, but their ability to stay updated remains slow. There is a large volume of queries whose responses and underlying concepts don't change much, and the current AI systems, through innovation and user-friendly interfaces, has managed to cast a wide net over such pre-learned responses.

However, unless these AI base model businesses create strong enough incentives for information owners to provide updates and engage in information pipelines that provide fresh data, the current moat of AI might not extend far enough to cover the full spectrum of queries.

The “fossil fuel” of LLMs-static public internet data-is running out.

Current efforts in RL help systems answer queries beyond their pre-learned knowledge by expanding on the user’s prompt, wherein the system ventures into unknown data territories through agents and self-talk, resulting in action-result memories. These, in turn, serve as a large enough context-rich prompt to have all the needles-in-hay-stack that form the final answer or response. This is made possible by large context windows.

For live internet queries, RL can work by expanding context with latest results fetched from the public web using a crawler. However, this is often done without the explicit consent from information providers and offers them little incentive beyond a link in the AI’s response summary. As a result, some providers have started withholding data, and many services now offer tools to block AI crawlers. Meanwhile, multimodal AI systems-capable of understanding text, visuals, and audio-are developing agents that can access content through simulated browser use, effectively bypassing traditional crawler firewalls.

This reality highlights the need for a good incentive system for information providers, one that encourages them to share dense, efficiently and ai-structured data. Certain domains have already begun embracing this and sharing their information in ai-native formats, since they have no moat in that information and rather see positive incentives - for example, certain documentation websites for tools and frameworks now provide formatted versions of their docs at /LLMs.txt links.

If the information is the resource exchanged on these internet pathways, businesses fundamentally operate either by generating this resource or by retrieving it once it exists, and the other businesses enable this whole endeavour. Ultimately, individuals and organizations will, seek, share and exchange information in ways that enables them to efficiently take decisions and their next actions. Therefore, the incentive to access the most up-to-date information becomes critical when those actions depend on accuracy and timeliness.

WA · 4h ago
Several factors are at play here, which are somewhat contradictive:

1. Publishers feel entitled to traffic, because Google send them traffic in recent years. See [1] for example:

> "Google's core search engine service is misusing web content for Google's AI Overviews in Google Search, which have caused, and continue to cause, significant harm to publishers, including news publishers in the form of traffic, readership and revenue loss," the document said.

> "Publishers using Google Search do not have the option to opt out from their material being ingested for Google's AI large language model training and/or from being crawled for summaries, without losing their ability to appear in Google's general search results page," the complaint said.

I have zero pity for these publishers.

2. Not all traffic is created equal. Does a business really lose customers, just because traffic goes down? Maybe that traffic wouldn't have converted anyways. This is basically the old piracy discussion revamped with businesses arguing that every single copy of a movie or a game would've been a customer. It's idiotic.

3. But: Google is now a content provider and different rules apply for those than being merely on a comparable level like an ISP. This has been discussed for years with social networks in minds. Hence, Google needs to be held accountable for providing harmful information (such as in another story here: the temperature for heating pork and when food safety is affected).

[1]: https://www.reuters.com/legal/litigation/googles-ai-overview...

feverzsj · 7h ago
The AI overview is wrong in most of the time when I check the referenced links. It's even worse than ads.
maxdo · 14h ago
Pay per click model Should die , it’s really ugly world where you need to fight through loads of ads to get tiny bit of information.

People will go to museums to see how complicated pre-ai era was

Gigachad · 14h ago
Yep, there’s so much hate for people who don’t read past the headline, but if you actually click on the articles the websites are almost unusable.
r33b33 · 4h ago
Good. About time search dies
skywhopper · 16h ago
Which is of course Google’s short-sighted goal. See also their push to switch to full “AI mode” search which doesn’t show results at all.
thewebguyd · 16h ago
It's a weird goal to me. Like, what's their end game here? Offer to manipulate the AI responses for ad money? Product placement in the summaries? I would hope those placements have to be disclosed as advertising, and it would immediately break trust in anything their AI outputs so surely that would only continue to harm them in the long run, no?

~57% of their revenue is from search advertising. How do they plan on replacing that?

mrheosuper · 10h ago
AI subscription would be my guess. Want better model ? Open your wallet.
gundmc · 9h ago
I'm pretty sure their endgame right now is just "don't lose all of your market share to companies that _will_ use LLMs".
xt00 · 15h ago
Yea it is tricky for them -- the old model of "search, see google text / link ad, scroll, click website, scroll, see some ads on that page as well, done" will be replaced with "search, see google text / link ad, read AI result, 'and here are some relevant websites'" -- where all of the incentives there will be to "go into more depth" on the websites that are linked there.
bugsMarathon88 · 12h ago
Total behavioral control through the augmentation of senses, emotions and all other sensibilities. Such political power is significantly more valuable than mere revenue.
landl0rd · 10h ago
Okay I think the question is still how they plan to convert this into cash, because political power can't buy food or pay your employees or be stored and quantified simply, which is why we invented money. Assuming this dystopian scenario is correct.
flashgordon · 16h ago
So youd be surprised and scared - the Ad PMs I know are totally salivating at this. Their angle is "SEO is no more - it is GEO now". GenAI Engine Optimization. Welcome to the Futurama Internet Future!
EarlKing · 14h ago
"Futurama does not endorse the COOOOOOL crime of fraudulent misrepresentation!"

Seriously, Futurama and Cyberpunk and 1984 were all supposed to be warnings... not how-to manuals.

wulfstan · 4h ago
If you're tired of this BS can I encourage you to make:

https://noai.duckduckgo.com/

Your default search engine. AI overviews begone.

protocolture · 7h ago
First they came for the stackoverflow comments, and I said nothing, for stackoverflow is notorious for being full of jerks.

Then they came for the search engine clicks, and I said nothing, because monetised search contributed heavily to the enshittification of the internet.

Great so far. Can AI kill off social media and internet surveillance next?

nicbou · 7h ago
Who will provide the training data for AI when Google has stolen everyone's audience and revenue?
HPsquared · 16h ago
I've seen blatantly wrong stuff in that overview too many times, I just ignore it now.
Jare · 15h ago
To be fair, the actual results are often even worse. I'm pretty sure we're close to the point where our favorite AI prompt replaces classic googling. While it will get a lot of the answer wrong, it will lead to the right result faster than plain searches. If nothing else, because refining our search at the AI prompt will be way easier than in classic google. Google knows and needs to stay on top of this paradigm change, but I guess doesn't know how to monetize AI search yet so it doesn't want to force the change (yet).
throaway5454 · 12h ago
Ai used in this way is going to replace gui as we know it. Why click when you can just tell ai what you want to to do.
thejohnconway · 11h ago
Because I usually don’t want to talk to computers in front of other people? It isn’t that it feels silly, but that it’s incredibly distracting for everyone to hear every interaction you have with a computer. This is true even at home.

Maybe we can type the commands, but that is also quite slow compared with tapping/clicking/scrolling etc.

Nursie · 11h ago
Sometimes search results don't contain the info you need, sometimes they are SEO-spam, or a major topic adjacent to what you need to know about floods the results.

But they're not often confidently wrong like AI summaries are.

ahartmetz · 12h ago
I recently typo'd something and the AI box just fabricated a semi-plausible story about how the ancient X did Y.
csomar · 7h ago
I am not sure how everyone is ignoring that or just getting along with it. I've had it return that the earth population is 10bn which is blatantly wrong and infuriating. It hallucinates harder than regular LLMs which are already bad for getting reliable answers.
anothernewdude · 4h ago
I thought the search clicks was a massive driver of data for the search engine. How is this not a bad thing for the search engine company?
jama211 · 5h ago
Well, yeah.
cess11 · 5h ago
I find this rather worrying, because the summaries I look at are frequently wrong or misleading.

Over the last years I've moved over a lot of initial search to Wikipedia, which either answers directly or provides useful links.

blibble · 14h ago
no reason not to block Googlebot now...
buyucu · 6h ago
AI overviews are search. It's the next iteration of how web searching works.
SwtCyber · 5h ago
But hey, I'm sure Google's quarterly earnings are doing just fine
tropicalfruit · 7h ago
if you can get answer in 1 step, why 2?

things like "gemini said" and "gpt said" will enter common lingo like "google it" did in the 2010s

ars · 15h ago
The AI overview doesn't (for me) cause a big drop in clicking on sites.

But AI as a product most certainly does! I was trying to figure out why a certain AWS tool stopped working, and Gemini figured it out for me. In the past I would have browsed multiple forums to figure out it.

josteink · 6h ago
I'm going to call good old "supply & demand" at this.

There was a demand for information, and people made websites to make that information easily accessible. The market was working.

Then cheap/free hosting combined with Google Ads came along incentivized the market to SPAM the internet with SEO-optimized websites which makes that information just as inaccessible as the market could possibly tolerate. Basically the market converged on enshittification, rather than excellence.

The result of that was that users were forced to search, the first SEO spam, then the second, then the third and maybe somewhere along the line they found what they were really looking for, from a real website. The market just barely supplied what there was a demand for.

But thanks to AI and LLMs, the power has now shifted back to the users, and we can sidestep all the Google Ads SEO SPAM enshittification nonsense. If this is done via ChatGpt, DDG Chat or Google AI overviews, I don't really care.

Once again we have gained easy access to information without any ill-incentivized middle men making the process cumbersome. This is absolutely a good thing. If this means the death of Google Ads funded content farms, I'm all for it.

The niche sites which contains real information was never funded or driven by Google Ads in the first place, so hopefully they won't buckle under during this minor AI apocalypse.

southernplaces7 · 7h ago
My question here is: why would Google try to actively sabotage, possibly even devastate the very backbone cash cow that is all those spammy, SEO-laden search and ad results right below its own AI overview, that's displacing them for engine users who don't bother with the rest (I thing I personally avoid because fuck you hallucinating AI overview)

Anyone have any ideas?

EDIT: Also, the irony of Google, which for years harped on and on about working to fight incorrect, erroneous and also even "misinformation-laden" search results, now making a hallucinating AI that often makes up outright bullshit with authority as its primary, first-line result for information. Truly a company that bathes in an ever deeper pool of shit the bigger it grows.

j45 · 15h ago
This means searches are still happening, just being routed elsewhere?

I noticed Google's new AI summary let's me click on a link in the summary and the links are posted to the right.

Those clicks are available, might not be discovered yet, curious though if those show up anywhere as data.

Google being able to create summaries off actual web search results will be an interesting take compared to other models trying to get the same done without similar search results at their disposal.

The new search engine could be google doing the search and compiling the results for us how we do manually.

thewebguyd · 15h ago
> Google being able to create summaries off actual web search results will be an interesting take compared to other models trying to get the same done without similar search results at their disposal.

And may get them in some anti-trust trouble once publishers start fighting back, similar to AMP, or their thing with Genius and song lyrics. Turns out site owners don't like when Google takes their content and displays it to users without forcing said users to click through to the actual website.

AnimeLife · 8h ago
They have decades of data to map queries to pages. They also have access to crawl the pages in advance.