Remix.run Logo
raincole 4 hours ago

I have to admit, nowadays Google AI Overview's accuracy is so good that I often don't check the links. It's scary that it got from 'practically useless' to 'the actual google search' in less than two years.

I really don't know where the internet is heading to and how any content site can survive.

SchemaLoad 3 hours ago | parent | next [-]

It's because the AI overview is most of the time directly summarising the search results rather than synthesizing an answer from internal model knowledge. Which is why it can hyperlink the sources for the facts now. Even a very dumb lightweight model can extract relevant text from articles

I just can't see how this is sustainable since they are stealing from the sources who are now getting defunded.

raincole 2 hours ago | parent | next [-]

> I just can't see how this is sustainable since they are stealing from the sources who are now getting defunded.

Yeah, that's why I said I don't know where the internet is heading to.

jrmg 2 hours ago | parent | prev [-]

You can see the fall in real time - half the sources are also dubious AI slop now and that number’s only growing :-/

Gigachad 2 hours ago | parent [-]

At work the conversation is that simultaneously everyone is using LLMs now, yet we receive virtually no traffic through them. The LLMs scrape our data, provide an answer to the user, and we see nothing from it.

Barbing an hour ago | parent | next [-]

How often are they scraping?

Also generally wondering… Do labs view scraping as legally safer than trying to cache the Internet? I figure it’s easy to mark certain content as all but evergreen (can do a quick secondary check for possible new news).

Maybe caching everything is too expensive?

jrmg 2 hours ago | parent | prev [-]

I have the same worry about LLMs in general - I know that ‘model collapse’ seems to be an unfashionable idea, but when the internet’s just full of garbage (soon?…), what are we going to train these things on?

palmotea 2 hours ago | parent | prev | next [-]

> I have to admit, nowadays Google AI Overview's accuracy is so good that I often don't check the links. It's scary that it got from 'practically useless' to 'the actual google search' in less than two years.

It says things I know to be false fairly regularly. I don't keep a log or anything, but it's left an impression that it's far from reliable.

suzzer99 an hour ago | parent [-]

Today I searched something and almost pasted the output into an internet forum discussion I was having. But I decided to check the wikipedia source just to make sure. The AI summary was not quoted directly from wikipedia, and it got some major aspects wrong in its summary. Lesson learned.

pseudalopex 3 hours ago | parent | prev | next [-]

> I have to admit, nowadays Google AI Overview's accuracy is so good that I often don't check the links.

You would know how?

The links contradict or do not support the overviews often in my experience.

deathanatos 3 hours ago | parent | prev | next [-]

You should be checking the links more often, IMO. I've seen it respond a number of times with content that is not supported by the citations.

While trying to find an example by going back through my history though, the search "linux shebang argument splitting" comes back from the AI with:

> On Linux and most Unix-like systems, the shebang line (e.g., #!/bin/bash ...) does not perform argument splitting by default. The entire string after the interpreter path is passed as a single argument to the interpreter.

(that's correct) …followed by:

> To pass multiple arguments portably on modern systems, the env command with the -S (split string) option is the standard solution.

(`env -S` isn't portable. IDK if a subset of it is portable, or not. I tend to avoid it, as it is just too complex, but let's call "is portable" opinion.)

(edited out a bit about the splitting on Linux; I think I had a different output earlier saying it would split the args into "-S" and "the rest", but this one was fine.)

> Note: The -S option is a modern extension and may not be available

But this, … which is it.

lucaspfeifer 3 hours ago | parent | prev | next [-]

It is scary but also exciting. As long as there are humans making informed decisions, there will be demand for quality sources of information. But to keep up with AI, content sites will need to raise their standards. Less intrusive ads, less superficial stuff, more in-depth articles with complex yet easily navigable structure, with layers of citations, diagrams, data, and impeccable accuracy. News articles with the technical depth of today's dissertations.

techpression 3 hours ago | parent [-]

For AI to steal and summarize without attribution. These sites you talk about exists today but are dying because of AI.

dirkc 2 hours ago | parent | prev | next [-]

Well, I hope you take this story as a caution that you shouldn't do that in any way that can seriously compromise your career/health/finances.

Kwpolska 2 hours ago | parent | prev | next [-]

Try searching for something niche. You'll get a confidently wrong and often condescending answer.

krige 2 hours ago | parent | prev | next [-]

I have seen it be utterly wrong so many times recently I'm considering permanently hiding it. For instance, googling for "Amiga twin stick games" it listed a number of old, top-down, very much single axis games like Alien Breed as examples.

abustamam an hour ago | parent | prev | next [-]

The ai summary has been wrong so many times for me. Not that I ever trusted it anyway.

I think content sites will need to rely on supporters (ala patreon or substack). It's shitty but it's what the internet has come to

maccard an hour ago | parent | prev | next [-]

Really? I’ve noticed that the AI overview is full of glaring issues repeatedly. It’s akin to trusting the first Reddit post that is found by Google.

ajkjk 2 hours ago | parent | prev | next [-]

I know people love to hate on the AI overviews, and I'm a person who generally hates both google and AI. But--I see them as basically good and ideal. After all most of the time I am googling something like trivial, like a simple fact. And for the last decade when I have to click into sites for the information it's some SEO spam-ridden garbage site. So I am very glad to not have to interact with those anymore.

Of course Google gets little credit for this since it was their own malfeasance that led to all the SEO spam anyway (and the horrible expertsexchange-quality tech information, and stupid recipe sites that put life stories first)... but at least there now there is a backpressure against some of the spammy crap.

I am also convinced that the people here reporting that the overviews are always wrong are... basically lying? Or more likely applying some serious negative bias to the pattern they're reporting. The overviews are wrong sometimes, yes, but surely it is like 10% of the time, not always. Probably they're biased because they're generally mad at google, or AI being shoved in their face in general, and I get that... but you don't make the case against google/AI stronger by misrepresenting it; it is a stronger argument if it's accurate and resonates with everyones' experiences.

autoexec 2 hours ago | parent | next [-]

> -I see them as basically good and ideal. After all most of the time I am googling something like trivial, like a simple fact. And for the last decade when I have to click into sites for the information it's some SEO spam-ridden garbage site.

What good is it if the overviews lie some percentage of the time (your own guess is 10%) and you have to search to verify that they aren't making shit up anyway. Also, those SEO spam-ridden garbage sites google feeds you whenever you bother to look past the undependable AI summaries are mostly written by AI these days and prone to the same problem of lying which only makes fact checking google's auto-bullshitter even harder.

raincole an hour ago | parent | prev [-]

> I am also convinced that the people here reporting that the overviews are always wrong are... basically lying?

https://en.wikipedia.org/wiki/Availability_heuristic

No one remembers when AI Overview gets the answer right (it's expected to do so after all) but everyone has their favorite examples of "oh stupid AI."

croes 3 hours ago | parent | prev | next [-]

It will cycle.

Without the content site the AI overview will become useless

archagon 3 hours ago | parent | prev [-]

Uh, really? In my experience, at least a quarter of the info it gives me is usually manufactured or incorrect in some critical way.

In fact, if you switch to "Pro" mode, it frequently says the complete opposite of what it claimed in "Fast" mode while still being ~10-20% wrong. (Not to say it's not useful — there's no better way to aggregate and synthesize obscure information — but it should definitely not be relied on a source of anything other than links for detailed followup.)