Remix.run Logo
falloutx 4 hours ago

Its not just AI scrappers doing it by themselves but now users are also being trained to put the link in the claude chat/chat gpt and ask it to summarise it. And off course that would show up on the website end as a scraper.

In fact firefox now allows you to preview the link and get key points without ever going to the link[1]

[1] https://imgur.com/a/3E17Dts

acatton 4 hours ago | parent | next [-]

> In fact firefox now allows you to preview the link and get key points without ever going to the link[1]

> [1] https://imgur.com/a/3E17Dts

This is generated on device with llama.cpp compiled to webassembly (aka wllama) and running SmolLM2-360M. [1] How is this different from the user clicking on the link? In the end, your local firefox will fetch the link in order to summarize it, the same way you would have followed the link and read through the document in reader mode.

[1] https://blog.mozilla.org/en/mozilla/ai/ai-tech/ai-link-previ...

ericd 4 hours ago | parent | next [-]

That’s awesome :-)

Like, can we all take a step back and marvel that freaking wasm can do things that 10 years ago were firmly in the realm of sci-fi?

I hope they’ll extend that sort of thing to help filter out the parts of the dom that represent attention grabbing stuff that isn’t quite an ad, but is still off topic/not useful for what I’m working on at the moment (and still keep the relevant links).

falloutx 4 hours ago | parent | prev [-]

I actually didnt know it was using a local model and that it fetches it locally.

DrewADesign 3 hours ago | parent [-]

They should advertise that. I pretty much reflexively avoid any mention of AI in interfaces because they usually mean "we're sending this all to openthropigoogosoft so I hope you don't have any secrets."

godelski 2 hours ago | parent [-]

  > They should advertise that
They did

   Previews can optionally include AI-generated key points, which are processed on your device to protect your privacy.
https://www.firefox.com/en-US/firefox/142.0/releasenotes/

I'll also add that if you go to the Labs page (in settings) you can enable another local model to semantically search your history

DrewADesign an hour ago | parent | next [-]

Ok, they should advertise it more .

an hour ago | parent | prev [-]
[deleted]
orbital-decay 4 hours ago | parent | prev | next [-]

It's three issues:

- AI shops scraping the web to update their datasets without respecting netiquette (or sometimes being unable to automate it for every site due to the scale, ironically).

- People extensively using agents (search, summarizers, autonomous agents etc), which are indistinguishable from scraper bots from website's perspective.

- Agents being both faster and less efficient (more requests per action) than humans.

TeMPOraL 4 hours ago | parent | prev [-]

Users are not being trained. Despite the seemingly dominant HN belief to the contrary, people use LLMs for interacting with information (on the web or otherwise) because they work. SOTA LLM services are just that good.