logoalt Hacker News

falloutxyesterday at 10:51 PM3 repliesview on HN

Its not just AI scrappers doing it by themselves but now users are also being trained to put the link in the claude chat/chat gpt and ask it to summarise it. And off course that would show up on the website end as a scraper.

In fact firefox now allows you to preview the link and get key points without ever going to the link[1]

[1] https://imgur.com/a/3E17Dts


Replies

acattonyesterday at 10:58 PM

> In fact firefox now allows you to preview the link and get key points without ever going to the link[1]

> [1] https://imgur.com/a/3E17Dts

This is generated on device with llama.cpp compiled to webassembly (aka wllama) and running SmolLM2-360M. [1] How is this different from the user clicking on the link? In the end, your local firefox will fetch the link in order to summarize it, the same way you would have followed the link and read through the document in reader mode.

[1] https://blog.mozilla.org/en/mozilla/ai/ai-tech/ai-link-previ...

show 2 replies
TeMPOraLyesterday at 11:21 PM

Users are not being trained. Despite the seemingly dominant HN belief to the contrary, people use LLMs for interacting with information (on the web or otherwise) because they work. SOTA LLM services are just that good.

orbital-decayyesterday at 11:32 PM

It's three issues:

- AI shops scraping the web to update their datasets without respecting netiquette (or sometimes being unable to automate it for every site due to the scale, ironically).

- People extensively using agents (search, summarizers, autonomous agents etc), which are indistinguishable from scraper bots from website's perspective.

- Agents being both faster and less efficient (more requests per action) than humans.