logoalt Hacker News

b65e8bee43c2ed0today at 5:09 AM3 repliesview on HN

doesn't it get tiring after a while? using the same (perceived) gotcha, over and over again, for three years now?

no one is ever going to release their training data because it contains every copyrighted work in existence. everyone, even the hecking-wholesome safety-first Anthropic, is using copyrighted data without permission to train their models. there you go.


Replies

necovektoday at 5:45 AM

There is an easy fix already in widespread use: "open weights".

It is very much a valuable thing already, no need to taint it with wrong promise.

Though I disagree about being used if it was indeed open source: I might not do it inside my home lab today, but at least Qwen and DeepSeek would use and build on what eg. Facebook was doing with Llama, and they might be pushing the open weights model frontier forward faster.

show 2 replies
Tepixtoday at 6:40 AM

Nvidia did with Nemo.

show 1 reply
fragmedetoday at 5:32 AM

it's not a gotcha but people using words in ways others don't like.

show 1 reply