We're in an era that presents some novel problems.
I've read a few people this week discuss the consideration that Anthropic's behavior itself will likely impact Claude's training.
The concern there is that if Claude ingests news articles that show Anthropic behaving in a manner that clashes significantly with the values they want to instill in Claude, it could make training less effective.
It's all very weird.
This comment is so deep, I fear to get lost in it.
If what you said was true, the only way to achieve a superior AI would be to incorporate the virtuous one is aiming at.
That would solve so many of the conundrums of the field, I wish it was true.