logoalt Hacker News

dragonwriteryesterday at 6:49 AM1 replyview on HN

> This is almost certainly because someone recently adjusted its prompt to tell it what its views on white genocide are meant to be

Well, someone did something to it; whether it was training, feature boosting the way Golden Gate Claude [0] was done, adjusting the system prompt, or assuring that it's internet search for contextual information would always return material about that, or some combination of those, is neither obvious nor, if someone had a conjecture as to which one or combination it was, easily falsifiable/verifiable.

[0] https://www.anthropic.com/news/golden-gate-claude


Replies

lolinderyesterday at 1:14 PM

Source [0]. The examples look pretty clearly like they stuck it in the context window, not trained it in. It consistently seems to structure the replies as though the user they're replying to is the one who brought up white genocide in South Africa, and it responds the way that LLMs often respond to such topics: saying that it's controversial and giving both perspectives. That's not behavior I would expect if they had done the Golden Gate Claude method, which inserted the Golden Gate Bridge a bit more fluidly into the conversation rather than seeming to address a phantom sentence that the user supposedly said.

Also, let's be honest, in a Musk company they're going to have taking the shortest possible route to accomplishing what he wanted them to.

[0] https://www.cnn.com/2025/05/14/business/grok-ai-chatbot-repl...

show 1 reply