logoalt Hacker News

gzreadtoday at 5:13 PM1 replyview on HN

And not in human-interpretable ways. An LLM was told to behave in a certain way and then output random numbers. When the numbers were pasted to another LLM instance, it also behaved that way. I wish I remembered more about that study or had a link to it - it was fascinating.


Replies