logoalt Hacker News

gmerctoday at 6:13 AM1 replyview on HN

This betrays a lack of understanding how inference works. You cannot categorically defeat prompt injection with instructions. It does not work. There are no privileged tokens.


Replies

lmeyerovtoday at 6:26 AM

Yep! One of my favorite attacks is just having a very long piece of a text so the LLM becomes unclear what's important and is happy to do something else