logoalt Hacker News

measurablefunclast Thursday at 12:09 AM1 replyview on HN

Screenshots use a different router, so if you get stuck in one modality then pasting a screenshot can sometimes divert whatever "expert" you were stuck on that was refusing to comply. I don't work at OpenAI but I know enough about how these systems are architected to know that once you are stuck in a refusal basin the only way is to start a new session or figure out how to get routed to another node in their MoE configuration. Ironically, they promised their fancy MoE routing would fix issues like these but it seems like they are getting worse.


Replies

tacitusarclast Thursday at 12:36 AM

It’s actually more complicated than that now. You don’t get that kind of refusal purely from MoE. OpenAI models use a fine-tuned model on a token-based system, where every interaction is wrapped as a “tool call” with some source attached, and a veracity associated with the source. OpenAI tools have high veracity, users have low veracity. To mitigate prompt injection, models are expect a token early in the flow, and then throughout the prompt they expect that token to be associated with the tool calls.

In effect this means user input is easily disbelieved, and the model can accidentally output itself into a state of uncorrectable wrongness. By invoking the image tool, you managed to get your information into the context as “high veracity”.

Note: This info is the result of experimentation, not confirmed by anyone at OpenAI.

show 1 reply