If you think that confusing message provenance is part of how thinking mode is supposed to work, I don't know what to tell you.
There is no "message provenance" in LLM machinery.
This is an illusion the chat UX concocts. Behind the scenes the tokens aren't tagged or colored.
There is no "message provenance" in LLM machinery.
This is an illusion the chat UX concocts. Behind the scenes the tokens aren't tagged or colored.