logoalt Hacker News

dnhkngyesterday at 5:32 PM4 repliesview on HN

Because its generally expected that models only work 'in distribution', i.e. they work on stuff they have previously seen.

They almost certainly have never seen regular conversations in Base64 in their training set, so its weird that it 'just works'.

Does that make sense?


Replies

dormentoyesterday at 5:36 PM

For all we know, AI tech companies could theoretically have converted all of the "acquired" (ahem!) training set material into base64 and used it for training as well, just like you would encode say japanese romaji or hebrew written in the english alphabet.

show 2 replies
fweimeryesterday at 9:01 PM

If you do not properly MIME-decode email, you end up with at least some base64-encoded conversations.

broDogNRGyesterday at 7:06 PM

[dead]