They're not blenders.
This is clear from the fact that you can distill the logic ability from a 700b parameter model into a 14b model and maintain almost all of it.
You just lose knowledge, which can be provided externally, and which is the actual "pirated" part.
The logic is _learned_
Are there any recent publications about it so I can refresh myself on the matter?
It hasn't learned any LOGIC. It has 'learned' patterns from the input.