logoalt Hacker News

franga200005/14/20251 replyview on HN

This is an LLM approximating a semantic calculator, based solely on trained-in knowledge of what that is and probably a good amount of sample output, yet somehow beating the results of a "real" semantic calculator. That's crazy!

The more I think about it the less surprised I am, but my initial thoughts were quite simply "now way" - surely an approximation of an NLP model made by another NLP model can't beat the original, but the LLM training process (and data volume) is just so much more powerful I guess...


Replies

CamperBob205/14/2025

This is basically the whole idea behind the transformer. Attention is much more powerful than embedding alone.

show 1 reply