logoalt Hacker News

pankajdohareyyesterday at 10:45 PM1 replyview on HN

Looks like a Tiny Analytic transformer, RNN is arguably a better choice if you are gonna handwire an architecture to mechanically do addition. Learning is about discovering the patterns and algorithm from data. Wiring a machine to follow a procedure defeats that purpose.


Replies

dnauticstoday at 1:57 AM

it proves that the algorithm is embeddable in a bigger transformer of ~similar architecture.