logoalt Hacker News

AstroBentoday at 3:47 AM1 replyview on HN

Our DNA does contain our pre-training, though. It's not true that we're an entirely blank slate.


Replies

davebrentoday at 4:10 AM

Pre-training is not a good term if you are trying to compare it to LLM pre-training. Closer would be the model's architecture and learning algorithms which has been designed through decades of PhD research, and my point on that is that the differences are still much greater than the similarities.