logoalt Hacker News

LiamPowelltoday at 6:19 AM0 repliesview on HN

The very simplified answer is that the models are first trained on everything and then are later trained more heavily on golden samples with perfect grammar, spelling, etc..