You'd be surprised how quickly improvement of autoregressive language models levels off with epoch count (though, admittedly, one epoch is a LOT). Diffusion language models otoh indeed keep profiting for much longer, fwiw.
Does this also apply to LLM training at scale? I would be a bit surprised if it does, fwiw.
Does this also apply to LLM training at scale? I would be a bit surprised if it does, fwiw.