I know how to do things by hand, man. But the writing is on the wall: that skill is going the way of writing programs on punchcards. And there's little we can do about it because the economics in favor of LLMs are like laws of physics.
Yes, model collapse is gonna suck. But LLMs are not just left to self-train, they are guided by human researchers who are going to find ways to groom and direct the models to avoid collapse. They can make billions by shipping better models, so why wouldn't they invest a lot of effort in that?
One of the amusing things about AI bros is how naively over-enthusiastic they are about the technology and its inevitability.
This is not terminals vs punchcards. This is like Windows ME over Windows 98. Or, maybe, the 286 over a 8086 when a 386 it's the proper path.