The most important two words in this article are the last two: for now.
Indeed, for now generative models generate triangle soup without much thought. The same was true for 2D illustrations where generative models like Deep Dream came up with horrendous images with eyes all over, dogs with multitudes of heads and oh did I mention the eyes? That was about 10 years ago. Things changed, models improved, the eyes were tamed. Yes, people had too many or too few fingers but that also changed. From nightmare fuelling imagery with many-eyed dog heads sticking out where you don't want them to fully animated hi-res video only took a decade and things are still speeding up. The triangle soup of current 3D generative models is like the eye soup of Deep Dream, something to remember somewhat fondly which is no longer relevant now.
If that's really true, then why is the logo texture on the handle still total slop?
One flaw with this assumption is that images are available in literally counts of trillions to train on. With 3D models there are virtually no production quality models freely available to train on. Even companies like ILM or Weta have nowhere near the number of models that would be needed to train a robust modelling AI