I agree, but with the caveat that it's probably a bad time to fall asleep at the wheel. I'm very much a "nothing ever happens" kind of guy, but I see a lot of people who aren't taking the time to actually understand how LLMs work, and I think that's a huge mistake.
Last week I showed some colleagues how to do some basic things with Claude Code and they were like "wow, I didn't even know this existed". Bro, what are you even doing.
There is definitely a lot of hype and the lunatics on Linkedin are having a blast, but to put it mildly I don't think it's a bad investment to experiment a bit with what's possible with the SOTA.
> the lunatics on Linkedin are having a blast
That’s a nice way to put it, made me chuckle. :)
I mean I didn’t find out about Claude code until like a week ago and it hasn’t materially changed my work, or even how I interact with LLMs. I still basically copy paste into claude on web most of the time.
It is ridiculously cool, but I think anybody developer who is out of the loop could easily get back into the loop at any moment without having to stay caught up most of the time.
> I see a lot of people who aren't taking the time to actually understand how LLMs work
The trouble is that the advice in the post will have very little impact on "understanding how LLMs work". The number of people who talk about LLMs daily but have never run an LLM local, and certainly never "opened it up to mess around" is very large.
A fun weekend exercise that anyone can do is to implement speculative decoding[0] using local LLMs. You'll learn a lot more about how LLMs work than reading every blog/twitter stream mentioned there.
0. https://research.google/blog/looking-back-at-speculative-dec...