This seems to be getting lost in the noise in the stampede for infrastructure funding
Deepseek v3 at $5.5M on compute and now r1 a few weeks later hitting o1 benchmark scores with a fraction of the engineers etc. involved ... and open source
We know model prep/training compute has potentially peaked for now ... with some smaller models starting to perform very well as inference improves by the week
Unless some new RL concept is going to require vastly more compute for a run at AGI soon ... it's possible the capacity being built based on an extrapolation of 2024 numbers will exceed the 2025 actuals
Also, can see many enterprises wanting to run on-prem -- at least initially