logoalt Hacker News

jaggsyesterday at 7:01 PM5 repliesview on HN

I think it's becoming clear that these mega AI corps are juggling with their models at inference time to produce unrealistically good results. By that it seems that they're just cranking up the compute beyond reasonable levels in order to gain PR points against each other.

The fact is most ordinary mortals never get access to a fraction of that kind of power, which explains the commonly reported issues with AI models failing to complete even rudimentary tasks. It's now turned into a whole marketing circus (maybe to justify these ludicrous billion-dollar valuations?).


Replies

andy12_yesterday at 7:44 PM

Models drop in price x10 each year. Us, common folk, getting access to these kinds of models is just a matter of time.

show 1 reply
Workaccount2yesterday at 9:54 PM

The bleeding edge behind closed doors token burning monsters of 2023 are bad compared to the free LLMs we have now.

I believe it was Sundar in an interview with Lex who said that the reason they haven't developed another Ultra model is because by the time it is ready to launch, the flash and pro versions will have already made it redundant.

show 1 reply
Davidzhengyesterday at 8:53 PM

Ok but if they can pump those compute and get science/math advancements it's worth something even if the costs are very high

show 1 reply
Davidzhengtoday at 1:32 AM

I think part of it depends on whether you see AI progress as research or product.

show 1 reply
twhyn2yesterday at 10:18 PM

"It's now turned into a whole marketing circus (maybe to justify these ludicrous billion-dollar valuations?)."

Yes theres an entire ecosystem being built up around language models that has to stay afloat for another 5 years at least, to hope for a significant breakthrough.