Why do you assume this?
I can produce total jibberish even faster, doesn’t mean I produce Einstein level thought if I slow down
Better models already exist, this is just proving you can dramatically increase inference speeds / reduce inference costs.
It isn't about model capability - it's about inference hardware. Same smarts, faster.
Not what he said.
Better models already exist, this is just proving you can dramatically increase inference speeds / reduce inference costs.
It isn't about model capability - it's about inference hardware. Same smarts, faster.