logoalt Hacker News

hasperdiyesterday at 8:31 PM3 repliesview on HN

and can be faster if you can get an MOE model of that


Replies

dormentoyesterday at 8:46 PM

"Mixture-of-experts", AKA "running several small models and activating only a few at a time". Thanks for introducing me to that concept. Fascinating.

(commentary: things are really moving too fast for the layperson to keep up)

show 2 replies
miohtamayesterday at 10:12 PM

All modern models are MoE already, no?

show 1 reply
bigyabaiyesterday at 9:33 PM

>90% of inference hardware is faster if you run an MOE model.