logoalt Hacker News

aziis98yesterday at 9:31 PM1 replyview on HN

I hope we get to good A1B models as I'm currently GPU poor and can only do inference on CPU for now


Replies

yowlingcatyesterday at 9:59 PM

It may be worth taking a look at LFM [1]. I haven't had the need to use it so far (running on Apple silicon on a day to day basis so my dailies are usually the 30B+ MoEs) but I've heard good things from the internet from folks using it as a daily on their phones. YMMV.

[1] https://huggingface.co/LiquidAI/LFM2.5-1.2B-Instruct