There already are a bunch of task-specific models running on their devices, it makes sense to maintain and build capacity in that area.
I assume they have a moderate bet on on-device SLMs in addition to other ML models, but not much planned for LLMs, which at that scale, might be good as generalists but very poor at guaranteeing success for each specific minute tasks you want done.
In short: 8gb to store tens of very small and fast purpose-specific models is much better than a single 8gb LLM trying to do everything.
Probably possible for pure coding models. I see on-device models becoming viable and usable in like 2-3 years on device