Unfortunately Apple appears to be blocking the use of these llms within apps on their app store. I've been trying to ship an app that contains local llms and have hit a brick wall with issue 2.5.2
ESET is blocking this site saying:
Threat found This web page may contain dangerous content that can provide remote access to an infected device, leak sensitive data from the device or harm the targeted device. Threat: JS/Agent.RDW trojan
It runs on Android too, with AI Core or even with llama.cpp
There are many apps to run local LLMs on both iOS & Android
For those who would like an example of its output, I'm currently working through creating a small, free (cc0, public domain) encyclopedia (just a couple of thousand entries) of core concepts in Biology and Health Sciences, Physical Sciences, and Technology. Each entry is being entirely written by Gemma 4:e4b (the 10 GB model.) I believe that this may be slightly larger than the size of the model that runs locally on phones, so perhaps this model is slightly better, but the output is similar. Here is an example entry:
Seems pretty good to me!
> edge AI deployment
Isn't the "edge" meant to be computing near the user, but not on their devices?
Is the output coherent though? I am yet to see a local model working on consumer grade hardware being actually useful.
Can we please ban content that is CLEARLY written by AI?
is there a comparison of it running on iPhone vs. Android phones?
Related: Gemma 4 on iPhone (254 comments) - https://news.ycombinator.com/item?id=47652561