logoalt Hacker News

Local LLM App by Ente

149 pointsby matthiaswhtoday at 12:49 PM56 commentsview on HN

Comments

VladVladikofftoday at 1:29 PM

Maybe I’m missing it but the page is really light on technical information. Is this a quantized / distilled model of a larger LLM? Which one? How many parameters? What quantization? What T/s can I expect? What are the VRAM requirements? Etc etc

show 3 replies
jasongilltoday at 3:08 PM

I love Ente Auth, but Ente (as a company/organization) does a somewhat poor job of calling out their non-photos apps in their branding and on their website. If you go to the "Download" button at the top of the page on this page about their LLM chat app, it downloads... their photo sharing application. If you click Sign Up, it takes you to a signup page with the browser title "Ente Photos" but the page text says "Private backups for your memories" with a picture of a lock - is that the Ente Auth signup, or the Ente Photos app signup?

A little bit of cleanup on their site to break out "Ente, our original photo sharing app" from the rest of their apps would do wonders, because I had to search around on the announcement to find the download for this app, which feels about like trying to find the popular Ente Auth app on their website

jubilantitoday at 2:33 PM

There's dozens of local inference apps that basically wrap llama.cpp and someone else's GGUFs. The decentralized sync history part seems new? Not much else. But the advertisement copy is so insufferably annoying in how it presents this wrapper as a product.

Have a comparison chart to Ollama, LMStudio, LocalAI, Exo, Jan.AI, GPT4ALL, PocketPal, etc.

show 1 reply
moqstertoday at 1:29 PM

Heard the first time about them (ente) yesterday in a discussion about "which 2FA are u using?". Directly switched to https://ente.com/auth/ on Android and Linux Desktop and very happy with it.

Going to give this a try...

show 3 replies
QubridAItoday at 3:18 PM

This is the most important part of local AI maturing not just better models, but better productization of on-device inference for normal people.

koehrtoday at 2:10 PM

I just tried it. It downloaded Qwen3.5 2B on my phone and it's pretty coherent in its sentences, but really annoying with the amount of Ente products mentioned in every occasion. Other than that it's fast enough to talk to and definitely an easy way to run a model locally on your phone.

netfl0today at 1:59 PM

Weird hype going on here in comments.

cdrnsftoday at 2:42 PM

I like Ente, but isn't their core product a photos application? Its offshoots like this and 2FA feel incongruous.

lone-cloudtoday at 2:56 PM

Any half capable engineer can vibe code this in a week. Who cares?

franzetoday at 2:08 PM

if you are into local LLMs check out apfel

https://github.com/Arthur-Ficial/apfel

Apple Ai on the command line

show 2 replies
imadchtoday at 2:58 PM

What do you mean by IA in your device ? is it a local LLM ? if yeas how much params 4B or 8B...?? device requirements not mentionned too

show 1 reply
maxlohtoday at 2:20 PM

There is also another app called Off Grid, which lets you run any model from Hugging Face (of course you need to choose one your phone can handle).

https://github.com/alichherawalla/off-grid-mobile-ai

FitchAppstoday at 2:42 PM

Have you tried WebLLM? Or this wrapper: CodexLocal.com Basically, you would have a rather simple but capable LLM right in your browser using WebLLM and GPU

mkageniustoday at 2:16 PM

Had used cactus before - https://news.ycombinator.com/item?id=44524544

Then moved to pocket pal now for local llm.

emehextoday at 2:48 PM

There are literally 1000s of these types of apps. Why is this on the Front Page?

talking_penguintoday at 2:23 PM

How is this any different from Ollama plus Open Web UI?

show 1 reply
tim-projectstoday at 3:12 PM

This app isn't very useful but it did get me thinking.

I have a phone in a drawer I could install termux and ollama on over tailscale and then I'd have an always on llm for super light tasks.

I do really long for a private chat bot but I simply don't have access to the hardware required. Sadly I think it's going to be years to get there..

dgb23today at 2:18 PM

The (hn) title is misleading (unlike the actual title): It's an LLM _App_ not an LLM.

post-ittoday at 2:20 PM

> This is not the beginning, nor is this the end. This is just a checkpoint.

Come onnnnnn. I would rather read a one line "Check out our offline llm" rather than a whole press release of slop.

This looks very neat. I'm not familiar with the nitty gritty of AI so I really don't understand how it can reply so quickly running on an iPhone 16. But I'm not even going to bother searching for details because I don't want to read slop.

nathan_comptontoday at 2:01 PM

Please god stop letting LLMs write your copy. My brain just slides right over this slop. Perhaps you have a useful product but christ almighty I cannot countenance this boring machine generated text.

Pythiustoday at 2:24 PM

[dead]

Arn_Thortoday at 2:11 PM

[dead]

chockstoday at 1:56 PM

This looks amazing! As I learn and experiment more with local LLMs, I'm becoming more of a fan of local/offline LLMs. I believe there's a huge gap between local LLM based apps and commercial models like Claude/ChatGPT. Excited to see more apps leveraging local LLMs.

juliushuijnktoday at 1:46 PM

I'm working on a rather simple idea; a Wordpress plugin that allows you to use a local LLM inside your wordpress CMS.

It requires a Firefox add-on to act as a bridge: https://addons.mozilla.org/en-US/firefox/addon/ai-s-that-hel...

There is honestly not much to test just yet, but feel free to check it out here, provide feedback on the idea: https://codeberg.org/Helpalot/ais-that-helpalot

The essence works, I was able to let it make a simple summary on CMS content. So next is making it do something useful, and making it clear how other plugins could use it.

show 2 replies