logoalt Hacker News

roughlyyesterday at 5:03 PM3 repliesview on HN

Looks like Confer is hosting its own inference: https://confer.to/blog/2026/01/private-inference/

> LLMs are fundamentally stateless—input in, output out—which makes them ideal for this environment. For Confer, we run inference inside a confidential VM. Your prompts are encrypted from your device directly into the TEE using Noise Pipes, processed there, and responses are encrypted back. The host never sees plaintext.

I don’t know what model they’re using, but it looks like everything should be staying on their servers, not going back to, eg, OpenAI or Anthropic.


Replies

JohnFenyesterday at 8:00 PM

> Looks like Confer is hosting its own inference

Even so, you're still exposing your data to Confer, and so you have to trust them that they'll behave as you want. That's a security problem that Confer doesn't help with.

I'm not saying Confer isn't useful, though. e2ee is very useful. But it isn't enough to make me feel comfortable.

show 1 reply
jeroadhdyesterday at 7:33 PM

That is a highly misleading statement: the GPU runs with real weights and real unencrypted user plaintext, since it has to multiply matrices of plain text, which is passed on to the supposedly "secure VM" (protected by Intel/Nvidia promises) and encrypted there. In no way is it e2e, unless you count the GPU as the "end".

show 2 replies
dangyesterday at 7:27 PM

We'll add that link to the toptext as well. Thanks!

(It got submitted a few times but did not get any comments - might as well consolidate these threads)