logoalt Hacker News

barbazootoday at 6:12 PM2 repliesview on HN

> What that means is that when you're looking to build a fully local RAG setup, you'll need to substitute whatever SaaS providers you're using for a local option for each of those components.

Even starting with having "just" the documents and vector db locally is a huge first step and much more doable than going with a local LLM at the same time. I don't know any one or any org that has the resources to run their own LLM at scale.


Replies

mips_avatartoday at 7:33 PM

It’s also just extremely viable to just host your own vector db. You just need a server with enough ram for your hnsw index.

procaryotetoday at 7:13 PM

Aren't there a bunch of models that run OK on consumer hardware now?

show 1 reply