logoalt Hacker News

ryeguy_24today at 1:52 AM3 repliesview on HN

How many proprietary use cases truly need pre-training or even fine-tuning as opposed to RAG approach? And at what point does it make sense to pre-train/fine tune? Curious.


Replies

mirekrusintoday at 5:14 AM

You can fine tune small, very fast and cheap to run specialized models ie. to react to logs, tool use and domain knowledge, possibly removing network llm comms altogether etc.

Shitty-kittytoday at 4:17 AM

rag basically gives the llm a bunch of documents to search thru for the answer. What it doesn't do is make the algorithm any better. pre-training and fine-tunning improve the llm abaility to reason about your task.

babytoday at 1:54 AM

RAG is dead

show 4 replies