logoalt Hacker News

tcsenpai10/11/20241 replyview on HN

Personally I use llama3.1:8b or mistral-nemo:latest which have a decent contex window (even if it is less than the commercial ones usually). I am working on a token calculator / division of the content method too but is very early


Replies

garyfirestorm10/12/2024

why not llama3.2:3B? it has fairly large context window too

show 1 reply