Personally I use llama3.1:8b or mistral-nemo:latest which have a decent contex window (even if it is less than the commercial ones usually). I am working on a token calculator / division of the content method too but is very early
why not llama3.2:3B? it has fairly large context window too
why not llama3.2:3B? it has fairly large context window too