can you elaborate? you can use quantized version, would context still be an issue with it?
A usable quant, Q5_KM imo, takes up ~26GB[0], which leaves around ~6-7GB for context and running other programs which is not much.
[0] https://huggingface.co/unsloth/Qwen3.5-35B-A3B-GGUF?show_fil...
context is always an issue with local models and consumer hardware.
A usable quant, Q5_KM imo, takes up ~26GB[0], which leaves around ~6-7GB for context and running other programs which is not much.
[0] https://huggingface.co/unsloth/Qwen3.5-35B-A3B-GGUF?show_fil...