How are you using that RAM with the GPU?
Llama.cpp with automatic offload to main memory. You can also use Ollama, it is easier, but slower.
Llama.cpp with automatic offload to main memory. You can also use Ollama, it is easier, but slower.