possible to run on RTX 2060 8GB + 32GB RAM DDR4?

#16
by krigeta - opened

it would be amazing and if it is possible then how much context size I can keep?

This model runned on my Ryzen 5 5600 with Vega 7 GPU and 32 GB RAM DDR4, in RTX 2060 + 32GB RAM it would run easily

quantized yes mainly with kobold or ollama or llm studio or kobold(endpoint)+open gui
kv 4 https://huggingface.co/unsloth/Qwen3.5-9B-GGUF

Your config is good for LLMs. I think i will buy a RTX 2060 to fast my training and inference on LLMs

Sign up or log in to comment