I have an GTX 1660 Super (6 GB)
Right now I have ollama with:
- deepseek-r1:8b
- qwen2.5-coder:7b
Do you recommend any other local models to play with my GPU?
I have an GTX 1660 Super (6 GB)
Right now I have ollama with:
Do you recommend any other local models to play with my GPU?
Mistral
I personally run models on my laptop. I have 48 GB of ram and a i5-12500U. It runs a little slow but usable
My gear is an old:
I7-4790 16GB RAM
How many tokens by second?
The biggest bottleneck is going to be memory. I would just stick with GPU only since your GPU memory has the most bandwidth.