I have an GTX 1660 Super (6 GB)

Right now I have ollama with:

  • deepseek-r1:8b
  • qwen2.5-coder:7b

Do you recommend any other local models to play with my GPU?

  • The Hobbyist
    link
    fedilink
    English
    arrow-up
    3
    arrow-down
    1
    ·
    8 days ago

    Deepseek is good at reasoning, qwen is good at programming, but I find llama3.1 8b to be well suited for creativity, writing, translations and other tasks which fall out of the scope of your two models. It’s a decent all arounder. It’s about 4.9GB in q4_K_M.

    • Disonantezko@lemmy.sdf.orgOP
      link
      fedilink
      English
      arrow-up
      1
      ·
      8 days ago

      It’s not out of my scope, I’m just learning what can I do locally with my current machine.


      Today I read about RAG, maybe I’m gonna try an easy local setup to chat with a PDF.