🇦🇺𝕄𝕦𝕟𝕥𝕖𝕕𝕔𝕣𝕠𝕔𝕕𝕚𝕝𝕖@lemm.ee to LocalLLaMA@sh.itjust.worksEnglish · 1 day agoHow much gpu do i need to run a 90b modelmessage-squaremessage-square13fedilinkarrow-up112arrow-down11file-text
arrow-up111arrow-down1message-squareHow much gpu do i need to run a 90b model🇦🇺𝕄𝕦𝕟𝕥𝕖𝕕𝕔𝕣𝕠𝕔𝕕𝕚𝕝𝕖@lemm.ee to LocalLLaMA@sh.itjust.worksEnglish · 1 day agomessage-square13fedilinkfile-text
minus-squareredlinkfedilinkEnglisharrow-up1·6 hours agothis is useless, llama.cpp already does that airllm does (offloading to CPU) but its actually faster. so just use ollama
this is useless, llama.cpp already does that airllm does (offloading to CPU) but its actually faster. so just use ollama