The Hobbyist@lemmy.ziptoTechnology@lemmy.world•Edward Snowden slams Nvidia's RTX 50-series 'F-tier value,' whistleblows on lackluster VRAM capacityEnglish
6·
5 hours agoYou can. I’m running a 14B deepseek model on mine. It achieves 28 t/s.
Just a stranger trying things.
You can. I’m running a 14B deepseek model on mine. It achieves 28 t/s.
Ollama, latest version. I have it setup with Open-WebUI (though that shouldn’t matter). The 14B is around 9GB, which easily fits in the 12GB.
I’m repeating the 28 t/s from memory, but even if I’m wrong it’s easily above 20.
Specifically, I’m running this model: https://ollama.com/library/deepseek-r1:14b-qwen-distill-q4_K_M