What's the largest deepseek-r1 model that a 4080 + 64GB DDR5 can handle?
I'm looking to run an AI on my personal machine. I wanted to ask; what's the largest parameter model my PC specs can handle (4080 + 64GB DDR5)? I imagine I can handle the 32B model with response times below 40 seconds (ideally), I haven't been able to find any kind of chart/graph depicting how model performance and response time scale with hardware.