Does anyone know what kind of HW is required to run it locally? There are instructions but nothing about HW required.
They released a bunch of different sized models and there are already quantized versions showing up on HF.
https://huggingface.co/unsloth/DeepSeek-R1-Distill-Llama-8B-... for example has versions that are 3GB, 4GB, 5GB, 8GB and 16GB.
That 3GB one might work on a CPU machine with 4GB of RAM.
To get good performance you'll want a GPU with that much free VRAM, or an Apple Silicon machine with that much RAM.
Deepseek v3 required about 1tb of VRAM / RAM so 10 A100.
There are various ways to run it with lower vram if you're ok with way worse latency & throughput
Edit: sorry this is for v3, the distilled models can be ran on consumer-grade GPUs
You can try something like this to get a rough estimate: https://huggingface.co/spaces/NyxKrage/LLM-Model-VRAM-Calcul...
But you really don't know the exact numbers until you try, a lot of it is runtime/environment context specific.
It's just a question of having enough VRAM+RAM to fit the model into memory.