Running DeepSeek R1 distills on Your Own Hardware – The fast and easy way

Published in LINUX-HOWTO.ORG • 01 February 2025

Prerequisites:

You’re root. On a spare PC. Or a VM with a dedicated GPU. Running a fresh Arch Linux install (though most Linux distros will work).
Because I never trust third-party software 100%, I would not recommend running this on your main system. Unless you like unnecessary risks. In that case, go ahead, genius.

 

Step 1: The Setup – Because We Love Clean Slates

This is a minimal VM setup named `deepseekvm` with a 100GB NVMe SSD:

[root@deepseekvm ~]# df -h
Filesystem      Size  Used Avail Use% Mounted on
/dev/sda2        91G   16G   75G  18% /

Step 2: Installing Ollama – The Magic Behind the Curtain

Why Ollama? Because it makes running large language models *actually* easy.
As seen on Ollama’s website, installation is a one-liner:

curl -fsSL https://ollama.com/install.sh | sh

One minute later, you should see something like:

>>> Installing ollama to /usr/local
>>> Downloading Linux amd64 bundle
>>> Adding ollama user to render group...
>>> Adding ollama user to video group...
>>> Creating ollama systemd service...
>>> Enabling and starting ollama service...
>>> NVIDIA GPU installed.

If it doesn’t work, fix your system. That’s not my problem.

 

Step 3: Starting Ollama – Because It Won’t Start Itself

ollama serve

If you get errors, maybe your GPU is fake. Or you’re using an Intel HD 3000. Good luck with that.

 

Step 4: Running DeepSeek R1 – The Fun Part

Change to another terminal and run:

ollama run deepseek-r1:14b

Or, if your GPU is a beast (20GB VRAM or more):

ollama run deepseek-r1:32b

 

If everything works, you’ll see something like:

 

pulling manifest
pulling 6e9f90f02bb3... 100% ▕█████████████████████▏ 9.0 GB                         
pulling 369ca498f347... 100% ▕█████████████████████▏  387 B                         
pulling 6e4c38e1172f... 100% ▕█████████████████████▏ 1.1 KB                         
pulling f4d24e9138dd... 100% ▕█████████████████████▏  148 B                         
pulling 3c24b0c80794... 100% ▕█████████████████████▏  488 B                         
verifying sha256 digest
writing manifest
success

 

If you don’t see this, double-check your GPU drivers. Or sacrifice a chicken to the Linux gods.

 

Step 5: Talking to Your AI – Because Why Not?

Let’s test it:

>>> hi there

An expected response from your local NVIDIA 3090 RTX (no internet required, as this llm runs directly from YOUR HARDWARE!):

<think>
</think>
Hello! How can I assist you today? 😊

 

You Now Have a Local AI – No Strings Attached!

No cloud nonsense. No subscription. No data leaks. Just raw AI power on your own hardware. What you do with it is up to you. Build something cool. Or don’t. Your call.