Skip to main content

Command Palette

Search for a command to run...

How to Run Ollama DeepSeek R1 on Your System

Updated
3 min read
How to Run Ollama DeepSeek R1 on Your System

Introduction

DeepSeek R1 is an advanced AI model that offers efficient and optimized performance for various AI-driven applications. Running it on your system using Ollama makes it seamless and straightforward. This guide will help you install and run Ollama DeepSeek R1 on your local machine.

Prerequisites

Before you start, make sure your system meets the following requirements:

  • Operating System: Windows, macOS, or Linux

  • RAM: Minimum 8GB (16GB+ recommended for larger models)

  • GPU: Optional but recommended for faster performance (NVIDIA CUDA-compatible for Linux/Windows)

  • Python: Python 3.8+ (if using additional AI utilities)

  • Docker (Optional): Required if you want to containerize the setup

Step 1: Install Ollama

Ollama is a streamlined way to run AI models efficiently. Install it by following the steps for your OS.

Windows

Download the installer from Ollama’s official website and follow the installation instructions.

macOS

Run the following command in the terminal:

curl -fsSL https://ollama.ai/install.sh | sh

Linux

For Debian-based distributions:

echo "deb [trusted=yes] https://ollama.ai/debian stable main" | sudo tee /etc/apt/sources.list.d/ollama.list
sudo apt update && sudo apt install ollama

For Arch-based distributions:

paru -S ollama

Step 2: Download and Run DeepSeek R1 Model

Once Ollama is installed, you can fetch the DeepSeek R1 model with the following command:

ollama pull deepseek-r1

This will download the model to your system. The first time you run it, it may take a while to set up.

Run DeepSeek R1

Now, execute the following command to start an interactive session:

ollama run deepseek-r1

Alternatively, you can pass a prompt directly:

ollama run deepseek-r1 "What is quantum computing?"

Step 3: Optimize Performance

To improve performance, consider the following:

Enable GPU Acceleration (Linux/Windows)

If you have an NVIDIA GPU, install CUDA and cuDNN:

sudo apt install nvidia-cuda-toolkit

Then, set Ollama to use GPU:

export OLLAMA_USE_GPU=1

Run DeepSeek R1 in a Container (Optional)

If you prefer running the model inside a Docker container:

docker run --rm -it ollama/ollama deepseek-r1

Step 4: Integrating DeepSeek R1 with Python

You can interact with DeepSeek R1 in Python using the requests library:

import requests

def query_ollama(prompt):
    response = requests.post("http://localhost:11434/api/generate", json={"model": "deepseek-r1", "prompt": prompt})
    return response.json()

print(query_ollama("Explain the theory of relativity."))

Ensure Ollama is running before executing this script.

Troubleshooting

1. Ollama command not found?

  • Restart your terminal or manually add Ollama to your system path.

2. Model not downloading?

  • Check your internet connection and try running ollama pull deepseek-r1 again.

3. Performance is slow?

  • Ensure you are using GPU acceleration (Linux/Windows) and have enough RAM allocated.

Conclusion

Running DeepSeek R1 with Ollama provides an efficient and user-friendly way to experiment with powerful AI models locally. Whether you're developing AI applications, testing model responses, or just curious about deep learning, this setup allows you to explore AI capabilities without relying on cloud-based services.