DEV Community

Cover image for Running DeepSeek-R1 with Ollama using Only Docker: A Complete Guide
Md Imran
Md Imran

Posted on

Running DeepSeek-R1 with Ollama using Only Docker: A Complete Guide

Introduction

DeepSeek-R1 is a powerful open-source LLM (Large Language Model) that can be easily run using Ollama inside Docker. This guide will walk you through setting up DeepSeek-R1 on a normal laptop with just Docker. If you have an NVIDIA GPU, an optional section will cover GPU acceleration.

By the end of this guide, you will be able to:

  • Run Ollama in Docker with just a normal laptop.
  • Pull and run DeepSeek-R1 using only CPU (no need for GPU).
  • Enable GPU acceleration if your system has an NVIDIA GPU.
  • Run the entire setup with a single command for ease of execution.
  • Optionally use a Web UI for a better experience instead of CLI.

Prerequisites (CPU Execution - Recommended for Most Users)

This guide is structured to prioritize CPU usage, ensuring that any normal laptop with Docker installed can run DeepSeek-R1 efficiently.

  • Only Docker is required (Install Docker).
  • No special hardware is needed—your normal laptop will work!
  • 16GB+ RAM recommended (for smooth performance).

Step 1: Pull and Run Ollama in Docker (CPU Only)

Ollama provides a convenient runtime for models like DeepSeek-R1. We will first run Ollama inside a Docker container.

docker run -d -v ollama:/root/.ollama -p 11434:11434 --name ollama ollama/ollama
Enter fullscreen mode Exit fullscreen mode

This will:

  • Start Ollama in a Docker container.
  • Expose it on port 11434.
  • Persist downloaded models using a volume (ollama:/root/.ollama).

To verify the container is running:

docker ps
Enter fullscreen mode Exit fullscreen mode

Step 2: Pull and Run DeepSeek-R1 Model (CPU Only)

Now that Ollama is running, we can pull and execute DeepSeek-R1.

Pull DeepSeek-R1 Model

docker exec -it ollama ollama pull deepseek-r1:8b
Enter fullscreen mode Exit fullscreen mode

Run DeepSeek-R1 (CPU Mode)

docker exec -it ollama ollama run deepseek-r1:8b
Enter fullscreen mode Exit fullscreen mode

Step 3: Running Everything in One Command (CPU Only)

docker run -d -v ollama:/root/.ollama -p 11434:11434 --name ollama ollama/ollama && \
docker exec -it ollama ollama pull deepseek-r1:8b && \
docker exec -it ollama ollama run deepseek-r1:8b
Enter fullscreen mode Exit fullscreen mode

Optional: Running DeepSeek-R1 with Web UI

If you prefer a graphical interface instead of using the command line, you can set up a Web UI for Ollama and DeepSeek-R1.

Step 1: Run Open WebUI with Ollama

docker run -d -p 3000:8080 -e OLLAMA_API_BASE_URL=http://host.docker.internal:11434 -v open-webui:/app/backend/data --name open-webui --restart always ghcr.io/open-webui/open-webui:main
Enter fullscreen mode Exit fullscreen mode

This will:

  • Start Open WebUI, which provides a web-based chat interface for DeepSeek-R1.
  • Expose it on http://localhost:3000.
  • Connect it to the running Ollama container.

Now, open your browser and navigate to http://localhost:3000 to chat with DeepSeek-R1 using an easy-to-use UI.

Running DeepSeek-R1 with Ollama using Only Docker: A Complete Guide


Optional: Running DeepSeek-R1 with GPU Acceleration

If you have an NVIDIA GPU, you can enable GPU acceleration for improved performance.

Prerequisites (GPU Execution)

  • NVIDIA GPU (with CUDA support).
  • NVIDIA Drivers installed (Check GPU Compatibility).
  • Docker with NVIDIA Container Toolkit installed.

Step 1: Run Ollama in Docker (With GPU Support)

docker run -d --gpus=all -v ollama:/root/.ollama -p 11434:11434 --name ollama ollama/ollama
Enter fullscreen mode Exit fullscreen mode

Step 2: Run DeepSeek-R1 with GPU

docker exec -it ollama ollama run deepseek-r1:8b --gpu
Enter fullscreen mode Exit fullscreen mode

Step 3: Running Everything in One Command (GPU Enabled)

docker run -d --gpus=all -v ollama:/root/.ollama -p 11434:11434 --name ollama ollama/ollama && \
docker exec -it ollama ollama pull deepseek-r1:8b && \
docker exec -it ollama ollama run deepseek-r1:8b --gpu
Enter fullscreen mode Exit fullscreen mode

Step 4: Verify GPU Utilization

To ensure DeepSeek-R1 is using your GPU, check NVIDIA System Management Interface (nvidia-smi):

docker exec -it ollama nvidia-smi
Enter fullscreen mode Exit fullscreen mode

You should see processes running under the GPU Memory Usage section.


Step 5: Stop and Remove Ollama Docker Container

If you ever need to stop and remove the container, use:

docker stop ollama && docker rm ollama
Enter fullscreen mode Exit fullscreen mode

This will:

  • Stop the running Ollama container.
  • Remove it from the system (but the model files will persist in the Docker volume).

Conclusion

In this guide, we covered how to:
✅ Set up Ollama in Docker.
✅ Pull and run DeepSeek-R1 using just a normal laptop (CPU only).
✅ Enable GPU acceleration if needed.
✅ Use a Web UI for a better experience.
✅ Execute everything in a single command.
✅ Verify GPU utilization (if applicable).

By following these steps, you can easily deploy DeepSeek-R1 in a Dockerized environment with minimal setup. 🚀

Image of Docusign

🛠️ Bring your solution into Docusign. Reach over 1.6M customers.

Docusign is now extensible. Overcome challenges with disconnected products and inaccessible data by bringing your solutions into Docusign and publishing to 1.6M customers in the App Center.

Learn more

Top comments (0)

A Workflow Copilot. Tailored to You.

Pieces.app image

Our desktop app, with its intelligent copilot, streamlines coding by generating snippets, extracting code from screenshots, and accelerating problem-solving.

Read the docs