DEV Community

Cover image for Running DeepSeek-R1 with Ollama using Only Docker: A Complete Guide
Md Imran
Md Imran

Posted on

1 1 1 1 1

Running DeepSeek-R1 with Ollama using Only Docker: A Complete Guide

Introduction

DeepSeek-R1 is a powerful open-source LLM (Large Language Model) that can be easily run using Ollama inside Docker. This guide will walk you through setting up DeepSeek-R1 on a normal laptop with just Docker. If you have an NVIDIA GPU, an optional section will cover GPU acceleration.

By the end of this guide, you will be able to:

  • Run Ollama in Docker with just a normal laptop.
  • Pull and run DeepSeek-R1 using only CPU (no need for GPU).
  • Enable GPU acceleration if your system has an NVIDIA GPU.
  • Run the entire setup with a single command for ease of execution.
  • Optionally use a Web UI for a better experience instead of CLI.

Prerequisites (CPU Execution - Recommended for Most Users)

This guide is structured to prioritize CPU usage, ensuring that any normal laptop with Docker installed can run DeepSeek-R1 efficiently.

  • Only Docker is required (Install Docker).
  • No special hardware is needed—your normal laptop will work!
  • 16GB+ RAM recommended (for smooth performance).

Step 1: Pull and Run Ollama in Docker (CPU Only)

Ollama provides a convenient runtime for models like DeepSeek-R1. We will first run Ollama inside a Docker container.

docker run -d -v ollama:/root/.ollama -p 11434:11434 --name ollama ollama/ollama
Enter fullscreen mode Exit fullscreen mode

This will:

  • Start Ollama in a Docker container.
  • Expose it on port 11434.
  • Persist downloaded models using a volume (ollama:/root/.ollama).

To verify the container is running:

docker ps
Enter fullscreen mode Exit fullscreen mode

Step 2: Pull and Run DeepSeek-R1 Model (CPU Only)

Now that Ollama is running, we can pull and execute DeepSeek-R1.

Pull DeepSeek-R1 Model

docker exec -it ollama ollama pull deepseek-r1:8b
Enter fullscreen mode Exit fullscreen mode

Run DeepSeek-R1 (CPU Mode)

docker exec -it ollama ollama run deepseek-r1:8b
Enter fullscreen mode Exit fullscreen mode

Step 3: Running Everything in One Command (CPU Only)

docker run -d -v ollama:/root/.ollama -p 11434:11434 --name ollama ollama/ollama && \
docker exec -it ollama ollama pull deepseek-r1:8b && \
docker exec -it ollama ollama run deepseek-r1:8b
Enter fullscreen mode Exit fullscreen mode

Optional: Running DeepSeek-R1 with Web UI

If you prefer a graphical interface instead of using the command line, you can set up a Web UI for Ollama and DeepSeek-R1.

Step 1: Run Open WebUI with Ollama

docker run -d -p 3000:8080 -e OLLAMA_API_BASE_URL=http://host.docker.internal:11434 -v open-webui:/app/backend/data --name open-webui --restart always ghcr.io/open-webui/open-webui:main
Enter fullscreen mode Exit fullscreen mode

This will:

  • Start Open WebUI, which provides a web-based chat interface for DeepSeek-R1.
  • Expose it on http://localhost:3000.
  • Connect it to the running Ollama container.

Now, open your browser and navigate to http://localhost:3000 to chat with DeepSeek-R1 using an easy-to-use UI.

Running DeepSeek-R1 with Ollama using Only Docker: A Complete Guide


Optional: Running DeepSeek-R1 with GPU Acceleration

If you have an NVIDIA GPU, you can enable GPU acceleration for improved performance.

Prerequisites (GPU Execution)

  • NVIDIA GPU (with CUDA support).
  • NVIDIA Drivers installed (Check GPU Compatibility).
  • Docker with NVIDIA Container Toolkit installed.

Step 1: Run Ollama in Docker (With GPU Support)

docker run -d --gpus=all -v ollama:/root/.ollama -p 11434:11434 --name ollama ollama/ollama
Enter fullscreen mode Exit fullscreen mode

Step 2: Run DeepSeek-R1 with GPU

docker exec -it ollama ollama run deepseek-r1:8b --gpu
Enter fullscreen mode Exit fullscreen mode

Step 3: Running Everything in One Command (GPU Enabled)

docker run -d --gpus=all -v ollama:/root/.ollama -p 11434:11434 --name ollama ollama/ollama && \
docker exec -it ollama ollama pull deepseek-r1:8b && \
docker exec -it ollama ollama run deepseek-r1:8b --gpu
Enter fullscreen mode Exit fullscreen mode

Step 4: Verify GPU Utilization

To ensure DeepSeek-R1 is using your GPU, check NVIDIA System Management Interface (nvidia-smi):

docker exec -it ollama nvidia-smi
Enter fullscreen mode Exit fullscreen mode

You should see processes running under the GPU Memory Usage section.


Step 5: Stop and Remove Ollama Docker Container

If you ever need to stop and remove the container, use:

docker stop ollama && docker rm ollama
Enter fullscreen mode Exit fullscreen mode

This will:

  • Stop the running Ollama container.
  • Remove it from the system (but the model files will persist in the Docker volume).

Conclusion

In this guide, we covered how to:
✅ Set up Ollama in Docker.
✅ Pull and run DeepSeek-R1 using just a normal laptop (CPU only).
✅ Enable GPU acceleration if needed.
✅ Use a Web UI for a better experience.
✅ Execute everything in a single command.
✅ Verify GPU utilization (if applicable).

By following these steps, you can easily deploy DeepSeek-R1 in a Dockerized environment with minimal setup. 🚀

Quadratic AI

Quadratic AI – The Spreadsheet with AI, Code, and Connections

  • AI-Powered Insights: Ask questions in plain English and get instant visualizations
  • Multi-Language Support: Seamlessly switch between Python, SQL, and JavaScript in one workspace
  • Zero Setup Required: Connect to databases or drag-and-drop files straight from your browser
  • Live Collaboration: Work together in real-time, no matter where your team is located
  • Beyond Formulas: Tackle complex analysis that traditional spreadsheets can't handle

Get started for free.

Watch The Demo 📊✨

Top comments (0)

Jetbrains Survey

Calling all developers!

Participate in the Developer Ecosystem Survey 2025 and get the chance to win a MacBook Pro, an iPhone 16, or other exciting prizes. Contribute to our research on the development landscape.

Take the survey

AWS Security LIVE!

Hosted by security experts, AWS Security LIVE! showcases AWS Partners tackling real-world security challenges. Join live and get your security questions answered.

Tune in to the full event

DEV is partnering to bring live events to the community. Join us or dismiss this billboard if you're not interested. ❤️