Local AI Hub
  • Compare Tools
  • Tutorials
  • Cloud Deploy
  • Blog
Run LLM on DigitalOcean — GPU Droplet Setup Guide
2026/04/17
Intermediate30 min

Run LLM on DigitalOcean — GPU Droplet Setup Guide

Step-by-step guide to running large language models on DigitalOcean GPU Droplets. Set up Ollama, deploy your first model, and keep cloud costs under control.

DigitalOcean GPU Droplets give you access to powerful GPUs with the same simple interface you're used to. This guide walks you through setting up Ollama and running large language models on DigitalOcean.

Why DigitalOcean?

  • Familiar interface if you already use DO
  • Reliable infrastructure with good uptime
  • Predictable billing through your existing DO account
  • Good documentation and support

Available GPUs

GPUVRAMPrice/hrBest For
A100 40GB40 GB~$1.10Models up to 30B
H100 80GB80 GB~$2.93Models up to 70B

DigitalOcean doesn't offer consumer GPUs like the RTX 4090. If you need cheaper GPU options, consider Runpod.

Step 1: Create a GPU Droplet

  1. Log in to your DigitalOcean account
  2. Click Create → Droplets
  3. Under Choose Image, select Ubuntu 22.04
  4. Under Choose Plan, select GPU Droplets
  5. Choose your GPU:
    • A100 40GB for models up to 30B parameters
    • H100 80GB for models up to 70B parameters
  6. Select a data center region
  7. Add your SSH key
  8. Click Create Droplet

Wait 2-3 minutes for the Droplet to provision.

Step 2: Install NVIDIA Drivers

Connect via SSH and install the GPU drivers:

# Update system
sudo apt update && sudo apt upgrade -y

# Install NVIDIA drivers
sudo apt install -y nvidia-driver-535
sudo reboot

After reboot, verify the GPU:

nvidia-smi

You should see your A100 or H100 listed with available VRAM.

Step 3: Install Ollama

curl -fsSL https://ollama.com/install.sh | sh

# Verify
ollama --version

Step 4: Download and Run Models

# For A100 40GB — run models up to 30B
ollama pull llama3.1:8b
ollama pull qwen2.5:14b
ollama pull deepseek-r1:8b

# For H100 80GB — can run 70B models
ollama pull llama3.1:70b
ollama pull qwen2.5:32b

# Start chatting
ollama run llama3.1:8b

Step 5: Expose the API (Optional)

To access Ollama from external applications:

# Set Ollama to listen on all interfaces
sudo systemctl edit ollama

# Add:
[Service]
Environment="OLLAMA_HOST=0.0.0.0:11434"

# Restart
sudo systemctl restart ollama

Set Up Firewall

# Allow HTTP/HTTPS
sudo ufw allow 22
sudo ufw allow 80
sudo ufw allow 443
sudo ufw allow 11434

sudo ufw enable

Set Up Nginx Reverse Proxy (Recommended)

For production use, put Ollama behind Nginx with HTTPS:

sudo apt install -y nginx certbot python3-certbot-nginx

# Create Nginx config
sudo tee /etc/nginx/sites-available/ollama << 'EOF'
server {
    listen 80;
    server_name your-domain.com;

    location / {
        proxy_pass http://localhost:11434;
        proxy_set_header Host $host;
        proxy_set_header X-Real-IP $remote_addr;
        proxy_set_header X-Forwarded-For $proxy_add_x_forwarded_for;
        proxy_set_header X-Forwarded-Proto $scheme;

        # Required for streaming
        proxy_buffering off;
        proxy_read_timeout 300s;
    }
}
EOF

sudo ln -s /etc/nginx/sites-available/ollama /etc/nginx/sites-enabled/
sudo nginx -t
sudo systemctl reload nginx

# Add HTTPS
sudo certbot --nginx -d your-domain.com

Step 6: Add Open WebUI (Optional)

docker run -d -p 3000:8080 \
  -e OLLAMA_BASE_URL=http://localhost:11434 \
  -v open-webui:/app/backend/data \
  --name open-webui \
  --restart always \
  ghcr.io/open-webui/open-webui:main

Access at http://your-droplet-ip:3000.

Cost Management

Tips to Keep Costs Low

  1. Power off when not in use — you only pay while the Droplet is running
  2. Use snapshots — create a snapshot before powering off to restore quickly
  3. Start with A100 — it's cheaper than H100 and sufficient for most models
  4. Monitor usage — set up billing alerts in the DO dashboard

Cost Estimates

GPUUsageMonthly Cost
A100 40GB2 hrs/day~$66
A100 40GB4 hrs/day~$132
H100 80GB2 hrs/day~$176
H100 80GB4 hrs/day~$352

DigitalOcean is more expensive than Runpod per hour. The trade-off is reliability, support, and ecosystem integration.

DigitalOcean vs Runpod

AspectDigitalOceanRunpod
Cheapest GPU$1.10/hr (A100)$0.20/hr (RTX 4000)
GPU SelectionLimited (A100, H100)Wide (consumer to enterprise)
One-Click TemplatesNoYes
SupportTicket-basedCommunity
ReliabilityHighGood
Best ForDO users, enterpriseMost users, best value

If cost is your primary concern, Runpod is significantly cheaper. If you value DO's reliability and ecosystem, GPU Droplets are a solid choice.

Summary

DigitalOcean GPU Droplets work well for running LLMs, especially if you're already in the DO ecosystem. Setup takes about 30 minutes including drivers, Ollama, and optional Nginx configuration. Costs are higher than Runpod, but you get DO's reliability and support.

Next Steps

  • Best GPU Cloud for LLM — compare all providers
  • Runpod Beginner Guide — cheaper alternative
  • Cheapest Way to Run LLM — cost optimization
Get started with DigitalOcean GPU Droplets — reliable cloud AI.
Get started with DigitalOcean for cloud GPU computing. No hardware upgrades needed — run any AI model on powerful remote GPUs.
Get Started with DigitalOcean

Partner link. We may earn a commission at no extra cost to you.

All Posts

Author

avatar for Local AI Hub
Local AI Hub

Categories

  • Cloud Deploy
  • Tutorials
Why DigitalOcean?Available GPUsStep 1: Create a GPU DropletStep 2: Install NVIDIA DriversStep 3: Install OllamaStep 4: Download and Run ModelsStep 5: Expose the API (Optional)Set Up FirewallSet Up Nginx Reverse Proxy (Recommended)Step 6: Add Open WebUI (Optional)Cost ManagementTips to Keep Costs LowCost EstimatesDigitalOcean vs RunpodSummaryNext Steps

More Posts

Open WebUI vs AnythingLLM — Which Local AI Interface Is Right for You?
Comparisons

Open WebUI vs AnythingLLM — Which Local AI Interface Is Right for You?

Comparison

Open WebUI and AnythingLLM both add chat interfaces to local AI, but serve very different needs. Compare features, RAG capabilities, and ease of use.

avatar for Local AI Hub
Local AI Hub
2026/04/12
How to Run Qwen Locally — Alibaba's Powerful Multilingual Model
Models & HardwareTutorials

How to Run Qwen Locally — Alibaba's Powerful Multilingual Model

Tutorial

Run Qwen 2.5 models on your own computer — one of the best open models for coding, multilingual tasks, and general use. Works on devices with 8GB RAM or more.

avatar for Local AI Hub
Local AI Hub
2026/04/13
Run Open WebUI on Runpod — Cloud ChatGPT in 10 Minutes
Cloud DeployTutorials

Run Open WebUI on Runpod — Cloud ChatGPT in 10 Minutes

Tutorial

Deploy Open WebUI with Ollama on Runpod for a private, ChatGPT-like experience on cloud GPU. Access your AI assistant from any device with a web browser.

avatar for Local AI Hub
Local AI Hub
2026/04/16
Local AI Hub

Run AI locally — fast, cheap, and private

Resources
  • Compare Tools
  • Tutorials
  • Cloud Deploy
  • Device Check
  • Blog
Company
  • About
  • Contact
Legal
  • Cookie Policy
  • Privacy Policy
  • Terms of Service
© 2026 Local AI Hub. All Rights Reserved.