Private • Local • Offline AI

Build Your
Private AI
Ollama + Open WebUI • Zero Cloud Dependency

A complete, step-by-step guide to running powerful large language models entirely on your own machine. No subscriptions, no data leaving your computer, no compromises on privacy. Your conversations stay yours.

💻
Your Machine
CPU / GPU / RAM
🦙
Ollama
Model Runtime
🌐
Open WebUI
Chat Interface
🔒
100% Private
Nothing Leaves

// 01 — Requirements

Hardware You'll Need

Local LLMs need memory. More RAM and VRAM means bigger, smarter models. Here's what works at each tier.

// 02 — Setup Guide

Install in 10 Minutes

From zero to your own private ChatGPT in four steps. Works on Windows, macOS, and Linux.

1

Install Ollama

Ollama is a lightweight runtime that downloads and runs LLMs locally. It handles model management, quantisation, and GPU acceleration automatically.

Terminal — Install Ollama
# macOS & Linux (one-line install) $ curl -fsSL https://ollama.com/install.sh | sh # Windows — download installer from: # https://ollama.com/download # Verify installation $ ollama --version ollama version 0.16.2 # The Ollama service starts automatically. # It runs on http://localhost:11434
2

Pull Your First Model

Choose a model from the Ollama Library. Start small to test your hardware, then scale up. Models download once and are stored locally.

Terminal — Download Models
# Pull a lightweight model to test (3.8B params) $ ollama pull llama3.2 pulling manifest... done pulling dde5aa3fc5ff... 100% 2.0 GB success # Try it out immediately $ ollama run llama3.2 >>> Hello! How can I help you today? # Pull a coding-focused model $ ollama pull qwen2.5-coder:7b # Pull a reasoning model $ ollama pull deepseek-r1:8b # List all downloaded models $ ollama list
3

Install Docker

Open WebUI runs in a Docker container. If you don't have Docker yet, install Docker Desktop — it takes 2 minutes and gives you a GUI to manage containers.

Terminal — Install Docker
# macOS & Windows: # Download Docker Desktop from https://docker.com # Linux (Ubuntu/Debian) $ sudo apt update && sudo apt install docker.io -y $ sudo systemctl enable --now docker $ sudo usermod -aG docker $USER # Verify Docker is running $ docker --version Docker version 27.4.0
4

Launch Open WebUI

One Docker command gives you a polished, ChatGPT-style interface that connects to Ollama. Your data is stored in a persistent volume — nothing is lost between restarts.

Terminal — Run Open WebUI
# Run Open WebUI (connects to local Ollama) $ docker run -d -p 3000:8080 \ --add-host=host.docker.internal:host-gateway \ -v open-webui:/app/backend/data \ --name open-webui --restart always \ ghcr.io/open-webui/open-webui:main # OR — all-in-one (Ollama + WebUI bundled): $ docker run -d -p 3000:8080 \ -v ollama:/root/.ollama \ -v open-webui:/app/backend/data \ --name open-webui --restart always \ ghcr.io/open-webui/open-webui:ollama # Open your browser: http://localhost:3000 # Create your admin account on first visit. # Select a model and start chatting!

// 03 — Choose Your Model

Recommended Models

Pick the right model for your hardware and use case. Smaller models are faster; larger ones are smarter.

// 04 — What You Can Do

Private AI Use Cases

Once running, your local AI becomes a Swiss Army knife for productivity — with zero data leaving your machine.

// 05 — Best Practices

Pro Tips

Get the most out of your local LLM setup with these expert recommendations.

// 06 — Troubleshooting

Common Issues

Quick fixes for the most frequent setup problems.