Welcome to local-ai.run

A self-hosted AI workspace that runs entirely on your own hardware. Chat with your files, swap model engines, transcribe audio — all without sending a single byte to a cloud provider.

New to local-ai.run? Head straight to Getting Started and run the one-command installer. Most people are up and running in under five minutes.

Start here

Pick a guide depending on what you want to do.

What local-ai.run is

A complete local AI workspace that runs in Docker on your own hardware. Six services on one Docker network, zero external dependencies after install:

Architecture at a glance

Seven containers — three application services, three infrastructure services, one updater — all behind a single Caddy reverse proxy:

┌──────────┐      ┌────────────┐      ┌─────────────┐
│ Next.js  │ ───▶ │   Django   │ ───▶ │  PostgreSQL │
│   UI     │      │  REST API  │      └─────────────┘
└──────────┘      └────────────┘
                        │
                        ├──▶ Ollama   (LLM inference)
                        ├──▶ RAG      (FastAPI + vector store)
                        └──▶ Whisper  (speech-to-text)

       All traffic routed through Caddy on port 80.

Why it exists

Cloud AI services route your conversations and uploaded files through third-party infrastructure. For privacy-sensitive work — legal documents, medical records, internal company data, compliance-bound industries — that's a non-starter.

local-ai.run gives you the same experience as a hosted chat tool, but every byte stays on your machine:

Need help?

Join the Discord community for live install help and feature discussions, or open an issue on GitHub.