back to home

Light-Heart-Labs / DreamServer

Local AI anywhere, for everyone — LLM inference, chat UI, voice, agents, workflows, RAG, and image generation. No cloud, no subscriptions.

View on GitHub
415 stars
107 forks
30 issues
PythonShellPowerShell

AI Architecture Analysis

This repository is indexed by RepoMind. By analyzing Light-Heart-Labs/DreamServer in our AI interface, you can instantly generate complete architecture diagrams, visualize control flows, and perform automated security audits across the entire codebase.

Our Agentic Context Augmented Generation (Agentic CAG) engine loads full source files into context on-demand, avoiding the fragmentation of traditional RAG systems. Ask questions about the architecture, dependencies, or specific features to see it in action.

Source files are only loaded when you start an analysis to optimize performance.

Embed this Badge

Showcase RepoMind's analysis directly in your repository's README.

[![Analyzed by RepoMind](https://img.shields.io/badge/Analyzed%20by-RepoMind-4F46E5?style=for-the-badge)](https://repomind.in/repo/Light-Heart-Labs/DreamServer)
Preview:Analyzed by RepoMind

Repository Overview (README excerpt)

Crawler view

Dream Server Own your AI. One person, one dream, one machine at a time. A handful of companies control the vast majority of global AI traffic — and with it, your data, your costs, and your uptime. Every query you send to a centralized provider is business intelligence you don’t own, running on infrastructure you don’t control, priced on terms you can’t negotiate. If AI is becoming critical infrastructure, it shouldn’t be rented. Self-hosting local AI should be a sovereign human right, not a career choice. **Dream Server is the exit.** A fully local AI stack — LLM inference, chat, voice, agents, workflows, RAG, image generation, and privacy tools — deployed on your hardware with a single command. No cloud. No subscriptions. No one watching. **New here?** Read the Friendly Guide or listen to the audio version — a complete walkthrough of what Dream Server is, how it works, and how to make it your own. No technical background needed. --- > **Platform Support — March 2026** > > | Platform | Status | > |----------|--------| > | **Linux** (NVIDIA + AMD) | **Supported** — install and run today | > | **Windows** (NVIDIA + AMD) | **Supported** — install and run today | > | **macOS** (Apple Silicon) | **Supported** — install and run today | > > **Tested Linux distros:** Ubuntu 24.04/22.04, Debian 12, Fedora 41+, Arch Linux, CachyOS, openSUSE Tumbleweed. Other distros using apt, dnf, pacman, or zypper should also work — open an issue if yours doesn't. > > **Windows:** Requires Docker Desktop with WSL2 backend. NVIDIA GPUs use Docker GPU passthrough; AMD Strix Halo runs llama-server natively with Vulkan. > > **macOS:** Requires Apple Silicon (M1+) and Docker Desktop. llama-server runs natively with Metal GPU acceleration; all other services run in Docker. > > See the Support Matrix for details. --- Why Dream Server? Because running your own AI shouldn't require a CS degree and a weekend of debugging CUDA drivers. Right now, setting up local AI means stitching together a dozen projects, writing Docker configs from scratch, and praying everything talks to each other. Most people give up and go back to paying OpenAI. We built Dream Server so you don't have to. • **One command** — detects your GPU, picks the right model, generates credentials, launches everything • **Chatting in under 2 minutes** — bootstrap mode gives you a working model instantly while your full model downloads in the background • **13 services, pre-wired** — chat, agents, voice, workflows, search, RAG, image generation, privacy tools. All talking to each other out of the box • **Fully moddable** — every service is an extension. Drop in a folder, run , done Open **http://localhost:3000** and start chatting. > **No GPU?** Dream Server also runs in cloud mode — same full stack, powered by OpenAI/Anthropic/Together APIs instead of local inference: > > **Port conflicts?** Every port is configurable via environment variables. See for the full list, or override at install time: > *The DREAMGATE installer handles everything — GPU detection, model selection, service orchestration.* Manual install (Linux) Windows (PowerShell) Requires Docker Desktop with WSL2 backend enabled. **Install Docker Desktop first and make sure it is running before you start.** The installer detects your GPU, picks the right model, generates credentials, starts all services, and creates a Desktop shortcut to the Dashboard. Manage with . macOS (Apple Silicon) Requires Apple Silicon (M1+) and Docker Desktop. **Install Docker Desktop first and make sure it is running before you start.** The installer detects your chip, picks the right model for your unified memory, launches llama-server natively with Metal acceleration, and starts all other services in Docker. Manage with . See the macOS Quickstart for details. --- What's In The Box Chat & Inference • **Open WebUI** — full-featured chat interface with conversation history, web search, document upload, and 30+ languages • **llama-server** — high-performance LLM inference with continuous batching, auto-selected for your GPU • **LiteLLM** — API gateway supporting local/cloud/hybrid modes Voice • **Whisper** — speech-to-text • **Kokoro** — text-to-speech Agents & Automation • **OpenClaw** — autonomous AI agent framework • **n8n** — workflow automation with 400+ integrations (Slack, email, databases, APIs) Knowledge & Search • **Qdrant** — vector database for retrieval-augmented generation (RAG) • **SearXNG** — self-hosted web search (no tracking) • **Perplexica** — deep research engine Creative • **ComfyUI** — node-based image generation Privacy & Ops • **Privacy Shield** — PII scrubbing proxy for API calls • **Dashboard** — real-time GPU metrics, service health, model management --- Hardware Auto-Detection The installer detects your GPU and picks the optimal model automatically. No manual configuration. NVIDIA | VRAM | Model | Example GPUs | |------|-------|--------------| | *The installer pulls all services in parallel. Downloads are resume-capable — interrupted downloads pick up where they left off.* Skip bootstrap: --- Switching Models The installer picks a model for your hardware, but you can switch anytime: If the new model isn't downloaded yet, pre-fetch it first: Already have a GGUF you want to use? Drop it in , update and in , and restart: Rollback is automatic — if a new model fails to load, Dream Server reverts to your previous model. --- Extensibility Dream Server is designed to be modded. Every service is an extension — a folder with a and a . The dashboard, CLI, health checks, and compose stack all discover extensions automatically. The installer itself is modular — 6 libraries and 13 phases, each in its own file. Want to add a hardware tier, swap a default model, or skip a phase? Edit one file. Full extension guide | Installer architecture --- dream-cli The CLI manages your entire stack: --- How It Compares Other tools get you part of the way. Dream Serv…