A self-contained AI box inside your device
Delta is a self-contained AI box inside your device. It includes AI models, search, document analysis, and other tools — all running offline, securely, and privately.
Everything you need for powerful AI, all running locally on your device
Use different LLM models from 0.5B to 8B+ parameters. Switch models as needed.
Analyze documents, PDFs, and text files locally. Your data stays on your device.
Terminal interface and web UI. Choose how you want to interact with Delta.
Your data stays on your device. All processing happens locally. No cloud required.
GPU support for CUDA, Metal, Vulkan, and ROCm. Faster inference on supported hardware.
One-command model downloads, automatic updates, and simple configuration.
One command to install your Personal AI Power Box
macOS 10.15+
Install via HomebrewWindows 10, 11
Install via WingetDebian, Ubuntu, Fedora, RHEL, Arch
Install via Install ScriptHomebrew (Recommended):
brew tap nile-agi/delta-cli && brew install --HEAD nile-agi/delta-cli/delta-cliWhat it does: Automatically clones repository, installs dependencies (including Node.js for web UI), builds from source (~40 seconds), builds custom web UI, and configures PATH. Users don't need to know about git!
Alternative - Installation Script:
curl -fsSL https://raw.githubusercontent.com/nile-agi/delta/main/install.sh | bashdeltaThis will auto-download the default model (qwen2.5:0.5b, ~400MB) if not installed, then start interactive mode.
delta pull llama3.1:8b # 4.7 GB - powerful, versatile\ndelta pull mistral:7b # 4.3 GB - great for codingdelta serverThen open http://localhost:8080 in your browser to use the web interface.
Delta CLI is released under the MIT License, which means it's free to use, modify, and distribute for any purpose, including commercial use.
Delta CLI is built on top of llama.cpp, a popular open-source project for running LLMs efficiently. This provides proven performance, active development, community support, model compatibility (all GGUF format models), and GPU acceleration (CUDA, Metal, Vulkan, ROCm).
View on GitHub