Linux Native Installation¶
Native Linux setup with full NVIDIA GPU acceleration for optimal performance.
🐧 Recommended for Linux This setup provides optimal performance with full NVIDIA GPU acceleration support.
Prerequisites¶
- Linux distribution (Ubuntu 20.04+, Fedora 35+, Arch, Debian, etc.)
- 8GB+ RAM (16GB+ recommended for GPU acceleration)
- 10GB free disk space
- Python 3.11 or higher
For GPU Acceleration (Optional)¶
- NVIDIA GPU (GTX 1060+ or RTX series recommended)
- NVIDIA drivers 470+ installed
- CUDA 11.7+ installed
Installation Methods¶
Script-Based Installation (Recommended)¶
- Run the setup script:
- Start all services:
- Install agent-cli:
- Test the setup:
NixOS Users¶
If you're using NixOS, see the dedicated NixOS Installation Guide for system-level service integration.
Option 3: Manual Installation¶
If you prefer manual setup:
# 1. Install dependencies
curl -LsSf https://astral.sh/uv/install.sh | sh
curl -fsSL https://ollama.ai/install.sh | sh
# 2. Start services individually
# Terminal 1: Ollama
ollama serve
# Terminal 2: Whisper (with GPU)
scripts/run-whisper.sh
# Terminal 3: Piper
scripts/run-piper.sh
# Terminal 4: OpenWakeWord
scripts/run-openwakeword.sh
Services Overview¶
| Service | Port | GPU Support | Auto-Detection |
|---|---|---|---|
| Ollama | 11434 | ✅ CUDA/ROCm | Automatic |
| Whisper | 10300 | ✅ CUDA | Automatic (falls back to CPU) |
| Piper | 10200 | N/A | N/A |
| OpenWakeWord | 10400 | N/A | N/A |
Session Management with Zellij¶
The scripts use Zellij for managing all services in one session (works on both Linux and macOS):
Starting Services¶
Zellij Commands¶
Ctrl-O d- Detach (services keep running)zellij attach agent-cli- Reattach to sessionzellij list-sessions- List all sessionszellij kill-session agent-cli- Stop all servicesAlt + arrow keys- Navigate between panesCtrl-Q- Quit (stops all services)
Automatic GPU Detection¶
The scripts automatically detect and use GPU acceleration:
- Whisper: Detects NVIDIA GPU and uses
large-v3model with CUDA, falls back totinyon CPU - Ollama: Automatically uses available GPU (CUDA/ROCm)
GPU Acceleration Setup¶
NVIDIA GPU (CUDA)¶
- Install NVIDIA drivers:
- Install CUDA toolkit:
- Verify GPU setup:
AMD GPU (ROCm)¶
- Install ROCm:
# Ubuntu/Debian
sudo apt install rocm-dev
# Configure for Ollama
export HSA_OVERRIDE_GFX_VERSION=10.3.0 # Adjust for your GPU
- Start Ollama with ROCm:
Why Native Setup?¶
- Full GPU acceleration - NVIDIA CUDA support
- Automatic configuration - Scripts detect and configure GPU
- Better performance - Direct system integration
Troubleshooting¶
GPU Not Working¶
Services Not Starting¶
General Issues¶
- Make sure you have enough RAM (8GB minimum)
- Services automatically download required models
Alternative: Docker¶
If you prefer Docker (with some performance limitations):
- Docker Setup Guide
- Note: May have reduced GPU acceleration support