Linux Native Installation
Native Linux setup with full NVIDIA GPU acceleration for optimal performance.
Tip
🐧 Recommended for Linux — Optimal performance with full NVIDIA GPU acceleration.
Prerequisites
- Linux distribution (Ubuntu 20.04+, Fedora 35+, Arch, Debian, etc.)
- 8GB+ RAM (16GB+ recommended for GPU acceleration)
- 10GB free disk space
- Python 3.11 or higher
- PortAudio development libraries (required for audio features):
- Ubuntu/Debian:
sudo apt install portaudio19-dev - Fedora:
sudo dnf install portaudio-devel - Arch:
sudo pacman -S portaudio - openSUSE:
sudo zypper install portaudio-devel
For GPU Acceleration (Optional)
- NVIDIA GPU (GTX 1060+ or RTX series recommended)
- NVIDIA drivers 470+ installed
- CUDA 11.7+ installed
Installation Methods
Script-Based Installation (Recommended)
- Run the setup script:
- Start all services:
- Install agent-cli:
- Test the setup:
NixOS Users
If you're using NixOS, see the dedicated NixOS Installation Guide for system-level service integration.
Option 3: Manual Installation
If you prefer manual setup:
# 1. Install dependencies
curl -LsSf https://astral.sh/uv/install.sh | sh
curl -fsSL https://ollama.ai/install.sh | sh
# 2. Start services individually
# Terminal 1: Ollama
ollama serve
# Terminal 2: Whisper (with GPU)
agent-cli server whisper
# Terminal 3: Piper
agent-cli server tts --backend piper
# Terminal 4: OpenWakeWord
scripts/run-openwakeword.sh
Services Overview
| Service | Port | GPU Support | Auto-Detection |
|---|---|---|---|
| Ollama | 11434 | ✅ CUDA/ROCm | Automatic |
| Whisper | 10300 | ✅ CUDA | Automatic (falls back to CPU) |
| Piper | 10200 | N/A | N/A |
| OpenWakeWord | 10400 | N/A | N/A |
Session Management with Zellij
The scripts use Zellij for managing all services in one session (works on both Linux and macOS):
Starting Services
Zellij Commands
Ctrl-O d- Detach (services keep running)zellij attach agent-cli- Reattach to sessionzellij list-sessions- List all sessionszellij kill-session agent-cli- Stop all servicesAlt + arrow keys- Navigate between panesCtrl-Q- Quit (stops all services)
Automatic GPU Detection
The scripts automatically detect and use GPU acceleration:
- Whisper: Detects NVIDIA GPU and uses
large-v3model with CUDA, falls back totinyon CPU - Ollama: Automatically uses available GPU (CUDA/ROCm)
GPU Acceleration Setup
NVIDIA GPU (CUDA)
- Install NVIDIA drivers:
- Install CUDA toolkit:
- Verify GPU setup:
AMD GPU (ROCm)
- Install ROCm:
# Ubuntu/Debian
sudo apt install rocm-dev
# Configure for Ollama
export HSA_OVERRIDE_GFX_VERSION=10.3.0 # Adjust for your GPU
- Start Ollama with ROCm:
Why Native Setup?
- Full GPU acceleration - NVIDIA CUDA support
- Automatic configuration - Scripts detect and configure GPU
- Better performance - Direct system integration
Troubleshooting
GPU Not Working
Services Not Starting
General Issues
- Make sure you have enough RAM (8GB minimum)
- Services automatically download required models
Alternative: Docker
If you prefer Docker (with some performance limitations):
- Docker Setup Guide
- Note: May have reduced GPU acceleration support