Files
wild-directory/open-webui

Open WebUI

Open WebUI is a comprehensive, open-source web interface for AI models. It features a user-friendly design similar to ChatGPT and can connect to local or hosted LLM backends.

Dependencies

None required, but works best with a local LLM backend like vLLM deployed on your cluster.

Configuration

Key settings configured through your instance's config.yaml:

  • domain - Where the UI will be accessible (default: chat.{your-cloud-domain})
  • vllmApiUrl - URL of your LLM backend (default: connects to vLLM on the cluster)
  • enableSignup - Whether to allow new account creation (default: false)
  • storage - Persistent volume size (default: 10Gi)

Access

After deployment, Open WebUI will be available at:

  • https://chat.{your-cloud-domain}

First-Time Setup

  1. Deploy a local LLM backend (e.g., vLLM) if you haven't already

  2. Add and deploy the app:

    wild app add open-webui
    wild app deploy open-webui
    
  3. Create your account and start chatting with your local AI models