Open WebUI
Open WebUI is a comprehensive, open-source web interface for AI models. It features a user-friendly design similar to ChatGPT and can connect to local or hosted LLM backends.
Dependencies
None required, but works best with a local LLM backend like vLLM deployed on your cluster.
Configuration
Key settings configured through your instance's config.yaml:
- domain - Where the UI will be accessible (default:
chat.{your-cloud-domain}) - vllmApiUrl - URL of your LLM backend (default: connects to vLLM on the cluster)
- enableSignup - Whether to allow new account creation (default:
false) - storage - Persistent volume size (default:
10Gi)
Access
After deployment, Open WebUI will be available at:
https://chat.{your-cloud-domain}
First-Time Setup
-
Deploy a local LLM backend (e.g., vLLM) if you haven't already
-
Add and deploy the app:
wild app add open-webui wild app deploy open-webui -
Create your account and start chatting with your local AI models