How to Self-Host Open WebUI
Self-hosted AI chat interface for LLMs
Open WebUI is a feature-rich, self-hosted UI for interacting with large language models. Connect to Ollama, OpenAI, or other LLM providers — keep your conversations private on your own server.
Open WebUI features
ChatGPT-like interface
Connect to Ollama, OpenAI, or custom endpoints
Conversation history
Model switching
RAG (document upload)
User management
Plugin system
Deploy Open WebUI in 5 steps
1
Sign up for TinyPod
Create a free account.
2
Deploy Open WebUI
Find Open WebUI in the App Catalog and deploy.
3
Connect your LLM
Point to an Ollama instance or enter your OpenAI API key.
4
Start chatting
Use the familiar chat interface for AI conversations.
5
Upload documents
Use RAG to chat with your own documents and data.
Common use cases
Private AI assistant
Team AI chat with shared models
Document Q&A with RAG
Local LLM hosting with Ollama
Replacing ChatGPT for sensitive data