0%
Navigate Content
January 1, 2026
Docker Compose file to run open WebUI and AI Agents
How to run open WebUI and download gguf files to run AI Agents
Docker Compose file to run open WebUI and AI Agents
Today I learned about Docker Compose file to run open WebUI and AI Agents just by downloading gguf files from platforms like hugging face.
What are gguf files?
GGUF files are a special type of file format used to store and run large language models efficiently. They help make these models faster to load and easier to use on regular computers.
Actual docker compose file
services:
llama-server:
image: ghcr.io/ggml-org/llama.cpp:server
container_name: llama-server
restart: unless-stopped
volumes:
- /home/user/models:/models
ports:
- "8080:8080"
command:
- "--models-dir"
- "/models"
- "--host"
- "0.0.0.0"
- "--port"
- "8080"
- "--ctx-size"
- "2048"
- "--parallel"
- "4"
open-webui:
image: ghcr.io/open-webui/open-webui:latest
container_name: open-webui
restart: unless-stopped
ports:
- "3001:8080"
environment:
- OPENAI_API_BASE_URL=http://llama-server:8080/v1
- OPENAI_API_KEY=sk-no-key-required
volumes:
- open-webui-data:/app/data
depends_on:
- llama-server
volumes:
open-webui-data:
Basic Explanation
This docker-compose file runs two services:
- llama-server: This service runs a llama.cpp server that provides access to the models stored in the
/modelsdirectory. - open-webui: This service runs the Open WebUI application, which provides a user interface for interacting with the models.
How to use
- Download the gguf files from hugging face.
- Place the gguf files in the
/modelsdirectory. - Run the docker-compose file.
- Open the Open WebUI application in your browser.
- Select the model you want to use.
- Start chatting!
You May Also Like
Category:
Devops
is a platform for developing, shipping, and running applications in containers, providing consistent environments across systems.