version: '3.9'
services:
ollama:
image: ollama/ollama:latest
restart: always
deploy:
replicas: 1
resources:
reservations:
devices:
- driver: nvidia
capabilities: ["compute", "gpu", "utility", "graphics"]
count: all # Adjust count for the number of GPUs you want to use
ports:
- 11434:11434
volumes:
- ollama:/root/.ollama
open-webui:
image: ghcr.io/open-webui/open-webui:main
- 3000:8080
- open-webui:/app/backend/data
environment:
- 'OLLAMA_BASE_URL=http://ollama:11434'
ollama: {}
open-webui: {}