services: ollama: deploy: resources: limits: cpus: 4 volumes: - ./data:/root/.ollama ports: - 9434:11434 container_name: ollama image: ollama/ollama restart: unless-stopped environment: - OLLAMA_ORIGINS="*" - OLLAMA=SERVE - OLLAMA_KEEP_ALIVE=-1 - OLLAMA_NEW_ENGINE=true - OLLAMA_MAX_LOADED_MODELS=2 - OLLAMA_CONTEXT_LENGTH=8129 open-webui: ports: - 9009:8080 volumes: - ./data:/root/.ollama - ./open-webui:/app/backend/data environment: - GLOBAL_LOG_LEVEL=ERROR - PAPERLESS_URL=http://192.168.178.183:9800 - PAPERLESS_API_TOKEN=e4581556a81b1ad64188c702ba713bb4bb531d86 container_name: open-webui depends_on: - ollama restart: always image: ghcr.io/open-webui/open-webui:ollama networks: {}