version: "3.3" services: text-generation-webui: build: context: . args: # specify which cuda version your card supports: https://developer.nvidia.com/cuda-gpus TORCH_CUDA_ARCH_LIST: ${TORCH_CUDA_ARCH_LIST:-7.5} WEBUI_VERSION: ${WEBUI_VERSION:-HEAD} env_file: .env ports: - "${HOST_PORT:-7860}:${CONTAINER_PORT:-7860}" - "${HOST_API_PORT:-5000}:${CONTAINER_API_PORT:-5000}" - "${HOST_API_STREAM_PORT:-5005}:${CONTAINER_API_STREAM_PORT:-5005}" stdin_open: true tty: true volumes: - ./characters:/app/characters - ./extensions:/app/extensions - ./loras:/app/loras - ./models:/app/models - ./presets:/app/presets - ./prompts:/app/prompts - ./softprompts:/app/softprompts - ./training:/app/training deploy: resources: reservations: devices: - driver: nvidia device_ids: ['0'] capabilities: [gpu]