services: litellm: image: ghcr.io/berriai/litellm:main-latest container_name: litellm-proxy ports: - "4000:4000" volumes: - ./litellm_config.yaml:/app/config.yaml env_file: - .env command: [ "--config", "/app/config.yaml", "--detailed_debug" ] restart: unless-stopped