Files
Timmy-time-dashboard/docker-compose.yml
2026-03-05 12:09:51 -05:00

205 lines
7.2 KiB
YAML

# ── Development Compose ───────────────────────────────────────────────────────
#
# Services
# dashboard FastAPI app (always on)
# taskosaur Taskosaur PM + AI task execution
# postgres PostgreSQL 16 (for Taskosaur)
# redis Redis 7 (for Taskosaur queues)
#
# Usage
# make docker-build build the image
# make docker-up start dashboard + taskosaur
# make docker-down stop everything
# make docker-logs tail logs
#
# ── Security note: root user in dev ─────────────────────────────────────────
# This dev compose runs containers as root (user: "0:0") so that
# bind-mounted host files (./src, ./static) are readable regardless of
# host UID/GID — the #1 cause of 403 errors on macOS.
#
# ── Ollama host access ──────────────────────────────────────────────────────
# By default OLLAMA_URL points to http://host.docker.internal:11434 which
# reaches Ollama running on the Docker host (macOS/Windows native).
#
# Linux: The extra_hosts entry maps host.docker.internal → host-gateway,
# which resolves to the host IP on Docker 20.10+.
services:
# ── Dashboard (FastAPI) ──────────────────────────────────────────────────
dashboard:
build: .
image: timmy-time:latest
container_name: timmy-dashboard
user: "0:0" # dev only — see security note above
ports:
- "8000:8000"
volumes:
- timmy-data:/app/data
- ./src:/app/src # live-reload: source changes reflect immediately
- ./static:/app/static # live-reload: CSS/asset changes reflect immediately
environment:
DEBUG: "true"
OLLAMA_URL: "${OLLAMA_URL:-http://host.docker.internal:11434}"
# Grok (xAI) — opt-in premium cloud backend
GROK_ENABLED: "${GROK_ENABLED:-false}"
XAI_API_KEY: "${XAI_API_KEY:-}"
GROK_DEFAULT_MODEL: "${GROK_DEFAULT_MODEL:-grok-3-fast}"
# Celery/Redis — background task queue
REDIS_URL: "redis://redis:6379/0"
# Taskosaur API — dashboard can reach it on the internal network
TASKOSAUR_API_URL: "http://taskosaur:3000/api"
extra_hosts:
- "host.docker.internal:host-gateway" # Linux: maps to host IP
depends_on:
taskosaur:
condition: service_healthy
networks:
- timmy-net
restart: unless-stopped
healthcheck:
test: ["CMD", "curl", "-f", "http://localhost:8000/health"]
interval: 30s
timeout: 5s
retries: 3
start_period: 30s
# ── Taskosaur — project management + conversational AI tasks ───────────
# https://github.com/Taskosaur/Taskosaur
taskosaur:
image: ghcr.io/taskosaur/taskosaur:latest
container_name: taskosaur
ports:
- "3000:3000" # Backend API + Swagger docs at /api/docs
- "3001:3001" # Frontend UI
environment:
DATABASE_URL: "postgresql://taskosaur:taskosaur@postgres:5432/taskosaur"
REDIS_HOST: "redis"
REDIS_PORT: "6379"
JWT_SECRET: "${TASKOSAUR_JWT_SECRET:-dev-jwt-secret-change-in-prod}"
JWT_REFRESH_SECRET: "${TASKOSAUR_JWT_REFRESH_SECRET:-dev-refresh-secret-change-in-prod}"
ENCRYPTION_KEY: "${TASKOSAUR_ENCRYPTION_KEY:-dev-encryption-key-change-in-prod}"
FRONTEND_URL: "http://localhost:3001"
NEXT_PUBLIC_API_BASE_URL: "http://localhost:3000/api"
NODE_ENV: "development"
depends_on:
postgres:
condition: service_healthy
redis:
condition: service_healthy
networks:
- timmy-net
restart: unless-stopped
healthcheck:
test: ["CMD", "curl", "-f", "http://localhost:3000/api/health"]
interval: 30s
timeout: 5s
retries: 5
start_period: 60s
# ── PostgreSQL — Taskosaur database ────────────────────────────────────
postgres:
image: postgres:16-alpine
container_name: taskosaur-postgres
environment:
POSTGRES_USER: taskosaur
POSTGRES_PASSWORD: taskosaur
POSTGRES_DB: taskosaur
volumes:
- postgres-data:/var/lib/postgresql/data
networks:
- timmy-net
restart: unless-stopped
healthcheck:
test: ["CMD-SHELL", "pg_isready -U taskosaur"]
interval: 10s
timeout: 5s
retries: 5
start_period: 10s
# ── Redis — Taskosaur queue backend ────────────────────────────────────
redis:
image: redis:7-alpine
container_name: taskosaur-redis
volumes:
- redis-data:/data
networks:
- timmy-net
restart: unless-stopped
healthcheck:
test: ["CMD", "redis-cli", "ping"]
interval: 10s
timeout: 5s
retries: 5
start_period: 5s
# ── Celery Worker — background task processing ──────────────────────────
celery-worker:
build: .
image: timmy-time:latest
container_name: timmy-celery-worker
user: "0:0"
command: ["celery", "-A", "infrastructure.celery.app", "worker", "--loglevel=info", "--concurrency=2"]
volumes:
- timmy-data:/app/data
- ./src:/app/src
environment:
REDIS_URL: "redis://redis:6379/0"
OLLAMA_URL: "${OLLAMA_URL:-http://host.docker.internal:11434}"
extra_hosts:
- "host.docker.internal:host-gateway"
depends_on:
redis:
condition: service_healthy
networks:
- timmy-net
restart: unless-stopped
profiles:
- celery
# ── OpenFang — vendored agent runtime sidecar ────────────────────────────
openfang:
build:
context: .
dockerfile: docker/Dockerfile.openfang
image: timmy-openfang:latest
container_name: timmy-openfang
profiles:
- openfang
environment:
OLLAMA_URL: "${OLLAMA_URL:-http://host.docker.internal:11434}"
OPENFANG_DATA_DIR: "/app/data"
extra_hosts:
- "host.docker.internal:host-gateway"
volumes:
- openfang-data:/app/data
networks:
- timmy-net
restart: unless-stopped
healthcheck:
test: ["CMD", "curl", "-f", "http://localhost:8080/health"]
interval: 30s
timeout: 5s
retries: 3
start_period: 15s
# ── Volumes ──────────────────────────────────────────────────────────────────
volumes:
timmy-data:
driver: local
driver_opts:
type: none
o: bind
device: "${PWD}/data"
openfang-data:
driver: local
postgres-data:
driver: local
redis-data:
driver: local
# ── Internal network ────────────────────────────────────────────────────────
networks:
timmy-net:
driver: bridge