Files

15 KiB

phase, plan, type, wave, depends_on, files_modified, autonomous, requirements, must_haves
phase plan type wave depends_on files_modified autonomous requirements must_haves
01-foundation 01 execute 1
pyproject.toml
docker-compose.yml
.env.example
packages/shared/models/message.py
packages/shared/models/tenant.py
packages/shared/models/auth.py
packages/shared/db.py
packages/shared/rls.py
packages/shared/config.py
packages/shared/redis_keys.py
packages/shared/__init__.py
packages/shared/models/__init__.py
migrations/env.py
migrations/versions/001_initial_schema.py
tests/conftest.py
tests/unit/test_normalize.py
tests/unit/test_tenant_resolution.py
tests/unit/test_redis_namespacing.py
tests/integration/test_tenant_isolation.py
true
CHAN-01
TNNT-01
TNNT-02
TNNT-03
TNNT-04
truths artifacts key_links
KonstructMessage Pydantic model validates and normalizes a Slack event payload into the unified internal format
Tenant A cannot query Tenant B's rows from the agents or channel_connections tables — enforced at the PostgreSQL layer via RLS
A channel workspace ID resolves to the correct Konstruct tenant ID via the channel_connections table
All Redis keys include {tenant_id}: prefix — no bare keys are possible through the shared utility
PostgreSQL and Redis are reachable via Docker Compose with TLS-ready configuration
path provides exports
packages/shared/models/message.py KonstructMessage, ChannelType, SenderInfo, MessageContent Pydantic models
KonstructMessage
ChannelType
SenderInfo
MessageContent
path provides exports
packages/shared/models/tenant.py Tenant, Agent, ChannelConnection SQLAlchemy models with RLS
Tenant
Agent
ChannelConnection
path provides exports
packages/shared/db.py Async SQLAlchemy engine, session factory, get_session dependency
engine
async_session_factory
get_session
path provides exports
packages/shared/rls.py current_tenant_id ContextVar and SQLAlchemy event hook for SET LOCAL
current_tenant_id
configure_rls_hook
path provides exports
packages/shared/redis_keys.py Namespaced Redis key constructors
rate_limit_key
idempotency_key
session_key
path provides contains
migrations/versions/001_initial_schema.py Initial DB schema with RLS policies and FORCE ROW LEVEL SECURITY FORCE ROW LEVEL SECURITY
path provides contains
tests/integration/test_tenant_isolation.py Two-tenant RLS isolation test tenant_a.*tenant_b
from to via pattern
packages/shared/rls.py packages/shared/db.py SQLAlchemy before_cursor_execute event hook on engine event.listens_for.*before_cursor_execute
from to via pattern
migrations/versions/001_initial_schema.py packages/shared/models/tenant.py Schema must match SQLAlchemy model definitions CREATE TABLE (tenants|agents|channel_connections)
from to via pattern
packages/shared/redis_keys.py Redis All key functions prepend tenant_id f"{tenant_id}:
Scaffold the Python monorepo, Docker Compose dev environment, shared Pydantic/SQLAlchemy models, PostgreSQL schema with RLS tenant isolation, Redis namespacing utilities, and the foundational test infrastructure.

Purpose: Establish the secure multi-tenant data layer that every subsequent plan builds on. Tenant isolation is the most dangerous failure mode in Phase 1 — it must be proven correct before any channel or LLM code exists.

Output: Working monorepo with uv workspaces, Docker Compose running PostgreSQL 16 + Redis 7 + Ollama, shared data models (KonstructMessage, Tenant, Agent, ChannelConnection), Alembic migrations with RLS, Redis key namespacing, and green isolation tests.

<execution_context> @/home/adelorenzo/.claude/get-shit-done/workflows/execute-plan.md @/home/adelorenzo/.claude/get-shit-done/templates/summary.md </execution_context>

@.planning/PROJECT.md @.planning/ROADMAP.md @.planning/STATE.md @.planning/phases/01-foundation/01-CONTEXT.md @.planning/phases/01-foundation/01-RESEARCH.md @CLAUDE.md Task 1: Monorepo scaffolding, Docker Compose, and shared data models pyproject.toml, docker-compose.yml, .env.example, packages/shared/__init__.py, packages/shared/models/__init__.py, packages/shared/models/message.py, packages/shared/models/tenant.py, packages/shared/models/auth.py, packages/shared/config.py, packages/shared/db.py, packages/shared/rls.py, packages/shared/redis_keys.py 1. Initialize the Python monorepo with `uv init` and configure `pyproject.toml` with workspace members for packages/shared, packages/gateway, packages/router, packages/orchestrator, packages/llm-pool. Add core dependencies: fastapi[standard], pydantic[email], sqlalchemy[asyncio], asyncpg, alembic, litellm, redis, celery[redis], slack-bolt, httpx, slowapi. Add dev dependencies: ruff, mypy, pytest, pytest-asyncio, pytest-httpx. Configure `[tool.pytest.ini_options]` with `asyncio_mode = "auto"` and `testpaths = ["tests"]`. Configure `[tool.ruff]` with line-length=120 and basic rules.
2. Create `docker-compose.yml` with services:
   - `postgres`: PostgreSQL 16 with `POSTGRES_DB=konstruct`, creates `konstruct_app` role via init script, port 5432
   - `redis`: Redis 7, port 6379
   - `ollama`: Ollama with GPU optional (deploy.resources.reservations.devices with count:all, but service starts regardless), port 11434
   - Shared `konstruct-net` bridge network

3. Create `.env.example` with all required environment variables: DATABASE_URL (using konstruct_app role, not postgres superuser), REDIS_URL, SLACK_BOT_TOKEN, SLACK_SIGNING_SECRET, ANTHROPIC_API_KEY, OPENAI_API_KEY, OLLAMA_BASE_URL, AUTH_SECRET.

4. Create `packages/shared/config.py` using Pydantic Settings to load all env vars with sensible defaults for local dev.

5. Create `packages/shared/models/message.py` with the KonstructMessage Pydantic model exactly as specified in RESEARCH.md: ChannelType (StrEnum: slack, whatsapp, mattermost), SenderInfo, MessageContent, KonstructMessage. The `tenant_id` field is `str | None = None` (populated by Router after tenant resolution).

6. Create `packages/shared/models/tenant.py` with SQLAlchemy 2.0 async models:
   - `Tenant`: id (UUID PK), name (str, unique), slug (str, unique), settings (JSON), created_at, updated_at
   - `Agent`: id (UUID PK), tenant_id (FK to Tenant, NOT NULL), name (str), role (str), persona (text), system_prompt (text), model_preference (str, default "quality"), tool_assignments (JSON, default []), escalation_rules (JSON, default []), is_active (bool, default True), created_at, updated_at
   - `ChannelConnection`: id (UUID PK), tenant_id (FK to Tenant, NOT NULL), channel_type (ChannelType enum), workspace_id (str, unique for the channel_type), config (JSON — stores bot tokens, channel IDs per tenant), created_at
   Use SQLAlchemy 2.0 `Mapped[]` and `mapped_column()` style — never 1.x `Column()` style.

7. Create `packages/shared/models/auth.py` with a `PortalUser` SQLAlchemy model: id (UUID PK), email (str, unique), hashed_password (str), name (str), is_admin (bool, default False), created_at, updated_at. This is for portal authentication (Auth.js v5 will validate against this via a FastAPI endpoint).

8. Create `packages/shared/db.py` with async SQLAlchemy engine (asyncpg driver) and session factory. Use `create_async_engine` with `DATABASE_URL` from config. Export `get_session` as an async generator for FastAPI dependency injection.

9. Create `packages/shared/rls.py` with `current_tenant_id` ContextVar and a `configure_rls_hook(engine)` function that registers a `before_cursor_execute` event listener to `SET LOCAL app.current_tenant = '{tenant_id}'` when `current_tenant_id` is set. CRITICAL: Use parameterized query for the SET LOCAL to prevent SQL injection — use `cursor.execute("SET LOCAL app.current_tenant = %s", (tenant_id,))`.

10. Create `packages/shared/redis_keys.py` with typed key constructor functions: `rate_limit_key(tenant_id, channel)`, `idempotency_key(tenant_id, message_id)`, `session_key(tenant_id, thread_id)`, `engaged_thread_key(tenant_id, thread_id)`. Every function prepends `{tenant_id}:`. No Redis key should ever be constructable without a tenant_id.

11. Create minimal `__init__.py` files for packages/shared and packages/shared/models with appropriate re-exports.
cd /home/adelorenzo/repos/konstruct && uv sync && python -c "from packages.shared.models.message import KonstructMessage; from packages.shared.models.tenant import Tenant, Agent, ChannelConnection; from packages.shared.redis_keys import rate_limit_key; print('imports OK')" - pyproject.toml configures uv workspaces with all dependencies - docker-compose.yml defines PostgreSQL 16, Redis 7, Ollama services - KonstructMessage, Tenant, Agent, ChannelConnection, PortalUser models importable - RLS hook configurable on engine - All Redis key functions require tenant_id parameter - .env.example documents all required env vars with konstruct_app role (not postgres superuser) Task 2: Alembic migrations with RLS and tenant isolation tests migrations/env.py, migrations/script.py.mako, migrations/versions/001_initial_schema.py, tests/conftest.py, tests/unit/__init__.py, tests/unit/test_normalize.py, tests/unit/test_tenant_resolution.py, tests/unit/test_redis_namespacing.py, tests/integration/__init__.py, tests/integration/test_tenant_isolation.py 1. Initialize Alembic with `alembic init migrations`. Modify `migrations/env.py` to use async engine (asyncpg) — follow the SQLAlchemy 2.0 async Alembic pattern with `run_async_migrations()`. Import the SQLAlchemy Base metadata from `packages/shared/models/tenant.py` so autogenerate works.
2. Create `migrations/versions/001_initial_schema.py`:
   - Create `konstruct_app` role: `CREATE ROLE konstruct_app WITH LOGIN PASSWORD 'konstruct_dev'` (dev password, .env overrides in prod)
   - Create tables: tenants, agents, channel_connections, portal_users — matching the SQLAlchemy models from Task 1
   - Apply RLS to tenant-scoped tables (agents, channel_connections):
     ```sql
     ALTER TABLE agents ENABLE ROW LEVEL SECURITY;
     ALTER TABLE agents FORCE ROW LEVEL SECURITY;
     CREATE POLICY tenant_isolation ON agents USING (tenant_id = current_setting('app.current_tenant')::uuid);
     ```
     Same pattern for channel_connections.
   - Do NOT apply RLS to `tenants` table itself (platform admin needs to list all tenants) or `portal_users` table.
   - GRANT SELECT, INSERT, UPDATE, DELETE on all tables to `konstruct_app`.
   - GRANT USAGE ON SCHEMA public TO `konstruct_app`.

3. Create `tests/conftest.py` with shared fixtures:
   - `db_engine`: Creates a test PostgreSQL database, runs migrations, yields async engine connected as `konstruct_app` (not postgres superuser), drops test DB after
   - `db_session`: Async session from the engine with RLS hook configured
   - `tenant_a` / `tenant_b`: Two-tenant fixture — creates two tenants, yields their IDs
   - `redis_client`: Creates a real Redis connection (or fakeredis if Docker not available) scoped to test prefix
   - Use `pytest.mark.asyncio` for async tests (auto mode from pyproject.toml config)
   - IMPORTANT: The test DB connection MUST use the `konstruct_app` role to actually test RLS. If using postgres superuser, RLS is bypassed and tests are worthless.

4. Create `tests/unit/test_normalize.py` (CHAN-01):
   - Test that a raw Slack `message` event payload normalizes to a valid KonstructMessage
   - Test ChannelType is set to "slack"
   - Test sender info is extracted correctly
   - Test thread_id is populated from Slack's `thread_ts`
   - Test channel_metadata contains workspace_id

5. Create `tests/unit/test_tenant_resolution.py` (TNNT-02):
   - Test that given a workspace_id and channel_type, the correct tenant_id is returned from a mock channel_connections lookup
   - Test that unknown workspace_id returns None
   - Test that workspace_id from wrong channel_type doesn't match

6. Create `tests/unit/test_redis_namespacing.py` (TNNT-03):
   - Test that `rate_limit_key("tenant-a", "slack")` returns `"tenant-a:ratelimit:slack"`
   - Test that `idempotency_key("tenant-a", "msg-123")` returns `"tenant-a:dedup:msg-123"`
   - Test that all key functions include tenant_id prefix
   - Test that no function can produce a key without tenant_id

7. Create `tests/integration/test_tenant_isolation.py` (TNNT-01) — THIS IS THE MOST CRITICAL TEST IN PHASE 1:
   - Uses the two-tenant fixture (tenant_a, tenant_b)
   - Creates an agent for tenant_a
   - Sets current_tenant_id to tenant_b
   - Queries agents table — MUST return zero rows (tenant_b cannot see tenant_a's agent)
   - Sets current_tenant_id to tenant_a
   - Queries agents table — MUST return one row
   - Repeat for channel_connections table
   - Verify with `SELECT relforcerowsecurity FROM pg_class WHERE relname = 'agents'` — must be True

8. Create empty `__init__.py` for tests/unit/ and tests/integration/.
cd /home/adelorenzo/repos/konstruct && docker compose up -d postgres redis && sleep 3 && alembic upgrade head && pytest tests/unit -x -q && pytest tests/integration/test_tenant_isolation.py -x -q - Alembic migration creates all tables with RLS policies and FORCE ROW LEVEL SECURITY - konstruct_app role exists and is used by all application connections - Unit tests pass for KonstructMessage normalization, tenant resolution logic, and Redis namespacing - Integration test proves tenant_a cannot see tenant_b's data through PostgreSQL RLS - `relforcerowsecurity` is True for agents and channel_connections tables - `docker compose up -d` starts PostgreSQL 16, Redis 7, and Ollama without errors - `alembic upgrade head` applies the initial schema with RLS - `pytest tests/unit -x -q` passes all unit tests (normalize, tenant resolution, redis namespacing) - `pytest tests/integration/test_tenant_isolation.py -x -q` proves RLS isolation - All imports from packages/shared work correctly - No Redis key can be constructed without a tenant_id

<success_criteria>

  • Green test suite proving tenant A cannot access tenant B's data
  • KonstructMessage model validates Slack event payloads
  • Docker Compose dev environment boots cleanly
  • All subsequent plans can import from packages/shared without modification </success_criteria>
After completion, create `.planning/phases/01-foundation/01-01-SUMMARY.md`