# Plan: Phase J (WebSocket) + Turntable Bug + Phase K (Asset Library) ## Kontext Analyse des aktuellen Codestands ergab: **Phasen F, G, H, I, L sind bereits vollständig implementiert.** | Phase | Status | Beleg | |-------|--------|-------| | F - Hash-Caching | DONE | `domains/products/cache_service.py` + migration 041 | | G - Billing | DONE | `domains/billing/` vollständig, WeasyPrint in Dockerfile | | H - Excel Sanity-Check | DONE | `domains/imports/service.py run_sanity_check()` + Upload.tsx Dialog | | I - Notification-Config | DONE | `notification_configs` migration 044, NotificationSettings.tsx | | L - Dashboard | DONE | AdminDashboard.tsx + ClientDashboard.tsx vollständig | | **J - WebSocket** | **FEHLT** | Kein `core/websocket.py`, alle Polls noch aktiv | Zusätzlich: **Kritischer Bug in `render_blender.py`** — ffmpeg-Overlay-Befehl haengt bei endlicher Frame-Sequenz (kein `shortest=1`) -> Timeout -> Turntable-Render schlaegt fehl. --- ## Bug Fix: Turntable ffmpeg Timeout **Root cause**: In `backend/app/services/render_blender.py:507`: ```python "-filter_complex", "[1:v][0:v]overlay=0:0", ``` Der `lavfi color`-Quell-Stream hat keine definierte Laenge. Ohne `shortest=1` wartet ffmpeg auf weitere Frames vom Farb-Stream nachdem die PNG-Sequenz endet -> haengt bis Timeout (300s). **Fix**: `overlay=0:0` -> `overlay=0:0:shortest=1` --- ## Phase J: WebSocket Backend + Frontend ### Architektur (ADR-05: FastAPI nativ + Redis Pub/Sub) ``` Backend Task/Router: -> redis.publish(f"tenant:{tenant_id}", json.dumps(event)) core/websocket.py: ConnectionManager: tenant_id -> set[WebSocket] background_task: asyncio.Task (redis subscribe loop) Frontend: useWebSocket() hook -> WebSocket('/api/ws') Empfaengt Events, invalidiert React Query caches ``` ### Events die gesendet werden: | Event | Sender | Daten | |-------|--------|-------| | `render_complete` | step_tasks.py | order_line_id, status, thumbnail_url | | `render_failed` | step_tasks.py | order_line_id, error | | `cad_processing_complete` | step_tasks.py | cad_file_id, status | | `order_status_change` | orders router | order_id, new_status | | `queue_update` | beat task (alle 10s) | depth per queue | --- ## Betroffene Dateien ### Neu erstellen: - `backend/app/core/websocket.py` -- ConnectionManager + Redis Pub/Sub Loop - `frontend/src/hooks/useWebSocket.ts` -- WebSocket hook mit Auto-Reconnect - `frontend/src/contexts/WebSocketContext.tsx` -- Context Provider ### Aendern: - `backend/app/services/render_blender.py` -- ffmpeg shortest=1 Bug-Fix - `backend/app/main.py` -- WebSocket-Endpoint registrieren (`/api/ws`) - `backend/app/tasks/step_tasks.py` -- WebSocket-Events emittieren - `backend/app/domains/orders/router.py` -- Order-Status-Events emittieren - `backend/app/tasks/celery_app.py` -- `broadcast_queue_status` Beat-Task hinzufuegen - `frontend/src/App.tsx` -- WebSocketProvider wrappen - `frontend/src/pages/WorkerActivity.tsx` -- polling durch WS ersetzen - `frontend/src/pages/OrderDetail.tsx` -- polling durch WS ersetzen - `frontend/src/pages/Orders.tsx` -- polling reduzieren - `frontend/src/components/layout/Layout.tsx` -- polling reduzieren - `frontend/src/components/layout/NotificationCenter.tsx` -- polling durch WS ersetzen ### Nach Phase J Commit -- Phase K: - `backend/alembic/versions/045_asset_libraries.py` -- asset_libraries Tabelle - `backend/app/domains/materials/models.py` -- AssetLibrary Model hinzufuegen - `backend/app/domains/materials/router.py` -- Asset Library CRUD + Upload - `render-worker/scripts/asset_library.py` -- Materialien + Node-Groups aus .blend laden - `render-worker/scripts/catalog_assets.py` -- Katalog aus .blend lesen - `render-worker/scripts/export_gltf.py` -- GLB Export mit Materialien - `render-worker/scripts/export_blend.py` -- .blend Export mit pack_all() - `backend/app/domains/rendering/workflow_builder.py` -- Asset Library Nodes - `frontend/src/pages/Admin.tsx` -- Asset Library Manager UI - `frontend/src/api/assetLibraries.ts` -- API Client --- ## Tasks (in Reihenfolge) ### Task 1: Bug-Fix ffmpeg Turntable Timeout [x] - **Datei**: `backend/app/services/render_blender.py:507` - **Was**: `"[1:v][0:v]overlay=0:0"` -> `"[1:v][0:v]overlay=0:0:shortest=1"` - **Akzeptanzkriterium**: Turntable-Render fuer Order f0436188 kann erneut gestartet werden und produziert MP4 - **Abhaengigkeiten**: keine ### Task 2: WebSocket Backend -- core/websocket.py [x] - **Datei**: `backend/app/core/websocket.py` (neu) - **Was**: ```python class ConnectionManager: _connections: dict[str, set[WebSocket]] # tenant_id -> sockets async def connect(ws, tenant_id) def disconnect(ws, tenant_id) async def broadcast_to_tenant(tenant_id, event: dict) async def start_redis_subscriber() # asyncio background task def publish_event_sync(tenant_id: str, event: dict): # Sync version fuer Celery tasks -- redis.publish() ``` - Redis Pub/Sub: subscribe auf `tenant:*` Channels - Bei Nachricht: alle WebSockets des Tenants benachrichtigen - Auto-Ping alle 30s gegen Disconnects - **Akzeptanzkriterium**: broadcast_to_tenant sendet an alle verbundenen WS des Tenants - **Abhaengigkeiten**: keine ### Task 3: WebSocket Endpoint in main.py [x] - **Datei**: `backend/app/main.py` - **Was**: ```python @app.websocket("/api/ws") async def ws_endpoint(websocket: WebSocket, token: str = Query(...)): user = await verify_ws_token(token) await manager.connect(websocket, str(user.tenant_id)) try: while True: await websocket.receive_text() # Keep-alive pings except WebSocketDisconnect: manager.disconnect(websocket, str(user.tenant_id)) ``` - Token-Auth via Query-Parameter (WS kann keinen Authorization-Header senden) - `verify_ws_token`: JWT decode, User laden (analog zu get_current_user) - `manager` als globale Instanz, gestartet im lifespan - **Akzeptanzkriterium**: `ws://localhost:8888/api/ws?token=` oeffnet Verbindung - **Abhaengigkeiten**: Task 2 ### Task 4: WebSocket Events in step_tasks.py [x] - **Datei**: `backend/app/tasks/step_tasks.py` - **Was**: In render_order_line_task und render_step_thumbnail nach Erfolg/Fehler: ```python from app.core.websocket import publish_event_sync # bei render complete: publish_event_sync(tenant_id, {"type": "render_complete", "order_line_id": str(line.id), "status": "completed"}) # bei render failed: publish_event_sync(tenant_id, {"type": "render_failed", "order_line_id": str(line.id), "error": str(exc)}) # bei CAD processing complete: publish_event_sync(tenant_id, {"type": "cad_processing_complete", "cad_file_id": str(cad_file.id), "status": "completed"}) ``` - tenant_id aus cad_file.tenant_id bzw. order_line -> order -> user.tenant_id laden - **Akzeptanzkriterium**: Render fertig -> WebSocket-Client empfaengt Event - **Abhaengigkeiten**: Task 2 ### Task 5: WebSocket Events in orders router [x] - **Datei**: `backend/app/domains/orders/router.py` - **Was**: Bei Order-Status-Aenderung (submit, complete, cancel): ```python from app.core.websocket import manager await manager.broadcast_to_tenant( str(current_user.tenant_id), {"type": "order_status_change", "order_id": str(order.id), "status": new_status} ) ``` - **Akzeptanzkriterium**: Order-Submit -> WebSocket-Event geht an alle Browser-Tabs des Tenants - **Abhaengigkeiten**: Task 2 ### Task 6: Queue-Update Beat-Task [x] - **Datei**: `backend/app/tasks/celery_app.py` - **Was**: Neuer Beat-Task alle 10s: ```python @shared_task(name="beat.broadcast_queue_status", queue="step_processing") def broadcast_queue_status(): from app.core.websocket import publish_event_sync from redis import Redis r = Redis.from_url(settings.redis_url) depths = { "step_processing": r.llen("step_processing"), "thumbnail_rendering": r.llen("thumbnail_rendering"), } # Broadcast an alle Tenants (broadcast_all) r.publish("__broadcast__", json.dumps({"type": "queue_update", "depths": depths})) ``` - `__broadcast__` Channel: wird an ALLE verbundenen WS gesendet (nicht tenant-spezifisch) - ConnectionManager subscribt auch auf `__broadcast__` - **Akzeptanzkriterium**: WorkerActivity-Queue-Tiefe aktualisiert alle 10s automatisch - **Abhaengigkeiten**: Task 2 ### Task 7: Frontend WebSocket Hook [x] - **Datei**: `frontend/src/hooks/useWebSocket.ts` (neu) - **Was**: ```typescript export function useWebSocketConnection() { // Verbindet zu ws://localhost:8888/api/ws?token= // Auto-Reconnect: 1s, 2s, 4s, 8s, ... max 30s // Emittiert Events via onMessage callback // Pings alle 25s (keep-alive) // Trennt Verbindung bei Logout } ``` - **Akzeptanzkriterium**: Verbindung bleibt offen, reconnected nach Netzwerktrennung - **Abhaengigkeiten**: keine ### Task 8: Frontend WebSocket Context [x] - **Datei**: `frontend/src/contexts/WebSocketContext.tsx` (neu), `frontend/src/App.tsx` aendern - **Was**: ```typescript export function WebSocketProvider({ children }) { const queryClient = useQueryClient() // on 'render_complete': invalidateQueries(['orders', order_line_id]) // on 'render_failed': invalidateQueries(['orders', order_line_id]) // on 'cad_processing_complete': invalidateQueries(['cad-activity']) // on 'order_status_change': invalidateQueries(['orders']) // on 'queue_update': queryClient.setQueryData(['queue-status'], ...) } // App.tsx: um wrappen ``` - **Akzeptanzkriterium**: render_complete Event -> OrderDetail aktualisiert ohne Poll-Interval - **Abhaengigkeiten**: Task 7 ### Task 9: Polling ersetzen -- WorkerActivity.tsx [x] - **Datei**: `frontend/src/pages/WorkerActivity.tsx` - **Was**: - `refetchInterval: 5000` entfernen -- bei `cad_processing_complete` invalidieren - `refetchInterval: 3000` fuer Queue-Status entfernen -- bei `queue_update` setQueryData - **Akzeptanzkriterium**: Keine automatischen HTTP-Requests im Network-Tab (nur WS-Frames) - **Abhaengigkeiten**: Task 8 ### Task 10: Polling ersetzen -- OrderDetail.tsx [x] - **Datei**: `frontend/src/pages/OrderDetail.tsx` - **Was**: - `refetchInterval: (query) => {...}` entfernen - Stattdessen: bei `render_complete` / `render_failed` fuer matching order_line_id -> invalidate - **Akzeptanzkriterium**: Render-Status in OrderDetail aktualisiert live ohne Poll - **Abhaengigkeiten**: Task 8 ### Task 11: Polling reduzieren -- Layout.tsx + NotificationCenter.tsx [x] - **Dateien**: `frontend/src/components/layout/Layout.tsx`, `NotificationCenter.tsx` - **Was**: - Layout: `refetchInterval: 8000` -> 60000 (1min) - NotificationCenter: `refetchInterval: 15_000` -> 60000; bei `order_status_change` zusaetzlich invalidieren - **Akzeptanzkriterium**: Signifikant weniger Poll-Requests im Network-Tab - **Abhaengigkeiten**: Task 8 ### Task 12: PLAN.md + LEARNINGS.md + Commit [x] - **Was**: - PLAN.md: Phase J als ABGESCHLOSSEN markieren, Status auf "Phase K als naechstes" - LEARNINGS.md: ffmpeg `shortest=1` Learning + WebSocket Auth via Query-Param Learning - `git commit -m "feat(J): WebSocket live-events + replace polling + fix ffmpeg turntable timeout"` - **Abhaengigkeiten**: Tasks 1-11 --- ## Phase K Tasks (nach Commit) ### Task K1: Migration 045 + AssetLibrary Model [ ] - **Datei**: `backend/alembic/versions/045_asset_libraries.py` (neu, autogenerate), `domains/materials/models.py` - **Was**: ```python class AssetLibrary(Base): id: UUID PK, tenant_id FK nullable, name VARCHAR(200) blend_file_key TEXT, # MinIO key catalog JSONB, # {materials: [...], node_groups: [...]} description TEXT, is_active BOOL, created_at TIMESTAMP ``` - `render_templates.asset_library_id` FK optional (nullable) - `output_types.asset_library_id` FK optional (nullable) - **Akzeptanzkriterium**: `alembic upgrade head` erfolgreich, `asset_libraries` Tabelle in DB ### Task K2: Asset Library CRUD Backend [ ] - **Datei**: `backend/app/domains/materials/router.py` + `service.py` + `schemas.py` - **Was**: - `POST /api/asset-libraries` -- .blend Upload -> MinIO `asset-libraries/{id}.blend` -> queut Katalog-Refresh - `GET /api/asset-libraries` -- Liste - `GET /api/asset-libraries/{id}/catalog` -- Materialien + Node-Groups - `DELETE /api/asset-libraries/{id}` -- nur wenn nicht in Verwendung (FK-Check) - `AssetLibraryOut` Schema mit `catalog` field - **Akzeptanzkriterium**: POST + GET funktionieren, .blend in MinIO gespeichert ### Task K3: Katalog-Refresh Celery Task + Blender Script [ ] - **Datei**: `backend/app/domains/materials/tasks.py` (neu), `render-worker/scripts/catalog_assets.py` (neu) - **Was**: - Celery Task `refresh_asset_library_catalog(asset_library_id)` auf Queue `thumbnail_rendering` - Laedt .blend aus MinIO in tmpdir - Startet `blender --background --python catalog_assets.py -- ` - `catalog_assets.py`: oeffnet .blend, liest alle markierten Assets: ```python import bpy, json, sys blend_path = sys.argv[sys.argv.index('--') + 1] bpy.ops.wm.open_mainfile(filepath=blend_path) catalog = { "materials": [m.name for m in bpy.data.materials if m.asset_data], "node_groups": [ng.name for ng in bpy.data.node_groups if ng.asset_data], } print(json.dumps(catalog)) ``` - Schreibt Katalog in `asset_libraries.catalog JSONB` - **Akzeptanzkriterium**: Nach .blend-Upload enthaelt `catalog` JSONB die Asset-Namen ### Task K4: Blender Asset Library Apply Script [ ] - **Datei**: `render-worker/scripts/asset_library.py` (neu) - **Was**: ```python def apply_asset_library_materials(blend_path: str, material_map: dict) -> None: """Laedt Materialien aus Asset-Library .blend, wendet auf Mesh-Parts an.""" with bpy.data.libraries.load(blend_path, link=True, assets_only=True) as (src, dst): dst.materials = [n for n in src.materials if n in material_map.values()] for obj in bpy.data.objects: if obj.type == 'MESH': for slot in obj.material_slots: resolved = material_map.get(slot.material.name if slot.material else '') if resolved and resolved in bpy.data.materials: slot.material = bpy.data.materials[resolved] def apply_asset_library_modifiers(blend_path: str, modifier_map: dict) -> None: """Laedt Geometry-Node-Gruppen, wendet als Modifier an.""" with bpy.data.libraries.load(blend_path, link=True, assets_only=True) as (src, dst): dst.node_groups = [n for n in src.node_groups if n in modifier_map.values()] for obj in bpy.data.objects: if obj.type == 'MESH': for part_name, mod_name in modifier_map.items(): if part_name.lower() in obj.name.lower(): mod = obj.modifiers.new(name=mod_name, type='NODES') mod.node_group = bpy.data.node_groups.get(mod_name) ``` - **Akzeptanzkriterium**: Render mit Asset-Library zeigt korrekte Produktionsmaterialien ### Task K5: export_gltf + export_blend Scripts [ ] - **Dateien**: `render-worker/scripts/export_gltf.py` (neu), `render-worker/scripts/export_blend.py` (neu) - **Was**: - `export_gltf.py`: 1. STL importieren (`bpy.ops.import_mesh.stl`) 2. Asset Library laden via `apply_asset_library_materials` + `apply_asset_library_modifiers` 3. `bpy.ops.export_scene.gltf(filepath=out, export_format='GLB', export_apply=True, export_draco_mesh_compression_enable=True)` 4. Output nach MinIO `production-exports/{cad_file_id}/{run_id}.glb` 5. MediaAsset-Record mit `asset_type=gltf_production` - `export_blend.py`: 1. STL + Asset Library laden (wie export_gltf) 2. `bpy.ops.file.pack_all()` 3. `bpy.ops.wm.save_as_mainfile(filepath=out, compress=True, copy=True)` 4. MediaAsset-Record mit `asset_type=blend_production` - **Akzeptanzkriterium**: GLB-Download oeffnet sich im Three.js Viewer mit Materialien ### Task K6: Workflow-Builder -- Asset Library Nodes [ ] - **Datei**: `backend/app/domains/rendering/workflow_builder.py` - **Was**: - Neue Celery Tasks: `apply_asset_library_materials_task`, `apply_asset_library_modifiers_task`, `export_gltf_task`, `export_blend_task` - Neuer Workflow-Typ `still_production`: ```python chain( convert_step.si(order_line_id), group( chain(apply_asset_library_materials.si(order_line_id), render_still.si(order_line_id)), chain(apply_asset_library_materials.si(order_line_id), export_gltf.si(order_line_id)), chain(apply_asset_library_materials.si(order_line_id), export_blend.si(order_line_id)), ), generate_thumbnail.si(order_line_id), publish_asset.si(order_line_id), ) ``` - **Akzeptanzkriterium**: Dispatch eines `still_production` Workflows -> PNG + GLB + .blend erzeugt ### Task K7: Asset Library Management UI [ ] - **Dateien**: `frontend/src/api/assetLibraries.ts` (neu), `frontend/src/pages/Admin.tsx` erweitern - **Was**: - API Client: `getAssetLibraries`, `uploadAssetLibrary` (multipart), `deleteAssetLibrary`, `getAssetLibraryCatalog` - Admin.tsx: neues Panel "Asset Libraries" (nach Render Templates) - Upload-Button + Drag-Drop - Tabelle: Name, Materialien-Anzahl, Node-Groups-Anzahl, Aktionen - Katalog-Detail: Material-Badge-Liste (gruen) + Node-Group-Badge-Liste (blau) - OutputTypeTable: Asset-Library-Dropdown-Spalte - **Akzeptanzkriterium**: Admin kann .blend hochladen, Katalog sehen, OutputType zuweisen ### Task K8: PLAN.md + LEARNINGS.md + Commit [ ] - **Was**: - PLAN.md: Phase K als ABGESCHLOSSEN markieren - LEARNINGS.md: Asset Library link=True Pattern, GLB-Export Blender API - `git commit -m "feat(K): Blender Asset Library + production exports (GLB + .blend)"` --- ## Migrations-Check | Migration | Phase | Status | |-----------|-------|--------| | 041 step_file_hash | F | existiert | | 042 invoices | G | existiert | | 043 import_validations | H | existiert | | 044 notification_configs | I | existiert | | **045 asset_libraries** | **K** | **fehlt** | --- ## Reihenfolge-Empfehlung ``` Task 1 (Bug-Fix, sofort) Tasks 2-6 parallel (Backend WebSocket) Tasks 7-8 parallel (Frontend Hook + Context) Tasks 9-11 (Polling ersetzen, nach 8) Task 12 (Commit) Tasks K1-K3 parallel (Datenmodell + Backend + Blender-Katalog) Tasks K4-K5 parallel (Blender Scripts) Tasks K6-K7 (Workflow + UI, nach K1-K5) Task K8 (Commit) ``` --- ## Risiken / Offene Fragen - **WebSocket Auth via Query-Param**: Token in Server-Logs sichtbar. Fuer v2 akzeptabel. In v3: kurzlebigen WS-Token (TTL 30s) aus JWT generieren. - **Redis Pub/Sub Skalierung**: Bei vielen Tenants/Tabs kann Subscriber-Loop Bottleneck werden. Fuer v2 OK. In v3: Redis Streams. - **Phase K -- MinIO Bucket**: `asset-libraries` Bucket muss beim Startup erstellt werden (lifespan in main.py). - **Phase K -- link=True** bedeutet .blend muss vor Render via MinIO heruntergeladen werden (in tmpdir). Bereits einkalkuliert in K3. - **Bestehende material_libraries**: Die alte `material_libraries` Tabelle/Feature bleibt parallel bestehen -- kein Breaking Change. Asset Libraries sind additiv.