Files
HartOMat/plan.md
T
Hartmut 59ce61098c feat: tenant AI chat agent with function calling
Actionable AI assistant that uses per-tenant Azure OpenAI credentials
to execute natural language commands against the render pipeline.

Backend:
- ChatMessage model + migration (session-based conversations)
- Chat service with 10 OpenAI function-calling tools:
  list_orders, search_products, create_order, dispatch_renders,
  get_order_status, set_material_override, set_render_overrides,
  get_render_stats, check_materials, query_database
- All tools tenant-scoped (queries filtered by tenant_id)
- Write operations use httpx to call backend API internally
- Chat API: POST /chat/messages, GET /chat/sessions, DELETE session
- Conversation history preserved in DB (last 50 messages per session)

Frontend:
- Slide-out ChatPanel (right side, w-96, animated)
- User/assistant message styling with avatars and timestamps
- Session management (new chat, session history, delete)
- Typing indicator while waiting for AI response
- Floating chat button in bottom-right corner
- Error state for unconfigured AI tenants

Example: "Render all Kugellager products as WebP at 1024x1024"
→ Agent calls search_products + create_order + dispatch_renders

Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
2026-03-15 12:46:21 +01:00

209 lines
9.7 KiB
Markdown

# Plan: Tenant AI Chat Agent (Actionable)
## Context
Each tenant has Azure OpenAI credentials stored in `tenant_config` JSONB. The goal is an **actionable AI agent** where users can type natural language commands to control the render pipeline — create orders, dispatch renders, check status, set overrides — scoped to their tenant.
Example interactions:
- "Render all Kugellager products as WebP at 1024x1024"
- "What's the status of my last order?"
- "Set material override to Steel-Bare on order SA-2026-00160"
- "How many renders failed this week?"
The agent uses **function calling** (Azure OpenAI tool use) — the LLM decides which API action to execute, the backend executes it, and returns the result. Tenants are fully isolated — each uses their own Azure API key and only sees their own data.
**What exists:**
- Per-tenant Azure OpenAI credentials in `tenant_config` JSONB
- WebSocket system scoped by tenant for real-time events
- `ai_validation` Celery queue (concurrency=8)
- Azure OpenAI integration boilerplate in `azure_ai.py`
## Affected Files
| File | Change |
|------|--------|
| `backend/app/models/chat.py` | **NEW** — ChatMessage model |
| `backend/app/models/__init__.py` | Import ChatMessage |
| `backend/app/api/routers/chat.py` | **NEW** — Chat API endpoints |
| `backend/app/services/chat_service.py` | **NEW** — Azure OpenAI chat + DB context |
| `backend/app/main.py` | Register chat router |
| `backend/alembic/versions/XXX_add_chat_messages.py` | Migration |
| `frontend/src/api/chat.ts` | **NEW** — Chat API types + functions |
| `frontend/src/components/chat/ChatPanel.tsx` | **NEW** — Chat UI component |
| `frontend/src/components/layout/Layout.tsx` | Add chat toggle button |
## Tasks (in order)
### [ ] Task 1: ChatMessage model + migration
- **File**: `backend/app/models/chat.py` (new)
- **What**: Create a ChatMessage model:
```python
class ChatMessage(Base):
__tablename__ = "chat_messages"
id: UUID PK
tenant_id: UUID FK → tenants.id (nullable, indexed)
user_id: UUID FK → users.id (nullable)
session_id: UUID (groups messages in a conversation, indexed)
role: String(20) — "user", "assistant", "system"
content: Text
context_type: String(50) nullable — "order", "product", "general"
context_id: UUID nullable — order_id or product_id
token_count: Integer nullable — track usage
created_at: DateTime
```
- **Also**: Import in `backend/app/models/__init__.py`
- **Migration**: `alembic revision --autogenerate -m "add chat_messages table"`
- **Acceptance gate**: Table exists in DB; model importable
- **Dependencies**: None
### [ ] Task 2: Chat service — Azure OpenAI with function calling
- **File**: `backend/app/services/chat_service.py` (new)
- **What**: Service with Azure OpenAI **tool use / function calling**:
1. Takes a user message + session_id + tenant_id + user_id
2. Loads tenant Azure credentials from `tenant_config`
3. Defines **tools** the LLM can call (JSON schema for each):
- `list_orders(status, limit)` — list tenant's orders
- `search_products(query, category, limit)` — search products
- `create_order(product_ids, output_type_name, render_overrides, material_override)` — create & submit
- `dispatch_renders(order_id)` — dispatch renders for an order
- `get_order_status(order_id)` — check render progress
- `set_material_override(order_id, material_name)` — batch material override
- `set_render_overrides(order_id, overrides)` — batch render overrides
- `get_render_stats()` — throughput stats
- `check_materials(order_id)` — unmapped materials check
- `query_database(sql)` — read-only SQL (SELECT only, tenant-scoped)
4. Calls Azure OpenAI with `tools` parameter — the LLM decides which tool to call
5. Executes the tool call internally (same functions as MCP server but tenant-scoped)
6. Returns tool result to LLM for a natural language response
7. Stores conversation in ChatMessage table
**Tenant isolation**: All DB queries filter by `tenant_id`. The `query_database` tool auto-appends `WHERE tenant_id = '{tenant_id}'` or validates tenant scope.
**Tool execution**: Uses the existing backend API functions directly (not HTTP calls) — import from the routers/services.
```python
tools = [
{
"type": "function",
"function": {
"name": "search_products",
"description": "Search products by name, PIM-ID, or category",
"parameters": {
"type": "object",
"properties": {
"query": {"type": "string"},
"category": {"type": "string"},
}
}
}
},
# ... more tools
]
response = client.chat.completions.create(
model=deployment,
messages=messages,
tools=tools,
tool_choice="auto",
)
# Handle tool_calls in response, execute, return result
```
- **Acceptance gate**: User can say "show my last 5 orders" and get real data back via function calling
- **Dependencies**: Task 1
### [ ] Task 3: Chat API endpoints
- **File**: `backend/app/api/routers/chat.py` (new)
- **What**: FastAPI router with endpoints:
- `POST /api/chat/messages` — send a message, get AI response
- Body: `{ message: str, session_id: str | None, context_type: str | None, context_id: str | None }`
- Creates session_id if not provided
- Returns: `{ session_id: str, message: ChatMessageOut, response: ChatMessageOut }`
- Auth: `get_current_user` — uses user's tenant AI config
- `GET /api/chat/sessions` — list user's chat sessions
- Returns: `[{ session_id, last_message, message_count, created_at }]`
- `GET /api/chat/sessions/{session_id}/messages` — get conversation history
- Returns: `[{ id, role, content, created_at }]`
- `DELETE /api/chat/sessions/{session_id}` — delete a conversation
- **Also**: Register router in `backend/app/main.py`
- **Acceptance gate**: POST /api/chat/messages returns an AI response using tenant credentials
- **Dependencies**: Task 2
### [ ] Task 4: Frontend — Chat API types
- **File**: `frontend/src/api/chat.ts` (new)
- **What**: TypeScript interfaces and API functions:
```typescript
interface ChatMessage { id: string; role: 'user' | 'assistant' | 'system'; content: string; created_at: string }
interface ChatSession { session_id: string; last_message: string; message_count: number; created_at: string }
interface ChatResponse { session_id: string; message: ChatMessage; response: ChatMessage }
function sendMessage(message: string, sessionId?: string, contextType?: string, contextId?: string): Promise<ChatResponse>
function getSessions(): Promise<ChatSession[]>
function getSessionMessages(sessionId: string): Promise<ChatMessage[]>
function deleteSession(sessionId: string): Promise<void>
```
- **Acceptance gate**: Types compile; functions callable
- **Dependencies**: Task 3
### [ ] Task 5: Frontend — ChatPanel component
- **File**: `frontend/src/components/chat/ChatPanel.tsx` (new)
- **What**: Slide-out chat panel (right side, similar to notification panels in modern apps):
1. **Header**: "AI Assistant" title + close button + session selector
2. **Message list**: Scrollable area with role-based styling:
- User messages: right-aligned, accent background
- Assistant messages: left-aligned, surface background, markdown support
- Timestamps below each message
3. **Input area**: Text input + send button (Enter to send)
4. **Loading state**: Typing indicator while waiting for AI response
5. **Session management**: "New conversation" button, session history dropdown
6. **Context awareness**: When opened from an order/product page, auto-includes context
**Styling**:
- Fixed right panel (w-96, full height)
- Backdrop overlay on mobile
- Smooth slide-in animation
- Use existing CSS variables (surface, content, accent)
- lucide-react icons (MessageSquare, Send, Loader2, X, Plus)
- **Acceptance gate**: Panel opens/closes, messages send and display, AI responds
- **Dependencies**: Task 4
### [ ] Task 6: Frontend — Chat toggle in Layout
- **File**: `frontend/src/components/layout/Layout.tsx`
- **What**: Add a chat toggle button:
1. Floating button in bottom-right corner (or in the sidebar)
2. Icon: `MessageSquare` from lucide-react
3. Badge with unread count (optional, for future)
4. Click toggles ChatPanel visibility
5. Only show when tenant has `ai_enabled = true`
- **Acceptance gate**: Button visible for users with AI-enabled tenant; clicking opens/closes ChatPanel
- **Dependencies**: Task 5
## Migration Check
**Yes** — one new table `chat_messages` with UUID PK, FK to tenants and users.
## Order Recommendation
1. Backend model + migration (Task 1)
2. Backend service (Task 2)
3. Backend API (Task 3)
4. Frontend types (Task 4)
5. Frontend chat UI (Task 5)
6. Frontend layout integration (Task 6)
## Risks / Open Questions
1. **Azure OpenAI availability**: If tenant hasn't configured AI credentials, the chat should show a helpful message ("AI not configured — ask your admin to set up Azure OpenAI in Tenant Settings")
2. **Token costs**: Each message uses Azure OpenAI tokens. Consider adding token counting and a configurable monthly limit per tenant.
3. **Context enrichment**: The system prompt could include live data (order counts, render status). This makes the AI more helpful but costs more tokens. Start simple, enhance later.
4. **Streaming responses**: Azure OpenAI supports streaming. V1 uses a simple request/response. V2 could stream via WebSocket for real-time typing effect.
5. **openai package**: The `openai` Python package must be installed in the backend container. Check if it's already a dependency (it may be via `azure_ai.py`).