feat(F-G-H-I): STL cache, invoices, import validation, notification settings
Phase F — STL Hash Cache:
- Migration 041: step_file_hash column on cad_files
- cache_service.py: SHA256 hash + MinIO-backed STL cache (check/store)
- render_step_thumbnail: compute+persist hash before render
- generate_stl_cache: check MinIO cache before cadquery conversion, store after
Phase G — Invoices:
- Migration 042: invoices + invoice_lines tables with RLS
- Invoice/InvoiceLine models + schemas
- billing service: generate_invoice_number (INV-YYYY-NNNN), create/list/get/delete/PDF
- WeasyPrint PDF generation; backend Dockerfile + pyproject.toml deps
- invoice_router with 6 endpoints; registered in main.py
- frontend: Billing.tsx page + api/billing.ts; route + nav link
Phase H — Import Sanity Check:
- Migration 043: import_validations table
- ImportValidation model + schemas
- run_sanity_check: material fuzzy-match (cutoff=0.8), STEP availability, duplicate detection
- validate_excel_import Celery task (queue: step_processing)
- uploads.py: create ImportValidation on /excel, fire task, expose GET /validations/{id}
- frontend: Upload.tsx polling ValidationDialog with Ampel status indicators
Phase I — Notification Settings:
- Migration 044: notification_configs table (user×event×channel toggles)
- NotificationConfig model + seeds (in_app=true, email=false)
- get/upsert/reset config endpoints on /notifications/config
- frontend: NotificationSettings.tsx page + api/notifications.ts extensions
Infrastructure:
- docker-compose.yml: add worker-thumbnail service (concurrency=1, Q=thumbnail_rendering)
- Fix Dockerfile: libgdk-pixbuf-2.0-0 (correct Debian bookworm package name)
Co-Authored-By: Claude Sonnet 4.6 <noreply@anthropic.com>
This commit is contained in:
@@ -0,0 +1,22 @@
|
||||
"""Add step_file_hash to cad_files.
|
||||
|
||||
Revision ID: 041
|
||||
Revises: 040
|
||||
"""
|
||||
import sqlalchemy as sa
|
||||
from alembic import op
|
||||
|
||||
revision = '041'
|
||||
down_revision = '040'
|
||||
branch_labels = None
|
||||
depends_on = None
|
||||
|
||||
|
||||
def upgrade():
|
||||
op.add_column('cad_files', sa.Column('step_file_hash', sa.String(64), nullable=True))
|
||||
op.create_index('ix_cad_files_step_file_hash', 'cad_files', ['step_file_hash'])
|
||||
|
||||
|
||||
def downgrade():
|
||||
op.drop_index('ix_cad_files_step_file_hash', table_name='cad_files')
|
||||
op.drop_column('cad_files', 'step_file_hash')
|
||||
@@ -0,0 +1,67 @@
|
||||
"""Add invoices and invoice_lines tables.
|
||||
|
||||
Revision ID: 042
|
||||
Revises: 041
|
||||
"""
|
||||
import sqlalchemy as sa
|
||||
from alembic import op
|
||||
from sqlalchemy.dialects.postgresql import UUID
|
||||
|
||||
revision = '042'
|
||||
down_revision = '041'
|
||||
branch_labels = None
|
||||
depends_on = None
|
||||
|
||||
|
||||
def upgrade():
|
||||
op.create_table(
|
||||
'invoices',
|
||||
sa.Column('id', UUID(as_uuid=True), primary_key=True, server_default=sa.text('gen_random_uuid()')),
|
||||
sa.Column('tenant_id', UUID(as_uuid=True), sa.ForeignKey('tenants.id', ondelete='CASCADE'), nullable=True),
|
||||
sa.Column('invoice_number', sa.String(20), nullable=False, unique=True),
|
||||
sa.Column('status', sa.String(20), nullable=False, server_default='draft'),
|
||||
sa.Column('issued_at', sa.Date, nullable=True),
|
||||
sa.Column('due_at', sa.Date, nullable=True),
|
||||
sa.Column('total_net', sa.Numeric(12, 2), nullable=True),
|
||||
sa.Column('total_vat', sa.Numeric(12, 2), nullable=True),
|
||||
sa.Column('vat_rate', sa.Numeric(5, 4), nullable=False, server_default='0.19'),
|
||||
sa.Column('currency', sa.String(3), nullable=False, server_default='EUR'),
|
||||
sa.Column('notes', sa.Text, nullable=True),
|
||||
sa.Column('pdf_key', sa.Text, nullable=True),
|
||||
sa.Column('created_at', sa.DateTime, nullable=False, server_default=sa.text('NOW()')),
|
||||
sa.Column('updated_at', sa.DateTime, nullable=False, server_default=sa.text('NOW()')),
|
||||
)
|
||||
op.create_index('ix_invoices_tenant', 'invoices', ['tenant_id'])
|
||||
op.create_index('ix_invoices_status', 'invoices', ['status'])
|
||||
|
||||
op.create_table(
|
||||
'invoice_lines',
|
||||
sa.Column('id', UUID(as_uuid=True), primary_key=True, server_default=sa.text('gen_random_uuid()')),
|
||||
sa.Column('invoice_id', UUID(as_uuid=True), sa.ForeignKey('invoices.id', ondelete='CASCADE'), nullable=False),
|
||||
sa.Column('order_line_id', UUID(as_uuid=True), sa.ForeignKey('order_lines.id', ondelete='SET NULL'), nullable=True),
|
||||
sa.Column('description', sa.Text, nullable=False),
|
||||
sa.Column('quantity', sa.Integer, nullable=False, server_default='1'),
|
||||
sa.Column('unit_price', sa.Numeric(10, 2), nullable=True),
|
||||
sa.Column('total', sa.Numeric(10, 2), nullable=True),
|
||||
)
|
||||
op.create_index('ix_invoice_lines_invoice', 'invoice_lines', ['invoice_id'])
|
||||
|
||||
# RLS
|
||||
op.execute("ALTER TABLE invoices ENABLE ROW LEVEL SECURITY")
|
||||
op.execute("""
|
||||
DO $$ BEGIN
|
||||
CREATE POLICY tenant_isolation ON invoices
|
||||
USING (tenant_id = current_setting('app.current_tenant_id', true)::uuid);
|
||||
EXCEPTION WHEN duplicate_object THEN NULL; END $$;
|
||||
""")
|
||||
op.execute("""
|
||||
DO $$ BEGIN
|
||||
CREATE POLICY admin_bypass ON invoices
|
||||
USING (current_setting('app.current_tenant_id', true) = 'bypass');
|
||||
EXCEPTION WHEN duplicate_object THEN NULL; END $$;
|
||||
""")
|
||||
|
||||
|
||||
def downgrade():
|
||||
op.drop_table('invoice_lines')
|
||||
op.drop_table('invoices')
|
||||
@@ -0,0 +1,33 @@
|
||||
"""Add import_validations table.
|
||||
|
||||
Revision ID: 043
|
||||
Revises: 042
|
||||
"""
|
||||
import sqlalchemy as sa
|
||||
from alembic import op
|
||||
from sqlalchemy.dialects.postgresql import UUID, JSONB
|
||||
|
||||
revision = '043'
|
||||
down_revision = '042'
|
||||
branch_labels = None
|
||||
depends_on = None
|
||||
|
||||
|
||||
def upgrade():
|
||||
op.create_table(
|
||||
'import_validations',
|
||||
sa.Column('id', UUID(as_uuid=True), primary_key=True, server_default=sa.text('gen_random_uuid()')),
|
||||
sa.Column('tenant_id', UUID(as_uuid=True), sa.ForeignKey('tenants.id', ondelete='CASCADE'), nullable=True),
|
||||
sa.Column('excel_path', sa.Text, nullable=False),
|
||||
sa.Column('status', sa.String(20), nullable=False, server_default='pending'),
|
||||
sa.Column('summary', JSONB, nullable=True),
|
||||
sa.Column('rows', JSONB, nullable=True),
|
||||
sa.Column('created_at', sa.DateTime, nullable=False, server_default=sa.text('NOW()')),
|
||||
sa.Column('completed_at', sa.DateTime, nullable=True),
|
||||
)
|
||||
op.create_index('ix_import_validations_tenant', 'import_validations', ['tenant_id'])
|
||||
op.create_index('ix_import_validations_status', 'import_validations', ['status'])
|
||||
|
||||
|
||||
def downgrade():
|
||||
op.drop_table('import_validations')
|
||||
@@ -0,0 +1,56 @@
|
||||
"""Add notification_configs table.
|
||||
|
||||
Revision ID: 044
|
||||
Revises: 043
|
||||
"""
|
||||
import sqlalchemy as sa
|
||||
from alembic import op
|
||||
from sqlalchemy.dialects.postgresql import UUID
|
||||
|
||||
revision = '044'
|
||||
down_revision = '043'
|
||||
branch_labels = None
|
||||
depends_on = None
|
||||
|
||||
# Standard events
|
||||
EVENTS = [
|
||||
"order.submitted",
|
||||
"order.completed",
|
||||
"render.completed",
|
||||
"render.failed",
|
||||
"excel.imported",
|
||||
]
|
||||
|
||||
|
||||
def upgrade():
|
||||
op.create_table(
|
||||
'notification_configs',
|
||||
sa.Column('id', UUID(as_uuid=True), primary_key=True, server_default=sa.text('gen_random_uuid()')),
|
||||
sa.Column('user_id', UUID(as_uuid=True), sa.ForeignKey('users.id', ondelete='CASCADE'), nullable=False),
|
||||
sa.Column('event_type', sa.String(100), nullable=False),
|
||||
sa.Column('channel', sa.String(20), nullable=False),
|
||||
sa.Column('enabled', sa.Boolean, nullable=False, server_default='true'),
|
||||
sa.Column('created_at', sa.DateTime, nullable=False, server_default=sa.text('NOW()')),
|
||||
)
|
||||
op.create_index('ix_notification_configs_user', 'notification_configs', ['user_id'])
|
||||
op.create_unique_constraint(
|
||||
'uq_notification_config_user_event_channel',
|
||||
'notification_configs',
|
||||
['user_id', 'event_type', 'channel']
|
||||
)
|
||||
# Seed defaults for admin user (in_app=true, email=false)
|
||||
for event in EVENTS:
|
||||
op.execute(f"""
|
||||
INSERT INTO notification_configs (user_id, event_type, channel, enabled)
|
||||
SELECT id, '{event}', 'in_app', true FROM users WHERE role = 'admin'
|
||||
ON CONFLICT DO NOTHING
|
||||
""")
|
||||
op.execute(f"""
|
||||
INSERT INTO notification_configs (user_id, event_type, channel, enabled)
|
||||
SELECT id, '{event}', 'email', false FROM users WHERE role = 'admin'
|
||||
ON CONFLICT DO NOTHING
|
||||
""")
|
||||
|
||||
|
||||
def downgrade():
|
||||
op.drop_table('notification_configs')
|
||||
Reference in New Issue
Block a user