Run #115 on main timed out after 30s on the Gitea runner under
concurrent-job load (writing 10001 rows via ExcelJS addRow + writeFile
is CPU-bound and CI contention pushed it past the previous threshold).
Locally these tests complete in ~1s, so doubling the budget removes
the flake without masking real regressions.
Co-Authored-By: Claude Opus 4.7 <noreply@anthropic.com>
Previous run's Build job failed but Gitea's actions log store didn't retain
the output (dbfs reports the file missing), so we can't diagnose from here.
Rerun to either reproduce the failure with a persisted log, or green-ify.
Co-Authored-By: Claude Opus 4.7 <noreply@anthropic.com>
Adds a one-time-use backup code set so users with a lost authenticator are not
locked out. Codes are Crockford base32 (XXXXX-XXXXX), hashed with argon2id, and
redeemed under a WHERE-guarded delete so a concurrent replay race fails closed.
- New MfaBackupCode model + migration
- Issue 10 codes inside the enable transaction; show plaintext exactly once
- Sign-in page accepts TOTP or backup code, reporting remaining count
- regenerateBackupCodes tRPC mutation wipes + reissues atomically
- Unit coverage for generator, normalizer, verify, redeem, and race path
Co-Authored-By: Claude Opus 4.7 <noreply@anthropic.com>
Two regressions surfaced after merging security/audit-2026-04-17:
1. **Build job** failed with `assertSecureRuntimeEnv` rejecting the CI
`NEXTAUTH_SECRET=ci-test-secret-minimum-32-chars-xx`. The CI placeholder
strings were added to `DISALLOWED_PRODUCTION_SECRETS` defensively, but
that list is only consulted when `NODE_ENV=production` — exactly the
mode `next build` runs in. The length + Shannon-entropy gates already
reject genuinely weak prod secrets (the CI value scores ~3.68 vs the
3.5 threshold), so removing the CI strings from the blocklist restores
the build without weakening prod protection.
2. **Unit-tests job** failed with `(0 , brace_expansion_1.default) is not
a function` from `minimatch@9` → `brace-expansion@5.0.5` (ESM-only)
loaded via CJS `require`. The blanket override `"brace-expansion":
"^5.0.5"` (added for CVE-2025-5889) was too broad. Switching to the
targeted `"brace-expansion@<2.0.2": ">=2.0.2"` patches the CVE while
leaving CJS consumers (test-exclude/glob/minimatch) on v2.
Drops the now-stale CI-placeholder unit test in `runtime-env.test.ts`.
Co-Authored-By: Claude Opus 4.7 <noreply@anthropic.com>
- dispo workbook imports are pinned to DISPO_IMPORT_DIR (default ./imports):
tRPC input rejects absolute paths and .. segments, runtime reader
re-validates containment via path.relative. Closes a path-traversal
class that reached ExcelJS CVEs through admin/compromised tokens.
- image validator now checks the full 8-byte PNG magic, enforces PNG IEND
and JPEG EOI trailers, scans the decoded buffer for markup polyglot
markers (<script, <svg, <iframe, javascript:, onerror=, ...), and
explicitly rejects SVG. Provider-generated covers (DALL-E, Gemini) run
through the same validator before persistence — an untrusted upstream
cannot smuggle a stored-XSS payload past us.
- added image-validation.test.ts and tightened documentation.
Co-Authored-By: Claude Opus 4.7 <noreply@anthropic.com>
- Auth.js authorize/signOut: await createAuditEntry on every branch so auth
events land in the audit store before the JWT is minted / session closes.
Previously these were fire-and-forget and would be dropped under DB load.
- Assistant chat: make appendPromptInjectionGuard async and await its own
SecurityAlert audit; add auditUserPromptTurn() that records every user
message turn as an AssistantPrompt entry containing conversationId, length,
SHA-256 fingerprint, pageContext and whether the injection guard fired.
Raw prompt text is intentionally not stored — the hash lets a responder
correlate a chat transcript with a forensic request without the audit
store accumulating a plain-text corpus of everything users typed.
- Replace bare crypto.* with explicit node:crypto imports.
- Document the retention posture in docs/security-architecture.md §6.
Fixes gitea #55.
Client-side validators (reset-password, invite-accept, first-admin setup,
user-create modal) previously checked password.length < 8 while every
server-side Zod schema required .min(12). External API consumers (or a
confused browser UI) could get past the client check but fail at the tRPC
boundary — or worse, quietly under-enforce policy compared to what
admins expect.
Fix: introduce PASSWORD_MIN_LENGTH (12) and PASSWORD_MAX_LENGTH (128) in
@capakraken/shared and import them from every pre-submit client validator
and every server Zod schema. Single source of truth; drift becomes a
compile error rather than a security finding.
Also hardens the AUTH_SECRET runtime check: in addition to the existing
placeholder-blacklist, production startup now rejects secrets shorter
than 32 chars OR with Shannon entropy below 3.5 bits/char. That covers
low-entropy-but-long values like "aaaa..." (38 chars, entropy 0) which
would have passed the previous checks.
Documented the rotation process for AUTH_SECRET + POSTGRES_PASSWORD in
docs/security-architecture.md §3.
Verified:
- pnpm test:unit — 396 files / 1922 tests passed
- pnpm --filter @capakraken/web exec tsc --noEmit — clean
- pnpm --filter @capakraken/api exec tsc --noEmit — clean
Co-Authored-By: Claude Opus 4.7 <noreply@anthropic.com>
- docker-compose.yml: require ${POSTGRES_PASSWORD} for the postgres service
and the app container's DATABASE_URL. No default — compose refuses to start
without it, mirroring the existing PGADMIN_PASSWORD pattern.
- Dockerfile.prod: move auth/db ENV assignments from persistent ENV lines into
an inline env prefix on the `pnpm build` RUN step. Placeholders are still
available to `next build` but no longer persist in the builder layer or in
the published migrator image (which is FROM builder).
- Dockerfile.dev: add HEALTHCHECK against /api/health and install curl for it.
- .dockerignore: cover nested **/.env*, **/*.pem, **/*.key, **/secrets/**.
- runtime-env.ts: add the CI build placeholder strings to the disallowed-secret
set so a misconfigured prod deploy using the baked-in ARG defaults fails
startup instead of silently running with a known-bad secret.
- .env.example: document the new POSTGRES_PASSWORD requirement.
- CI: write POSTGRES_PASSWORD into the Fresh-Linux Docker Deploy job's .env
(must match docker-compose.ci.yml's hardcoded DATABASE_URL), and provide a
dummy value in the E2E job where compose validates all services' interp.
Co-Authored-By: Claude Opus 4.7 <noreply@anthropic.com>
- shrink roleDefaults cache TTL from 60s to 10s (safety-net staleness bound)
- publish/subscribe on capakraken:rbac-invalidate so peer instances drop
their local role-defaults cache on mutation (ioredis pub/sub; lazy init
so idle test files don't open connections)
- after updateUserRole/setUserPermissions/resetUserPermissions: delete
all ActiveSession rows for that user so the next request re-auths via
tRPC's jti check, and invalidate the role-defaults cache
- tests: peer-instance invalidation via FakeRedis pub/sub fan-out; mutation
side-effects assert session deletion + cache invalidation on each path
Without this, demoted admins kept their JWT valid until expiry and peer
instances kept serving stale role defaults for up to the TTL window.
Co-Authored-By: Claude Opus 4.7 <noreply@anthropic.com>
Five helper error mappers (timeline / project-creation / resource-creation
/ vacation-creation / task-action-execution) fell through to
`return { error: error.message }` for BAD_REQUEST and CONFLICT cases. When
the TRPCError wrapped a Prisma error, the message contained column names,
relation paths, and the offending unique-constraint value — all of which
would reach the LLM in chat context and, via audit_log.changes JSONB, the DB.
Add `sanitizeAssistantErrorMessage()` that regex-detects Prisma and raw
Postgres signatures (P2002/P2003/P2025, not-null, FK, check-constraint,
duplicate-key) and replaces them with a generic "Invalid input". Also caps
messages at 500 chars to defend against stack-trace-like payloads. Wire
the helper into all five call-sites; the developer-constructed
`AssistantVisibleError` branch in `normalizeAssistantExecutionError` is
left untouched since those strings are hand-written.
Coverage: 11 new tests in assistant-tools-error-sanitiser.test.ts; existing
vacation / task-action / resource-creation / project-creation error tests
(12 tests, 5 files) all remain green.
Co-Authored-By: Claude Opus 4.7 <noreply@anthropic.com>
Admin-editable blueprint field patterns go through `new RegExp(pattern).test(userValue)`
— a classic ReDoS sink if the admin account is compromised or the
permission is ever delegated. A pattern like `^(a+)+$` against 30
'a's followed by '!' freezes the event loop for seconds per request.
Three layers of defence:
- Save-time: FieldValidationSchema.pattern now has `.max(200)` and a
`.refine()` that rejects nested-quantifier shapes like `(x+)+`,
`(?:x*)+`, `(x{2,})*`.
- Runtime (engine/blueprint/validator.ts):
- isSuspectRegexPattern() runs the same heuristic. If it fires, the
field fails validation outright — regex is never compiled.
- Input strings are sliced to 4096 chars before .test() so even a
benign pattern against a 10 MB payload returns in < 50 ms.
- RegExp compile failures are caught and treated as validation
errors rather than crashing the request.
Tests: 10 cases in packages/engine/src/__tests__/blueprint-validator-redos.test.ts,
including the canonical `^(a+)+$` attack — completes in < 50 ms.
Co-Authored-By: Claude Opus 4.7 <noreply@anthropic.com>
Adds a regression suite asserting that the read-only Prisma proxy is
still in effect after a tool's executor forwards ctx.db into a scoped
tRPC caller (helpers.ts::createScopedCallerContext). Covers all three
attack surfaces: model writes, raw-SQL escape hatches, and interactive
$transaction / $runCommandRaw calls.
These tests pin the behaviour enforced by 1ff5c33; any future refactor
that unwraps the proxy during forwarding will fail this suite.
Co-Authored-By: Claude Opus 4.7 <noreply@anthropic.com>
createAuditEntry now deep-walks before/after/metadata and replaces
values of password, newPassword, currentPassword, passwordHash, token,
accessToken, refreshToken, sessionToken, apiKey, authorization, cookie,
secret, totpSecret, backupCode(s) with "[REDACTED]" before the JSONB
write.
The pino logger already redacts these paths for stdout (see
lib/logger.ts), but DB writes had no equivalent guard — the AI chat
loop at assistant-chat-loop.ts:265 blindly stores parsedArgs from tool
calls (e.g. set_user_password, create_user) into the AuditLog table.
Matching is case-insensitive; nested objects and arrays are recursed to
a depth of 8. Diffs are computed post-redaction so UPDATE entries that
only changed a sensitive field are correctly collapsed to no-op.
Co-Authored-By: Claude Opus 4.7 <noreply@anthropic.com>
Expand the SSRF blocklist from IPv4-only to IPv6 loopback/ULA (fc00::/7)/
link-local (fe80::/10)/multicast/IPv4-mapped, plus the missing IPv4 ranges
0.0.0.0/8, 100.64.0.0/10 CGNAT, and TEST-NET/benchmark ranges. Replace the
single-lookup SSRF guard with resolveAndValidate(): resolves all DNS records
(lookup { all: true }) so a hostname returning "public + private" is
rejected, and returns the first validated address for connection pinning.
The webhook dispatcher now switches from plain fetch() to https.request()
with a custom Agent.lookup that returns the pre-validated IP. A DNS rebind
between the guard check and the TCP connect() can no longer redirect the
dial to an internal address. Hostname still flows through for SNI and
certificate validation.
Co-Authored-By: Claude Opus 4.7 <noreply@anthropic.com>
The previous SELECT → compare → UPDATE sequence let two concurrent login
requests with the same valid 6-digit code both observe a stale lastTotpAt,
both pass the in-JS replay check, and both succeed. A stolen TOTP (shoulder-
surf, phishing-proxy replay) was usable twice within its 30 s window.
Replace the three callsites (login authorize, self-service enable, self-
service verify) with a shared consumeTotpWindow() helper: a single
updateMany() expresses "window unused" as a SQL WHERE clause, so Postgres'
row lock serialises concurrent writers and whichever commits second sees
count=0 and is treated as a replay.
Backup codes (ticket part 2) are tracked as follow-up work.
Co-Authored-By: Claude Opus 4.7 <noreply@anthropic.com>
Browser code never calls OpenAI/Azure/Gemini directly; all AI traffic is
server-side tRPC. connect-src is now locked to 'self'. Added object-src 'none',
frame-src 'none', media-src 'self', and worker-src 'self' blob:. style-src
keeps 'unsafe-inline' for React + @react-pdf/renderer (documented residual
risk — script-src is nonce-based so CSS injection cannot escalate to JS).
Added three regression tests covering connect-src no-wildcards, object/frame-src
'none', and worker-src scope.
Co-Authored-By: Claude Opus 4.7 <noreply@anthropic.com>
Previously middleware.ts listed /api/ as a public prefix, so any new
API route added under /api/** was served without a session check
unless the developer remembered to self-authenticate it. The
middleware now returns 404 for any /api path not explicitly
allowlisted (auth, trpc, sse, cron, reports, health, ready, perf) —
adding a new API route is a deliberate allowlist edit. verifyCronSecret
was already fail-closed when CRON_SECRET is unset; added unit tests.
Co-Authored-By: Claude Opus 4.7 <noreply@anthropic.com>
Three related fixes:
- Cookie secure flag now tracks AUTH_URL scheme (https → Secure),
not NODE_ENV — staging over HTTPS with NODE_ENV!=production used
to ship Set-Cookie without Secure. Cookie name gains __Host-
prefix when Secure is on.
- jwt() callback no longer swallows session-registry write failures;
concurrent-session cap is now fail-closed.
- Session callback no longer copies token.sid onto session.user.jti.
The tRPC route handler reads the JTI directly from the encrypted
JWT via getToken() so it stays server-side.
Co-Authored-By: Claude Opus 4.7 <noreply@anthropic.com>
Both auth.ts and trpc.ts now delegate the E2E_TEST_MODE-in-production
check to a single shared helper (packages/api/src/lib/runtime-security.ts).
trpc.ts used to only console.warn; it now throws at module load time,
matching the behaviour already enforced by assertSecureRuntimeEnv on the
auth side. A future refactor can no longer silently drop the guard on
either side.
Co-Authored-By: Claude Opus 4.7 <noreply@anthropic.com>
checkPromptInjection now NFKD-normalises, strips zero-width / combining
chars, and folds common Cyrillic / Greek homoglyphs before matching. 10
documented bypass examples (fullwidth, ZWJ, ZWSP, soft-hyphen, Cyrillic
е/о, combining marks, LRM, BOM) are covered by unit tests. Security
docs explicitly mark the guard as defense-in-depth — real boundary is
per-tool requirePermission.
Co-Authored-By: Claude Opus 4.7 <noreply@anthropic.com>
Prevent user-enumeration via login-response timing and audit-log content.
All failing branches now run argon2.verify against a precomputed dummy
hash (discarding the result), and emit a single "Login failed" audit
summary. Detailed reason stays in the server-only pino logger.
Co-Authored-By: Claude Opus 4.7 <noreply@anthropic.com>
`messages[].content` and `pageContext` had no `.max()` — a single chat
turn could ship 50 MB / 200 messages and OOM JSON.parse, balloon prompt
assembly, and burn arbitrary AI-provider cost. Separately, the
project-cover image-generation path concatenated user free-text into
the DALL-E / Gemini prompt without any injection check, so a manager
could pivot the image model into "ignore previous instructions" /
role-override style attacks against downstream prompt-aware infra.
- assistant-procedure-support: add `.max(10_000)` per message,
`.max(2_000)` on pageContext, and a `.superRefine` aggregate cap
(200 KB total bytes across all messages + page context). Constants
exported so call sites and tests share one source of truth.
- project-cover.generateCover: run `checkPromptInjection` over the
user-supplied `prompt` field; reject with BAD_REQUEST on match.
- 7 schema-bound tests covering per-message, page-context, aggregate,
message-count, and happy-path cases.
Covers EAPPS 3.2.7 (input bounds) / EGAI 4.6.3.2 (prompt-injection
detection on user inputs).
Co-Authored-By: Claude Opus 4.7 <noreply@anthropic.com>
batchUpdateCustomFields used $executeRaw to merge a manager-supplied
record straight into Resource.dynamicFields with no key whitelist —
so a manager could pollute the JSONB namespace with arbitrary keys
(e.g. ones admin tools later interpret). Separately, several user-facing
JSONB fields (allocation/demand metadata, dynamicFields) were typed as
unbounded z.record(z.string(), z.unknown()), letting clients ship
multi-MB payloads that flow into DB writes, audit logs, and SSE frames.
- Add BoundedJsonRecord helper (shared) — 64 keys / depth 4 /
8 KB strings / 32 KB serialized total. Conservative defaults; call
sites needing more should use a strict object schema.
- Apply BoundedJsonRecord to the highest-traffic untrusted JSONB inputs:
allocation metadata (Create/CreateDemandRequirement/CreateAssignment),
resource & project dynamicFields, and the createDemand router input.
- batchUpdateCustomFields:
* Tighten input schema (key length, value bounds, max 100 keys).
* Fetch each target resource and verify all input keys are in the
union of (specific blueprint defs) ∪ (active global RESOURCE
blueprint defs) for that resource. Empty whitelist → reject all
keys (stricter than create/update, but appropriate for a bulk
escape-hatch endpoint).
* Run the existing per-key value validator afterwards.
* 404 if any requested id does not exist (was silently skipped).
- New helper getAllowedDynamicFieldKeys() in blueprint-validation.
- 7 new BoundedJsonRecord tests, 2 new batchUpdateCustomFields tests
covering the whitelist-rejection and not-found paths.
Covers EAPPS 3.2.7 (input bounds) / OWASP A03 (injection / mass assignment).
Co-Authored-By: Claude Opus 4.7 <noreply@anthropic.com>
The read-only proxy previously wrapped model delegates to block writes,
but left client-level raw/escape hatches ($transaction, $executeRaw,
$executeRawUnsafe, $queryRawUnsafe, $runCommandRaw) intact. A read-tool
could smuggle DML via raw SQL, or open an interactive $transaction whose
tx-scoped client (unproxied by construction) accepts writes.
- read-only-prisma: block $transaction, $executeRaw, $executeRawUnsafe,
$queryRawUnsafe, $runCommandRaw at the client level. Template-tagged
$queryRaw stays allowed (read-only by API contract).
- assistant-tools: add create_estimate to MUTATION_TOOLS — it uses
$transaction internally and was previously bypassing the proxy only
because $transaction wasn't blocked.
- shared: document isReadOnly flag on ToolContext so any scoped tRPC
caller a tool spawns keeps the proxied client.
- helpers: note the runtime wrap at assistant-tools.ts:739 is
authoritative; forwarding ctx.db verbatim is correct.
- tests: cover model writes, raw escapes, and the allowed $queryRaw
path (7 cases, all pass).
- loosen one estimate-detail test that compared the exact db instance
(fails once that instance is a proxy; the assertion's intent is the
estimate id).
Covers EGAI 4.1.1.2 / IAAI 3.6.22.
Co-Authored-By: Claude Opus 4.7 <noreply@anthropic.com>
Rate-limiter now accepts string | string[] so callers can key on
multiple buckets simultaneously. If any bucket is exhausted the
request is denied, which lets login/TOTP/reset-password throttle on
BOTH user identifier and source IP without either becoming a bypass.
Fail-closed: empty/whitespace-only keys now deny by default instead
of silently allowing unbounded attempts (was CWE-307 gap).
Degraded-fallback divisor reduced from /10 to /2 — the old aggressive
clamp forced-logged-out legitimate users during brief Redis outages;
/2 still meaningfully slows distributed brute-force.
Callers updated:
- auth.ts (login): both email: and ip: buckets
- auth router requestPasswordReset: email + IP
- auth router resetPassword: IP before lookup, email-reset after
- invite router getInvite/acceptInvite: IP
- user-self-service verifyTotp: userId + IP
TRPCContext now carries clientIp; web tRPC route extracts it from
X-Forwarded-For / X-Real-IP.
Co-Authored-By: Claude Opus 4.7 <noreply@anthropic.com>
#36 CRITICAL: add .max(128) to all password Zod schemas to prevent
Argon2-based DoS from unbounded password strings.
#46 HIGH: configure pino redact paths so passwords/tokens/cookies/TOTP
secrets are never serialized in logs.
#58 MEDIUM: upgrade dompurify to ^3.4.0 and add pnpm overrides for
brace-expansion (>=5.0.5) and esbuild (>=0.25.0) to patch known CVEs.
Vite moderate (path traversal, dev-only) remains — requires vitest 3.x
major upgrade, deferred.
Co-Authored-By: Claude Opus 4.7 <noreply@anthropic.com>
The app container is attached to both `default` and `gitea_gitea` networks.
Both have a container answering to "postgres" (ours on default, Gitea's
core on gitea_gitea). Docker's embedded DNS returns IPs from all attached
networks, so the app startup script's `prisma db push` and the seed
script's `prisma.user.count()` cached different IPs and hit different
postgres instances. The seed then saw "table public.users does not exist"
even though `/api/health` reported db:ok.
Override DATABASE_URL and REDIS_URL in docker-compose.ci.yml to use the
unique compose container names (capakraken-postgres-1, capakraken-redis-1)
so resolution is unambiguous.
Co-Authored-By: Claude Opus 4.6 <noreply@anthropic.com>
next build collects page data for /api/auth/[...nextauth] and aborts
when NEXTAUTH_URL/SECRET/DATABASE_URL are unset. The CI Build job
sets these as env vars; Dockerfile.prod did not, so the prod image
build failed during Release Images even though plain build worked.
Add ARG defaults that mirror the CI placeholders. Real values are
injected at container start, so build-time placeholders are inert.
Co-Authored-By: Claude Opus 4.6 <noreply@anthropic.com>
The QNAP host kernel rejects fchmodat2 AT_EMPTY_PATH calls that newer
buildkit's runc emits, breaking docker/build-push-action@v5. The
docker-deploy-test job already builds the same Dockerfile.prod via
plain docker build (DooD) and works, so do the same here: drop the
buildx setup and use docker build + docker push directly against the
host daemon.
Co-Authored-By: Claude Opus 4.6 <noreply@anthropic.com>
The auto-provisioned GITHUB_TOKEN in Gitea Actions does not carry
package-registry write permission. Use a personal access token stored
as a repo secret instead.
GITHUB_SERVER_URL inside act_runner resolves to gitea:3000 (internal
docker hostname) which is not reachable from the build job container.
Use the externally-resolvable hostname instead.
act_runner v0.3.1 occasionally cleans the action checkout dir between
the main and post step; v4.0.4's post step then errors on the missing
.gitignore ("remove ... .gitignore: no such file") and fails the job.
Floating to v4 picks up the more defensive cleanup in v4.1+.
The release-images job failed on every run because GHCR_USERNAME and
GHCR_TOKEN are not configured on the Gitea repo — and they don't need
to be: Gitea has its own container registry at the same host, reachable
with the auto-provisioned GITHUB_TOKEN.
- Derive the registry host from GITHUB_SERVER_URL (the Gitea base URL)
- Log in with $GITHUB_TOKEN + ${{ github.actor }}
- Tag images as <gitea-host>/<owner>/<repo>-{app,migrator}:sha-<commit>
- Add packages: write permission
- Drop the workflow_call secrets block — no external secrets needed
Consumers (deploy-staging.yml, deploy-prod.yml) that previously pulled
from ghcr.io/<owner>/<repo>-app will need to be updated to pull from
the Gitea registry next; flagging separately.
Chromium on the QNAP act_runner intermittently raises ERR_CONNECTION_
REFUSED on page.goto('/') even when curl on the same pinned IP returns
307 a second earlier and the other four smoke tests (api/health,
/auth/signin, login, nav) all pass against the same container. The
smoke suite has blocked release-images on two successive docker-deploy
failures (bee5bbf, e2982a8) and a shell-level suite retry didn't help
— the Chromium refusal is reproducible per run.
Switch this one test to Playwright's HTTP request API with
maxRedirects: 0 and assert on status + Location. Semantically
equivalent (it verifies middleware wires / to /auth/signin) and
bypasses whatever Chromium-specific quirk is refusing the navigation.
Next.js dev mode on the QNAP runner intermittently drops its listening
socket for ~1-2s during route-transition compiles — smoke test #2
(page.goto('/')) has hit ERR_CONNECTION_REFUSED despite both warm-ups
and the immediately preceding health test succeeding. Playwright's
in-process retry fires while the socket is still down.
Wrap the playwright invocation in a shell-level retry: if the first
full run fails, re-warm / aggressively (up to 10 probes waiting for
307) and rerun the whole suite once.
The 'rejects worksheets that exceed the row limit' test took 6599ms on
the QNAP act_runner, overflowing the default 5000ms vitest timeout.
Writing and parsing MAX_DISPO_WORKBOOK_ROWS+1 rows via ExcelJS is slow
on constrained hardware. Extend timeout for all three writeWorkbook-
dependent tests (row limit, column limit) to 30s, matching the fix
already applied to excel.test.ts and workbook-export.test.ts.
The 'app' hostname on gitea_gitea collides with foreign containers from
other stacks that also answer /api/health. Previous logic picked the first
IP whose health check returned 200 — sometimes a neighbor whose process
died mid-test, producing ERR_CONNECTION_REFUSED on smoke test #2.
Use 'docker compose ps -q app' + docker inspect to read our own
container's gitea_gitea IP. Zero DNS ambiguity.
Same pattern as excel.test.ts and skillMatrixParser.test.ts:
ExcelJS dynamic import + writeBuffer exceeds the default 5s vitest
timeout on the QNAP CI runner.
Co-Authored-By: Claude Opus 4.6 <noreply@anthropic.com>
The initial warm-up runs ~4 minutes before the smoke tests (seed,
Node setup, Playwright install all take real time on the QNAP
runner). Between those steps, Next.js dev server can evict or
recompile routes under memory pressure — test #2 kept hitting
ERR_CONNECTION_REFUSED on / (139ms, consistently) while /auth/signin,
login, and authed nav all passed cleanly in the same run.
Re-warm both routes right before Playwright starts so the server
is guaranteed hot at the moment smoke test #2 navigates.
Co-Authored-By: Claude Opus 4.6 <noreply@anthropic.com>
Smoke test #2 kept hitting ERR_CONNECTION_REFUSED on the root path
even though curl warm-ups of the same path succeeded. Root cause is
the same split-brain bug we just fixed for e2epg: the 'app' hostname
on the shared gitea_gitea network resolves to multiple IPs (leftover
containers from concurrent runs), and curl vs Chromium picked
different ones.
Probe each resolved IP for /api/health, pin the winner as APP_BASE_URL
via GITHUB_ENV, and route health check, warm-up, and the Playwright
smoke run through that explicit IP.
Co-Authored-By: Claude Opus 4.6 <noreply@anthropic.com>
The 'e2epg' service-container hostname resolves to 3 IPs on the
shared gitea_gitea network (leftover containers from concurrent /
crashed runs). Prisma picked one IP, psql picked another — push
reported success but the verification query saw an empty schema.
Probe every resolved IP with our credentials and lock onto the one
that accepts them, then rewrite DATABASE_URL / PLAYWRIGHT_DATABASE_URL
via GITHUB_ENV so every subsequent step (prisma push, seed, E2E
webServer, Playwright fixtures) hits the same postgres instance.
Co-Authored-By: Claude Opus 4.6 <noreply@anthropic.com>
Previous warm-up used curl -L, which followed the 307 from / to a
Location target the runner could not reach (the curl output was
'307000' — root redirected, follow-up connection refused). That
meant the warm-up never exited early on a ready server, and smoke
test #2 still hit an uncompiled root occasionally.
Replace with two independent warm-ups (/ expecting 307, /auth/signin
expecting 200) that compile each route without following the
redirect.
Co-Authored-By: Claude Opus 4.6 <noreply@anthropic.com>
psql's \\dt meta-command interpreted 'public.*' as a literal pattern
on the runner's psql build, returning 'Did not find any relation
named public.*' even though prisma db push had succeeded. Replace
with a direct query against pg_tables so the verification reflects
actual schema state.
Co-Authored-By: Claude Opus 4.6 <noreply@anthropic.com>
ExcelJS dynamic import + workbook writeBuffer exceeds the default 5s
vitest timeout on the constrained QNAP CI runner, matching the same
pattern already applied to skillMatrixParser.test.ts.
Co-Authored-By: Claude Opus 4.6 <noreply@anthropic.com>
Dockerfile.dev serves via 'pnpm dev', so Next.js JIT-compiles routes on
first hit. On the QNAP runner, the cold compile of the root page +
middleware can take >10s and occasionally OOM-kills a worker, causing
test #2 (unauthenticated / → signin) to hit ERR_CONNECTION_REFUSED
while the other smoke tests (which target /auth/signin, pre-warmed via
admin-login steps) pass fine. Add an explicit curl warm-up loop so
Playwright only runs against a ready server.