
The EU AI Act isn’t a PDF you “comply with” — it’s a set of control objectives you design into your product: evaluation, documentation, monitoring, and provable safety boundaries.
Axel Domingues
Regulation is usually framed as paperwork.
But if you’ve shipped real systems, you know the truth:
regulation is architecture — with deadlines and penalties.
The EU AI Act is the first time many teams will be forced to operationalize that idea for AI:
This article is about turning that into something an engineering org can actually execute.
Not with vibes.
With controls and evidence.
If you’re shipping into regulated domains, involve legal/compliance early — but don’t outsource architecture to them.
The goal
Turn “EU AI Act compliance” into an implementable control map + evidence pipeline.
The mental model
Treat the Act like a set of control objectives — not a checklist.
The output
A living system: model registry, eval harness, audit logs, incident response.
The trap
If you add compliance after shipping, you’ll rebuild everything under pressure.
Most teams get stuck because they start in the wrong place: “What does the Act say?”
Start here instead:
What risk category are we in — and what obligations follow?

At a high level, the mental buckets look like:
A regulation with phased application dates creates a new kind of technical debt:
time-bound debt.
If you treat this like “we’ll fix later”, “later” arrives on a calendar.
Engineering takeaway
Your compliance program must be staged: inventory first, then controls, then evidence automation.
Product takeaway
If AI is on your critical path, “legal deadlines” become architecture milestones.
Here’s the move that makes this tractable:
A regulation is a set of control objectives.
Compliance is the ability to show evidence that your controls are operating.
That’s it.
So the question is not:
“Are we compliant?”
It’s:
This is exactly the mindset shift we learned in 2022 (operational architecture):
AI compliance is the same category of work.
Most organizations fail compliance because they can’t answer basic questions:
So the first deliverable is boring — and essential:
an AI system registry.
Decide what you include in scope:
For each AI feature, record:
Even if you’re “just integrating” a model, you still own:
So assign internal owners:
Once you have an inventory, the next move is to translate “requirements language” into engineering controls.
A practical translation for high-impact AI systems looks like this:
Control objective: you identify harms, mitigate them, and verify mitigations.
Engineering controls:
Evidence:
Control objective: your training/finetuning/evaluation data is lawful, relevant, and controlled.
Engineering controls:
Evidence:
Control objective: users understand when they’re interacting with AI and what limitations exist.
Engineering controls:
Evidence:
Control objective: humans can supervise, intervene, and override when required.
Engineering controls:
Evidence:
Control objective: the system is resilient, secure, and monitored for drift and failures.
Engineering controls:
Evidence:
Here’s the engineering secret:
If your AI product is operable, compliance evidence becomes cheap.
If your AI product is not operable, compliance evidence becomes impossible.
So you want an architecture where:
Think of this as “CI/CD for trust”.

A model registry
A single source of truth for model versions, prompts, policies, and tool permissions.
An eval harness
Repeatable tests that run on every change — with stored results.
Observability + audit logs
Telemetry for safety, drift, cost, and misuse — plus immutable audit trails.
An audit packet generator
One click: “show me the evidence” for this system, this version, this date range.
You don’t need a mega-platform.
You need a few boring, durable primitives.
Store:
Design note: treat prompts as code. PRs. Reviews. Rollback.
A service that can:
Design note: evaluations are not just “accuracy”. They are controls verification.
At inference time:
Design note: guardrails must be measurable. If you can’t measure them, you can’t prove they operate.
You want:
And you want playbooks:
A prompt is not a control.
Controls are things that fail safely when the model misbehaves.A prompt is a policy suggestion.
If you’re leading an engineering org, you need a plan that decomposes.
Here’s a minimal backlog that actually ships.
EU AI Act — official text (Regulation (EU) 2024/1689)
The source of truth: risk categories, obligations, timelines, and definitions — start here before translating requirements into controls.
NIST AI RMF 1.0 (risk → controls → monitoring)
A practical engineering-friendly framework for turning “risk” into control objectives, metrics, and governance you can actually operate.
No.
Even if your system falls under lighter obligations, the Act will push:
Building the control/evidence spine early makes future classification changes survivable.
Treating compliance as a document project.
If you don’t have:
…then you don’t have compliance — you have a folder of hope.
Evidence is anything that can answer: “prove your controls were active.”
Concrete examples:
You can buy parts of the stack (logging, model gateways, risk management tooling).
But your architecture still needs:
Tools help. They do not replace design.
This month was about making regulation operational:
Next month we go deeper into the runtime itself:
Reasoning Budgets: fast/slow paths, verification, and when to “think longer”.
Because once you’re operating within compliance constraints, the next question becomes:
How do you spend “thinking time” like a budget — and prove the system used it wisely?
Reasoning Budgets: fast/slow paths, verification, and when to “think longer”
“Think step by step” isn’t an architecture. In production you need budgets, routing, and verifiers so the system knows when to go fast, when to slow down, and when to refuse.
Tool Use with Open Models: function calling, sandboxes, and “capability boundaries”
Tool use is where LLMs stop being “text generators” and start being integration surfaces. With open weights, reliability isn’t a given — you have to engineer it with contracts, sandboxes, and explicit capability boundaries.