EU AI Act compliance
Art. 4 AI Act competence duty: what you need to know now.
In force since February 2025. Applies to almost every business in the EU. Fines up to €15m for non-compliance. Our step-by-step explainer.
What does Art. 4 of the AI Act actually say?
'Providers and deployers of AI systems shall take measures to ensure, to the best of their extent, a sufficient level of AI literacy of their staff and other persons dealing with the operation and use of AI systems on their behalf …'
Art. 4 EU AI Act
In plain language: if anyone in your company uses an AI tool — ChatGPT, Copilot, Claude, or internal models — you must be able to prove that this person is 'sufficiently competent' with that AI.
'Sufficiently competent' is deliberately open-ended. In practice that means: documented training, clear usage guidelines, demonstrable risk awareness.
Who is affected by Art. 4?
Definitely affected
- Any business providing AI tools (Copilot licenses, ChatGPT Enterprise, internal models)
- Any business whose employees use AI at work — even without an official license
- All industries, including non-tech
In the grey zone
- One-off freelance engagements
- Purely private AI use with no work connection
- AI models for research purposes only
Realistic assessment: 95% of German mid-market firms are affected — and don't know it.
What happens on non-compliance.
15M €
or 3% of global annual revenue
The AI Act uses tiered fines. The maximum rates apply to prohibited AI practices (Art. 5). Violations of Art. 4 (competence duty) carry lower but still material penalties: up to €7.5m or 1.5% of annual revenue.
Important: German regulators will begin active enforcement mid-2026. First reference cases are expected in Q3/Q4 2026.
How you get compliant — in 90 days.
Step 1 — Inventory
Which AI tools are currently in use? (Usually more than you think.) We provide an AI-inventory template.
Step 2 — Risk categorisation
AI systems classified by risk tier (prohibited, high-risk, transparency, limited). Checklist in the download.
Step 3 — Measure competence baseline
Who on your team knows what? Who needs what? Our assessment tool produces individual scores.
Step 4 — Structured training
Not 'a mandatory e-learning module', but role-based hands-on training on real tickets.
Step 5 — Documentation
Everything auditable: training records, AI policy, usage guidelines, regular refreshers.
Our Agentic University transformation covers all five steps — and delivers AI Act compliance as a side-effect of the productivity transformation.
The 5 most common misconceptions about Art. 4.
Art. 4 only applies to high-risk AI.
False. Art. 4 applies to ANY AI usage in a business context — regardless of whether the system is classified as high-risk.
#ai-act-faq-1We don't build AI ourselves — this doesn't apply to us.
False. Once your employees use ChatGPT, Copilot or Claude, you are a deployer under the Act — even without your own models.
#ai-act-faq-2A brief e-learning module is enough.
Dangerous. 'Sufficiently competent' means role- and risk-appropriate training, not generic video-clicking. Regulators will examine evidence in depth.
#ai-act-faq-3We have until 2027.
False. Art. 4 has been in full force since February 2025. German regulators will begin active enforcement mid-2026, with first reference cases expected in Q3/Q4 2026.
#ai-act-faq-4The effort is manageable.
For large enterprises, yes. For mid-market firms without a structured L&D function: significant. Structured inventory, categorisation and training will tie up internal staff over several months.
#ai-act-faq-5Resources.
The EU AI Act compliance checklist with 12 questions — fill it in once, and you know exactly where you stand.
EU AI Act compliance checklist (PDF)
12 questions in 5 minutes.
- Structured self-assessment — auditable.
- Delivered as a PDF by email — print-ready.
- Sent once, no spam sequence.
You have no time to lose.
Regulators will begin active enforcement mid-2026. We get you compliant in 90 days — and build measurable productivity gains along the way.