Overview

The EU AI Act is a risk-based product safety regulation. It entered into force on 1 August 2024. Prohibitions take effect from February 2025, and general-purpose AI (GPAI) rules from August 2025. These dates appear in the European Commission’s overview and the Official Journal publication.

The Act allocates obligations by role (provider, deployer, importer, distributor). It empowers the EU AI Office for oversight. It also requires CE marking for high-risk AI to demonstrate conformity.

This playbook is for product, legal, data, quality, and security leaders accountable for implementation. It covers scope tests, risk classification, conformity routes and CE marking, Annex IV documentation, procurement essentials, and budgeting benchmarks. Each step includes references to authoritative sources for verification.

Who the EU AI Act applies to and extraterritorial scope

The Act applies to anyone placing AI systems or GPAI models on the EU market or putting them into service. It includes non-EU organizations when outputs are used in the EU. Territorial scope is extraterritorial: if your AI affects people in the EU, obligations can apply regardless of establishment, per the Commission’s summary.

Roles determine duties. A “provider” (often the developer or brand owner) is primarily responsible for conformity. “Deployers” (users) hold governance and transparency duties. “Importers” and “distributors” must verify documentation and labeling. Non-EU providers may appoint an authorised representative in the EU to manage regulatory correspondence and vigilance.

For example, a US speech-to-text vendor serving EU hospitals is a provider under the Act. It must comply even without an EU office. Map your roles per channel and appoint an EU representative where applicable.

Enforcement is coordinated by national market surveillance authorities and the AI Office for GPAI. Administrative fines can reach up to €35 million or 7% of global annual turnover for prohibited practices. Lower ceilings apply for SMEs and startups, and €15 million or 3% for other non-compliance, as defined in Regulation (EU) 2024/1689. Build your accountability map early. Assign provider/deployer/importer responsibilities and track which entity owns CE marking.

How to classify your system: unacceptable, high-risk, transparency obligations, or minimal risk

Classification follows four risk tiers set in the Act. Unacceptable risk systems are banned. High-risk systems require CE marking and comprehensive controls. Some systems carry transparency duties. The remainder are minimal risk with voluntary measures.

Make a first-pass scope test by use case, sector, and functionality. Then validate your result against the Act’s annexes.

Use this quick classification sequence:

Borderlines matter. An AI that scores candidates for shortlisting in recruitment is typically high-risk. A chatbot that merely schedules interviews likely falls under transparency duties. Document your classification logic and link it to Annexes II and III for auditability. Pass it through regulatory counsel review before finalizing obligations and build plans.

Sector-specific edge cases and classification pitfalls

In health, a diagnostic triage model embedded in a radiology workflow is commonly a safety component of a medical device. This pulls you into high-risk obligations via MDR. A wellbeing chatbot offering general health tips without diagnosis may trigger transparency duties if it interacts with users. This is especially true where it might be mistaken for human advice.

In finance, fraud detection that gates access to essential services (e.g., payment or credit) can be high-risk due to potential denial of service. A back-office anomaly detector used solely for analyst triage may remain minimal risk.

In HR, automated ranking or scoring of candidates or workers is high-risk. Pure scheduling assistants are typically non-high-risk but may still require user disclosure.

In education, proctoring and automated grading used for significant assessments are high-risk. A study guide generator is more likely to carry transparency duties if it could be mistaken for a human tutor.

When in doubt, analyze the impact on rights and safety. Check whether the system is decisive for access to services or a safety component of a regulated product. Record the rationale in your technical file.

GPAI models: obligations, code of practice, and oversight

GPAI providers must meet transparency, documentation, and model-level governance obligations. Additional duties apply when models pose systemic risks. Oversight sits with the Commission’s EU AI Office, supported by an AI Board and scientific panels.

Core expectations for GPAI providers include:

Use the Codes of Practice shared by the Commission and AI Office as interim guidance. Transition to harmonized standards when cited in the Official Journal. Document GPAI controls with references to recognized AI standards (e.g., ISO/IEC JTC 1/SC 42). Maintain a changelog aligned with model releases.

Timelines, compliance windows, and transition planning

Obligations are staged to allow implementation. The Act entered into force on 1 August 2024. Prohibitions apply from February 2025. GPAI rules apply from August 2025. Most high-risk obligations, CE marking, and market surveillance follow longer transition windows defined in the text and Commission guidance.

Expect further guidance and standards through 2025–2026 as the AI Office and Member States stand up sandboxes and designate notified bodies. Plan remediation by mapping obligations to product release trains.

Freeze requirements and design decisions 6–9 months before high-risk launches. This allows for pre-release testing, data governance improvements, and Annex IV drafting. For GPAI, align public model cards and training data summaries with the August 2025 date. Backfill red-teaming evidence. Monitor national regulatory sandboxes and the Commission’s AI Pact to pilot controls early with supervisory feedback.

Conformity assessment in practice: choosing routes, CE marking, and post-market monitoring

Conformity assessment demonstrates EU AI Act compliance and enables CE marking for high-risk AI. Depending on use case and harmonized standards availability, you may self-assess or require a third-party notified body (NB). You must also maintain post-market monitoring with incident reporting for the lifecycle.

Choose your route by answering three questions.

Is your AI a safety component of a regulated product (e.g., under MDR/IVDR or Machinery) that already mandates third-party assessment? If yes, you will generally go through an NB with AI scope.

Are applicable, cited harmonized standards or common specifications fully applied? If yes, provider self-assessment under internal control may be possible for certain standalone high-risk AI, subject to legal conditions.

Does your system include novel or high-uncertainty risks (e.g., biometric identification, critical infrastructure, or no standards coverage)? If yes, expect third-party assessment.

After passing the assessment, affix CE marking, register where required, and maintain vigilance plans. Plan for periodic audits and change-controlled model updates.

Post-market monitoring is mandatory. Establish key risk indicators, human oversight effectiveness checks, drift monitoring, and an incident reporting process to competent authorities. When in doubt, consult the EU NANDO database of notified bodies for designations and scopes. Align your quality management system with the risk management and documentation cadence anticipated by Annex IV.

Notified bodies landscape and readiness

Notified bodies will be designated by Member States and listed in NANDO with explicit AI Act scopes. Many will build on existing product safety designations (e.g., MDR/IVDR, Machinery). Early in the regime, expect limited capacity and lead times measured in months. Bottlenecks are likely in biometrics, medical, and critical infrastructure, where specialist competence is needed.

Select an NB with domain experience, AI competence, and availability that matches your release plan. Ask about audit scope (documentation-only versus witnessing of tests), model update policies, and experience with ML lifecycle controls. Given likely bottlenecks, initiate scoping discussions 6–12 months before your target CE marking to avoid late-stage delays.

Annex IV technical documentation and data governance requirements

Annex IV is the technical file. It is a comprehensive, auditable record of how the system meets high-risk AI requirements. It must be complete before CE marking and maintained over the lifecycle.

Treat it as a single source of truth covering design intent, data, testing, risk controls, and human oversight.

Use this Annex IV checklist as your baseline:

Anchor each section to harmonized standards as they are cited. In the interim, map controls to recognized frameworks from ISO/IEC JTC 1/SC 42. Include evidence artifacts (plots, summaries, SOPs). Make the file easy to navigate for auditors and engineers.

Human oversight, testing, and model transparency, including content provenance and deepfake disclosures

Human oversight must be effective. Operators should understand system status, intervene, and override with adequate time and context. Pre-release testing should verify accuracy and safety under stress, fairness across relevant subgroups, and resilience to misuse. Deployers must be trained and informed about limitations and intended use.

For content-generating systems, transparency and provenance are key. Deepfakes and materially manipulated content must be clearly disclosed to affected persons. Providers should implement technical measures to support labelling at scale.

Consider adopting the C2PA standard to embed tamper-evident provenance in images, audio, and video. Add UI indicators for end-users. Build your testing plan to include human-in-the-loop latency tests, red-team scripts for harmful outputs, and logging reviews. Verify that oversight mechanisms are actually used in practice.

Open-source and research exemptions: what is covered and what is not

The Act includes targeted exemptions to avoid chilling effects on open research and open-source components. Obligations reattach when systems are placed on the market or put into service for regulated uses. Publishing a model under an open-source license can be exempt from certain provider duties. Integrating that model into a high-risk application for EU users brings you into scope as the provider of the final system.

Two practical limits matter. First, transparency duties (e.g., deepfake notices) and certain GPAI obligations may still apply irrespective of licensing if the functionality reaches users. Second, downstream integrators inherit obligations when they control the final intended purpose, data, and deployment context. If you redistribute or fine-tune an open-source GPAI and market it for high-risk use (e.g., employment screening), prepare a full Annex IV. Run a conformity assessment and affix CE marking.

Interplay with sectoral regimes: GDPR/Data Act, MDR/IVDR, Machinery Regulation, NIS2, and DORA

The AI Act complements—rather than replaces—sectoral frameworks. For medical AI, MDR/IVDR continue to govern clinical safety and performance. The AI Act layers AI-specific risk management, data governance, and human oversight on top.

If your AI is a safety component in machinery, align with the Machinery Regulation and the AI Act simultaneously. This will often occur via a single conformity assessment covering both scopes.

GDPR continues to govern personal data, including lawful bases, DPIAs, and data subject rights. The Data Act adds data access and fairness provisions in B2B/B2C contexts. Operational resilience duties under NIS2 and DORA (financial services) remain and dovetail with AI Act cybersecurity and logging.

Build a harmonized control set mapping AI Act Articles and Annex IV to GDPR DPIAs, DORA ICT controls, and sectoral testing standards. Keep legal references anchored to Regulation (EU) 2024/1689 for traceability.

Roles and responsibilities for non-EU providers, importers, distributors, and authorised representatives

Accountability follows the economic operator chain. Providers (often developers or branded owners) own risk management, Annex IV, conformity assessment, CE marking, and post-market monitoring. Deployers must use systems as intended, ensure human oversight and logging, and perform context-specific impact assessments when required.

Importers ensure that non-EU products meet EU requirements before market entry and that documentation and CE marking are correct. Distributors must not place non-compliant systems on the market and should stop distribution and inform authorities if risks emerge.

Non-EU providers typically appoint an authorised representative in the EU to field authorities’ requests and hold technical documentation. Clarify these duties in contracts. Specify who maintains Annex IV, who handles incidents, and who pays for surveillance audits. Mirror these in supplier quality agreements.

Public procurement and buyer obligations: model clauses, due diligence, and supplier attestations

Public buyers and large private buyers should bake AI Act requirements into procurement to avoid downstream non-conformities. Treat vendors of high-risk AI as critical suppliers subject to documentation, testing, and audit rights. Align RFPs with risk classification and transparency obligations.

Add these clauses and checks to your procurement pack:

During due diligence, request sample outputs and bias/robustness metrics. Verify notified body involvement where required. Check the vendor’s scope via the NANDO database. Maintain a vendor scorecard tied to your internal risk taxonomy and update it with each release.

Budgeting and staffing for compliance: benchmarks by risk class and company size

Budget scales with risk, novelty, and sector. High-risk systems typically require six-figure investments in year one to stand up risk management, testing infrastructure, documentation, and audits. GPAI providers add red-teaming and provenance pipelines.

SMEs can reduce costs via standards reuse, automation, and sandboxes. They also benefit from simplified fines ceilings compared to large undertakings under the AI Act.

Expect cost drivers in these categories: standards adoption. Data curation and documentation. Testing for fairness, robustness, and red-teaming. Notified body fees where applicable. CE marking and registration. Post-market monitoring and incident response. Procurement and legal work for role allocation.

A pragmatic staffing pattern for a mid-sized high-risk program includes an AI compliance lead, ML safety/reliability engineer, data governance lead, QA/test engineer, security engineer, and legal/regulatory counsel. Add SMEs from product and UX. ROI improves when you reuse controls across products, adopt harmonized standards early, and automate documentation and test reporting so Annex IV stays “always current.”

National implementation tracker and authoritative sources

Keep your plan anchored to primary sources and evolving guidance. Track EU-level updates, standards work, and notified body designations. Monitor your national competent authority for sandbox and enforcement notices.

Start with these authoritative links:

With these sources and the steps in this playbook, your team can classify systems accurately, select the right conformity route, and produce Annex IV documentation that stands up to scrutiny. You can align procurement to vendor duties and operationalize EU AI Act compliance on time.