In This Guide
EU AI Act Overview
The EU AI Act (Regulation 2024/1689) is the world's first comprehensive AI legislation. It establishes harmonized rules for AI systems placed on the EU market or affecting people in the EU. The Act uses a risk-based approach, with stricter requirements for higher-risk AI systems.
August 2024: AI Act entered into force
February 2025: Prohibited AI practices banned
August 2025: GPAI model obligations apply
August 2026: High-risk AI requirements fully applicable
August 2027: Certain Annex I high-risk systems deadline
Who Does It Apply To?
- Providers: Organizations developing AI systems or placing them on the EU market
- Deployers: Organizations using AI systems (previously called "users")
- Importers: Entities bringing non-EU AI systems into the EU market
- Distributors: Entities making AI systems available on the market
The Act applies regardless of where the provider is established if the AI system is placed on the EU market or its output is used in the EU.
Compliance Timeline
| Date | What Applies | Action Required |
|---|---|---|
| Feb 2025 | Prohibited AI practices | Stop prohibited AI uses (social scoring, emotion recognition in workplace/education, etc.) |
| Aug 2025 | General-purpose AI models | GPAI providers must comply with transparency and documentation requirements |
| Aug 2025 | AI literacy (Article 4) | Ensure staff have sufficient AI competence |
| Aug 2026 | High-risk AI systems | Full compliance with Chapter 2 requirements |
| Aug 2027 | Annex I high-risk systems | Existing high-risk systems in Annex I areas must comply |
AI Act Risk Categories
Prohibited AI Practices (Unacceptable Risk)
The following AI practices are banned outright:
- Social scoring by governments
- Exploitation of vulnerable groups
- Subliminal manipulation causing harm
- Real-time remote biometric identification in public (with exceptions)
- Emotion recognition in workplace and educational settings
- Biometric categorization based on sensitive attributes
- Facial recognition databases from untargeted scraping
High-Risk AI Systems
AI systems in specific areas face strict requirements:
- Annex I: AI systems that are safety components of products covered by EU harmonization legislation (medical devices, machinery, vehicles, etc.)
- Annex III: Standalone high-risk AI in areas including:
- Biometric identification and categorization
- Critical infrastructure management
- Education and vocational training
- Employment, worker management, self-employment access
- Access to essential services (credit, insurance, public assistance)
- Law enforcement
- Migration, asylum, border control
- Administration of justice and democratic processes
Limited Risk
AI systems with transparency obligations:
- Chatbots and conversational AI (must disclose AI interaction)
- Emotion recognition systems (must inform subjects)
- Deep fakes and generated content (must label as AI-generated)
Minimal Risk
All other AI systems—no mandatory requirements but voluntary codes of conduct encouraged.
High-Risk AI System Requirements
High-risk AI systems must meet requirements in Chapter 2 of the AI Act:
Article 9: Risk Management System
- Establish and maintain a risk management system throughout the lifecycle
- Identify and analyze known and foreseeable risks
- Estimate and evaluate risks
- Adopt risk management measures
- Test to identify most appropriate measures
Article 10: Data and Data Governance
- Training, validation, and testing data subject to governance practices
- Data quality criteria (relevance, representativeness, error-free, completeness)
- Examination of possible biases
- Appropriate data preparation procedures
Article 11: Technical Documentation
- Comprehensive technical documentation before market placement
- Documentation kept up-to-date
- Content per Annex IV requirements
Article 12: Record-Keeping
- Automatic logging capabilities
- Logs enabling tracing of AI system operation
- Logs retained for appropriate period
Article 13: Transparency and Information
- Designed for transparency enabling deployer interpretation
- Instructions for use provided
- Information on capabilities, limitations, and risks
Article 14: Human Oversight
- Designed to enable effective human oversight
- Human ability to understand system capabilities and limitations
- Ability to interpret outputs and make decisions
- Override or intervention capability
Article 15: Accuracy, Robustness, Cybersecurity
- Appropriate levels of accuracy, robustness, cybersecurity
- Resilient against attempts to exploit vulnerabilities
- Technical redundancy solutions where appropriate
ISO 42001 to EU AI Act Mapping
ISO 42001 provides substantial coverage of EU AI Act requirements. Here is a practical mapping for high-risk AI systems:
| EU AI Act Article | ISO 42001 Coverage | Alignment Level |
|---|---|---|
| Art. 9 Risk Management | Clause 6.1.2 AI Risk Assessment, Clause 8.2 | Strong |
| Art. 10 Data Governance | Annex A.6 Data for AI Systems (quality, provenance, preparation) | Strong |
| Art. 11 Technical Documentation | Annex A.5.8 AI System Documentation, A.7 | Strong |
| Art. 12 Record-Keeping | Clause 7.5 Documented Information, A.5.7 Monitoring | Moderate |
| Art. 13 Transparency | Annex A.7 AI System Information, A.7.3 | Strong |
| Art. 14 Human Oversight | Annex A.8 Use of AI Systems, A.8.3 | Moderate |
| Art. 15 Accuracy/Robustness | A.5.5 Verification & Validation, A.5.7 Monitoring | Moderate |
| Art. 4 AI Literacy | Clause 7.2 Competence, 7.3 Awareness | Strong |
Where ISO 42001 Helps - And Where It Doesn't
Strong Alignment Areas
- Risk Management: ISO 42001's AI risk assessment aligns well with Article 9
- Data Quality: Annex A.6 addresses data governance requirements
- Documentation: ISO 42001 requires comprehensive documentation throughout
- Competence: Clause 7.2 addresses AI literacy requirements
- Third-Party Management: Annex A.9 covers supply chain requirements
Gaps Requiring Additional Work
- Conformity Assessment: AI Act requires specific conformity assessment procedures not covered by ISO 42001
- CE Marking: High-risk AI systems need CE marking - a regulatory process beyond ISO certification
- EU Database Registration: High-risk systems must be registered in EU database
- Post-Market Monitoring: AI Act has specific post-market surveillance requirements
- Serious Incident Reporting: Mandatory reporting to authorities within specific timeframes
- Instructions for Use: AI Act specifies detailed content requirements
ISO 42001 certification does not automatically mean EU AI Act compliance. However, it provides a strong foundation that covers many requirements and demonstrates organizational commitment to responsible AI. Additional work is needed for full regulatory compliance.
Action Plan for EU AI Act Preparation
Step 1: AI System Classification (Now)
- Inventory all AI systems
- Classify each system per AI Act risk categories
- Identify prohibited practices (address immediately)
- Flag high-risk systems for priority attention
Step 2: Implement ISO 42001 AIMS
- Establish AI governance framework
- Implement risk assessment and treatment processes
- Deploy Annex A controls relevant to your AI systems
- Build documentation and record-keeping capabilities
Step 3: Address AI Act Gaps
- Map ISO 42001 coverage to specific AI Act articles
- Identify gaps requiring additional controls
- Develop conformity assessment approach
- Prepare for EU database registration
- Establish incident reporting procedures
Step 4: Prepare for Deadlines
- Prohibited practices: Ensure stopped by February 2025
- AI literacy: Ensure staff competence by August 2025
- High-risk compliance: Full compliance by August 2026
Think of ISO 42001 as building the governance foundation and operational muscle for AI management. The EU AI Act adds specific regulatory requirements on top. Starting with ISO 42001 makes AI Act compliance significantly more achievable.