Annex A Overview

ISO 42001 Annex A provides a reference set of control objectives organized into domains. Unlike prescriptive controls, these are objectives - the organization determines how to achieve them based on their context and risk assessment.

Using Annex A

During risk treatment (Clause 6.1.3), compare your selected controls against Annex A to ensure nothing relevant has been overlooked. Document your Statement of Applicability (SoA) explaining why each control is included or excluded.

Annex A Structure

Annex A is organized into the following domains:

  • A.2 - AI Policies
  • A.3 - Internal Organization
  • A.4 - Resources for AI Systems
  • A.5 - AI System Lifecycle
  • A.6 - Data for AI Systems
  • A.7 - AI System Information
  • A.8 - Use of AI Systems
  • A.9 - Third-Party Relationships

A.2 AI Policies

This domain ensures organizational policies address AI-specific governance requirements.

A.2.2 AI Policy

Objective: Establish management direction and commitment for responsible AI through documented policies.

Implementation Example:

  • Documented AI policy approved by executive leadership
  • Policy addresses fairness, transparency, human oversight, and safety
  • Commitment to compliance with applicable AI regulations
  • Policy communicated to all relevant personnel
  • Annual policy review and update process

A.2.3 Review of AI Policies

Objective: Ensure AI policies remain suitable and effective over time.

Implementation Example:

  • Scheduled annual policy review
  • Triggered reviews when significant changes occur (new regulations, major incidents)
  • Review considers effectiveness, stakeholder feedback, and emerging best practices

A.3 Internal Organization

Controls ensuring appropriate organizational structure and accountability for AI governance.

A.3.2 Roles and Responsibilities

Objective: Clearly define and assign responsibilities for AI governance.

Implementation Example:

  • AI Governance Committee with executive representation
  • Designated AI Ethics Officer or equivalent role
  • RACI matrix for AI development, deployment, and monitoring
  • Clear escalation paths for AI-related decisions

A.3.3 Reporting

Objective: Establish reporting mechanisms for AI governance matters.

Implementation Example:

  • Monthly AI governance dashboard to leadership
  • Quarterly reports on AI risk status and incidents
  • Channels for reporting AI ethics concerns

A.4 Resources for AI Systems

Controls addressing human, technical, and financial resources for AI governance.

A.4.2 Human Resources

Objective: Ensure personnel have necessary competencies for their AI-related roles.

Implementation Example:

  • Competency requirements defined for AI roles
  • Mandatory AI ethics training for developers and operators
  • Specialized training on bias detection and mitigation
  • Background checks for personnel with access to sensitive AI systems

A.4.4 Computing Resources

Objective: Provide adequate computing infrastructure for AI systems.

Implementation Example:

  • Dedicated AI development and production environments
  • Scalable infrastructure for model training and inference
  • Security controls for AI infrastructure
  • Capacity planning for AI workloads

A.5 AI System Lifecycle

Controls spanning the entire AI system lifecycle from design through retirement.

A.5.2 AI System Design

Objective: Incorporate responsible AI principles into system design.

Implementation Example:

  • Design review checklist including fairness, explainability, and safety
  • Architecture documentation addressing AI-specific concerns
  • Human oversight mechanisms designed into systems

A.5.4 AI System Development

Objective: Implement responsible development practices.

Implementation Example:

  • Code review requirements for AI components
  • Version control for models and training code
  • Testing procedures including bias and fairness testing
  • Model cards documenting intended use and limitations

A.5.5 AI System Verification and Validation

Objective: Verify AI systems meet requirements and perform as intended.

Implementation Example:

  • Test plans covering functional and non-functional requirements
  • Validation against fairness metrics and thresholds
  • User acceptance testing for AI-driven features
  • Performance testing under realistic conditions

A.5.6 AI System Deployment

Objective: Control the deployment of AI systems into production.

Implementation Example:

  • Deployment approval process with governance sign-off
  • Staged rollout (canary, blue-green) for high-risk systems
  • Rollback procedures documented and tested
  • Post-deployment monitoring activation

A.5.7 AI System Operation and Monitoring

Objective: Maintain oversight of AI systems in production.

Implementation Example:

  • Real-time performance monitoring dashboards
  • Model drift detection and alerting
  • Fairness metric monitoring over time
  • Incident response procedures for AI issues

A.5.8 AI System Documentation

Objective: Maintain comprehensive documentation of AI systems.

Implementation Example:

  • Model cards for each deployed model
  • System architecture and data flow documentation
  • Training data documentation and lineage
  • User guides and operational runbooks

A.5.9 AI System Retirement

Objective: Properly decommission AI systems when no longer needed.

Implementation Example:

  • Retirement approval process
  • Data retention and disposal procedures
  • Notification to dependent systems and users
  • Archive of documentation for regulatory purposes

A.6 Data for AI Systems

Controls addressing data quality, management, and governance for AI systems.

A.6.2 Data Quality

Objective: Ensure data used for AI systems meets quality requirements.

Implementation Example:

  • Data quality metrics defined (completeness, accuracy, timeliness)
  • Data profiling and quality assessment before use
  • Data cleaning and preprocessing standards
  • Quality monitoring for ongoing data pipelines

A.6.3 Data Provenance

Objective: Maintain records of data origin and transformations.

Implementation Example:

  • Data lineage tracking from source to model
  • Documentation of data sources and acquisition methods
  • Version control for training datasets
  • Audit trail of data transformations

A.6.4 Data Preparation

Objective: Properly prepare data for AI system use.

Implementation Example:

  • Documented data preparation procedures
  • Bias assessment of training data
  • Data augmentation with documented rationale
  • Train/test/validation split procedures

A.7 AI System Information

Controls ensuring transparency and information availability about AI systems.

A.7.2 Recording Information about AI Systems

Objective: Document relevant information about AI systems.

Implementation Example:

  • AI system registry with key metadata
  • Purpose and intended use documentation
  • Known limitations and constraints recorded
  • Performance characteristics documented

A.7.3 Providing Information to Interested Parties

Objective: Communicate appropriate information to stakeholders.

Implementation Example:

  • User-facing disclosures when AI is involved in decisions
  • Transparency reports for high-impact systems
  • Information available for regulatory inquiries
  • Customer-facing documentation about AI capabilities

A.8 Use of AI Systems

Controls governing the responsible use and operation of AI systems.

A.8.2 Intended Use

Objective: Ensure AI systems are used for intended purposes.

Implementation Example:

  • Documented intended use statements for each AI system
  • Prohibited use cases explicitly defined
  • User agreements acknowledging appropriate use
  • Monitoring for misuse or scope creep

A.8.3 Processes Regarding AI System Output

Objective: Appropriately handle and use AI system outputs.

Implementation Example:

  • Guidance on interpreting AI outputs
  • Human review requirements for high-stakes decisions
  • Procedures for challenging or overriding AI recommendations
  • Documentation of how AI outputs inform decisions

A.9 Third-Party Relationships

Controls for managing external parties in the AI ecosystem.

A.9.2 Monitoring and Review of Third-Party Services

Objective: Maintain oversight of third-party AI services.

Implementation Example:

  • Inventory of third-party AI services in use
  • Due diligence assessments before procurement
  • Contractual requirements for responsible AI practices
  • Periodic review of third-party performance and compliance
  • Incident notification requirements in contracts

A.9.4 AI System Component Sourcing

Objective: Govern sourcing of AI components (models, data, tools).

Implementation Example:

  • Assessment of pre-trained models before adoption
  • Licensing and intellectual property review
  • Security assessment of AI tools and frameworks
  • Vendor risk assessment for AI suppliers

Annex A controls are not one-size-fits-all. Your Statement of Applicability should reflect your organization's specific AI systems, risks, and context. A control that is critical for one organization may be non-applicable for another.