Claude-Skills iso42001-ai-management

install
source · Clone the upstream repo
git clone https://github.com/borghei/Claude-Skills
Claude Code · Install into ~/.claude/skills/
T=$(mktemp -d) && git clone --depth=1 https://github.com/borghei/Claude-Skills "$T" && mkdir -p ~/.claude/skills && cp -r "$T/ra-qm-team/iso42001-ai-management" ~/.claude/skills/borghei-claude-skills-iso42001-ai-management && rm -rf "$T"
manifest: ra-qm-team/iso42001-ai-management/SKILL.md
source content

ISO 42001 AI Management System

Tools and guidance for ISO/IEC 42001:2023 — the first international standard for AI Management Systems (AIMS).


Table of Contents


Tools

AIMS Readiness Checker

Assesses organizational readiness against all ISO 42001 clauses and Annex A controls. Scores each clause on a 0-100 scale and identifies gaps for certification preparation.

# Assess readiness from a JSON profile
python scripts/aims_readiness_checker.py --input org_profile.json

# Generate a blank input template
python scripts/aims_readiness_checker.py --template > org_profile.json

# JSON output for automation
python scripts/aims_readiness_checker.py --input org_profile.json --json

# Export report to file
python scripts/aims_readiness_checker.py --input org_profile.json --output report.json

Assessment Areas:

ClauseAreaKey Checks
Clause 4ContextScope defined, interested parties, AIMS boundaries
Clause 5LeadershipAI policy, governance structure, management commitment
Clause 6PlanningRisk assessment methodology, AI objectives, impact assessments
Clause 7SupportResources, competence, awareness, documentation
Clause 8OperationAI lifecycle, data management, risk treatment, third-party controls
Clause 9PerformanceMonitoring, internal audit, management review
Clause 10ImprovementCorrective actions, continual improvement, incident management
Annex AControlsA.2-A.10 control implementation status

Output:

  • Overall readiness score (0-100)
  • Per-clause scores with maturity level (Initial/Developing/Defined/Managed/Optimized)
  • Annex A control implementation status (Implemented/Partial/Not Implemented/Not Applicable)
  • Gap analysis with prioritized recommendations
  • Certification readiness assessment (Ready/Near Ready/Significant Gaps)

AI Impact Assessor

Generates comprehensive AI impact assessments evaluating fairness, transparency, safety, privacy, and security dimensions. Maps impacts to interested parties and provides risk treatment recommendations.

# Assess an AI system from a JSON description
python scripts/ai_impact_assessor.py --input ai_system.json

# Generate a blank input template
python scripts/ai_impact_assessor.py --template > ai_system.json

# Export assessment report
python scripts/ai_impact_assessor.py --input ai_system.json --output assessment.json

# Generate markdown report
python scripts/ai_impact_assessor.py --input ai_system.json --format markdown --output assessment.md

Assessment Dimensions:

DimensionEvaluatesKey Factors
FairnessBias, discrimination, equityTraining data diversity, protected attributes, outcome parity
TransparencyExplainability, interpretabilityModel complexity, decision documentation, user disclosure
SafetyReliability, robustness, harm preventionFailure modes, edge cases, human oversight, fallback mechanisms
PrivacyData protection, consent, minimizationPI processing, consent mechanisms, data retention, anonymization
SecurityAdversarial resilience, access controlAttack vectors, model integrity, access management, audit logging
AccountabilityGovernance, responsibility, auditabilityDecision ownership, audit trails, escalation procedures

Features:

  • Risk scoring per dimension (Low/Medium/High/Critical)
  • Interested party impact mapping (users, affected individuals, society, regulators)
  • Risk treatment options (Avoid, Mitigate, Transfer, Accept)
  • Regulatory mapping (EU AI Act risk tier, ISO 42001 Annex A controls)
  • Residual risk calculation after treatment
  • Markdown and JSON report generation

Reference Guides

ISO 42001 Clause Guide

references/iso42001-clause-guide.md

Comprehensive clause-by-clause guidance:

  • All clauses (4-10) with requirements and implementation steps
  • Annex A controls (A.2-A.10) detailed with evidence requirements
  • Audit questions per clause for internal audit preparation
  • Common nonconformity findings and how to avoid them
  • Required documented information per clause
  • Cross-references to ISO 27001, ISO 9001, and EU AI Act

AI Lifecycle Management

references/ai-lifecycle-management.md

End-to-end AI system lifecycle guidance:

  • Lifecycle stages: design, development, testing, deployment, monitoring, retirement
  • Design and development controls (requirements, architecture, coding standards)
  • Testing and validation requirements (functional, bias, robustness, performance)
  • Deployment procedures (staging, canary, rollback, approval gates)
  • Monitoring and maintenance (drift detection, performance degradation, retraining)
  • Retirement and decommissioning (data disposal, model archival, stakeholder notification)
  • Data management across lifecycle (quality, provenance, bias assessment, lineage)
  • Model versioning and change management (version control, change impact, approval workflows)

Workflows

Workflow 1: ISO 42001 Readiness Assessment

Step 1: Define AIMS scope
        → Identify AI systems in scope
        → Determine organizational boundaries
        → Document interested parties and requirements

Step 2: Generate assessment template
        → python scripts/aims_readiness_checker.py --template > org_profile.json
        → Fill in organizational details and current state

Step 3: Run readiness assessment
        → python scripts/aims_readiness_checker.py --input org_profile.json

Step 4: Review results
        → Address critical gaps (Clauses 5, 6, 8 typically weakest)
        → Prioritize Annex A controls by risk
        → Develop remediation roadmap

Step 5: Conduct AI impact assessments
        → python scripts/ai_impact_assessor.py --template > ai_system.json
        → Assess each in-scope AI system
        → python scripts/ai_impact_assessor.py --input ai_system.json

Step 6: Plan implementation
        → See references/iso42001-clause-guide.md for requirements
        → See references/ai-lifecycle-management.md for operational controls

Workflow 2: AI System Impact Assessment

Step 1: Identify AI system for assessment
        → Document system purpose, inputs, outputs, and decisions
        → Identify affected individuals and groups

Step 2: Generate assessment template
        → python scripts/ai_impact_assessor.py --template > ai_system.json
        → Complete all sections (model details, data sources, deployment context)

Step 3: Conduct assessment
        → python scripts/ai_impact_assessor.py --input ai_system.json --format markdown --output report.md

Step 4: Review dimension scores
        → Fairness: check for bias in training data and outcomes
        → Transparency: verify explainability mechanisms
        → Safety: validate failure modes and human oversight
        → Privacy: confirm data protection measures
        → Security: assess adversarial resilience

Step 5: Implement risk treatments
        → Apply recommended mitigations per dimension
        → Document residual risk acceptance decisions
        → Assign treatment owners and timelines

Step 6: Monitor and review
        → Schedule periodic reassessment (quarterly minimum)
        → Track treatment implementation progress
        → Update assessment when system changes materially

Workflow 3: AIMS Certification Preparation

Step 1: Gap analysis
        → python scripts/aims_readiness_checker.py --input org_profile.json
        → Target overall score of 80+ for certification readiness

Step 2: Document AIMS
        → AI policy (Clause 5.2)
        → AIMS scope (Clause 4.3)
        → Risk assessment methodology (Clause 6.1)
        → Statement of Applicability for Annex A controls
        → AI objectives (Clause 6.2)

Step 3: Implement operational controls
        → AI lifecycle procedures (Clause 8)
        → Data management processes (Annex A.7)
        → Third-party management (Annex A.10)
        → Impact assessments for all AI systems (Annex A.5)

Step 4: Conduct internal audit
        → Use references/iso42001-clause-guide.md audit questions
        → Document findings and corrective actions
        → Verify closure of nonconformities

Step 5: Management review
        → Present AIMS performance to top management
        → Review AI objectives achievement
        → Obtain commitment for continual improvement

Step 6: Stage 1 and Stage 2 audits
        → Stage 1: Documentation review (readiness check)
        → Stage 2: Implementation effectiveness audit
        → Address any nonconformities from audit

Standard Overview

ISO 42001:2023 Overview

ISO/IEC 42001:2023 is the world's first international standard for AI Management Systems (AIMS). Published in December 2023, it provides a framework for organizations to responsibly develop, provide, and use AI systems. The standard follows the ISO Harmonized Structure (Annex SL) for management system standards, enabling integration with ISO 27001, ISO 9001, and ISO 14001.

Key Characteristics:

  • Certifiable management system standard
  • Technology-neutral (applies to any AI approach)
  • Risk-based approach to AI governance
  • PDCA (Plan-Do-Check-Act) cycle
  • Applicable to organizations of any size and sector

AIMS Framework (Plan-Do-Check-Act)

Context of the Organization (Clause 4)

RequirementSectionDescription
Organization context4.1Internal/external issues relevant to AI objectives
Interested parties4.2Stakeholders, their requirements, and expectations
AIMS scope4.3Boundaries and applicability of the AIMS
AIMS establishment4.4Establish, implement, maintain, and improve the AIMS

Leadership (Clause 5)

RequirementSectionDescription
Leadership commitment5.1Top management demonstrates commitment to AIMS
AI policy5.2Responsible AI principles, ethical guidelines, organizational values
Roles and responsibilities5.3Clear assignment of AIMS roles, authority, and accountability

AI Policy Must Include:

  • Commitment to responsible AI development and use
  • Ethical principles guiding AI decisions
  • Alignment with applicable legal and regulatory requirements
  • Commitment to continual improvement of the AIMS
  • Framework for setting AI objectives

AI Governance Structure:

  • AI governance board or committee
  • AI system owners with defined accountability
  • Data stewards for AI data management
  • Ethics review function
  • Incident response roles

Planning (Clause 6)

RequirementSectionDescription
Risks and opportunities6.1Actions to address AI-specific risks and opportunities
AI risk assessment6.1.2Methodology for identifying and evaluating AI risks
AI objectives6.2Measurable objectives for responsible AI
Impact assessment6.1.4Assessment of AI system impacts on individuals and society

AI Risk Assessment Must Cover:

  • Fairness and non-discrimination risks
  • Transparency and explainability gaps
  • Safety and reliability concerns
  • Privacy and data protection risks
  • Security vulnerabilities
  • Accountability gaps
  • Societal and environmental impacts

Support (Clause 7)

RequirementSectionDescription
Resources7.1Compute, data, expertise, and infrastructure
Competence7.2Required skills for AI roles, training plans
Awareness7.3AI literacy across the organization
Communication7.4Internal/external communication on AI matters
Documented information7.5Document creation, control, and retention

Operation (Clause 8)

RequirementSectionDescription
Operational planning8.1Planning and controlling AI processes
AI risk assessment8.2Executing risk assessments per methodology
AI risk treatment8.3Implementing risk treatment plans
AI system lifecycle8.4Managing AI systems through all lifecycle stages

AI System Lifecycle Stages:

  1. Design: Requirements, architecture, ethical review
  2. Development: Data preparation, model training, coding standards
  3. Testing: Functional, bias, robustness, performance validation
  4. Deployment: Staging, approval, monitoring setup
  5. Operation: Performance monitoring, drift detection, incident response
  6. Retirement: Decommissioning, data disposal, stakeholder notification

Data Management for AI:

  • Data quality assessment and improvement
  • Data provenance and lineage tracking
  • Bias assessment in training and evaluation data
  • Data governance and access controls
  • Personal data protection measures
  • Data retention and disposal procedures

Third-Party and Supplier Management:

  • AI component supplier evaluation
  • Third-party AI service agreements
  • Supply chain risk assessment
  • Ongoing supplier monitoring

Performance Evaluation (Clause 9)

RequirementSectionDescription
Monitoring and measurement9.1AI system performance metrics and KPIs
Internal audit9.2Planned audits of the AIMS
Management review9.3Top management review of AIMS effectiveness

AI Performance Metrics:

  • Model accuracy, precision, recall
  • Fairness metrics (demographic parity, equalized odds)
  • Latency and availability
  • Drift indicators (data drift, concept drift)
  • Incident frequency and severity
  • Consumer complaint rates

Improvement (Clause 10)

RequirementSectionDescription
Nonconformity10.1Corrective actions for nonconformities
Continual improvement10.2Ongoing enhancement of the AIMS
AI incident management10.3Handling AI system incidents and near-misses

Annex A Controls

ControlTitleDescription
A.2AI PoliciesPolicies for responsible AI aligned with organizational objectives
A.3Internal OrganizationRoles, responsibilities, segregation of duties for AI
A.4Resources for AI SystemsCompute, data, tools, and expertise management
A.5Assessing AI System ImpactImpact assessment processes for AI systems
A.6AI System LifecycleControls across design, development, deployment, retirement
A.7Data for AI SystemsData quality, provenance, bias, governance, protection
A.8Information for Interested PartiesTransparency, disclosure, and communication
A.9Use of AI SystemsAcceptable use policies, human oversight, user guidance
A.10Third-Party RelationshipsSupplier management, outsourced AI, component evaluation

Annex B — Implementation Guidance

Annex B provides non-normative guidance for implementing Annex A controls:

  • Practical examples for each control objective
  • Scalability guidance for different organization sizes
  • Sector-specific considerations
  • Integration points with existing management systems

Annex C — AI Risk Sources and Objectives

AI-specific risk sources organized by category:

  • Technical risks: Model failure, data quality, adversarial attacks, drift
  • Ethical risks: Bias, discrimination, lack of transparency, autonomy erosion
  • Legal risks: Regulatory non-compliance, liability, intellectual property
  • Societal risks: Job displacement, misinformation, environmental impact
  • Organizational risks: Skill gaps, dependency, reputation damage

AI-specific control objectives:

  • Ensure fairness and non-discrimination
  • Maintain transparency and explainability
  • Guarantee safety and reliability
  • Protect privacy and data
  • Secure AI systems against threats
  • Enable accountability and governance

Annex D — Use of AIMS Across Domains

Sector-specific considerations:

  • Healthcare: Patient safety, clinical validation, regulatory approval (FDA, MDR)
  • Finance: Algorithmic trading, credit scoring, anti-money laundering
  • Autonomous systems: Safety-critical decisions, human override, fail-safe design
  • Human resources: Hiring bias, employee monitoring, fairness
  • Public sector: Citizen impact, democratic values, public trust

Relationship to Other Standards

StandardRelationshipIntegration Points
ISO 27001Information securityRisk assessment, access controls, incident management
ISO 9001Quality managementProcess approach, document control, continual improvement
ISO 14001Environmental managementImpact assessment, lifecycle thinking
ISO 31000Risk managementRisk framework, assessment methodology
ISO 22989AI concepts/terminologyFoundational definitions
ISO 23894AI risk managementRisk management guidance

Relationship to EU AI Act

EU AI Act RequirementISO 42001 Mapping
Risk management system (Art. 9)Clause 6.1, 8.2, 8.3, Annex A.5
Data governance (Art. 10)Clause 8.4, Annex A.7
Technical documentation (Art. 11)Clause 7.5, Annex A.6
Transparency (Art. 13)Annex A.8
Human oversight (Art. 14)Annex A.9
Accuracy, robustness, security (Art. 15)Clause 9.1, Annex A.6
Quality management system (Art. 17)Full AIMS (Clauses 4-10)
Conformity assessmentCertification process

Certification Process

PhaseActivityDuration
PreparationGap analysis, implementation, internal audit6-12 months
Stage 1 AuditDocumentation review, readiness assessment1-2 days
Gap RemediationAddress Stage 1 findings1-3 months
Stage 2 AuditImplementation effectiveness assessment2-5 days
CertificationCertificate issued (3-year validity)Upon passing
SurveillanceAnnual surveillance audits1-2 days/year
RecertificationFull reassessment every 3 years2-4 days

Implementation Roadmap

Phase 1 — Foundation (Months 1-3):

  • Define AIMS scope and boundaries
  • Establish AI governance structure
  • Develop AI policy
  • Conduct initial AI system inventory
  • Define risk assessment methodology

Phase 2 — Core Implementation (Months 4-6):

  • Conduct AI risk assessments for all in-scope systems
  • Perform impact assessments (Annex A.5)
  • Implement AI lifecycle controls (Annex A.6)
  • Establish data management processes (Annex A.7)
  • Develop third-party management procedures (Annex A.10)

Phase 3 — Operationalize (Months 7-9):

  • Deploy monitoring and measurement (Clause 9.1)
  • Train personnel on AIMS roles and responsibilities
  • Implement incident management procedures
  • Conduct awareness programs for AI literacy
  • Establish communication processes

Phase 4 — Verify and Certify (Months 10-12):

  • Conduct internal audit (Clause 9.2)
  • Hold management review (Clause 9.3)
  • Address nonconformities
  • Prepare for Stage 1 certification audit
  • Compile evidence packages per clause

Troubleshooting

ProblemPossible CauseResolution
Readiness score low on Clause 5 (Leadership) despite executive sponsorshipAI policy does not include ethical principles, responsible AI commitment, or framework for setting AI objectivesUpdate AI policy to explicitly address all required elements: ethical principles, responsible AI, legal alignment, continual improvement commitment, and AI objectives framework; obtain formal management sign-off
AI impact assessment returns High/Critical risk across all dimensionsAI system processes sensitive personal data, makes autonomous decisions, and affects large populations without safeguardsImplement targeted mitigations per dimension: human-in-the-loop for safety, bias testing for fairness, explainability mechanisms for transparency, data protection for privacy; re-run assessment after mitigation
Annex A controls scored as "Not Implemented" despite operational practicesPractices exist informally but are not documented per ISO 42001 requirementsDocument all existing AI practices as formal procedures; create evidence artifacts (policy documents, meeting minutes, risk registers, training records); map to specific Annex A control objectives
Certification body auditor questions AI risk assessment methodologyRisk assessment does not cover all seven required risk categories (fairness, transparency, safety, privacy, security, accountability, societal)Update risk assessment methodology to explicitly address all ISO 42001 risk categories; use
ai_impact_assessor.py
template to ensure comprehensive coverage; document risk criteria and tolerance levels
Third-party AI components lack governance controlsOrganization uses third-party AI models or APIs without formal evaluation or supplier managementImplement Annex A.10 (Third-Party Relationships) controls; evaluate all third-party AI components; establish contractual requirements for AI service providers; monitor supplier AI practices
Data management procedures incomplete for AI lifecycleData quality, provenance, and bias assessment not systematically performed for training and evaluation dataImplement Annex A.7 (Data for AI Systems) controls; establish data quality assessment procedures; document data provenance and lineage; conduct bias assessments per dataset; define retention and disposal procedures
Stage 1 audit finds AIMS documentation insufficientDocumentation follows generic QMS structure without AI-specific elementsRestructure documentation to address all ISO 42001 clauses (4-10) and Annex A controls (A.2-A.10); include AI-specific policies, risk assessments, impact assessments, and lifecycle procedures

Success Criteria

  • Overall readiness score of 80+ for certification readiness -- as measured by
    aims_readiness_checker.py
    , with all clauses at Defined maturity level or above
  • AI policy established and communicated -- including ethical principles, responsible AI commitment, legal compliance alignment, continual improvement, and framework for AI objectives, with formal management approval
  • AI impact assessments completed for all in-scope AI systems -- covering all six dimensions (fairness, transparency, safety, privacy, security, accountability) with risk treatments documented and residual risk accepted by management
  • AI risk assessment methodology covers all required categories -- fairness, transparency, safety, privacy, security, accountability, and societal/environmental impacts, with defined risk criteria and tolerance levels
  • Annex A controls implemented with evidence -- A.2 (Policies) through A.10 (Third-Party) with documented procedures, records, and evidence artifacts suitable for certification audit
  • Internal audit conducted against all AIMS clauses -- with findings documented, corrective actions tracked to closure, and management review completed with documented improvement decisions
  • AI lifecycle procedures operational -- covering design, development, testing, deployment, monitoring, and retirement stages with documented controls at each gate

Scope & Limitations

In Scope:

  • ISO 42001:2023 readiness assessment across all clauses (4-10) and Annex A controls (A.2-A.10)
  • AI impact assessment across six dimensions (fairness, transparency, safety, privacy, security, accountability)
  • AIMS certification preparation including gap analysis, implementation roadmap, and audit readiness
  • AI lifecycle management guidance (design through retirement)
  • Data management for AI systems (quality, provenance, bias, governance)
  • Third-party AI supplier management and evaluation
  • Regulatory mapping to EU AI Act requirements
  • Integration guidance with ISO 27001, ISO 9001, and ISO 14001

Out of Scope:

  • Actual AI model development, training, testing, or deployment -- this skill provides governance frameworks, not ML engineering
  • Certification body selection, audit scheduling, or certification fee negotiation
  • Ethical review board establishment or ethical decision-making beyond procedural guidance
  • Specific AI fairness algorithm implementation (e.g., adversarial debiasing, calibrated equalized odds) -- use
    eu-ai-act-specialist
    bias detector for technical testing
  • Environmental impact measurement or carbon footprint calculation for AI training

Important Notes:

  • ISO 42001 certification follows a 3-year cycle with annual surveillance audits at 12-month intervals
  • Major certification bodies (BSI, DNV, TUV, LRQA) have operationalized ISO 42001 audit services as of 2025-2026
  • Many organizations pursue dual alignment: ISO 42001 certification for governance controls plus EU AI Code of Practice for regulatory expectations
  • The standard's Annex SL structure enables direct integration with ISO 27001, reducing redundant documentation and audit effort

Integration Points

SkillIntegrationWhen to Use
eu-ai-act-specialist
ISO 42001 AIMS maps directly to EU AI Act requirements; certification demonstrates Art. 17 QMS complianceWhen building AI governance satisfying both ISO 42001 and EU AI Act obligations
information-security-manager-iso27001
ISO 27001 security controls integrate with AIMS via shared Annex SL structure; risk assessment methodologies alignWhen implementing joint ISMS + AIMS covering both information security and AI governance
gdpr-dsgvo-expert
AIMS data management (Annex A.7) aligns with GDPR data protection requirements; AI processing requires DPIAWhen AI systems process personal data and require both AIMS and GDPR compliance
isms-audit-expert
Internal audit methodology and finding management shared between ISO 27001 and ISO 42001When conducting internal audits covering both ISMS and AIMS

Tool Reference

aims_readiness_checker.py

Assesses organizational readiness against all ISO 42001:2023 clauses and Annex A controls.

FlagRequiredDescription
--input <file>
Yes (unless
--template
)
Path to JSON organizational profile for assessment
--template
NoGenerate blank input template to stdout
--json
NoOutput results in JSON format for automation
--output <file>
NoExport report to specified file path

Assessment Scope: Clause 4 (Context), Clause 5 (Leadership), Clause 6 (Planning), Clause 7 (Support), Clause 8 (Operation), Clause 9 (Performance), Clause 10 (Improvement), and Annex A controls (A.2-A.10).

Output: Overall readiness score (0-100), per-clause scores with maturity level (Initial/Developing/Defined/Managed/Optimized), Annex A control implementation status, gap analysis with prioritized recommendations, and certification readiness assessment (Ready/Near Ready/Significant Gaps).

ai_impact_assessor.py

Generates comprehensive AI impact assessments across six risk dimensions with regulatory mapping.

FlagRequiredDescription
--input <file>
Yes (unless
--template
)
Path to JSON AI system description for assessment
--template
NoGenerate blank AI system template to stdout
--format <fmt>
NoOutput format:
json
(default) or
markdown
--output <file>
NoExport assessment report to specified file path

Assessment Dimensions: Fairness (bias, discrimination, equity), Transparency (explainability, interpretability), Safety (reliability, robustness, harm prevention), Privacy (data protection, consent, minimization), Security (adversarial resilience, access control), Accountability (governance, responsibility, auditability).

Output: Per-dimension risk scoring (Low/Medium/High/Critical), interested party impact mapping, risk treatment options (Avoid/Mitigate/Transfer/Accept), regulatory mapping (EU AI Act risk tier, ISO 42001 Annex A controls), residual risk calculation, and markdown or JSON report.