ISO/IEC 42005 Impact Assessment

Build audit-ready AI governance aligned with international standards

A strong audit framework is the foundation of accountable, fair, and transparent AI.

  • Reinforce responsible AI with recognised audit practices

  • Detect and address risks before external scrutiny

  • Demonstrate fairness, accountability, and compliance in practice

Embedding Accountability in AI Governance

ISO/IEC 42005:2023 provides guidance on how AI governance should be audited. Unlike ISO/IEC 42001, which sets requirements for an AI Management System (AIMS), 42005 defines the methods to evaluate and assure that system. While it is not a certifiable standard, it helps organisations prove that their AI governance is fair, defensible, and accountable. By applying 42005, organisations can strengthen trust, ensure readiness for ISO/IEC 42001 certification, and reinforce alignment with the EU AI Act.

Our Approach

1

Clarify

We define the scope, context, and purpose of the AI system. Stakeholders, intended outcomes, and foreseeable risks are identified.

2

Assess

We evaluate positive and negative impacts, including unintended consequences and misuse scenarios, across the AI lifecycle.

3

Document

We capture findings in a structured record that aligns with ISO 42005. Risks, mitigations, monitoring, and oversight are clearly presented.

4

Integrate

We embed the assessment into your governance processes and ensure it is updated when systems change or risks evolve.

The Result: Responsible AI Assurance

Fairness in focus

Highlight and mitigate risks of bias, discrimination, or rights impacts through structured audits.

Certification readiness

Strengthen preparation for ISO/IEC 42001 certification with clear evidence and audit practices.

Defensible accountability

Demonstrate responsible AI governance with auditable proof for regulators and stakeholders.

Two people analyzing graphs and data on a large screen with charts, graphs, and a heart rate monitor in the background.

Frequently Asked Questions

  • ISO/IEC 42005 provides a structured approach to assessing the impact of AI systems on individuals, organisations, and society.

    It focuses on identifying risks related to fairness, transparency, accountability, and potential harm, helping organisations understand not just how AI works, but how it affects people and decision-making in practice.

  • ISO/IEC 42001 defines the overall management system for AI governance.

    ISO/IEC 42005 complements it by focusing specifically on impact assessments, providing a method to evaluate how individual AI systems create risk.

    In practice:

    • ISO 42001 = governance framework

    • ISO 42005 = risk and impact analysis within that framework

  • An AI impact assessment should be applied when systems:

    • Influence decisions affecting individuals

    • Process sensitive or large-scale data

    • Introduce automation in critical processes

    • Create potential legal, ethical, or reputational risk

    In many cases, this aligns with situations that would also trigger a DPIA under the GDPR.

  • The standard goes beyond technical risk and considers:

    • Impact on individuals’ rights and freedoms

    • Bias and fairness in decision-making

    • Transparency and explainability

    • Operational and organisational risks

    • Broader societal and ethical implications

    This makes it particularly relevant for AI systems used in decision-making or automation.

  • A structured impact assessment report with findings and recommendations, providing assurance that AI systems align with responsible governance and regulatory expectations.

  • The EU AI Act requires organisations to identify, assess, and manage risks associated with AI systems, particularly for high-risk use cases.

    ISO/IEC 42005 provides a practical method for conducting these assessments, helping organisations demonstrate that risks have been systematically identified and addressed.

  • AI impact assessments require cross-functional input, including:

    • Legal and compliance teams

    • Data protection and security professionals

    • Technical and product teams

    • Business stakeholders responsible for outcomes

    This ensures that both technical and real-world impacts are properly understood.