top of page

Adopt the new, auditable standard for global AI governance

ISO/IEC 42001 is positioned to be the first global auditable standard to enable organizations to demonstrate trustworthiness in their AI systems.

What is ISO 42001?

ISO/IEC 42001 is a voluntary standard for organizations to implement an artificial intelligence (AI) management system. Under ISO 42001, an organization’s AI management system sets policies, procedures, and objectives for their AI systems. ISO 42001 is intended to set a baseline governance standard for all AI systems within an organization, rather than focusing on specific types of AI systems (i.e., high-risk AI).

Key Requirements of ISO 42001

Establishing Context

Requirement

Create an AI Management System (AIMS), and clearly document the internal and external context of your organization related to AI. This includes identifying relevant stakeholders, regulations, and scope of AI use for your organization.

How Trustible Helps

Trustible helps collect and document all the information related to AI systems, and offers policy templates aligned with ISO 42001 to help organizations bootstrap their AI Management System.

Leadership

Requirement

Identify the relevant executive leaders, and map out roles and responsibilities for AI governance. These roles should be reflected in organizational policies and clearly communicated throughout the organization.

How Trustible Helps

Trustible offers policy templates aligned with ISO 42001 to help organizations accelerate their path to compliance, as well as out-of-the-box workflows, dashboards, and reports to efficiently inform senior leadership about relevant AI systems.

Planning

Requirement

Requires organizations to implement processes for AI governance to ensure AI risks are properly captured and documented. This includes performing risk assessments, impact assessments, and building risk treatment plans.

How Trustible Helps

Trustible offers out-of-the-box workflows for risk and impact assessments, and offers risk and mitigation recommendations to accelerate building and implementing risk treatment plans. In addition, Trustible can deliver the latest best practices on AI risk management to help organizations stay on top of the fast-moving AI environment.

Support

Requirement

Organizations must allocate the appropriate resources towards AI governance, build an internal knowledge base about AI systems, and ensure allocated staff is appropriately trained and educated on AI risks.

How Trustible Helps

Trustible offers both AI compliance training, as well as continuously updated AI risk insights including risk measurement guidance, recommended mitigations, model risk ratings, and updates on AI regulatory compliance practices to help keep AI resources up-to-date on the latest AI best practices. In addition, Trustible supports the industry leading AI inventory solution that integrates across the tech stack to help track all necessary information about AI use cases, models, and vendors.

Operation

Requirement

Organizations must have clear paper trails, and processes for maintaining their AI management systems over time.

How Trustible Helps

Trustible helps organizations create auditable paper trails of AI system proposals, risk assessments, deployment approvals, etc and can generate reports to evaluate how efficiently an organization is governing their AI.

Performance Evaluation

Requirement

Organizations need to identify which AI systems require additional monitoring, and how effective their own management system is through internal audits and reviews.

How Trustible Helps

Trustible helps organizations identify which AI uses cases that require regular reviews, and has guided workflows for conducting internal audits or periodic reviews.

Governance Improvements

Requirement

Organizations need to continuously improve their AI governance processes and structure, and have formal plans for identifying and fixing any gaps or non-compliance instances.

How Trustible Helps

Trustible helps organizations continuously iterate on their AI governance practices, and helps automatically detect non-compliance with internal AI policies, or non-compliance with regulations based on up-to-date regulatory insights and best practices.

Navigate ISO 42001 with Trustible

trustible-icon-stats-monitor-wire.png
Risk & Impact Assessments

Identify, manage, measure, and mitigate potential risks or harms in your AI systems.

trustible-icon-legal-wire.png
AI Policies

Develop and enforce AI policies that protect your organization, users, and society.

trustible-icon-chart-wire.png
Documentation

Centralize your AI documentation in a single source of truth.

FAQs

Who is the intended audience?

ISO 42001 was designed for any organization that develops, deploys, and/or uses AI. While the requirements are more prescriptive than other existing AI governance and risk management frameworks, ISO 42001 is meant to be scalable for organizations of any size.

Is this an enforceable standard?

While ISO 42001 is voluntary, it is not uncommon for components of voluntary standards to become legal requirements. Policymakers, especially in the EU, may gravitate towards ISO 42001 as an enforceable standard for AI governance.

How can I comply with ISO 42001?

ISO 42001 includes annexes that map the requirements to a series of controls and implementing guidance. The additional guidance is meant to provide a more granular roadmap for organizations seeking to comply with the standard.

bottom of page