Skip to content
English
  • There are no suggestions because the search field is empty.

How to complete your AI Register

This article outlines how to complete the AI Register (AI Inventory)

Purpose of the AI Register

The AI Register is the core evidence source for all AI systems your organisation develops or uses. It supports risk management, lifecycle oversight, transparency to interested parties, and performance monitoring as required by the Artificial Intelligence Management System (AIMS) and ISO/IEC 42001:2023.

The register captures information about every AI system, including its purpose, risk, oversight mechanisms, data sources, monitoring activities, and how information is shared with stakeholders. The AI Register is used in conjunction with your AIMS Manual, risk processes, and governance activities.

The type of AI system (for example, in-house developed or third-party provided) may affect how much information is available and where supporting evidence is maintained. However, it does not change the governance expectations. All AI systems in use — including third-party or off-the-shelf tools — require appropriate oversight, risk and impact assessment, monitoring, and due diligence, which should be reflected across the relevant fields in the AI Register.


Scope

Include all AI/ML tools, such as:

  • Internal models and algorithms (e.g. predictive scoring, NLP tools)

  • Third-party tools or APIs using AI (e.g. OpenAI, AWS Rekognition)

  • Off-the-shelf AI features in SaaS products

  • AI tools or services used by employees or contractors as part of organisational activities, whether formally deployed or informally adopted.


How to Complete Each Field

 

Field What to Enter Why It Matters
System / Tool Name

Enter the name of the AI tool, model, or product (e.g. "ChatGPT", "Internal Risk Model v2")

Identifies the AI system clearly in audits and reports

AI System Type

Select how the AI system is provided (In-house, Third-party, Hybrid, Open-source)

Supports proportionate governance and oversight

Business Use Case

Briefly describe what the tool is used for (e.g. “automated CV screening”, “customer churn prediction”, "customer support", "fraud detection")

Helps determine risk level, transparency needs, and regulatory scope

Owner

Select the individual responsible for the AI system

Assigns accountability for oversight, risk, and data governance

Review date

Select the date the AI system is next due for review

Supports periodic review and ongoing suitability

Vendor / Source

Enter the vendor or source of the AI system, if applicable

Clarifies third-party reliance and due-diligence needs

Status

Select the current lifecycle status: Proposed, Active, Suspended or Retired

Indicates whether the AI system is in use and its lifecycle stage

Risk Category

Use the drop down to select from:Minimal, Limited, High or Unacceptable

Helps to identify and prioritise controls required.

Data Types Processed Describe the types of data processed (e.g. personal, financial) Supports privacy, security, and regulatory assessment
 System Impact Assessment

Describe whether the AI makes, assists, or autom ates decisions with material impact

Identifies potential operational and business impact

Individual Impact Assessment

Describe potential impacts on individuals or groups

Supports ethical and legal risk assessment

Societal Impact Assessment

Describe potential wider societal impacts

Helps assess broader ethical considerations

Privacy Risk Level

Select Low, Medium, or High

Indicates privacy risk requiring management

Bias Risk Level

Select Low, Medium, or High

Highlights potential fairness or discrimination risks

Training Data Source(s)

Select applicable sources (Internal, Public, Third-party, Synthetic)

Supports data governance and traceability

Level of Autonomy

Select the level of autonomy (Fully autonomous, Human-in-the-loop, Human-in-command)

Clarifies decision authority and oversight needs

Human Oversight Defined

Select Yes or No

Confirms escalation and operator controls exist

Explainability Required

Select Yes or No

Identifies transparency obligations for stakeholders

Monitoring in Place

Select Yes or No

Confirms ongoing performance and risk monitoring

Last Model Update

Enter the date of the most recent update

Supports lifecycle and change tracking

Information Provided to Interested Parties

Select Yes, No, or Not applicable

Indicates whether transparency information is communicated externally

Regulatory Relevance

Describe applicable regulations (e.g. GDPR, ISO/IEC 42001)

Supports compliance assessment

Comments / Notes

Enter any additional relevant information

Provides context or supporting detail

Best Practices

  • Assign clear ownership
    Ensure each AI system has a named owner responsible for oversight, risk management, and ongoing review.
  • Keep the register up to date
    Add AI systems as soon as they are proposed or adopted, and regularly review entries to suspend or retire systems that are no longer in use.
  • Use risk categorisation effectively
    Use the Business Use Case, Risk Category, and Impact Assessment fields to prioritise oversight and review frequency, particularly for systems affecting people, finances, or critical operations.
  • Review systems periodically
    Use the Review Date and Status fields to support ongoing suitability, ensuring AI systems remain appropriate as risks, usage, or regulatory requirements change.
  • Record oversight and monitoring consistently
    Ensure fields relating to autonomy, human oversight, explainability, and monitoring are completed accurately to demonstrate responsible and controlled use of AI systems.

Benefits of Completion

  • Helps identify AI systems that may be subject to heightened regulatory obligations, including potential classification under emerging AI regulations such as the EU AI Act

  • Supports data-protection compliance by identifying AI systems that process personal or sensitive data

  • Provides a structured foundation for AI risk and impact assessment in line with ISO/IEC 42001 requirements

  • Creates a clear audit trail to support internal reviews, external audits, and regulatory enquiries

  • Strengthens accountability by clearly identifying ownership and responsibility for each AI system, supporting effective governance and oversight