AI compliance glossary

Understand key terms in AI compliance

Navigate the complex world of AI compliance with ease. This glossary breaks down essential terms and concepts to help AI professionals and enterprise buyers alike gain clarity on regulations, standards, and practices that shape today’s AI landscape. Stay informed and stay compliant.

AI Compliance Glossary
Thank you! Your submission has been received!
Oops! Something went wrong while submitting the form.

AI (Artificial intelligence)

A machine-based system capable of making predictions, recommendations, or decisions influencing environments. AI systems operate autonomously or semi-autonomously, using inference or pattern recognition to achieve specific goals within defined parameters.

AI Accountability

The responsibility of providers and deployers to ensure AI systems are developed, deployed, and monitored in compliance with regulations and ethical standards, ensuring that AI’s use does not infringe on fundamental rights and operates as intended.

AI Accuracy

The degree to which an AI system’s outcomes align with the intended purpose, minimizing errors and biases. For high-risk systems, accuracy must be continuously monitored to maintain system integrity and prevent risks to users and affected persons.

AI Autonomy

The capacity of an AI system to function with minimal human intervention, making decisions based on programmed objectives and evolving inputs. This autonomy allows AI to operate independently, adapting to changing environments and inputs.

AI Bias Audit

A structured evaluation aimed at identifying, analyzing, and mitigating biases within AI systems. Bias audits ensure that AI models operate fairly, produce non-discriminatory outcomes, and comply with ethical and regulatory standards to protect affected groups.

AI Bias Detection and Mitigation

Processes to identify and reduce biases in AI, which may arise from systemic, computational, or human sources, potentially leading to unfair outcomes. Bias mitigation ensures AI systems work equitably across diverse groups and contexts.

AI Buyer’s Risk Assessment

A process through which potential buyers of AI assess risks associated with an AI system, including compliance, security, and ethical factors. The assessment aids buyers in making informed decisions by evaluating operational risks and impact.

AI Compliance Framework

A structured set of guidelines and best practices designed to ensure that AI systems meet legal, ethical, and operational standards. The framework addresses accountability, transparency, risk management, and data security requirements.

AI Deployers

Individuals or entities that operate or use AI systems in various applications. Deployers are responsible for ensuring that systems function as intended and in compliance with regulatory standards, especially when using high-risk AI systems.

AI Distributors

Natural or legal persons within the supply chain, other than providers or importers, who make AI systems available on the EU market. They must ensure that systems meet all regulatory standards and provide essential documentation upon request.

AI Ethics

Principles guiding the responsible development and use of AI to align with human values and rights, including transparency, fairness, and accountability. AI ethics frameworks aim to prevent harm, discrimination, and misuse of AI systems in various contexts.

AI Importer

A natural or legal person who places an AI system from a third country on the EU market. Importers are responsible for verifying compliance with EU regulations and ensuring all necessary documentation accompanies the system.

AI Interpretability

The extent to which the meaning of AI outputs is clear and understandable within the system’s designed functional purpose. Interpretability helps operators or overseers effectively use and govern the AI system in practice.

AI Product Manufacturers

Entities responsible for integrating AI systems as components of physical products. They must ensure these systems comply with safety and compliance requirements before releasing the products on the market.

AI Provider

Any entity that develops or markets an AI system in the EU. Providers are accountable for system compliance, including meeting technical, safety, and documentation standards, especially for high-risk applications.

AI System

‘AI system’ means a machine-based system designed to operate with varying levels of autonomy, that may exhibit adaptiveness after deployment and that, for explicit or implicit objectives, infers, from the input it receives, how to generate outputs such as predictions, content, recommendations, or decisions that can influence physical or virtual environments

AI Transparency

Obligations ensuring that an AI system’s functions, capabilities, and limitations are clear to users and regulators, promoting informed use and adherence to compliance requirements, particularly in high-risk cases.

AI auditing

The process of systematically reviewing AI systems to ensure adherence to regulatory, ethical, and operational standards, focusing on system accuracy, fairness, and accountability. Audits identify risks and verify compliance with established guidelines.

AI compliance

The adherence of AI systems to established legal, ethical, and technical standards to ensure safe and trustworthy deployment. Compliance encompasses data governance, accountability, and user rights in alignment with regulatory requirements.

AI explainability

The degree to which an AI system’s mechanisms can be understood by humans. Explainability allows stakeholders to gain insights into how the system generates outputs, fostering trust, facilitating debugging, and enabling accountability.

AI governance

Oversight structures and policies ensuring that AI systems are developed, deployed, and monitored in accordance with legal, ethical, and organizational standards, including data security, fairness, transparency, and risk management protocols.

AI literacy

The ability of users and stakeholders to understand basic AI concepts, including its risks, limitations, and implications on decision-making. AI literacy enables informed use, promotes transparency, and enhances trust in AI-driven outcomes.

AI monitoring and Reporting

Continuous oversight of AI system performance and compliance, ensuring it adheres to regulatory standards and operates as intended. Monitoring includes documenting system outputs and reporting deviations or incidents affecting user rights or safety.

AI policies

Internal organizational guidelines that establish standards for AI development, deployment, and oversight, focusing on ethics, accountability, and data governance to align AI use with regulatory and organizational objectives.

AI questionnaires

Standardized sets of questions assessing AI system risks, performance, and compliance. These questionnaires help AI buyers evaluate AI reliability, fairness, and adherence to ethical and regulatory standards, providing insights for decision-makers.

AI reliability

The ability of an AI system to perform as required, without failure, over a specified time interval and under defined conditions. Reliable AI systems consistently operate correctly, promoting user trust and reducing risk in critical applications.

AI resilience

The capacity of an AI system to maintain functionality and performance in the face of unexpected challenges, including adversarial attacks or environmental shifts. Resilient AI systems are robust and recover gracefully from disruptions.

AI risk management

Coordinated activities to identify, assess, and mitigate risks throughout the AI lifecycle. AI risk management enhances understanding of potential impacts, reduces harm, and improves the trustworthiness of AI systems across contexts.

AI robustness

The ability of an AI system to sustain performance across a range of scenarios, including those not initially anticipated. Robust systems demonstrate reliability and minimize harm, even in varied or unexpected environments.

AI safety

Assurance that an AI system will not, under defined conditions, cause harm to human life, health, property, or the environment. Safe AI systems incorporate design practices, rigorous testing, and controls to avoid dangerous states or failures.

Adversarial Attacks

Techniques used to exploit vulnerabilities in AI systems by manipulating data or inputs, causing the AI to perform unintended actions or produce incorrect outputs. Resilient AI systems are designed to withstand such attacks and maintain functionality.

Algorithmic Auditing

A thorough examination of algorithms within AI systems to assess their accuracy, fairness, and compliance with legal and ethical standards. Algorithmic audits aim to identify and mitigate potential biases and ensure responsible AI deployment.

Algorithmic Bias

A tendency of an AI algorithm to produce discriminatory outcomes, often due to skewed training data or flawed model design. Algorithmic bias audits address disparities, ensuring outcomes align with fairness principles and anti-discrimination laws.

Algorithmic Fairness

A principle ensuring AI systems do not perpetuate or amplify existing biases, particularly against protected groups. This includes detecting and mitigating discriminatory outcomes to align with fundamental rights and anti-discrimination laws.

Algorithmic Transparency

The principle that AI systems’ decision-making processes should be clear and understandable, enabling users to comprehend how outputs are derived. Transparency fosters accountability and builds user trust by providing insight into AI operations.

Authorized representative

A natural or legal person based in the EU, mandated by providers from outside the European Union, to fulfill specific compliance obligations and represent them in all regulatory matters concerning their AI system in the European market.

Automated AI Compliance

The use of AI-driven tools to automatically monitor and enforce compliance with regulatory, ethical, and operational standards. Automated compliance tools ensure continuous adherence to requirements without manual oversight.

Automated Decision-Making

Processes in which AI systems independently make decisions or recommendations impacting individuals or groups. Automated decisions must align with fairness, transparency, and accuracy standards, particularly in high-risk applications.

Biometric Categorization

The process of assigning individuals to specific groups or categories based on biometric data, such as age, gender, or behavior, excluding any direct identification purposes or consent-based verification processes.

Biometric Data

Data derived from physical, physiological, or behavioral characteristics enabling unique identification of a natural person, including facial images, fingerprints, and gait, processed for either identity verification or categorization purposes.

Biometric Identification

The automated process of identifying individuals by comparing their biometric data with reference data, typically without active participation from the individual, to establish identity in various environments, including security settings.

Black Box AI Model

AI models with complex internal operations that are opaque or difficult to interpret by users. These models pose challenges for transparency and accountability, often requiring explainability measures to ensure trust and understanding.

Business Continuity Plan

A strategic approach to maintaining AI system functionality during disruptions, ensuring ongoing operations. Plans include measures for data backup, disaster recovery, and contingency actions to mitigate risks of operational interruptions.

Business Impact Analysis

An assessment process evaluating the potential effects of AI system failure or disruption on business functions. This analysis identifies critical operations, quantifies impacts, and guides risk management and continuity planning efforts.

Compliance Documentation

Records and materials that detail an AI system’s design, functionality, and risk assessments, ensuring conformity with applicable EU regulations, including technical specifications, testing protocols, and operational guidance for responsible use.

Compliance Mechanism

Structured approaches within an organization to ensure AI systems align with applicable legal and regulatory requirements. Compliance mechanisms encompass policies, processes, and monitoring activities to manage AI risks effectively.

Conformity Assessment

A systematic evaluation process to verify that high-risk AI systems meet all EU compliance requirements, ensuring their safe deployment and adherence to standards that protect users and affected persons’ health, safety, and fundamental rights.

Continuous Monitoring

Ongoing observation and assessment of an AI system’s performance, identifying and addressing emergent risks and system failures over time. Continuous monitoring supports adaptation to evolving conditions and stakeholder needs.

Cybersecurity

Measures implemented to protect AI systems from unauthorized access, manipulation, or malicious attacks. These measures ensure the integrity, confidentiality, and reliability of the AI system’s performance across its operational lifecycle.

Data Governance

Policies and procedures ensuring data used in AI systems is accurate, representative, and privacy-compliant, particularly in high-risk contexts. Effective data governance prevents biases, protects user data, and maintains the integrity of AI processes.

Data Minimization

The principle that only the necessary amount of data should be collected, processed, and retained to achieve the AI system’s intended purpose, particularly for personal and sensitive data, to safeguard privacy and prevent unauthorized data use.

Data Provenance

The tracking and documentation of the origin, history, and quality of data used in AI systems. Maintaining data provenance ensures transparency, accountability, and compliance, supporting risk management and trustworthiness.

Data Quality Standards

Standards ensuring that the data used for training and operating AI systems is complete, accurate, and free from biases that could affect system performance, especially in high-risk applications where data quality directly impacts decision integrity.

Data Sovereignty

A principle ensuring that data collected, processed, and stored by AI systems complies with applicable jurisdictional regulations, particularly for cross-border data transfers, and respects individuals’ rights to control their personal data.

Due diligence questionnaires

Standardized surveys evaluating AI systems’ compliance, performance, and ethical adherence. These questionnaires help AI buyers assess potential risks, including data governance, transparency, and regulatory compliance, prior to deployment.

EU Database for High-Risk AI

A centralized EU database where high-risk AI systems are registered to enhance transparency, facilitate regulatory oversight, and provide accessible information on compliance status and risk assessments associated with these systems.

Emergent Properties

Unanticipated behaviors or effects that arise in complex AI systems, often as a result of interactions among components. These properties may lead to unintended consequences and require careful monitoring and management.

Emotion Recognition System

AI systems designed to infer emotions or intentions from biometric data, such as facial expressions or voice intonation, potentially impacting users’ privacy. Such systems are subject to stricter regulations in sensitive environments like workplaces.

Enterprise AI Compliance

Policies and standards applied within organizations to ensure AI systems meet regulatory and ethical standards, fostering trust and accountability. Enterprise AI compliance encompasses data security, risk assessment, and stakeholder engagement.

Ethical AI

The development and use of AI systems that prioritize human values, transparency, and fairness to protect individuals’ rights and freedoms. Ethical AI practices are intended to prevent discrimination, misuse, and adverse impacts on society.

Explainability

The capability of an AI system to make its decisions or outcomes understandable to users or regulators, particularly in high-risk applications, supporting transparency and allowing individuals to assess system impact on their rights and interests.

Explainable AI (XAI)

AI systems designed to provide clear, interpretable explanations of their decision-making processes. Explainable AI enhances transparency and accountability, allowing users to understand and trust AI-driven outcomes in complex applications.

Fail-Safe Mechanism

Built-in measures allowing AI systems to revert to a safe state in the event of errors, technical malfunctions, or unexpected behavior, thereby preventing harm to users, the public, or affected environments.

Fairness in AI

A characteristic that addresses equity and bias in AI systems, ensuring outcomes do not discriminate against specific groups. Fair AI systems promote inclusivity, mitigate unintended harm, and align with societal standards of justice.

Fundamental Rights Impact Assessment

A pre-deployment assessment of high-risk AI systems to identify and mitigate risks to individuals’ rights and freedoms, ensuring AI use complies with EU human rights protections and does not adversely affect vulnerable groups.

General Data Protection Regulation (GDPR)

A European Union regulation ensuring the protection of individuals’ personal data. AI systems processing personal data must adhere to GDPR principles, including consent, transparency, and accountability, to safeguard privacy rights.

General-Purpose AI

AI models or systems capable of performing a wide range of functions across multiple contexts. These models are typically trained on extensive datasets and may be adapted or fine-tuned for specific applications.

General-Purpose AI with Systemic Risks

General-purpose AI models deemed to pose broad risks due to their capabilities and potential for widespread impact. They are subject to additional regulatory requirements for transparency, risk management, and impact assessment.

High-Risk AI Systems

AI systems identified by the EU as having significant potential to impact health, safety, or fundamental rights. These systems must meet strict compliance standards, including risk assessments, documentation, and human oversight.

Human Oversight

Mechanisms that allow humans to monitor and, where necessary, intervene in AI system operations to prevent or mitigate harmful outcomes, particularly relevant for high-risk AI systems in sensitive sectors.

Human Rights Impact Assessment

A review assessing the impact of an AI system on fundamental human rights, including privacy and non-discrimination. This assessment identifies and mitigates potential harms, ensuring AI deployment aligns with EU rights protections.

Human-in-the-Loop (HITL)

Involvement of human oversight in AI decision-making, allowing humans to interpret, adjust, or override AI outputs. HITL configurations enhance safety, accountability, and the ethical deployment of AI systems.

ISO 42001

An international standard for managing AI systems, focusing on guidelines for responsible AI development, deployment, and oversight. ISO 42001 outlines best practices for transparency, accountability, and continuous risk management.

Impact Assessment

A pre-deployment evaluation of potential risks and impacts associated with an AI system, focusing on user safety, privacy, and rights. This assessment informs risk management strategies and enhances compliance with regulatory standards.

Intellectual Property Compliance

Adherence to copyright, patent, and other intellectual property laws in the development and deployment of AI systems, ensuring that any protected content or methods used in AI are authorized or properly licensed.

Interpretability

The ability of AI systems to generate outcomes that can be understood by end users and regulators, especially in high-risk contexts, to promote transparency, accountability, and informed decision-making.

Limited risk AI systems

AI systems with moderate potential for adverse impact, subject to transparency and information requirements but not classified as high-risk. Users must be informed they are interacting with AI, ensuring responsible and transparent usage.

Machine Learning

An AI technique where systems improve their performance by learning from data without explicit programming. Machine learning models require proper documentation and compliance to minimize risks of bias and errors.

Machine Learning Fairness (ML Fairness)

Techniques and practices aimed at promoting equitable outcomes in machine learning systems by addressing potential biases. ML Fairness seeks to prevent discrimination and ensure AI models are fair across demographics.

Minimal risk AI systems

AI systems with low potential to impact users’ health, safety, or fundamental rights. These systems typically require no specific regulatory oversight and are presumed safe for general use without additional compliance obligations.

Model Accountability

Ensuring AI models operate in line with ethical and regulatory standards, with responsibility for outcomes assigned to providers. Model accountability includes transparency, risk management, and compliance with legal obligations.

Model Generalizability

The extent to which an AI model can maintain performance on new or diverse data that differs from its training data. Generalizable models perform effectively across varied contexts, enhancing reliability and reducing risk.

Model Interpretability

The ability of AI systems to produce results that are understandable to users, enabling them to comprehend the reasoning behind AI-driven outcomes. Interpretability is key for transparency, trust, and effective decision-making.

NIST Compliance Standards

Guidelines established by the National Institute of Standards and Technology for ethical, transparent, and secure AI systems. NIST standards support regulatory compliance, risk management, and protection of user rights.

Post-Market Monitoring

Continuous monitoring of an AI system’s performance and compliance after its deployment, enabling the provider to detect and correct any issues or unintended consequences impacting safety, accuracy, or compliance.

Privacy Impact Assessment

An evaluation to determine how AI systems handle personal data, assessing compliance with data protection standards, including GDPR. The assessment helps identify privacy risks and implement safeguards to protect user information.

Privacy-Enhanced AI

AI designed to respect user privacy by limiting data access, ensuring confidentiality, and allowing individuals to control their data. Privacy-enhanced AI promotes autonomy and reduces risks of privacy intrusion.

Prohibited AI systems

AI systems whose applications are banned within the EU due to their capacity to harm fundamental rights, including systems using subliminal techniques, social scoring, or remote biometric identification in public spaces for surveillance purposes.

Provider’s Quality Management System

Internal processes established by providers to maintain compliance, quality, and safety standards throughout an AI system’s lifecycle, including during development, testing, and post-market monitoring.

Quantitative Bias Testing

A method for analyzing statistical biases in AI model outputs, ensuring they do not disproportionately affect specific groups. Quantitative bias testing supports fairness and compliance with anti-discrimination standards.

Quantitative Risk Assessment

A structured process quantifying potential risks posed by AI systems, evaluating the likelihood and severity of adverse impacts. This assessment informs risk mitigation strategies and promotes responsible AI deployment.

Real-Time AI Systems

AI systems capable of generating instant or near-instantaneous responses, often in high-stakes scenarios. These systems must be robust, secure, and reliable to mitigate risks associated with time-sensitive applications.

Regulatory Compliance

The requirement for AI systems to adhere to applicable EU regulations governing safety, transparency, and user protection. Regulatory compliance ensures AI systems meet legal standards for operation within the European market.

Remote Biometric Identification System

AI systems used to identify individuals from a distance, often in public spaces, by comparing live biometric data to reference databases, with stringent requirements for transparency, data protection, and authorization.

Responsible AI

The development and use of AI systems that prioritize ethical principles, transparency, and human rights. Responsible AI practices minimize harm, enhance accountability, and support compliance with regulatory standards.

Risk Assessment Frameworks

Structured methodologies for evaluating and mitigating risks associated with AI systems, ensuring that systems operate within acceptable risk levels and are equipped to handle potential adverse outcomes.

Risk Management System

A continuous, lifecycle-wide process to identify, assess, and mitigate risks posed by AI systems, ensuring their safe, ethical, and compliant operation within their intended use environments.

Risk Tolerance

The level of risk an organization or stakeholder is prepared to accept to achieve objectives. In AI, risk tolerance varies based on legal requirements, organizational priorities, and the specific context of system deployment.

Socio-Technical Factors

The interplay of human, organizational, and technical influences on AI system design, development, and deployment. Socio-technical factors shape AI risks and benefits, impacting fairness, interpretability, and system acceptance.