AI Cards: Towards an Applied Framework for Machine-Readable AI and Risk Documentation Inspired by the EU AI Act

Conference
to be presented Annual Privacy Forum (APF)
Delaram Golpayegani* , Isabelle Hupont , Cecilia Panigutti , Harshvardhan J. Pandit , Sven Schade , Declan O'Sullivan , Dave Lewis
Description: AI Cards is a novel holistic framework for representing intended uses of AI systems by encompassing information regarding technical specifications, context of use, and risk management, both in human- and machine-readable formats.
published version 🔓open-access archives: arXiv , harshp.com , OSF
📦resources: repo , AI Cards: Towards an Applied Framework for Machine-Readable AI and Risk Documentation Inspired by the EU AI Act

Affiliations:
Delaram Golpayegani, Declan O'Sullivan, Dave Lewis - ADAPT Centre, Trinity College Dublin, Dublin, Ireland
Isabelle Hupont, Cecilia Panigutti, Sven Schade - European Commission, Joint Research Centre (JRC), Ispra, Italy
Harshvardhan J. Pandit - ADAPT Centre, Dublin City University, Dublin, Ireland [email protected]

Abstract

Introduction

Artificial Intelligence (AI) has become a centrepiece of reshaping many aspects of individual life, society, and public and private sectors for the better [1][3]. Yet, concerns have been raised regarding the ethical implications and potential risks associated with the use of AI, such as biases in decision-making algorithms [4], [5] and privacy issues [6], [7]. These concerns have prompted a global wave of trustworthy AI guidelines [8], [9] and some preliminary regulatory efforts aimed at mitigating the potential harms of AI, ensuring its development and use are aligned with safety standards and respect human rights and freedoms. The European Union’s (EU) AI Act [10]1 is the first comprehensive legal framework on AI, through which AI systems are subjected to a set of regulatory obligations according to their risk level—that can be, from highest to lowest, unacceptable, high-risk, limited-risk, or minimal-risk.

With the risk level being the yardstick for determining the regulatory requirements that AI systems need to satisfy, the AI Act is intent on promoting secure, trustworthy, and ethical use of AI, with a particular emphasis on management and mitigation of potential harms that high-risk AI systems may cause. The risk management system provisions, set in Art. 9, thereby play a pivotal role in the implementation of the Act. Effective AI risk management requires information regarding the AI system, its incorporating components, its context of use, and its potential risks to be maintained and communicated in the form of documented information. While maintaining and sharing information regarding AI systems and their risks promote transparency and in turn trustworthiness [11], it is additionally a legal obligation for providers of high-risk AI systems to draw up technical documentation to demonstrate compliance with the Act’s requirements (Art. 11). The elements of technical documentation, which includes risk management system, are described in Annex IV at a high-level. However, to serve its purpose technical documentation needs to be extensive and detailed. Being limited to defining essential requirements, the AI Act relies on European harmonised standards for technical specifications to help with implementation and enforcement of its legal requirements, including technical documentation [12], [13]. Notably, the European Commission’s draft standardisation request [14] refrains from calling for European standards in relation to data, model, or AI system documentation [15]. Consequently, the European Standardisation Organisations’ work plan does not include standards specifically addressing these aspects [16]. In parallel, the existing body of work on AI, model, and data documentation approaches pays little attention to documentation for legal compliance, fails to provide a strong connection between the documented information and responsible AI implications, and does not take into consideration aspects related to risk management [11], [17].

Currently, there is a critical need for technical documentation to support and be in sync with the AI system development and usage practices. Moreover, with the involvement of several entities across the AI value chain, the documentation must be generated and maintained in a manner that ensures consistency and interoperability. This is also crucial for investigation of technical documentation by AI auditors and conformity assessment bodies for compliance checking and certification. The motivation for this work is therefore to address this current need for consistent, uniform, and interoperable specifications to support effective implementation of the AI Act. This paper aims to address the current lack of unified technical documentation practices aligned with the AI Act, with a threefold contribution:

  1. we provide an in-depth analysis of the provisions of the EU AI Act in regard to documentation with a focus on technical documentation and risk management system documentation (Section 3);

  2. we propose AI Cards—a novel framework providing a human-readable overview of a given use of an AI system and its risks (Section 4); and

  3. we present a machine-readable specification for AI Cards that enables generating, maintaining, and updating documentation in sync with AI development. It further supports querying and sharing information needed for tasks such as compliance checking, comparing multiple AI specifications for purposes such as AI procurement, and exchanging information across the value chain (Section 4.3).

We demonstrate a practical application of AI Cards using an illustrative example of an AI-based proctoring system (Section 5), validate its usefulness through a survey (Section 6), and discuss its benefits (Section 7).

Related Work

Previous Studies on the AI Act’s Documentation and Risk Management Requirements

Since the release of the AI Act’s proposal by the European Commission in April 2021, the research community, international organisations, and industrial actors have been exploring the new avenues it opened by analysing the AI Act’s contents for suggesting clarifications, identifying potential gaps, or giving critique. In this section, we refer to some existing studies on the provisions of the AI Act in regard to documentation and risk management.

Panigutti et al. explore the role of the AI Act’s transparency and documentation requirements in addressing the opacity of high-risk AI systems [18]. Gyevnara et al. discuss compliance-oriented transparency required to satisfy the AI Act’s requirements in regard to risk and quality management systems [19]. Schuett presents a comprehensive analysis of the AI Act’s risk management provisions, without delving into the details of risk management system documentation [20]. Soler et al.’s assessment of international standards in regard to addressing AI Act’s requirements shows that these standards are insufficient to meet the provisions of the AI Act related to risk management [16].

In regard to technical documentation (Art. 11 and Annex IV), the most comprehensive studies, to the best of our knowledge, are the work of Hupont et al. [11], which identifies 20 information elements needed in documentation of AI systems and their constituting datasets, and the analysis of Annex IV provided by Golpayegani et al. [21], which identifies 50 information elements. Nevertheless, both studies remain short in being fully comprehensive in terms of covering both technical and risk management system documentation requirements laid down in Art. 11, Annex IV, and Art. 9. Building on our previous work, presented in [21] and [11], this paper provides an in-depth analysis of technical documentation with a focus on risk management system documentation.

Alignment of Existing Documentation Practices with the AI Act

Compliance with the AI Act’s documentation requirements requires guidelines, standards, tools, and formats to assist with generation, communication, and auditing of technical documentation. With transparency being widely recognised as a key factor in implementing trustworthy AI [9], several documentation frameworks proposed by the AI community, such as Datasheets for Datasets [22] and Model Cards [23], have become de-facto practices. As these documentation approaches are widely-adopted, a key question is the extent to which they could be leveraged for regulatory compliance tasks. This question is investigated by the studies mentioned in the following. Pistilli et al. discuss the potential of Model Cards as a compliance tool and anticipate its adoption—among other existing documentation practices originated from the AI community—for compliance with the AI Act’s documentation obligations [24]. The work by Hupont et al. [11], which investigates the 6 most widely-used AI and data documentation approaches for their alignment with the implementation of documentation provisions, concludes that AI Factsheets [25] offers a higher overall degree of information coverage, followed by Model Cards and the AI Classification Framework proposed by the Organization for Economic Cooperation and Development (OECD) [8]. The research also demonstrates that while data-related information elements are well-covered by most documentation approaches, particularly Datasheets for Datasets [22], the Dataset Nutrition Label [26], and the Accountability for Machine Learning framework [27], they still do not cover technical information needs related to AI systems. This finding is further strengthened in a follow-up comparative analysis of 36 AI system, model, and/or dataset documentation practices, which unravels the overall alignment of documentation practices with transparency requirements of the EU AI Act and other recent EU data and AI initiatives, and spots a gap in representing the information related to AI systems in its entirety and its context of use [28]. Interestingly, to date, the only documentation methodology conceived from design for the AI Act is Use Case Cards [29], which primarily focuses on documenting the intended use of AI systems to assess their risk level as per the AI Act, without mentioning technical information and risk management details.

Machine-Readable Data, Model, and AI System Documentation

While documentation approaches are acknowledged as instruments for improving transparency, and in turn enhancing trustworthiness, there has been little attention to the barriers in generation, maintenance, assessment, and exchange of conventional text-based documentation. Providing machine-readable specifications is an idea taken up by some recent work to support adaptable and interoperable documentation. In the area of data documentation, Open Datasheets2, proposed by Roman et al., provides a metadata framework for documenting open datasets in a machine-readable manner [30]. For machine-understandable documentation of machine learning models, Model Card Report Ontology (MCRO)3, developed by Amith et al., offers the metadata for the content of Model Card reports [31]. Linked Model and Data Cards (LMDC) present a schema for integration of Model and Data Cards in a data space to provide a holistic view of a model or an AI service [32]. The Realising Accountable Intelligent Systems (RAInS) ontology4, created by Naja et al., models the information relevant to AI accountability traces [33].

In the context of documentation for regulatory compliance, the W3C Data Privacy Vocabulary (DPV) [34]5 is developed based on the requirements of the EU GDPR (General Data Protection Regulation) and has been applied for representing Data Protection Impact Assessment (DPIA) information [35], documenting data breach reports [36], and representing Register of Processing Activities (ROPA) [37]. In our previous work, we proposed the AI Risk Ontology (AIRO)6 for describing AI systems and their risks based on the AI Act and ISO 31000 family of standards [21]. In this work, AIRO is employed to provide the basis for the machine-readable representation of the AI Cards. However, it is important to note that ISO 31000 family of standards is not aligned with the AI Act due to some fundamental differences [16]. AIRO is used while awaiting future standards to be developed in response to the European Commission’s standardisation request. Therefore, it needs to be updated with future publication of harmonised standards related to AI risk management.

Provisions of the EU AI Act Regarding Documentation of AI Systems

Documentation of AI systems is highly tied to the principles of transparency and accountability [29], [33], providing an appropriate degree of information to different stakeholders to enable several types of assessments. In a similar vein, documentation is essential for assessing legal compliance. Figure 1 shows the AI Act’s documentation requirements for high-risk AI systems and illustrates how they are related to each other. Within the AI Act, technical documentation and quality management system documentation are key documents in the conformity assessment procedure, containing detailed information required for self- and third-party assessments. Given the central role of technical documentation and presence of related guidance on its content (Annex IV), this work focuses on identifying its concrete information elements. However, among its incorporating documents, our analysis expands on risk management system documentation considering that the risk-centred nature of the AI Act makes the risk management system requirement (Art. 9) taking the lead in ensuring that the potential risks of high-risk AI systems are reduced to an acceptable level. In the following, we focus on identifying the information elements required to be featured in technical (Section 3.1) and risk management (Section 3.2) documentation.

High-risk AI systems’ documentation requirements as per the AI Act.

Article 11 and Annex IV - Technical Documentation

As discussed earlier, technical documentation is a key artefact for demonstrating compliance with the AI Act. According to Art. 11, it is the high-risk AI provider who should draw up the technical documentation ex-ante and keep it updated. From this obligation, three critical challenges arise: (i) when development of AI systems includes integrating third-party components into the system, e.g. training data and pre-trained models, high-risk AI providers need to dispense details regarding these components in technical documentation, e.g. information about how data is labelled (Annex IV(2)(d)). Thus, generating technical documentation to fulfill the requirements of the AI Act can become a collective activity, which requires communication and exchange of often sensitive and confidential information. This opens up a myriad of questions concerning trust, accountability, and liability, that lie beyond the scope of this discussion. (ii) Real-world testing environments and sandboxes assist in determining what might be difficult to ascertain ex-ante. However, specificities of the context of use, which significantly influence AI risks and impacts, might not be revealed until the system’s deployment. This suggests a need for ex-ante involvement of potential users and ex-post feedback loops. (iii) Finally, keeping technical documentation up-to-date and re-assessing legal compliance are hard-to-manage tasks. A key question here is “what changes do trigger the update process?”. A conspicuous, albeit not comprehensive, answer is substantial modifications—changes that affect the system’s compliance with the Act or its intended purpose (Art. 3(23)). However, lack of clarity regarding what modifications are deemed as substantial adds to the complexities of this challenge.

To help with the generation and auditing of technical documentation, the minimum set of its incorporating information elements is outlined in Annex IV, which is subject to the European Commission’s potential amendments (Art. 11(3)). Annex IV serves as a primary template wherein information elements are described with varying degrees of detail, with the majority articulated at a high-level. This implies a need for harmonised standards to support the implementation of Art. 11. However, as mentioned earlier, to the best of our knowledge there is no ongoing standardisation activities to address this need. Aiming to provide clarification on the content of technical documentation, we conducted an in-depth analysis of Annex IV, with a particular emphasis on the risk management system. The analysis is an initial step in establishing a collective understanding of the technical documentation’s content and could inspire the development of prospective European standards, guidelines, and templates.

Article 9 - Risk Management System Documentation

The AI Act will rely on harmonised standards to operationalise legal requirements at the technical level, and risk management system will be no exception. At the time of writing, however, European Standardisation Organisations are still working on deliverables planned in response to the European Commission’s standardisation request. Furthermore, existing ISO/IEC standards on AI risk management and AI management system, namely ISO/IEC 238947 and ISO/IEC 420018, have been found insufficient in meeting the requirements of the AI Act [16].

Indeed, within the current AI standardisation realm, the two main international standards covering AI risk management aspects that have been analysed in regard to AI risk management in the technical report published by the European Commission’s Joint Research Centre [16] are: (i) ISO/IEC 42001 “Information technology — Artificial intelligence — Management system” that offers a framework to assist with implementation of AI management systems, and (ii) ISO/IEC 23894 “Artificial intelligence — Guidance on risk management”, which offers practical guidance on managing risks, albeit defined as the effect of uncertainty, without relation to potential harms to individuals. Due to this and other reasons, the requirements defined in these standards are not useful to comply with the provisions of the AI Act in regard to risk management [16], however they may be useful to identify relevant technical information elements for the development of AI Cards while we await harmonised standards for the AI Act. In this section, we elaborate on information elements for the documentation of an AI risk management system by obtaining insights from these two standards.

We used the overall structure of AI management systems, which follows the ISO/IEC’s harmonised structure for management system standards9, as a resource for extracting the key organisational activities within an AI risk management system. For each activity, documentation needs, in terms of information elements, were identified from both ISO/IEC 42001 and 23894 (for a summary of the analysis see Figure 2). These information elements can be categorised into four overall groups:

  • Information about the context of the AI system and the organisation, for example, the AI system’s intended purpose and the role of the organisation in relation to the system.

  • Details of the risk management system in place, e.g. the policies, responsibilities, and resources required for implementation of the risk management system itself. This category of information is relevant to the ISO/IEC 23894’s AI risk management framework, whose intention is to help with AI governance and integration of AI risk management activities into an organisation’s existing processes.

  • Documentation of risk management processes across different phases: planning (ex-ante), operation (ongoing), and post-operation (ex-post).

  • Results of AI risk management, which can be represented in artefacts produced throughout the risk management process, e.g. risk assessment documentation that lists AI risks identified, their likelihood, severity, sources, consequences, and impacts.

Information elements identified for AI risk management system documentation (note they are inspired by ISO/IEC 42001 and 23894, and therefore will be updated according to the AI Act’s future official harmonised standards).

AI Cards

Extensiveness of technical documentation and confidentiality of its content may hinder collaboration and communication with AI stakeholders—many of whom are not necessarily technical AI experts. To address this challenge, we propose the AI Cards framework, as a structured information sheet, which offers an overview of a use of an AI system and its related trustworthy AI concerns by inclusion of crucial information about technical specification, context of use, risks, and compliance. With an intuitive representation, the AI Cards framework introduces a simple, transparent, and comprehensible summary card laying out a holistic picture of a specific AI use case and its risks without disclosing sensitive information. The framework encompasses machine-readable specifications that enhance interoperability, enable automation, and support extensibility.

Development Process

In shaping the framework, we first specified its requirements, which are summarised in Table 1. Based on the defined scope, we re-examined our analysis of Annex IV. The information elements within the scope were considered as candidates for inclusion in the AI Cards. However, many of them were too detailed to be represented in a summary card. Therefore, we propose condensed views for these information elements through visual aids.

The AI Cards framework was defined and iteratively refined in consultation with researchers involved in digital policymaking within the European Commission and further validated by conducting an online anonymous survey with law and technology researchers (see Section 6 for more details).

AI Cards framework requirements
Purpose Providing an overview of technical documentation that effectively conveys key information
Scope

In the scope: AI system, as a whole, its context of use, and information relevant to trustworthy AI concerns including risk management system.

Out of the scope: organisational processes, details of management systems, and documentation of AI components, including data and model. The framework is horizontal and does not take sector-specific nuances into account.

Key audience AI users, end-users, subjects, providers, developers, auditors, and policymakers
Representation formats An easy-to-understand visual human-readable representation accompanied with a machine-readable specification

Information Elements

Figure 3 shows a visualisation of the AI Card which condenses the information elements in 9 sections. To enable representing AI Cards in a time series and to allow version control required to reflect the evolving nature of AI systems, the Card’s metadata includes essential information describing its version, issuance date, language, publisher, and contact information. Further, the URL of the machine-readable specification is included.

Visual representation of AI Cards.

(1) General Information

This section provides the key information about the AI system including its version, modality, e.g. standalone software or safety component of a product, main AI techniques used, provider(s), and developer(s).

(2) Intended Use

The AI Act put a considerable emphasis on intended use of the system, given its profound effect on risks and impacts. For describing the intended use, we propose using the combination of the 5 concepts identified in [15], which are:

  • the domain in which the system is intended to be used within,

  • the purpose, i.e. end goal, of using the system in a specific context,

  • the AI capability that enables realisation of the purpose,

  • the AI deployer which is the entity using the AI system,

  • the AI subject which is the entity subjected to the outputs, e.g. decisions, generated by the AI system.

Noteworthy, the subsequent sections heavily rely of the information represented in this section, considering that they should be defined in the view of the intended use.

(3) Key Components

Many AI systems are built through integration of multiple AI models, datasets, general-purpose AI systems, and other software elements, each of which has an effect on the system’s behaviour and in turn its risks [28]. This section provides the system’s high-level architecture in terms of incorporating components. For each component, its name, version, and link to documentation or ID are presented. For detailed information about components, we rely on their documentation provided presumably by the components’ providers in the form of information sheets, e.g. Datasheets, Model Cards, and AI Factsheets.

(4) Data Processing

AI systems that do not process data, if they exist at all, are rare. Within the EU digital acquis, the GDPR [38], which protects the fundamental right to privacy by regulating personal data, is applicable to those AI systems that process natural persons’ data. Therefore, having knowledge of whether a given use of an AI system involves processing of personal data is crucial to correctly interpret the resulting legal compliance obligations. This section specifies inclusion of processing personal data and shows the category of the data, e.g. biometric data, and represents whether data protection impact assessment (DPIA) is conducted. It further specifies inclusion of non-personal, anonymised, and licenced data.

(5) Human Involvement

Involvement can take different forms depending on the phase of AI development, the role of human actors, and the system’s level of automation—which has a range from fully autonomous to fully human-controlled according to ISO/IEC 2298910. The level of automation also has a substantial effect on the safeguards, including human oversight measures, required for controlling AI risks. This section provides an overview of involvement of two specific human actors: AI end-users, who use the AI system’s output and AI subjects, who are subjected to the outputs of the system. For these actors, we look into the following aspects of involvement:

  • Intended involvement: represents whether the involvement of a specific actor is as intended. An example of an intended AI subject in an AI-based proctoring system is a student sitting an online test. In this case, other persons present in the room are unintended AI subjects.

  • Active involvement: shows whether a specific actor actively interacts with the AI system.

  • Informed involvement: represents whether a specific actor was informed that an AI system is in place; for example, in cases where a decision affecting a person’s education is made using AI-based solutions.

  • Control over AI outputs: shows the extent to which AI subjects and end-users have control over AI outputs, in particular decisions made by the AI system and their impacts. Inspired by OECD’s modes of operationality [8], we consider the following six levels of control:

    • An AI subject/end-user can opt in the system’s output.

    • An AI subject/end-user can opt out of the system’s output.

    • An AI subject/end-user can challenge the system’s output.

    • An AI subject/end-user can correct the system’s output.

    • An AI subject/end-user can reverse the system’s output ex-post.

    • An AI subject/end-user cannot opt out of the system’s output.

(6) Risk Profile

This section provides a high-level summary of risk management results, which includes an overview of likelihood, severity and residual risk associated with risks that have impact on areas of health and safety, fundamental rights, society, and environment. Further, this section shows whether any technical, monitoring (human oversight), cybersecurity, transparency, and logging measures applied to control the risks.

(7) Quality

With many AI incidents rooted in poor quality, ensuring that the system is of high quality by participating in AI regulatory sandboxes, benchmarking, and performing tests is necessary. We propose illustrating key AI system qualities using a radar chart. We recommend including Accuracy, robustness, and cybersecurity, which are the key qualities explicitly mentioned in the Act (Art. 15). Further, the relevant product quality and quality in use, described by ISO/IEC 25059 on AI SQuaRE (Systems and software Quality Requirements and Evaluation)11 can be considered to be included in the quality section.

(8) Pre-determined Changes

This section provides a list of pre-determined changes to the system and its performance in terms of subject and frequency of change as well as the potential impacts of change on performance and risks.

(9) Regulations & Certification

This section lists the main digital regulations the AI system is compliant with, key standards to which the system or the provider(s) conform, and codes of conduct followed in development or use of the AI system.

Machine-Readable Specification

The visual representation assists stakeholders in gaining an understanding of an AI system, its context, and the associated trustworthy AI concerns without delving into the extensive details of technical documentation. However, these are not sufficient to support some of the desirable features of documentation, including search and tracking capabilities, conducting meta-analysis, comparing multiple AI systems, and automating generation and update of documentation [17]. To include these features, the AI Cards framework supports machine-readable representation of information by leveraging the standards, methods, and tools provided by the World Wide Web Consortium (W3C)12, motivated by the body of work discussed in Section 2.3 as well as the rise in uptake of open data formats for documentation, reporting, and sharing information by the AI community, e.g. HuggingFace’s use of JSON Model Cards13, and by the EU, e.g. machine-readable regulatory reporting [39], DCAT-AP open data portals14, and EU vocabularies and ontologies15. The machine-readable representation, which relies on Semantic Web technologies, fosters openness and interoperability—both essential for exchanging information across the AI value chain. This representation is extensible and therefore enables accommodating sector-specific information requirements and allows adaptation to highly-anticipated guidelines from authorities including the AI office, the potential amendments to the Act via delegated Acts, and case law. Grounded on formal logic, it also assists in ensuring that the information is complete, correct, and verifiable.

For provision of machine-readable specifications for AI Cards, an ontology is needed to provide a common semantic model that explicitly represents terms, their definitions, and the semantic relations between them, ensuring consistency and interoperability. For this, building upon our previous work [21], we further extend AIRO to support modelling information elements featured within the AI Cards. This extension borrows concepts and relations from Data Quality Vocabulary (DQV)16 [40] for expressing AI quality and Data Privacy Vocabulary (DPV)17 [34] for modelling personal data processing. Figure 4 depicts key classes and relations required for representing information featured in each section of the AI Card.

An extension of AIRO for AI Cards.

With the ontology providing the schema, the information incorporated in the AI Cards can be represented as an RDF graph. This opens the door to leveraging the power of Semantic Web standards and technologies across a variety of tasks, among them are:

  • Querying to retrieve information about an AI system and information related to demonstration and investigation of legal compliance, for example verifying presence of measures to address potential AI impacts on the fundamental right to non-discrimination in support of compliance with Art. 9 of the AI Act. The information retrieval is implemented using SELECT queries in the SPARQL18, which the W3C standardised query language, as shown in [21].

  • Updating information about an AI application, for example when a mitigation measure is no longer effective and is replaced with a new measure, which is implemented using DELETE/INSERT SPARQL queries.

  • Integrating information provided in documentation of incorporating components (demonstrated in the “Key Componets” section of the AI Card), for example a Model Card that documents a system’s incorporating model, when provided in standardised linked data formats such as JSON-LD or RDF.

  • Reasoning that might support conformity assessment and legal compliance tasks, for example defining semantic rules to suggest the regulatory risk profile associated with an AI application. Semantic rule-checking using the Shapes Constraint Language (SHACL)19, which is a W3C standardised language, is valuable in validation of information against a set of rules (see prior work on determining high-risk AI applications using SHACL [15]).

  • Expressing intended use policies in agreements between different parties in the AI value chain, for example agreements between AI providers and deployers describing conditions of using or modifying an AI application. In this, the Open Digital Rights Language (ODRL)20—the W3C recommended language for representing policies—can be used for describing intended and precluded uses as permission and prohibition statements respectively.

Proof-of-Concept: AI Cards for an AI-Based Proctoring System

To demonstrate the potential scalability, and applicability of the proposed framework, and inspired by the use cases described in [18] and [29], we take an AI-based student proctoring tool called Proctify as an illustrative proof-of-concept. It is important to note that this system is merely selected to demonstrate the applicability of the AI Cards framework in real-world cases. The AI Act generally considers such proctoring systems as high-risk in the education domain when used for monitoring and detecting suspicious behaviour of students during tests [7].

Proctify is intended to detect suspicious behaviour during online exams by analysing facial behaviour from a student’s facial video captured throughout the exam using a webcam. Prior to this, students have explicitly consented to be recorded during the exam and informed that they must be alone in the room. The system incorporates a graphic interface displaying an analysis of the student’s face including the head pose, gaze direction, and face landmarks’ positions. This extracted information is then provided as an input to SusBehavedModel, which has been trained in-house by the system’s provider using SusBehavedDataset, to determine whether the student is displaying suspicious behaviour, e.g. looking away from the screen, leaving the room, or a third person detected in the room. Detection of suspicious behaviour raises an alarm in the interface to inform and let the human oversight actors, e.g. human instructors, take appropriate actions, e.g. communicating with the student.

Throughout the risk management process, the risks and impacts of the system are identified and assessed by the provider, including the following: the system may have lower accuracy for students with darker skin tones and a higher rate of false-positive alarms for students wearing glasses. Further, false-negatives and false-positives are more frequent for students with health issues or disabilities that affect their facial behaviour. There is also a chance of over-reliance of human instructors on the system’s output (automation bias). These events have the potential to negatively impact students’ mental health, future career, and their rights to dignity and non-discrimination. Some of the measures applied to address the system’s risks and impacts are: ensuring the dataset is representative and diverse in demographic terms, conducting rigours and frequent testing of accuracy, assigning expert human proctors, and creating clear protocols to act upon when an alarm is raised. The Proctify’s AI Card is visualised in Figure 5 and its machine-readable specification is available online21. As shown in the figure, the visual representation of the AI Card provides a summary of risk management information without disclosing the details. However, the machine-readable specification is capable of modelling the details.

An example of AI Card for an AI-based proctoring system.

Validation

As mentioned earlier, to identify essential information elements of the AI Cards framework, we consulted experts involved in EU digital policymaking, in particular the AI Act, whose backgrounds are in AI transparency and documentation, explainability, cybersecurity, standardisation, and Semantic Web. Throughout the development process, our analysis of the AI Act’s provisions, the visual representation of AI Cards, and its machine-readable specification were discussed to ensure alignment with EU digital policies, adoption of correct terminology, and suitability of the AI Cards framework for addressing common concerns of AI stakeholders. After reaching a solid structure for the AI Cards, an online anonymous survey has been conducted to assess the usefulness of the AI Cards framework.

Survey Design

In addition to questions regarding the respondent’s background, the survey includes questions about usefulness of (i) (only) the visual representation, (ii) (only) the machine-readable representation, and (iii) the overall AI Cards framework (both human- and machine-readable representations). In the latter, to assess how stakeholders envisage usability of the AI Card the System Usability Scale (SUS) [41] is used. As the SUS is originally designed for assessing perceived usability, minor wording changes were applied to fit it with the purpose of AI Cards’ assessment and to provide more clarity. Table 2 shows the list of SUS-based questions for assessing the AI Cards’ usability, each answered on a 5-point Likert scale ranging from 1 (strongly disagree) to 5 (strongly agree).

Recruitment

In the first phase, a student cohort of participants (N=23) were recruited from the Data Governance module in Dublin City University attended by students enrolled in Masters of Art (MA) in Data Protection and Privacy (MDPP) and the European Master in Law, Data and Artificial Intelligence (EMILDAI). It is important to highlight that, because of the nature of this module, some of the participants (30 percent) have notable positions in public organisations, industry, and NGOs. It should be noted that a second phase of evaluation with policymakers and industry stakeholders is currently underway.

Preliminaries

Prior to carrying out the survey, it was reviewed and approved by the ethics committee within the school of computer science and statistics of Trinity College Dublin. To start, in a preliminary interactive session, information about the AI Act and the AI Cards was presented to participants and they were asked to sign the informed consent form. Then, they could start completing the survey. It should be noted that the participation in the survey was optional.

Findings

23 responses were collected in this first phase of the evaluation from the student cohort. Regarding the usefulness of the visual representation, inclusion of the following information was suggested in the survey results:

  • information related to the registry process,

  • link to the GDPR compliance including summary of the requirements for data processing and retention period of data if personal data processed,

  • limitations of the system,

  • information regarding tests including tests, bugs, issues, and date of testing,

  • a legend or key to explain the abbreviations.

In the context of implementation and enforcement of the AI Act, the participants agreed on the usefulness of the human-readable representation in compliance checking, creating technical and risk management system documentation, and creating guidelines. The participants further referred to the following potential uses of the AI Cards: ensuring traceability and transparency, monitoring the system, reviewing periods for pre-determined changes, raising awareness, and explaining the core concepts to stakeholders with different levels of AI literacy and technical knowledge. In regard to usefulness of AI Cards as a tool to facilitate exchange of information regarding an AI system and its risks, while all participants acknowledged its usefulness to some extent for communication within the AI development ecosystem and with authorities, a minority (4 out of 23) did not perceive it valuable for communication with the public.

Regarding machine-readable specifications, there was general consensus on its usefulness for automated tools, establishing a common language, and structuring the EU registry of AI systems. Further, the participants indicated that the machine-readable specifications could assist in determining the legal risk category, providing a better understanding of models, and building models with better compliance capabilities.

Regarding the potential target users of the framework, one participant referred to “AI Brokers for comparing AI Systems - would require more metrics but a good start for comparison” and another participant mentioned “Members of the public, in fostering awareness and understanding of AI systems”. The results of SUS usability test are shown in Table 2, where due to the mixed tone of the items a higher score for odd-numbered items and a lower score for even-numbered items are desired. The average SUS score, calculated using the formula proposed in [42], is 66.30.

SUS questions used for evaluating usability of AI Cards
No. Question Score (mean)
1 I think that AI stakeholders would like to use the AI Cards framework frequently. 4.21
2 I find the AI Cards framework unnecessarily complex. 2.13
3 I think the AI Cards framework is easy to use. 3.96
4 I think that AI stakeholders would need the support of a technical person to be able to use the AI Cards framework. 3.35
5 I find the various aspects (i.e. information elements and human- and machine-readable formats) in the AI Cards framework are well integrated. 4.22
6 I think there is too much inconsistency in the AI Cards framework. 1.83
7 I would imagine that most people would learn to use the AI Cards framework very quickly. 3.91
8 I find the AI Cards framework very cumbersome to use. 2.26
9 I feel very confident using the AI Cards framework. 3.43
10 I need to learn a lot of things before I could get going with the AI Cards framework. 3.30

Another key aspect in evaluation of effectiveness and usefulness of an AI documentation framework is the extent of its adoption by the AI community, as evidenced in widely-adopted documentation approaches, e.g. Datasheets [22], Model Cards [23], and Factsheets [25]. Given that the AI Act was only adopted recently and the final text has not been published yet, the extent of AI Cards’ adoption will need to be assessed over time. As indicated by one the participants, “The real world application would be interesting to see and may require changing based on different Member State authorities requiring other information be made available for documentation purposes”.

Benefits and Potential Applications

The AI Cards framework distinguishes itself from existing documentation approaches through:

  • Its alignment with the EU AI Act’s provisions in regard to technical and risk management system documentation. It should be noted that though the framework is designed based on the requirements of the EU AI Act, its modular and semantic nature makes it scalable to be applied to document a wide range of AI applications, regardless of the jurisdiction they are used in.

  • Its dual approach towards information representation, which makes the framework accessible to both humans and machines, facilitates communication, promotes interoperability, and enables automation.

  • Its holistic AI system-focused approach, which allows inclusion of information regarding the context of use, risk management, and compliance in addition to technical specifications. While most existing documentation practices address potential risks or ethical issues at a very high and unstructured level [11], an AI Card draws a picture of the risk management system that is already established by illustrating an overview of identified risks and applied mitigation measures.

  • Its modular and future-proof design that enables reconfiguration and extension of the framework to address documentation requirements arising from forthcoming AI regulations, policies, and standards.

AI Cards’ machine-readable specification adds an extra level of potential for application. In terms of documentation management, this specification facilitates frequent modification, version control, and integration of data from different sources. Further, it lays the ground for development of supporting tools and RegTech solutions. Within the context of the EU AI Act, these supporting tools can be employed to assist AI providers in documentation and exchange of information required for compliance, help conformity assessment bodies in tasks related to auditing and certification, and assist AI deployers in conducting fundamental rights impact assessment (FRIA). Further, AI Cards might become useful for the AI Office in the development of automated Semantic Web-based tools to assist with FRIA, which has an overlap with technical documentation in terms of information requirements. Currently, there is little guidance for how to conduct risk assessment on the impact of AI systems on fundamental rights and the state of the art in taking fundamental rights seriously in the context of AI technical standardisation is new and fast evolving. Therefore, the locating, consumption and reuse of examples of AI risks assessment and accompanying technical documentation is urgently needed to establish a public knowledge base from which AI providers and deployers, especially less well-resourced SMEs and public bodies, can draw legal certainty. Such a sharing of AI risks and the documentation in AI card may therefore accelerate regulatory learning both between regulatory bodies and between prospective AI providers and deployers. Where confidentiality concern may impede the open access sharing of such information, official regulatory learning structures such as sandboxes and testing with human subjects could use this mechanism to maximise sharing and learning between participating stakeholders. Shared searchable repositories of AI risks and corresponding AI Cards may also offer support in identifying risks of reasonably foreseeable misuse of AI systems.

The semantic model, provided as an open-source ontology, can function as a basis for a pan-European AI vocabulary22 to help establish a common language across different and multidisciplinary actors involved in the AI ecosystem. This consistency in the language, accompanied by a machine-readable representation, promotes interoperability and streamlines the information exchange required for incident reporting, compliance checking, and sharing best practices. It also promotes broader participation of stakeholders needed in development of standards for the AI Act [43]. In addition, it can be utilised for structuring the EU database of high-risk AI systems (Art. 60) and incident reports (Art. 62). As an added advantage, the semantic model can further be improved to evolve into a multilingual ontology supporting official EU languages.

In a broader context, the semantic model is helpful for AI providers and users that operate in different jurisdictions in addressing challenges of cross-border compliance and interoperability by providing an extensible and adaptable structure to maintain information. Provided as an open resource, the schema can be reused and enhanced by the AI community to fit their needs in regard to documentation and risk management.

Conclusion and Further Work

With this paper, we took an initial step to address the current lack of standardised and interoperable AI documentation practices in alignment with the EU AI Act. We proposed AI Cards, a novel framework for documentation of uses of AI systems in two complementary human- and machine-readable representations. We envision this contribution as a valuable input for future EU policymaking and standardisation efforts related to technical documentation, such as Implementing Acts, European Commission-issued guidelines, and European standards. We also hope that this work promotes use of standardised and interoperable Semantic Web-based formats in AI documentation.

In an ongoing effort to enhance AI Cards’ usability and wider applicability, an in-depth user experience study (second phase of the evaluation) involving AI practitioners, auditors, standardisation experts, and policymakers is underway. Our future work also takes the direction towards development of automated tools for generating AI Cards from users’ input and existing AI documents provided in textual formats. Further, alignment of AI Cards with documentation and reporting requirements of EU digital regulations, including the GDPR, Digital Services Act (DSA) [44], Interoperability Act [45], and Data Governance Act (DGA) [46], as well as well-known risk management frameworks, e.g. NIST Artificial Intelligence Risk Management Framework (AI RMF 1.0) [47], is a key future step in refining the AI Cards.


This project has received funding from the European Union’s Horizon 2020 research and innovation programme under the Marie Skłodowska-Curie grant agreement No 813497 (PROTECT ITN), as part of the ADAPT SFI Centre for Digital Media Technology is funded by Science Foundation Ireland through the SFI Research Centres Programme and is co-funded under the European Regional Development Fund (ERDF) through Grant#13/RC/2106_P2.


The views expressed in this article are purely those of the authors and should not, under any circumstances, be regarded as an official position of the European Commission.


References

[1]
R. Gruetzemacher and J. Whittlestone, “The transformative potential of artificial intelligence,” Futures, vol. 135, p. 102884, 2022.
[2]
A. Balahur et al., “Data quality requirements for inclusive, non-biased and trustworthy AI,” 2022.
[3]
G. Maragno, L. Tangi, L. Gastaldi, and M. Benedetti, “Exploring the factors, affordances and constraints outlining the implementation of artificial intelligence in public sector organizations,” International Journal of Information Management, vol. 73, p. 102686, 2023.
[4]
M. Miron, S. Tolan, E. Gómez, and C. Castillo, “Evaluating causes of algorithmic bias in juvenile criminal recidivism,” Artificial Intelligence and Law, vol. 29, no. 2, pp. 111–147, 2021.
[5]
T. Araujo, N. Helberger, S. Kruikemeier, and C. H. De Vreese, “In AI we trust? Perceptions about automated decision-making by artificial intelligence,” AI & society, vol. 35, pp. 611–623, 2020.
[6]
Y. Zhang, M. Wu, G. Y. Tian, G. Zhang, and J. Lu, “Ethics and privacy of artificial intelligence: Understandings from bibliometrics,” Knowledge-Based Systems, vol. 222, p. 106994, 2021.
[7]
I. Hupont, S. Tolan, H. Gunes, and E. Gómez, “The landscape of facial processing applications in the context of the european AI act and the development of trustworthy systems,” Scientific Reports, vol. 12, no. 1, p. 10688, 2022.
[8]
OECD, OECD Framework for Classification of AI Systems: a tool for effective AI policies.” Available: https://oecd.ai/en/classification, 2022.
[9]
European Commission and Directorate-General for Communications Networks, Content and Technology, Ethics guidelines for trustworthy AI. Publications Office, 2019 [Online]. Available: https://data.europa.eu/doi/10.2759/346720
[10]
“Proposal for a regulation of the european parliament and of the council laying down harmonised rules on artificial intelligence (artificial intelligence act) and amending certain union legislative acts) and amending certain union legislative acts.” European Commission, April 2021 [Online]. Available: https://eur-lex.europa.eu/legal-content/EN/TXT/?uri=CELEX:52021PC0206
[11]
I. Hupont, M. Micheli, B. Delipetrev, E. Gómez, and J. S. Garrido, “Documenting high-risk AI: A european regulatory perspective,” Computer, vol. 56, no. 5, pp. 18–27, 2023.
[12]
G. Mazzini and S. Scalzo, “The proposal for the artificial intelligence act: Considerations around some key concepts,” Camardi (a cura di), La via europea per l’Intelligenza artificiale, 2023.
[13]
A. Tartaro, “Towards european standards supporting the AI act: Alignment challenges on the path to trustworthy AI,” in Proceedings of the AISB convention, 2023, pp. 98–106.
[14]
E. Commission, “Draft standardisation request to the european standardisation organisations in support of safe and trustworthy artificial intelligence,” 2022. [Online]. Available: https://ec.europa.eu/docsroom/documents/52376. [Accessed: 07-Dec-2023]
[15]
D. Golpayegani, H. J. Pandit, and D. Lewis, “To be high-risk, or not to be—semantic specifications and implications of the AI act’s high-risk AI applications and harmonised standards,” in Proceedings of the 2023 ACM conference on fairness, accountability, and transparency, 2023, pp. 905–915.
[16]
J. Soler Garrido et al., “Analysis of the preliminary AI standardisation work plan in support of the AI act,” Joint Research Centre (Seville site), 2023.
[17]
A. K. Heger, L. B. Marquis, M. Vorvoreanu, H. Wallach, and J. Wortman Vaughan, “Understanding machine learning practitioners’ data documentation perceptions, needs, challenges, and desiderata,” Proceedings of the ACM on Human-Computer Interaction, vol. 6, no. CSCW2, pp. 1–29, 2022.
[18]
C. Panigutti et al., “The role of explainable AI in the context of the AI act,” in Proceedings of the 2023 ACM conference on fairness, accountability, and transparency, 2023, pp. 1139–1150.
[19]
B. Gyevnara, N. Fergusona, and B. Schaferb, “Get your act together: A comparative view on transparency in the AI act and technology,” arXiv preprint arXiv:2302.10766, 2023.
[20]
J. Schuett, “Risk management in the artificial intelligence act,” European Journal of Risk Regulation, pp. 1–19, 2023.
[21]
D. Golpayegani, H. J. Pandit, and D. Lewis, AIRO: An ontology for representing AI risks based on the proposed EU AI act and ISO risk management standards,” Towards a Knowledge-Aware AI, pp. 51–65, 2022.
[22]
T. Gebru et al., “Datasheets for datasets,” Communications of the ACM, vol. 64, no. 12, pp. 86–92, 2021.
[23]
M. Mitchell et al., “Model cards for model reporting,” in Proceedings of the conference on fairness, accountability, and transparency, 2019, pp. 220–229.
[24]
G. Pistilli, C. Muñoz Ferrandis, Y. Jernite, and M. Mitchell, “Stronger together: On the articulation of ethical charters, legal tools, and technical documentation in ML,” in Proceedings of the 2023 ACM conference on fairness, accountability, and transparency, 2023, pp. 343–354.
[25]
M. Arnold et al., FactSheets: Increasing trust in AI services through supplier’s declarations of conformity,” IBM Journal of Research and Development, vol. 63, no. 4/5, pp. 6–1, 2019.
[26]
S. Holland, A. Hosny, S. Newman, J. Joseph, and K. Chmielinski, “The dataset nutrition label,” Data Protection and Privacy, vol. 12, no. 12, p. 1, 2020.
[27]
B. Hutchinson et al., “Towards accountability for machine learning datasets: Practices from software engineering and infrastructure,” in ACM conference on fairness, accountability, and transparency, 2021, pp. 560–575.
[28]
M. Micheli, I. Hupont, B. Delipetrev, and J. Soler-Garrido, “The landscape of data and AI documentation approaches in the european policy context,” Ethics and Information Technology, vol. 25, no. 4, p. 56, 2023.
[29]
I. Hupont, D. Fernández-Llorca, S. Baldassarri, and E. Gómez, “Use case cards: A use case reporting framework inspired by the european AI act,” Ethics and Information Technology, vol. 26, no. 2, 2024.
[30]
A. C. Roman et al., “Open datasheets: Machine-readable documentation for open datasets and responsible AI assessments,” arXiv preprint arXiv:2312.06153, 2023.
[31]
M. T. Amith et al., “Toward a standard formal semantic representation of the model card report,” BMC bioinformatics, vol. 23, no. 6, pp. 1–18, 2022.
[32]
A. Donald et al., “Towards a semantic approach for linked dataspace, model and data cards,” in Companion proceedings of the ACM web conference 2023, 2023, pp. 1468–1473.
[33]
I. Naja, M. Markovic, P. Edwards, W. Pang, C. Cottrill, and R. Williams, “Using knowledge graphs to unlock practical collection, integration, and audit of AI accountability information,” IEEE Access, vol. 10, pp. 74383–74411, 2022.
[34]
H. J. Pandit, B. Esteves, G. P. Krog, P. Ryan, D. Golpayegani, and J. Flake, “Data privacy vocabulary (DPV)–version 2,” arXiv preprint arXiv:2404.13426, 2024.
[35]
H. J. Pandit, “A semantic specification for data protection impact assessments (DPIA),” in Towards a knowledge-aware AI: SEMANTiCS 2022—proceedings of the 18th international conference on semantic systems, 13-15 september 2022, vienna, austria, IOS Press, 2022, pp. 36–50.
[36]
H. J. Pandit, P. Ryan, G. P. Krog, M. Crane, and R. Brennan, “Towards a semantic specification for GDPR data breach reporting,” in Legal knowledge and information systems, IOS Press, 2023, pp. 131–136.
[37]
P. Ryan, R. Brennan, and H. J. Pandit, DPCat: Specification for an interoperable and machine-readable data processing catalogue based on GDPR,” Information, vol. 13, no. 5, p. 244, 2022.
[38]
“Regulation (EU) 2016/679 of the european parliament and of the council of 27 april 2016 on the protection of natural persons with regard to the processing of personal data and on the free movement of such data, and repealing directive 95/46/EC (general data protection regulation),” Official Journal of the European Union, vol. L119. 2016 [Online]. Available: http://eur-lex.europa.eu/legal-content/EN/TXT/?uri=OJ:L:2016:119:TOC
[39]
European Commission and Directorate-General for Financial Stability, Financial Services and Capital Markets Union, MRER proof of concept – assessing the feasibility of machine-readable and executable reporting for EMIR. Publications Office of the European Union, 2022.
[40]
R. Albertoni and A. Isaac, “Introducing the data quality vocabulary (DQV),” Semantic Web, vol. 12, no. 1, pp. 81–97, 2021.
[41]
J. Brooke, SUS-a ‘quick and dirty’ usability scale,” Usability evaluation in industry, pp. 189–194, 1996.
[42]
J. R. Lewis, “The system usability scale: Past, present, and future,” International Journal of Human–Computer Interaction, vol. 34, no. 7, pp. 577–590, 2018, doi: 10.1080/10447318.2018.1455307.
[43]
M. Veale and F. Zuiderveen Borgesius, “Demystifying the draft EU artificial intelligence act—analysing the good, the bad, and the unclear elements of the proposed approach,” Computer Law Review International, vol. 22, no. 4, pp. 97–112, 2021.
[44]
“Regulation (EU) 2022/2065 of the european parliament and of the council of 19 october 2022 on a single market for digital services and amending directive 2000/31/EC (digital services act).” October 2022 [Online]. Available: http://data.europa.eu/eli/reg/2022/2065/oj
[45]
“Proposal for a regulation of the european parliament and of the council laying down measures for a high level of public sector interoperability across the union (interoperable europe act).” European Commission, November 2022 [Online]. Available: https://eur-lex.europa.eu/legal-content/EN/TXT/?uri=CELEX:52022PC0720
[46]
“Regulation (EU) 2022/868 of the european parliament and of the council of 30 may 2022 on european data governance and amending regulation (EU) 2018/1724 (data governance act).” May 2022 [Online]. Available: http://data.europa.eu/eli/reg/2022/868/oj
[47]
National Institute of Standards and Technology, “Artificial intelligence risk management framework (AI RMF 1.0).” 2023 [Online]. Available: https://doi.org/10.6028/NIST.AI.100-1

  1. For this work, we examined multiple AI Act mandates published since April 2021, in particular the agreed provisional text. However, the references to the AI Act within this paper shall be interpreted as references to the European Commission’s proposal; as at the time of writing, the final content of the AI Act is not published in the official journal of the European Union.↩︎

  2. https://github.com/microsoft/opendatasheets-framework↩︎

  3. https://github.com/UTHealth-Ontology/MCRO↩︎

  4. https://w3id.org/rains↩︎

  5. https://w3id.org/dpv/↩︎

  6. https://w3id.org/airo↩︎

  7. https://www.iso.org/standard/77304.html↩︎

  8. https://www.iso.org/standard/81230.html↩︎

  9. https://www.iso.org/sites/directives/current/consolidated/index.html↩︎

  10. https://www.iso.org/standard/74296.html↩︎

  11. https://www.iso.org/standard/80655.html↩︎

  12. https://www.w3.org/↩︎

  13. https://huggingface.co/docs/hub/model-cards↩︎

  14. https://op.europa.eu/en/web/eu-vocabularies/dcat-ap↩︎

  15. https://op.europa.eu/en/web/eu-vocabularies/controlled-vocabularies↩︎

  16. https://www.w3.org/ns/dqv↩︎

  17. https://w3id.org/dpv/↩︎

  18. https://www.w3.org/TR/sparql11-query/↩︎

  19. https://www.w3.org/TR/shacl/↩︎

  20. https://www.w3.org/TR/odrl-model/↩︎

  21. https://github.com/DelaramGlp/airo/blob/main/usecase/proctify.ttl↩︎

  22. See existing examples of EU vocabularies here: https://op.europa.eu/en/web/eu-vocabularies↩︎

Prior Publication Attempts

This paper was published after 2 attempts. Before being accepted at this venue, it was submitted to: FAccT