Abstract
Keywords
Introduction
Technological advancement and the production of large volumes of data (so-called big data) have led to the increasing use of machine learning (ML) and artificial intelligence (AI) in all aspects of everyday life, including science, technology and education. These advances in computing and data processing have sparked discussions on digitising Research and Innovation (R&I) by industry 1 and the European Commission. 2
The rapid growth of AI has raised ethical concerns3–5 about its responsible use and potential unintended impacts on people and the environment. AI system behaviour is unpredictable, and an unreliable predictor of future actions, as learning and new data can change future responses. 3 AI can also have unintended consequences, such as re-identifying anonymised sensitive data in previously unforeseen ways. 3
Data governance 6 is key to ethical AI, given the challenges of managing large, dynamic and diverse datasets. Data governance ensures availability, usability, integrity and security, becoming a core focus across all data-driven fields. The 2016 introduction of FAIR (Findable, Accessible, Interoperable, Reusable) principles 7 has supported ongoing improvements in scientific, and wider, data governance and management. Similarly, the extension of FAIR to Computational Workflows 8 and research software (FAIR4RS), 9 supports FAIR practices for ML/AI models and related webtools.
Thus, the following questions emerge:
Should the FAIR data principles be part of ethical AI development? Do existing AI-ethics frameworks align with the FAIR principles?
This alignment can potentially transform key pillars of AI-ethics frameworks like transparency, auditability, accountability and remedy of harmful outcomes into testable and machine-actionable processes using FAIR's infrastructure, for example, Globally Unique, Persistent, and Resolvable Identifiers (GUPRIs), standardised metadata, and structured vocabularies. Thus, data stewards can utilise existing tools and develop roadmaps for FAIR-aligned ethical AI development and assessment.
Overview of AI ethical frameworks
A wide range of AI-ethics frameworks have been published by public, private and research organisations. Although these frameworks address similar ethical issues like transparency, fairness, non-maleficence (or do-no-harm), responsibility and privacy, their approaches vary significantly. 10 The key high-level ethical guidelines are the EU's Ethics Guidelines for Trustworthy AI, 11 the OECD AI Principles, 12 UNESCO's Recommendation on the Ethics of AI, 13 the IEEE Ethically Aligned Design, 14 Singapore's Model AI Governance Framework, 15 The Montreal Declaration for Responsible AI, 16 the Toronto Declaration, 17 the Beijing AI Principles 18 and the G20 AI Principles. 19 A qualitative analysis was performed to evaluate how these frameworks align with the FAIR, FAIR for Computational Workflows and FAIR4RS principles, through critical text review and analysis. Each identified, to the authors’ opinion, FAIR alignment was evaluated as Strong, Moderate, Partial, Implicit or Weak based on predefined criteria (Table 1), for example, whether mechanisms like GUPRIs, structured vocabularies, licensing, are mentioned or a compatible term was stated, for example, transparency, auditability. The results of this analysis are presented as a heatmap in Figure 1, while the tables corresponding to the heatmap can be found in reference. 20 Terms similar to the FAIR principles’ requirements, like ‘open and shared’, ‘traceability’, ‘access’ and ‘documentation’ are common across ethical AI guidelines. The EU Guidelines for Trustworthy AI emphasise data quality, integrity and mitigating implicit bias through careful dataset governance. 11 The OECD AI Principles 12 require robust risk management and traceability to address (meta)data reliability and bias. UNESCO 13 promotes non-discrimination, traceability and interoperability, while Beijing Principles 18 require fairness, bias handling, and traceable and auditable systems, and open and shared platforms. The Montreal 16 and Toronto 17 declarations warn against data-related discrimination, with the latter outlining specific prevention measures. The G20 AI Principles 19 promote transparent, explainable AI, aligning with FAIR principles F2 and R1 on rich metadata. Furthermore, G20 19 calls for digital AI ecosystems based on free data flow and interoperable resources, echoing FAIR aspects of findability and accessibility (FAIR principles F4, A1, A1.1, A1.2, I1, I2, I3).

Alignment heatmap of the major AI ethical frameworks with the FAIR (A), FAIR for Computational Workflows (B), and FAIR for Research Software (FAIR4RS) (C) principles.
Alignment level between the FAIR, FAIR for Computational Workflows and FAIR4RS principles and the major ethical AI frameworks and their respective requirements.
Given their high-level, policy-focused and domain-agnostic nature, these guidelines do not mention technical details or requirements. Terms like open data or traceable systems lack concrete solutions, like GUPRIs or (meta)data schemas, that align with FAIR, 7 FAIR for Computational Workflows 8 and FAIR4RS recommendations. 9 The EU Guidelines 11 and OECD AI Principles 12 mention accountability and documentation, but omit practical solutions like version control or licensing. While preventing discrimination through balanced data is a shared goal, the AI-ethics frameworks lack practical measures like using (meta)data schemas or controlled vocabularies. Calls for clarity, transparency and accountability in AI data align with the FAIR principles and the ‘as open as possible, as closed as necessary’ FAIRness approach. References to auditability, explainability and risk management align with FAIR's emphasis on organised, discoverable and reproducible data, while calls for robust metadata, oversight and data quality echo FAIR's core goals.
The FAIR, 7 Computational Workflow 8 and FAIR4RS 9 principles guide data, workflow and software management to support digitisation and improve machine actionability of all (meta)data. Rich metadata aim to support community consensus on enhancing (meta)data accountability, reproducibility and transparency. FAIR data are often mistaken for open data, but they primarily mean well-documented, securely managed and machine-actionable datasets. This distinction is significant, as digital R&I and AI-driven process optimisation, including health and safety aspects, rely, in part, on managing sensitive human and commercial data. As a result, the line between FAIR and AI ethics becomes blurred.
Operationalising FAIR through tools like GUPRIs, rich metadata, standard formats, licensing and provenance supports machine actionability and aligns closely with ethical AI concerns, which together form the basis of data governance. Furthermore, FAIR's metadata richness, sharing, long-term preservation and transparency, are reflected in frameworks like the EU Guidelines for Trustworthy AI, 11 the OECD AI Principles 12 and Singapore's AI Governance Framework. 15
The EU Guidelines 11 stress the need for data governance to ensure ‘data quality and integrity’, highlighting traceability and verifiability as prerequisites for transparency and accountability. Like FAIR, the OECD AI Principles 12 recommend robust data governance to enable risk assessment, reproducibility and user confidence. Singapore's framework 15 addresses data provenance, lifecycle management and internal and external audit mechanisms, directly aligning with FAIR's Accessible, Interoperable and Reusable pillars.
Another way to understand the connection between the FAIR principles and AI-ethics frameworks is the nomenclature used. Traceability and openness are explicitly referenced by the Beijing 18 and the G20 19 AI Principles, while rich metadata are implicit via transparency, disclosure and auditability in the Toronto Declaration. 17 UNESCO's recommendations 13 advocate, in practice, FAIR-aligned data governance, stressing safeguarded accessibility, traceability and auditability, increased interoperability of tools and data, and the use of robust datasets for scientific, R&I and societal benefit. These values align with the FAIR principles and promote AI transparency and accountability, allowing validation, explainability and ethical AI monitoring.
Considering this correspondence, FAIR, FAIR for Computational Workflows and FAIR4RS principles can be mapped to major AI-ethics frameworks (presented as a heatmap in Figure 1), based the authors’ critical qualitative evaluation deeming the alignment to be strong, moderate, partial, implicit or weak. For example, the FAIR principles emphasis on findable well-documented data align with the transparency and auditability requirements of the EU Guidelines 11 and Toronto Declaration. 17 The emphasis on controlled and standardised access, using free and open protocols and, where required, authentication and authorisation, supports the AI-ethics frameworks’ calls for independent review.12,13,19 FAIR's focus on data reusability aligns with the IEEE design 14 and EU Guidelines, 11 which mandate robust documentation and provenance. Similarly, the FAIR for Computational Workflows and FAIR4RS follow the same approach regarding GUPRIS, versioning, licensing and execution provenance, which are prerequisites for AI reproducibility and auditability.
Challenges and opportunities
Although the FAIR principles and their extensions can substantially support ethical AI, explicit mention and utilisation remain absent from current AI-ethical frameworks. Like the FAIR principles, most AI-ethics frameworks promote traceability, transparency, auditability and accountability, without dictating ways to achieve these. Likewise, the FAIR principles are not intended, by design, to be part of ethical AI development. 21
A major gap across many current AI-ethics frameworks is the lack of specificity in data governance, providing limited technical solutions, for example, GUPRIS, metadata schemas. In line with the FAIR, 7 FAIRification of computational workflows, 8 and FAIR4RS 9 principles, the EU Guidelines for Trustworthy AI 11 stress (meta)data auditability and integrity but lack specifics on metadata standards, GUPRIs, or curation processes required to assure these. The G20 AI Principles 19 promote transparency, traceability and accountability but offer no technical guidance for managing or structuring data to ensure interoperability or reusability. The Beijing AI Principles 18 mention data openness and sharing but lack guidance on controlled access, licensing or use of GUPRIs. Singapore's Model AI Governance Framework 15 offers risk-tiered guidance on data traceability (provenance, quality and lineage) and human oversight (human-in-the-loop, human-over-the-loop, human-out-of-the-loop), but omits explicit FAIR practices like domain-specific metadata standards, ontologies and version control.
Despite existing gaps, ethical AI frameworks conceptually align with the FAIR principles on data traceability, auditability and accountability. While not explicitly endorsing FAIR, the UNESCO Recommendation on AI Ethics 13 strongly supports robust data governance, transparency and scientific reproducibility. The OECD AI Principles 12 and Toronto Declaration 17 emphasise transparency, auditability and remedy of harmful outcomes, including from algorithms, which are facilitated by Findable and Reusable data. The IEEE Ethically Aligned Design guidelines 14 support long-term data stewardship and human-understandable records, aligning with human actionability and FAIR principle A2 on persistent metadata even when the data no longer exist.
In summary, although current AI-ethics frameworks don’t explicitly reference FAIR, FAIR for Computational Workflows, or FAIR4RS, they are closely aligned (Figure 1). Although alignment varies by framework, there are many opportunities for more descriptive and practical alignment to accelerate ethical AI adoption in data, model and software production. Metadata standards, traceability, interoperability protocols and responsible access models are just some of the opportunities. Adopting FAIR and its extensions to bridge AI ethics gaps offers a promising path forward, even for AI system architecture.
Implementing FAIR into ethical AI frameworks and practice faces several barriers. Legal constraints, for example, GDPR, can limit data accessibility. Interoperability challenges include missing metadata standards, limited ontologies and the need for community consensus. AI ethics challenges arise when reusability or accessibility conflict with confidentiality, IP rights or community data ownership. Practical solutions include secure data enclaves, differential privacy, data stewards and structured access and data governance, while interdisciplinary ethics boards can help resolve conflicts, ensuring FAIR practices balance innovation with ethical safeguards.
A starting point for data stewards, responsible for ethical AI management, could be the solutions identified and presented in Table 2, which links technical solutions to the respective FAIR-related principles and the ethical AI pillar they would enable. A practical data steward roadmap for implementation should (i) include GUPRIs, for traceability and unambiguous referencing, (ii) distinct machine-actionable and domain-specific metadata records, for persistence, explainability and reproducibility, licencing, human- and machine-actionability, to enable lawful and independent review, (iii) community-related vocabularies, for cross-domain bias checks and linkage and (iv) versioning, for change-auditing and comparability. The Research Data Alliance FAIR4ML group, which implements domain-specific and agnostic FAIR in ML by extending the FAIR4RS principles, 22 converged on a practical top 10 for FAIRifying ML models using a Delphi process. 23 Their findings align well and inform our data steward roadmap (Table 2), which also considers the FAIR for Computational Workflows and FAIR4RS principles, as reported in Figure 1.
Technical solutions to operationalise the verification of ethical-AI pillars using the FAIR, FAIR for Computational Workflows, and FAIR4RS principles, targeted to data stewards.
Technical solutions to operationalise the verification of ethical-AI pillars using the FAIR, FAIR for Computational Workflows, and FAIR4RS principles, targeted to data stewards.
This paper examines the convergence between the FAIR data principles, and their extensions for computational workflows and research software, and widely adopted ethical AI frameworks. While not explicitly referenced, there is strong alignment with FAIR in principles like transparency, traceability and responsible data governance. The practical and ethical benefits of integrating FAIR, for example, enhanced accountability, reproducibility and inclusivity will benefit the development of responsible AI systems. To strengthen this alignment, we recommend actions including the development and adoption of standardised metadata schemas for ML and AI models, FAIR-aligned technical systems, and the explicit reference of FAIR, FAIRification of computational workflows and FAIR4RS in ethical AI guidelines. Future work should explore measures to integrate these principles and how their adoption impacts AI system trustworthiness and sustainability.
