Abstract
Introduction
The prohibition or restriction of the use of black-box artificial intelligence (‘AI’) in healthcare, namely AI that lacks causal explanations, through regulation is not uncommon. For example, the Australian Therapeutics Good Administration (TGA) (2024) defines black-box AI as insufficiently transparent, and therefore prohibited from receiving regulatory approval for use in healthcare. The
We believe that blanket prohibitions on black-box AI in healthcare are excessive, and the need for and type of explanation required for ethical use of black-box AI in healthcare is contextual and relative to other reasons for using it in healthcare. That is, the justifiability of black-box AI ought not to be reduced to mere explainability; the moral justifiability of medical black-box AI is contingent on many factors. Whilst explainability importantly contributes to transparency, accountability, and trustworthiness (Hamon et al., 2020; Rudin, 2019; Rueda et al., 2024), black-box AI algorithms are valuable
Following the recent and consistently improving research on the beneficial nature of healthcare black-box AI, we will claim that the lexical priority afforded to causal explainability as necessary for the ethical and lawful uses of black-box AI warrants investigation. This article seeks to draw this literature together and make an original contribution by analysing how the various object-given reasons and considerations of using black-box AI in healthcare interplay. We want to understand what considerations are at play that make explainability necessary or unnecessary. This article is therefore not a literature review of current leading considerations of the use of black-box AI in healthcare, but a normative analysis for understanding how such considerations intersect and potentially extenuate each other according to the contextual use of an algorithm. In Section ‘What is a justification?’, we will therefore define and defend an account of justifiability by which to measure whether black-box AI has met a sufficient ethical threshold. Section ‘What reasons are relevant in justifying black-box AI?’ will highlight varying reasons that can be used to justify black-box AI, namely, the role of AI explainability and accuracy, the seriousness of the patient’s condition, the role of human intervention, the effect of bias, and the nature of the context. In Section ‘Balance and justifiability’, we will provide a summary of how we can foreseeably use these casuistic lessons drawn from case studies to explain when and how to justify the use of black-box AI in healthcare.
What is a justification?
Determining whether it is justifiable to use black-box AI in healthcare requires us to define justifiability. Broadly defined, justifiable decisions are those with robust
However, justifications are contextual depending on the person or situation for
In our attempt, therefore, to assess the justifiability of black-box AI in healthcare, we will adopt a
What reasons are relevant in justifying black-box AI?
Explainability
Given the substantial literature on the importance of explanations in justifying black-box AI, we should begin with understanding the nature, role, and limits of explanations in justifications. Explanations are generally viewed as both intrinsically valuable as an epistemic good and instrumentally valuable for achieving transparency (Bjerring and Busch, 2021; Ordish et al., 2020). This is first because if we can understand why a prediction was made, the model will be seen as trustworthy and thus used more widely in healthcare (Balasubramaniam et al., 2023; Bjerring and Busch, 2021; Hois et al., 2019). Second, transparency protects the rule of law by ensuring accountability for decisions that are not consistent with public norms of justice (e.g. bias) or for wrongful harm (Binns, 2018; Kempt et al., 2022a; Maclure, 2021; Steging et al., 2021). C
However, providing
This leads us, however, to a more important point: there are various definitions and standards of what is required by an explanation. This is because explainability is a broad concept – it is not solely whether we can explain the
But within healthcare there are different stakeholders seeking explanations of different tools and phenomena for different reasons (Durán, 2021; Kempt et al., 2022a; Malgieri and Pasquale, 2022; Ordish et al., 2020). For example, within a single context involving a patient determining which treatment option to follow, the patient may want explanations of success rates, side-effects, and standards versus alternative treatments; the doctor may want an explanation of which symptoms are indicative of a specific diagnosis; and, regulatory approval bodies may want explanations of the risks, costs, benefits, and side effects of treatments for patients populations.
Evidently, the definition and threshold of a satisfactory explanation is not a one-size-fits-all. The users and beneficiaries of medical AI are a distinct community with distinct needs and as such, various explanations may be justifiable for various stakeholders. The threshold explanation required to justify an algorithmic decision
In this case study, there is no explanation for the patient's chronic pain, but the black-box AI model is more likely to reduce patient pain than the doctor's professional opinion. There are several relevant facts to highlight from this model that give us reasons to justify its use in these circumstances.
First, the doctor provides the patient the option to decide
An objection might be made here by pointing to the regulatory or accountability requirement for an explanation in medical decision making, particularly for ensuring the decision was just and fair. For example, a patient might be subject to a black-box diagnosis and mistakenly
In the presence of this ‘“good” chain of reasoning’, in which the doctor considers how the factors might have been weighed and measures this against professional standards of care, the decision is not only justifiable to the patient, but if adopted, will also have be justified
Objections have been made to the use of XAI or post-hoc rationalisations of black-box AI on the grounds that they are ‘not faithful’ (Afnan et al., 2021), ‘misleading’ (Rudin, 2019), ‘inherently insincere’ (Babic and Cohen, 2023), and ‘fool's gold’ for giving false trust, false impressions, and false confidence (Babic et al., 2021; Peters, 2023; Rudin, 2019). Some of these concerns can be alleviated with modifying the terminology, such as by labelling these as ‘interpretative’ rather than ‘explanatory’. A better response, however, is to highlight the unique nature and purpose of medical knowledge and care that renders uncertainties in knowledge secondary in importance to reliable and beneficial outcomes. Modern medical knowledge is wide, complex, and deep – no one person holds, or could ever possibly hold, all that is known (Ferreira, 2021). Cohen (2020) argues that doctors are ‘likely quite ignorant of the underlying trial design or results that led the FDA to believe that the drug was safe and effective, but her knowledge that it has been FDA-approved supplies the necessary epistemic warrant’ to justify recommending the treatment. Similarly, professional standards do not require physicians to explain technologies such as MRI machines despite relying on these devices for significant decisions (Sand et al., 2021). Inscrutability of some expert decision making is inherent in such justifiable divisions of labour, and we cannot expect all practitioners to equally scrutinise all medical decisions.
One difference, however, between AI and the division of labour in medical decision making, is that the causal reason can be known by the latter but not in the former (Bjerring and Busch, 2021; Ferreira, 2021). Kempt et al. (2022b) argue that whilst patients generally do not care about technology specifics, such as in diagnostics or details, ‘the point of the conditional good of explainability is its potentiality.’ If a doctor cannot explain how exactly a new machine works, the researcher, technician, or developer can. In comparison, this is not similarly possible with black-box AI.
However, not all explanations are known for medical phenomena. The opacity in medical decision making is far more routine in medicine than critics realise (Kawamleh, 2023; London, 2019). Medical research and patient treatment planning are often heuristic and medical certification processes are primarily justified through evidence of safety and efficacy rather than causal explainability (Kempt et al., 2022b). For example, paracetamol is causally unexplainable but is available over the counter to the public because of how low-risk, beneficial, and useful it is for a variety of conditions. Shavhasi (2016) argues that what justifies unexplainable but useful treatments is that we can make hypotheses as to their causality that are consistent with accepted bodies of scientific knowledge. London (2019) thus argues that:
‘As counterintuitive and unappealing as it may be, the opacity, independence from an explicit domain model, and lack of causal insight associated with some of the most powerful machine learning approaches are not radically different from routine aspects of medical decision-making.’
This is likely because, when all else is equal, we care more about saving a life, treating disease, and reducing pain than providing
Accuracy
A widely accepted compromise across AI, legal, and ethical literature is that if transparency must be sacrificed, it should at least be for improved performance (London, 2019; Maclure, 2021). Some argue that in the absence of explanations, the success rate expected of medical AI ought to be much higher than professional standards (Sand et al., 2021). That is, medical AI ought to be clinically proven through reliable testing to produce results as well as, or better than, physicians. The accuracy or reliability of the black-box model entirely depends on the rigour of the clinical trial process, as well as the kind of data collected (e.g. according to race, sex, and age).
Fortunately, medical AI is proving to be highly accurate in certain contexts (Mota et al., 2024; Rodriguez-Ruiz et al., 2019; Ting et al., 2019; Zeltzer et al., 2023). This is due to the increased availability of data from which AI can make decisions, and the increased capacity of AI to process mass data, and to learn and adapt in real time. This increased accuracy in comparison to standard diagnostic testing and treatment predictions is intrinsically valuable in medical decision making. The overarching goal of medical care is after all, to improve patient well-being specifically by addressing their needs with medical care.
1
That black-box AI can be a highly reliable and accurate decision-making tool renders its use important and perhaps required in certain contexts. We are willing to go so far as to say that being as accurate in diagnostic testing or reliable in reducing patient symptoms as doctors is
One reason is if an explanation would increase a patient's well-being due to the nature of the condition or treatment history. For example, the average diagnostic delay for endometriosis is 6.6 years (but can be up to 27 years in the UK) (Fryer et al., 2024). Women also report general bias and prejudice against them from the medical community, specifically, they experience epistemic distrust against them as knowledge-holders of their chronic pain and symptoms (Jackson, 2019). In such cases of systematic opacity that undermines their autonomy, some patients may prefer explanations to opacity even at the cost of slightly or even moderately improved accuracy.
2
For example, in our above
We could make similar arguments about some diseases like acne that are persistent, ongoing, non-life threatening, and in which patients have often tried a range of methods to treat it, such as prescriptions and self-treatment through facial products. Indeed, acne treatments are commodified and arguably exploited by pharmaceutical companies with patients spending hundreds on products with no evidence or information whether it is targeted toward their specific kind of acne or suited to their skin type, mainly because patients generally do not understand the causes and corresponding treatments of acne (Tan et al., 2001). A patient may have been promised treatments by practitioners, and none of them have worked. Even if black-box AI provides a highly accurate answer, if we cannot explain it the patient may lack trust in the unexplainable decision. As such, a patient may prefer a causal explanation for a recommended treatment of their acne with slightly or moderately lower accuracy, over an unexplainable but slightly higher accurate treatment. Research also indicates that patients are
Seriousness and urgency
An important consideration in determining whether there is a robust reason to justify the use of black-box AI is the urgency of a medical situation (London, 2019). Babic et al. (2021) acknowledge that prediction explanations might not be so necessary if the algorithm is consistently shown to be more accurate in situations with higher and serious risks. However, some scholars make claims that the more serious the implication, the more reason to ensure supervisory efforts and human involvement (Ding et al., 2022). This might be because the more serious the risk of harm, the greater the need for an explanation if it goes wrong, or because increased seriousness requires greater trust from the patient and such trust requires transparency (Hois et al., 2019).
A better understanding of this condition, therefore, is to discern between
However, this ‘last resort’ option does not apply to all contexts. For example, a person recently diagnosed with Huntington's disease may prefer to try standard or recommended treatment options as suggested by their doctor. Whether a condition is serious does not necessitate object-given reasons to use black-box AI; instead, the seriousness of a condition impacts the justifiability of black-box AI in two ways. First, if black-box AI is being used to
Context of the decision
The nature or context of the decision can also impact the justifiability of using black-box AI in medical decision making (Babic and Cohen, 2023; Kempt et al., 2022a; Ordish et al., 2020). Some medical decisions are particularly well suited to black-box AI (not all decisions are made inherently better by an increase in data). For example, diagnostic and prognostic decisions must be rooted in empirical data and patterns, and so long as the initial algorithm is trained and approved by relevant regulatory authorities, these models can meet and even supersede human decision-makers due to their capacity to (efficiently) process significantly more data. As such, the more data computable by the decision-maker, the more accurate the decision becomes (up to a point). However, not all types of healthcare decisions should be made by unexplainable decision-makers. To illustrate this claim, we will consider how the context of a black-box AI decision can vary the nature of justifiability despite all involving distributive justice issues.
Liver transplantation allocations are difficult decisions to make due to the complexity of calculating success rates, the limited number of donors, and the difficulty of fairly distributing organs according to normative values. If a black-box AI model is used to determine which patients should receive an organ, a patient who is rejected or ranked lower than others may wish to know why to ensure that it was done fairly (Rueda et al., 2024). In this context, in which interpersonal comparisons and highly normative valuations are being made to determine who will live according to ethical theories of justice, we have strong reason to be concerned about irrelevant factors that may impact decision making. Unlike medical diagnosis and prognosis predictions which are based in scientific data and patient history, organ distribution is a question based on predictive accuracy
Emergency triage situations are similarly fraught with distributive justice issues but are contextually distinct because how quickly a person is treated is itself a determinative factor in survival and quality of life outcomes. The under resourced, urgent, and serious nature of emergency triage compounds a need for quick, accurate, and efficient healthcare. Given reduced time for deliberation, standards of care are also different: emergency doctors deal with ‘variable degrees of uncertainty, with less-than-ideal information, and under severe time constraints’ (Iserson, 2006). Physicians make ‘efficiency-thoroughness trade-offs’: not by choice, but by necessity (Baartmans et al., 2022), and regulatory and health law reduces standard requirements, such as the need for informed consent (Boyle and Stepanov, 2021). The priority of triage decision making is improving as many patient outcomes as quickly as possible (Savulescu et al., 2020). This utilitarian framework is generally consistently applied in times of resource shortages and health crises (Vearrier and Henderson, 2021). Furthermore, given the time and resource pressures, emergency decisions are often justified through post-hoc rationalisations, and as such, are arguably indistinguishable from the post-hoc rationalisations of black-box AI.
Nevertheless, we might distinguish between standard or consistent resource rationing (such as in emergency departments), and extraordinary rationing (such as during natural disasters or pandemics). In the former, AI might be considered more reliable given the consistency of data available in emergency departments. In extraordinary circumstances, however, the nature of the emergency is unpredictable and often impacts disadvantaged populations disproportionately more heavily than privileged ones. We might argue that worries about distribution of resources in extraordinary circumstances might be more akin to organ donation cases, in which following utilitarian principles might further perpetrate bias. At the same time, however, the
A final example of the importance of context in justifiability is the role of black-box AI in addressing the inequality in access to healthcare experienced by regional and remote patients. Regional and remote patients experience both inequality in
We ought to be careful, however, of lowering standards of care in ways that lead to relativism between neighbouring communities and even within individual hospitals (Kempt et al., 2022a). 5 Nevertheless, balancing the value of increasing access to healthcare with the value for in-person medical care or explainability is an (unfortunately) necessary compromise. Addressing healthcare inequalities with sufficiently accurate, accessible, and efficient healthcare can outweigh the moral concern of a lack of causal explainability or unequally accurate black-box AI in comparison to metro areas (although not unacceptable standards). Arguably, therefore, accurate black-box AI may be justifiable in circumstances when it addresses or improves inequalities in access to healthcare and healthcare outcomes.
Shared decision making
Another reason to justify certain uses of black-box AI in medical decision making is when these decisions are reviewed, interpreted, tested, and confirmed by a human, and thus, decision making is still shared between doctor and patient (Amann et al., 2020; Holzinger et al., 2019; Maclure, 2021; Robbins, 2019). Just because black-box AI is inscrutable, does not mean that all aspects of AI development, implementation, and use are inscrutable (Ferreira, 2021). As Ehsan et al. (2021) argue, ‘the “ability” in explainability does not lie exclusively in the guts of the AI system’. Black-box AI need not simply apply through a copy-and-paste exercise, but how and when human involvement in black-box AI use is
First, justifiably using black-box AI
Second, doctors ought to be involved in matters that require medical judgment and provide post-hoc justifications of AI decisions through shared decision making with the patient (Coeckelbergh, 2020). Physicians are trained according to medical knowledge and should explain the reasons why the AI decision is consistent with medical knowledge in relevant and meaningful ways to the patient. For example, they can explain why the treatment is recommended for the patient's condition, what the side-effects and benefits of treatment are, and why the diagnosis could rationally explain the patient's symptoms. Furthermore, in the same way technicians are relied upon for their technological expertise, a professional (not necessarily the physician) can also be trained to know some technological details of the black-box AI they used, such as the accuracy rate, the data it was trained on, the range of output values, how to monitor the device for decline, and the benefits and limits of the algorithm (Durán, 2021; Sand et al., 2021).
Shared decision making between doctors and patients regarding the use of black-box AI also protects patient autonomy in two ways (Prince and Lim, 2025). First, ensuring that the black-box AI decision is rationally consistent with medical science, the patient's capacity for rational decision making is respected. Second, doctor oversight can provide the epistemic warrant for the patient's trust in the decision-making process and empower the patient with more accurate diagnostic and prognostic information. The patient may then provide reasons and justifications for adopting the recommended decision or indicate why they would prefer an alternative. Such shared decision making would also ensure the patient justifies the decision to themselves (Muralidharan et al., 2024).
Arguably, shared decision-making and thus doctor involvement is necessary to the justifiability of black-box AI, but we believe there are some limited exceptions to the rule. The first is that doctor involvement does not always have to be immediate or concurrent. For example, black-box AI could be used to provide some forms of healthcare to remote and regional patients who would otherwise completely lack, or have extremely limited access to, healthcare (Kantipudi et al., 2021; Kothamali et al., 2023). A more contextual and useful position would be ensuring these patients have the option for review and post-hoc rationalisation by doctors upon request. We ought to be careful of justifying relativist and discriminatory healthcare and safeguard against the worsening of healthcare quality and forcing already discriminated groups from sacrificing explainable medical care out of mere necessity. Nevertheless, this concern can be proactively addressed by ensuring all patients accessing black-box AI algorithms have access to a human (e.g. doctor) to review the AI decisions and by only implementing algorithms that have been sufficiently regulated, trialled, and tested for bias and accuracy.
Some fears, however, have arisen that involving AI in medical decision making necessitates ‘machine paternalism’ in which physicians simply copy-paste AI decisions and inflict the decision on the patient (Afnan et al., 2021). Bjerring and Busch (2021) argue that in deploying reliable and accurate AI, practitioners will have an epistemic obligation to align their medical verdicts with those of AI systems in a way akin to general practitioners relying on the opinions of experts. Doctors might feel that they must accept the AI decision making for fear of litigation. However, if black-box AI is sufficiently accurate, reliable,
Bias
We might consider a possible extenuating factor: that is, the worry that black-box AI can hide harmful biases against already marginalised and disadvantaged populations. The justifiability of black-box AI indeed depends in part on how biased the algorithm is, and how harmful the consequences of any bias (Amann et al., 2020). The most obvious harm is that black-box AI will be less accurate and therefore reliable for such groups owing to poor data (Klugman, 2021; Robbins, 2019). Some have argued that bias is only relevant insofar as it reduces accuracy (Akinrinmade et al., 2023) and some argue that black-box AI bias is not morally different to human bias in medical decision making (Kawamleh, 2023; Kempt et al., 2022a; London, 2019). However,
However, there are several considerations that are relevant when discussing bias. First, in cases of serious bias that results in inaccurate decision making in certain demographics, there will be corresponding decreases in accuracy and reliability. Given this lowered accuracy and thus increased harm, it may not be justifiable to use such algorithms. After all, if an algorithm intended to be used in a healthcare context fails to adequately work for approximately 50% of the population (e.g., women), then it cannot really be similarly accurate to existing standards. Some scholars accept this consequence and justify the use of healthcare that only works for a certain demographic or population (Vandersluis and Savulescu, 2024). Indeed, it is generally difficult to produce therapeutics that work for all patients 100% of the time. However, we cannot resort to defending oppression and discrimination as a reason for a reduction in the quality of medical care; black-box AI is not justifiable if it is only accurate for a privileged demographic.
Second, we can proactively address bias through regulatory standards in data testing and establish pathways for legal redress as and when mistakes inevitably occur. Some scholars and studies show that black-box AI, although causally unexplainable, is, or at least perceived to be, more easily screened for biases than humans (Drezga-Kleiminger et al., 2023; Kawamleh, 2023). After all, black-box AI has no internal reason or unconscious bias against certain people; any inherent bias is due to missing data or learned patterns, and such, can be systematically tested for bias and modified following such testing. Although we ought to be cautious of overstating the ease of regulating and identifying bias, the mere
Furthermore, if medical AI devices become regulated through clinical trials or other regulatory processes, any criticisms of the role of bias and error should be redirected to the regulatory process, rather than black-box AI alone. If black-box AI is permitted to be biased, this is not inherent to the algorithm so much as it is in the clinical trial and regulatory process. However, clinical trials do not demand perfection, and as a society, we have long accepted the limits of medical treatments and diagnostics. As such, we have, or at least ought to have, regulatory systems that have clear principles for enforcing accountability when developers intentionally or negligently fail to proactively address bias, and redress for persons who are subject to bias or discriminatory treatment. Nevertheless, whether black-box AI is being introduced in such regulatory systems with laws and policies contextualised to AI will impact whether using such black-box AI is justified.
Balance and justifiability
We have thus far argued that we are unjustified in affording lexical priority to the explainability of medical decision making. The complexity of medical decision making means we cannot recommend a uniform normative framework for recommending the most justifiable means of using black-box AI, but we can identify the variety of factors at play and use casuistic reasoning to understand how these factors can be weighed to make its use justifiable. As we have argued, there are many good reasons to justify the use of black-box AI – and these reasons can interact and justify the absence of other factors. We have argued that accuracy is a necessary, although not sufficient, condition in any justification of its use. We have also considered how black-box AI can be more easily justified in urgent and serious cases because the need to make quick decisions directly impacts the patient's outcome, and that black-box AI is more likely to be justifiable in clinical decision making rather than normative decisions given that need humans to be accountable for inherently ethical decisions. Furthermore, we have argued that
In contrast, sometimes a patient's unique medical history and experience of uncertainty, inexplicability, and poor treatment by the medical system may create reason to use explainable over accurate decision-making processes. The use of black-box AI may be unjustified to make a decision that is necessarily subjective, such as in accordance with the patient's values, and if it is implemented without informed consent and shared decision making with the patient. Finally, the presence of bias does not provide a blanket prohibition of black-box AI insofar as such bias has been proactively regulated and mitigated through clinical trials, testing, and regulation.
Justifiability is about reasonability and defensibility, and we can balance a variety of reasons in different contexts to determine when it is justifiable to use black-box AI. Importantly, however, it is not necessary to have all factors
Conclusion
We have argued that accurate and reliable black-box AI can be justifiably used in medical practice under certain conditions (see Table 1 for a summary). The proposition that black-box AI is a ‘trade off’ of explainability for accuracy is thus arguably misconceived. Nevertheless, guidelines and policies must be sensitive to the interplay of considerations that changes according to each new medical context. The use of accurate and reliable black-box AI can therefore be justified in healthcare, and we must be careful of ousting it for the reason that regulators could more easily attribute accountability if we knew the reasons for which decisions were made.
A summary of the principle normative factors relevant when determining whether the use of black-box AI in medical decision-making is justifiable, and examples of how they can be balanced and used in potential ethical guidelines in medical decision-making.
