Abstract
Keywords
Introduction
Artificial intelligence (AI) has emerged as a groundbreaking innovation with transformative potential across diverse industries, including healthcare. Its integration into medical practices has catalyzed significant advancements in diagnostics, personalized medicine, and patient care management, underscoring its versatility and impact.1,2 AI-driven tools excel at processing and analyzing vast datasets, uncovering patterns that often surpass human capabilities in specific medical tasks. For example, a study by Kim et al. 3 demonstrated that an AI algorithm trained on extensive mammography data achieved superior diagnostic accuracy in detecting breast cancer compared to traditional radiologists. Similarly, research has shown that AI systems can significantly reduce false positives and negatives in mammogram interpretations, further enhancing diagnostic precision. 4 Comparable breakthroughs have been documented in the detection of pneumonia through chest radiographs, 5 cardiac arrhythmias, 6 diabetic retinopathy,7,8 and skin cancer.9,10 Beyond diagnostics, AI-powered healthcare systems have proven highly effective in formulating evidence-based treatment plans. By synthesizing patient records and medical literature, these systems provide tailored treatment recommendations that often align with or exceed the expertise of human clinicians. IBM's Watson for Oncology, for instance, leverages vast clinical datasets to deliver personalized cancer treatment options, achieving high levels of agreement with oncologists’ recommendations. 11 Studies further underscore AI's role in optimizing treatment strategies for various conditions, improving patient outcomes, and reducing medical errors.12,1 Moreover, AI is reshaping hospital operations by automating administrative tasks, enhancing efficiency, and allowing healthcare providers to dedicate more time to patient care, thereby improving the overall quality of healthcare delivery.13,14
In Africa, the adoption of AI in healthcare, although slower than in other regions, is gaining momentum through innovative initiatives, such as leveraging AI to revolutionize cancer diagnostics through cervical cell image analysis, tissue evaluation, disease prediction, and teleradiology.
15
Research has also highlighted AI's efficacy in diagnosing and predicting cervical cancer, colorectal cancer, leukemia, breast cancer, lung cancer, and pulmonary tuberculosis.15,16 In Tanzania, Mwanga et al.
17
introduced an innovative approach to malaria screening by integrating the Logistic Regression model with mid-infrared (MIR) spectroscopy, utilizing human dried blood spots (DBS) as the primary sample medium. This method leverages the unique MIR spectra captured by a spectrometer and trains a model to classify DBS samples as either malaria-positive or malaria-negative. The proposed method is promising for enabling rapid, high-throughput screening of
Despite these promising initiatives, the widespread adoption of AI in healthcare remains limited in Africa due to several persistent challenges, including data scarcity. The shortage of curated clinical datasets limits both the training of robust AI models and their translation into practice in low-resource settings. This scarcity undermines progress toward the African Union's Continental AI Strategy targets and the WHO Global Strategy on Digital Health, both of which emphasize data quality, accessibility, and equity. Specifically, this article focuses on the issue of limited access to high-quality clinical data, which significantly impedes the development and deployment of AI models across the continent. While there are notable advancements that illustrate the growing global reliance on synthetic data to overcome clinical data limitations, Africa's engagement with this emerging approach remains underexplored. To this end, this study examines the potential of synthetic data for healthcare AI in Africa with a dual focus on technical and governance dimensions. The contributions of this study are threefold: first, it synthesized evidence on how synthetic data performs across technical dimensions such as fidelity, generalization, and domain adaptation in African healthcare contexts. Second, it examined governance, regulatory readiness, and institutional constraints that influence the safe use of synthetic data in low-resource settings. Third, it developed a conceptual framework that integrates technical and policy considerations to guide future evaluation, deployment, and oversight of synthetic data-based healthcare AI systems.
The remainder of this article is organized as follows. The second section examines the nature of clinical data scarcity within African health systems and provides an overview of synthetic data. The third section outlines the study's methodology, including the review design, search strategy, and selection criteria. The fourth section presents the study's results, while the fifth section offers a detailed discussion of the technical findings, adoption barriers, and existing knowledge gaps. Finally, the sixth section concludes the article with policy recommendations for the safe and effective integration of synthetic data into healthcare AI in Africa.
Data scarcity and application of synthetic data
Clinical data scarcity
Clinical data serves as the foundation for machine learning (ML) algorithms to create computational models that generate predictive insights for informed medical decision making. 20 Nevertheless, limited access to such data poses a significant obstacle to the practical application of ML tools in healthcare. 21 Africa, in particular, faces substantial challenges due to low levels of digitization, which limit the availability of locally generated clinical data, a crucial resource for developing AI-driven healthcare solutions. 22 For example, many healthcare facilities in sub-Saharan Africa still rely on fragmented health information systems, ranging from paper-based records to isolated digital platforms. This fragmentation leads to inconsistencies in data quality and a lack of interoperability between different systems.23,24 Notably, the paucity of health data is a significant barrier to Africa's adoption of digital health solutions and evidence-based clinical practices. 25 Although Africa faces the highest burden of both communicable and noncommunicable diseases worldwide, it accounts for only about 1% of global research output. This disparity is partly due to challenges in accessing and maintaining medical data. 26 The AU Continental Artificial Intelligence Strategy 2024 27 highlights a significant gap in the inclusiveness, quality, and availability of data to support African AI innovations. The strategy entails several actions to enhance data availability on the continent, including cross-border data sharing among AU member states. Moreover, it is pertinent to note that even when clinical data is available, a significant portion remains underutilized due to issues related to data complexity, storage, and poor management systems. 28 This challenge is further intensified by the prevalence of imbalanced and low-quality data, which not only hampers the effective implementation of ML models29,30,31 but also amplifies issues such as model underfitting, overfitting, and the risk of biased outcomes.32,15 Data scarcity and incompleteness have been reported in several studies,30,33,34 demonstrating how it can lead to reliance on algorithms predominantly developed outside Africa. 35 Deploying such externally developed tools in new settings can result in several issues, including poor model performance, biased outcomes, and misaligned clinical relevance. Some notable examples highlighting the impact of data unavailability include the study by Mwanga et al. 17 While their proposed AI-driven malaria screening tool demonstrated significant potential, its real-world application was hindered by the lack of diverse and representative datasets, which are essential for comprehensive validation and refinement. In addition, Manson et al. 36 emphasize that the limited availability of sufficient radiological images has been a major obstacle to the adoption of AI in radiotherapy, as these images are critical for training deep learning models.
Synthetic data
Synthetic data has emerged as a solution to address the challenges of limited or inaccessible real-world data. 37 In healthcare, real-world data can be costly, difficult to access, or pose privacy risks. Synthetic data is artificially created to mimic the patterns and properties of real data. Synthetic data generation enables the creation of high-quality, diverse datasets for training and validating machine learning models, providing a valuable resource for healthcare applications where data availability is often limited. For instance, to address the scarcity of chest X-rays and brain magnetic resonance imaging (MRIs), critical for diagnosing respiratory and neurological diseases, respectively, Dhawan and Nijhawan 38 developed a specialized Generative Adversarial Network (GAN) architecture to generate synthetic imaging data. By integrating synthetic data with real datasets, their approach significantly improved classification performance, achieving up to 85.9% accuracy on a brain MRI task using the EfficientNet v2 model. Moreover, Montes et al. 39 demonstrated that deep generative models, such as GANs and Restricted Boltzmann Machines, can effectively learn and replicate complex genomic data distributions, generating high-quality artificial genomes (AGs) with minimal risk to privacy. Their findings demonstrate that AGs preserve key genetic characteristics, enhance imputation accuracy, and support supervized learning tasks in genomic studies. Schmitz et al. 40 developed SynthMD, a hierarchical data generation tool that produces synthetic datasets for rare diseases using publicly available statistics, addressing the critical data scarcity challenge in rare disease research. Their open-source approach enables researchers to simulate patient data for conditions such as Sickle Cell Disease, Cystic Fibrosis, and Duchenne Muscular Dystrophy, facilitating method development and the development of privacy-preserving innovations. These examples demonstrate the value of synthetic data in developing healthcare AI models when real data is limited.
Various techniques are employed to generate synthetic data, including GANs, simulation models, and data augmentation methods. GANs work by training two neural networks against each other: a generator that creates synthetic data, and a discriminator that evaluates how closely the synthetic data matches real data. 41 Over time, the generator improves its ability to produce realistic data, making it highly effective for generating synthetic images, texts, and other data types. In healthcare, GANs have been utilized to generate synthetic medical images, including computed tomography (CT), MRI, and X-ray scans, which are crucial for training deep learning models for disease detection. 38 These synthetic medical images can be created from limited real-world data, reducing the need for extensively annotated datasets that they often expensive and tedious to produce.
On the other hand, simulation models are widely used to generate synthetic data by producing datasets based on specified rules, parameters, and variables. 42 In healthcare, simulation models can create synthetic patient records, medical histories, and even time-series data, such as cardiac blood volume pulse. 43 With such models, researchers can simulate complex patient scenarios that would be difficult or impractical to observe in real life. Finally, data augmentation techniques are employed to replicate real-world phenomena, including variations in object and scene appearance resulting from changes in pose, viewpoint, lens distortion, deformation, and other camera-related artifacts. 44 This is particularly useful in image-based tasks, where augmenting a limited dataset with slightly altered images can provide a more extensive, varied dataset for training AI models. For example, medical image augmentation can help train AI models for diagnostic purposes by generating variations of existing scans, allowing the model to learn from a broader range of conditions. 45 Readers seeking an in-depth analysis of synthetic data generation methods in healthcare are encouraged to consult the study by Pezoulas et al. 46
The practical applications of synthetic data have been demonstrated globally. For example, in Europe, the EU-funded SYNTHIA project exemplifies large-scale synthetic data integration, generating multimodal datasets that include imaging, genomics, and clinical notes to train AI models for diseases such as lung cancer, Alzheimer's disease, and type 2 diabetes. 47 In Germany, the NFDI4Health initiative has developed SYNDAT, a web-based tool that utilizes AI models, such as VAMBN and MultiNODEs, to generate and assess synthetic health data. It enables the evaluation of data utility and privacy risks, with practical applications demonstrated using real-world datasets for Alzheimer's disease and cancer. 48
Despite its promise, synthetic data presents several challenges that must be addressed to ensure its effectiveness in real-world healthcare applications. Key concerns include maintaining the authenticity and fidelity of generated data, minimizing algorithmic bias, and rigorously validating models trained on synthetic datasets. 46 Additionally, ethical considerations, particularly regarding privacy, fairness, and transparency, remain crucial in the deployment of synthetic data within sensitive domains, such as healthcare. 49 A recent study proposing an evaluation framework for synthetic data in health noted that ensuring a robust fidelity-utility-privacy tradeoff is challenging, as synthetic data generators must balance realism with privacy guarantees. 50 Kaabachi et al. 51 suggest that there is no universally accepted set of benchmarks or evaluation metrics, which hampers the validation, comparison, and adoption of synthetic datasets at scale. Moreover, synthetic health data can reproduce or even amplify biases present in the original data, raising concerns about fairness. 52
Methodology
Using a Critical Literature Review (CLR), this study interrogated relevant scholarly and technical literature focused on the use of synthetic data for AI in healthcare across African settings. The review was guided by three research questions: (1) How has synthetic data been used to support healthcare AI development in Africa and comparable low-resource settings? (2) What technical performance issues arise concerning fidelity, generalization, and domain adaptation? (3) What governance, regulatory, and institutional factors influence the adoption and safe use of synthetic data in African healthcare? Policy-relevant studies, including those discussing regulation, governance, and health technology assessment, were included to align the analysis with health policy priorities. Rather than simply summarizing findings, the CLR method provided a systematic process to reveal conceptual assumptions, empirical inconsistencies, and research gaps.53,54 The approach also enabled an assessment of technical claims and socioinstitutional barriers, which are often overlooked in traditional narrative reviews.54,55
A systematic search was conducted across Scopus, Web of Science, PubMed, and Google Scholar, targeting peer-reviewed journal articles, technical reports, and institutional publications. The following search string was applied across all databases for consistency and transparency: (“synthetic data” AND (“healthcare” OR “clinical”) AND (“artificial intelligence” OR “machine learning”) AND (“Africa” OR “low-resource” OR “low-income”)). Search terms used to identify articles related to (i) synthesized data, healthcare AI, (ii) Africa, (iii) domain adaptation, and (iv) data scarcity. The search covered literature published between 2010 and 2025, reflecting the period during which synthetic data approaches gained relevance in healthcare AI. We included studies based on their relevance to synthetic data generation, model training processes, and adoption in low-resource healthcare systems.
The quality and relevance of retrieved literature were assessed through predefined inclusion and exclusion criteria focusing on methodological rigor, clarity of synthetic data evaluation, and applicability to healthcare AI in low-resource settings. Studies were further appraised using a structured checklist that examined credibility, transparency of methods, and alignment with the review's objectives to ensure that only robust and contextually relevant evidence was included. The inclusion criteria were: (1) studies involving synthetic data for healthcare or clinical AI; (2) studies presenting technical performance outcomes; (3) research relevant to Africa or other low-resource health systems; and (4) peer-reviewed articles or high-quality institutional reports. Exclusion criteria were: (1) studies unrelated to healthcare; (2) papers lacking methodological detail on synthetic data generation; (3) commentaries without empirical or technical content; and (4) non-English publications. Insights were synthesized using thematic analysis techniques, allowing for critical engagement with the dominant discourses present within them and the situational applicability of these discourses.
The review followed Critical Literature Review principles, combined with PRISMA (

PRISMA flow diagram of the search process for the studies for review (source: authors’ construct).
Results
Technical performance and model reliability
Synthetic data appears to be a promising solution to address the issue of insufficient data in healthcare AI; however, a thorough analysis reveals significant performance limitations.64,65 Many models developed solely from synthetic data showed significantly lower generalizability when tested for clinical use, especially in heterogeneous, low-resource settings.57,56 Generative techniques, such as medGAN, yielded more realistic synthetic data, but fidelity gaps remained, impacting generalizability and, ultimately, the reliability of diagnostics. The difference is crucial because it highlights the need to recalibrate expectations regarding training with synthetic data only. The prominence of GAN-based methods in Table 1 reflects the current state of published research rather than reviewer or author bias. GANs remain the dominant approach used in medical synthetic data generation, which explains their strong representation in the reviewed studies.
Summary of key studies on synthetic data use in healthcare AI (source: authors’ construct).
3D: three-dimensional; AI: artificial intelligence; EHR: electronic health records; GAN: Generative Adversarial Networks ; ICU: intensive care unit; ML: machine learning; MRI: magnetic resonance imaging.
Significantly, the most common method used by studies has been to assess models using (often overly optimistic) internal synthetic test sets. 60 Only a fraction of the studies in our review, summarized in Table 1, compared synthetic-only with synthetic–fundamental hybrid training approaches. In all of these comparisons, models incorporating both synthetic and real data consistently outperformed pure real-only and synthetic-only models in terms of generalization (predictive performance on held-out data) and reduced errors in prediction within underrepresented cohorts, such as specific ethnicities and ethnic groups. Even in the presence of such evidence, hybrid modeling continues to be underutilized in the African context of health AI projects, signaling a lack of coherence between technical advancements and the processes applied towards implementation. These findings have direct policy implications for health technology assessment (HTA) processes, suggesting that regulatory bodies should require hybrid validation protocols before approving synthetic data-based AI tools for clinical use.
Current techniques for generating synthetic healthcare data include generative adversarial networks, diffusion models, probabilistic models, simulation-based generators, transformer-based language models for clinical text, and classical augmentation pipelines. These techniques produce structured records, medical images, clinical notes, and time-series signals, each with varying fidelity and resource requirements. Recent advances in diffusion models and multimodal generators have expanded the range of synthetic data modalities available for clinical AI development.
Availability of synthetic datasets varies globally. High-income regions host numerous open-source repositories containing synthetic medical images, structured clinical records, and genomic datasets. In contrast, Africa has very limited publicly available synthetic datasets due to infrastructure constraints, fragmented health information systems, and restrictive data governance environments. The lack of shared synthetic datasets impedes benchmarking, validation, and model reproducibility in African healthcare AI projects.
Data fidelity and risks of bias
Data fidelity emerged as a decisive factor influencing the clinical utility of synthetic datasets. In this review, fidelity is defined by two dimensions. First is distributional fidelity, which captures how closely the synthetic data match the overall statistical properties of real datasets. Second is clinical fidelity, which assesses whether rare events, comorbidity patterns, and sociodemographic variations are represented realistically. Both dimensions are essential for safe clinical deployment, as low fidelity increases the risk of biased predictions and diagnostic misclassification. However, the fidelity limitation was significant across many studies, where generative models accurately reproduced general features of new patients but failed to capture comorbidities, rare conditions, and sociodemographic variations.58,59 This is likely to have increased the risk of diagnostic bias and misclassification.
Globally, key challenges include inconsistent fidelity across modalities, limited validation protocols, unclear privacy guarantees, and the absence of benchmarking standards. In Africa, these challenges are intensified by weak digital infrastructure, limited computational capacity, scarce local datasets, and complex regulatory environments. These factors combine to reduce the practical utility of synthetic data in resource-limited clinical settings, particularly where rare conditions and diverse comorbidity patterns remain underrepresented.
In several contexts with limited or incomplete baseline datasets, generative models constructed from incomplete records perpetuated systemic bias instead of countering it, raising concerns about fidelity. 60 In Table 1, we demonstrate that for state-of-the-art applications of synthetic datasets, certain limits of fidelity performance remain constant. Such would necessitate standardized fidelity assessment metrics from now on. Importantly, critical systems such as healthcare systems must ensure that they do not release synthetic-data-driven AI for patient care purposes without proper validation methods that are specific to their use-case configuration.
Domain adaptation for low-resource health systems
Domain adaptation techniques received increasing attention as a bridge between synthetic and real-world environments. Synthetic data remained unable to be directly bridged to the real world due to the domain gap; thus, contextual adaptation techniques began to gain additional focus. Transfer learning, adversarial fine-tuning, and domain randomization approaches have also been used to increase model robustness in this scenario by addressing the distribution shift.62,61 Such techniques facilitate better diagnostic fairness and allow classifiers to generalize across historically underrepresented population segments in training datasets. Nonetheless, only a few studies have empirically explored domain adaptation in resource-constrained scenarios. Notably, Zhu et al. 63 demonstrated that their research was effective in an low and medium income countries using only a small amount of real data for recalibration, provided the recalibration threshold was minimal. Beyond imaging and structured electronic health records (EHR)s, synthetic data have also been applied to clinical notes, diagnostic reports, and time-series data such as electrocardiogram, photoplethysmography, and intensive care unit monitoring signals. These modalities introduce additional challenges due to temporal dependencies, linguistic variation, and the need to preserve clinical coherence. Emerging transformer-based models and physiologically grounded simulators have shown promise in generating higher-fidelity text and time-series data suitable for downstream clinical tasks. Domain adaptation, then, in theory, would work. Because the capacities of those health systems are roughly the same, there are hints of applicability when discussing the implementation of a model in a new region or population; however, unfortunately, in Africa, those health systems lack these capacities. Hence, the direct application is at best problematic. Achieving adaptation benefits requires policymakers to facilitate targeted investments in infrastructure and capacity building.
Adoption barriers and policy readiness in Africa
While there have certainly been strides in technology, structural challenges to adoption remain. The lack of adequate infrastructure, particularly erratic electricity, limited data storage capacity, and poor internet access, will still prevent the use of synthetic data in many African health systems. 69 Such restrictions hinder the generation and effective deployment of AI systems. Moreover, as shown in Table 2, institutional fragmentation and inadequate collaboration between health and information and communication technologies stakeholders hinder the implementation. Shortages in skills and human capital compound these. 70 Existing studies have made assumptions about high technical capabilities for model deployment and domain adaptation; neither assumption is a common characteristic across continents. 61 Ethical concerns also complicate the adoption process. Without rigorous windows of opportunity for validation, synthetic data-based models could operate with clinical freedom, without oversight or accountability. In particular, the lack of transparency in model development is one of the many reasons behind clinician skepticism. 68 On the path to making it the norm, synthetic data must be embedded in a broader digital health approach with explicit regulatory clarity and stakeholder alignment.
Adoption barriers of synthetic data in African healthcare contexts (source: authors’ construct).
Evaluation of synthetic data
To support safe and responsible use, a practical framework for evaluating and releasing synthetic healthcare data is essential. Generally, when evaluating synthetic data, the most common aspects that are examined include Fidelity, Utility, and Privacy.
71
Fidelity is evaluated by examining how closely synthetic data matches real data at both the individual-feature level and across multiple features or sequences. First, feature-wise fidelity can be assessed using descriptive statistics, such as mean, variance, and range, to ensure basic alignment between the real and synthetic distributions. This can be supplemented with statistical drift measures and Goodness-of-Fit tests, such as the Kolmogorov–Smirnov test, which provide statistical evidence on how well each feature's distribution is reproduced. For a more quantitative measure, a classifier-based test can be employed: a model is trained to distinguish between real and synthetic data, where high accuracy indicates poor fidelity (the data are easily distinguishable), while accuracy near 50% suggests strong fidelity.
72
The recent study by Adams et al.
73
evaluated fidelity using three complementary measures. The
On the other hand, developers should conduct utility assessments to ensure that synthetic datasets preserve clinically meaningful patterns. Evaluating predictive model performance on both real and synthetic health records helps detect loss of clinical signal before deployment. Achterberg et al. 72 suggest that the utility of synthetic data is usually evaluated by determining whether it can replace real data in standard analytical tasks while maintaining comparable performance. This evaluation entails comparing the performance of models trained on synthetic and real data when tested on a real test set. This approach is commonly referred to as the Train Synthetic Test Real (TSTR).50,72 For example, in assessing the utility of synthetic data, Hernandez et al. 50 trained several machine learning models on both real and synthetic tabular datasets and evaluated them using real data. They then compared the performance differences between models trained on real and synthetic data, calculating the statistical significance of these differences. Smaller performance gaps indicate higher utility of the synthetic data.
Equally important is the evaluation of privacy risks, which aims to uncover how effectively an external party could infer sensitive information from the original dataset if they had access to the synthetic data. It is worth noting that high-fidelity synthetic data can still leak identifiable attributes if not properly assessed. Privacy in synthetic data is a multidimensional concept, reflected in the many metrics used to quantify it. Common approaches include similarity tests, counting the number of identical records between the original and synthetic datasets, or distance-based metrics such as the mean absolute error between real and synthetic records.74,75 However, Giomi et al. 76 suggest that these metrics often provide limited insight into the real-world risks associated with individual records. Consequently, their study proposed evaluating privacy through concrete attack scenarios, focusing on singling out, linkability, and inference attacks, which are legally mandated under general data protection regulation. Singling out occurs when the original dataset contains exactly one record with a specific, uniquely identifiable combination of attributes. Whereas linkability measures the risk of connecting multiple records to the same individual(s), which can facilitate de-anonymization. Inference involves deducing whether a real record was used in training (membership inference) by finding its closest synthetic counterpart, and/or inferring the unknown sensitive attribute of that record (attribute inference) by comparing the sensitive value of its synthetic neighbor to the real value. Moreover, to ensure privacy, the Differential Privacy (DP) 77 framework is often used to add carefully calibrated noise during data generation so that the synthetic data reflects overall patterns while guaranteeing that no single person's information can be accurately inferred.
While these aspects can be individually assessed, maintaining the tradeoff between them, for instance, between utility and privacy, remains a challenge. If not well assessed, synthetic data may lead to unpredictable utility loss and highly unpredictable privacy risks. 78 The study by Adams et al. 73 evaluated the tradeoffs across five synthetic data models and three patient-level datasets in terms of privacy, fidelity, and utility. The study shows that when data fidelity declines, particularly when correlation patterns are not well preserved, it negatively impacts the overall utility of the data. Moreover, achieving high levels of fidelity can also introduce significant privacy risks.
Figure 2 illustrates the proposed conceptual Framework for the Critical Evaluation of Synthetic Data in African Healthcare settings. It outlines a multilayered process for evaluating synthetic data within the African healthcare context. Evaluation begins with technical components, where three core dimensions are assessed, data fidelity, which entails how closely synthetic data replicates real-world distributions and relationships, model performance, which defines how well models are trained on synthetic data perform compared to those trained on real data, and domain adaptation, which explains how effectively synthetic data supports clinical tasks in local healthcare settings. These evaluations uncover fidelity challenges, performance limitations, and adaptation requirements.

Conceptual framework for critical evaluation of synthetic data in African healthcare AI (source: authors’ construct).
Next, the framework underscores that synthetic data must be evaluated within the African healthcare context, which is characterized by resource constraints, fragmented health systems, and limited baseline data. These factors directly impact the feasibility and reliability of synthetic data generation and validation. A hybrid feedback loop combines both synthetic and real data iteratively. Finally, a robust evaluation of synthetic data must address adoption challenges, including technical infrastructure limitations, institutional fragmentation, human capacity gaps, and ethical and trust-related regulatory issues. These contextual factors shape how synthetic data is evaluated not only in terms of technical merit but also with respect to safety, privacy, acceptability, and suitability for real-world healthcare deployment. Henceforth, it is recommended that synthetic data releases should be accompanied by transparent documentation, such as data cards or model cards, detailing generation methods, intended use cases, and known limitations to guide safe adoption within African healthcare ecosystems.
Discussion
This study underscores that synthetic data offer both opportunities and constraints for healthcare AI in African contexts. Technical evaluations across the reviewed studies highlighted consistent performance gains when synthetic data were combined with small, real-world datasets, confirming that hybrid training strategies delivered superior generalization and reduced predictive bias. These indicated that synthetic-only approaches remained insufficient for most clinical use-cases in low-resource health systems where distributional gaps and population heterogeneity exist. Furthermore, persistent fidelity gaps, incomplete representation of rare conditions, and variable model stability underscore the need for stronger empirical evidence before large-scale clinical deployment in African settings.
Several knowledge gaps emerged from the synthesis of evidence. First, only a small subset of studies provided disaggregated performance metrics for clinically important subgroups, despite the well-known risks of bias amplification when synthetic datasets lack demographic diversity. Second, very few studies examined longitudinal and multimodal data beyond imaging and structured EHRs, leaving limited evidence for use-cases involving clinical notes, time-series signals, or multisource health records. Third, the governance literature revealed fragmented regulatory environments in African health systems, with minimal guidance on privacy-preserving synthetic data generation, quality control, or health technology assessment procedures. These gaps limit the ability of developers, regulators, and clinicians to evaluate the safety and fitness of synthetic data models.
The review also, identified broader set of trends, for examples the diffusion models and transformer-based generators were significantly reported to have potential of producing higher-fidelity medical text, synthetic waveforms, and pathology images. Domain adaptation methods gained traction as practical mechanisms for recalibrating models trained outside Africa to new population groups, although empirical evidence from African settings remained scarce. Also, the rise of open-source platforms for synthetic data generation signals a shift toward more accessible tooling; however, these platforms have not embraced synthetic data effectively.
The review results have implications for clinical governance and policy. Hybrid validation using real and synthetic data should be considered a minimum requirement for health technology assessment processes in Africa, particularly where synthetic data may obscure clinically relevant patterns. Investments in computational infrastructure, digital health records, and workforce capacity are necessary to support safe integration of synthetic data pipelines in routine care. Regulatory frameworks should also incorporate explicit standards for privacy evaluation, quality assurance, and dataset documentation. Without these foundations, synthetic data is risked in reinforcing existing inequities rather than mitigating data scarcity.
Priorities areas for future research are threefold. First, empirical studies must evaluate synthetic data across diverse African populations to quantify fidelity at the level of rare diseases, comorbidities, and regional variations. Second, the field requires scalable and low-cost tools for context-aware domain adaptation, allowing models pretrained elsewhere to be safely recalibrated using minimal local data. Third, governance scholars and health system actors should co-design regulatory mechanisms that integrate technical benchmarks with ethical, legal, and operational considerations. Advancing these research priorities areas will support a coherent pathway for safe and equitable adoption of synthetic data in African healthcare AI ecosystems.
Conclusion
Synthetic data has been gaining attention as a potential solution to the problem of data scarcity, particularly in resource-constrained settings. However, its usefulness is limited due to issues such as data fidelity, model generalizability, and domain adaptation. While there has been an increased focus on generating synthetic data, the performance difference between synthetic-only models and hybrid models persists. Limited infrastructure and institutional fragmentation maintain adoption barriers to synthetic-data-based AI systems in African healthcare contexts. The synthetic data could be yet another unrealized innovation without investment in digital infrastructure, interdisciplinary collaboration, and ethical oversight. Contextualized approaches, such as hybrid frameworks that incorporate synthetic and real data, complemented by robust validation procedures, seem to be a more viable solution moving forward. We recommend that policymakers: (1) mandate hybrid synthetic–real data validation in HTA processes; (2) establish national and regional regulatory standards for synthetic data in healthcare AI; (3) invest in infrastructure and capacity to enable fidelity benchmarking and domain adaptation; and (4) embed synthetic data strategies into health information systems to ensure interoperability, transparency, and trust. These steps will support the integration of equitable, effective, and sustainable AI in African healthcare.
