Abstract
In late 2017, a Reddit user calling themselves “DeepFakes” posted pornographic videos in which the faces of celebrities were realistically superimposed on those of porn actors using open-source face-swapping technology. Since then, the concept of deepfakes has entered the public eye and received increasing attention. Deepfakes, a portmanteau of “deep learning” and “fake,” are produced using artificial intelligence (AI), particularly machine-learning techniques that “merge, combine, replace, and superimpose images and video clips to create fake videos that appear authentic” (Westerlund, 2019: 39). While deepfakes are among the most recognized examples, they belong to a broader category known as synthetic media. This term encompasses a diverse range of digital audiovisual artifacts generated from new or preexisting digital information. The outputs of synthetic media are often substantially distinct from their original data sources, but to the average viewer, they can be almost indistinguishable from authentic content (Barnes & Barraclough, 2020). Deepfakes thus exemplify the potential of synthetic media to produce highly realistic yet fabricated audiovisual content that depicts someone saying or doing something that never actually occurred.
Deepfake technology is a double-edged sword. On the positive side, it can provide benefits to many industries, such as entertainment and business. For example, it can be employed to produce more realistic visual effects in films, television programs, and other media forms, thereby delivering a more immersive and engaging experience for audiences. It also enables brands to provide highly personalized recommendations to meet consumer needs and lowers the cost of producing video campaigns for marketers, as in-person actors may no longer be required. However, this technology brings serious challenges to assessing visual authenticity and could become a major threat to individuals and society as a whole. For instance, deepfakes have been widely used to produce nonconsensual pornographic content, with 96% of online deepfakes being porn targeting specific women as subjects (Ajder et al., 2019). In addition, deepfakes open up new possibilities for the production and spread of disinformation, which is defined as false information
Faced with the opportunities and challenges posed by deepfakes, a growing number of studies have investigated deepfakes in a wide range of contexts, such as politics (Appel & Prietzel, 2022; Hameleers et al., 2022), entertainment (Feffer et al., 2023; Singh et al., 2023), business (Powers et al., 2023; Sivathanu & Pillai, 2023), and sexuality (Fido et al., 2022; Flynn et al., 2022). These studies have evolved across disciplines and have become increasingly diverse in terms of concepts, theoretical perspectives, and methodologies. For instance, some scholars highlight the element of deceptive intent in their definitions of deepfakes, referring to them as “AI-enabled multimodal disinformation” (Lee & Shin, 2022: 533) or “advanced forms of visual disinformation” (Weikmann & Lecheler, 2024: 4). In contrast, other researchers take a more neutral perspective, defining deepfakes as synthetic media that present highly realistic content depicting individuals saying or doing things that never actually occurred, without necessarily implying malicious intent or deliberate deception (Preu et al., 2022; Westerlund, 2019). The lack of consensus on a definition of deepfakes was almost inevitable, given that researchers from various disciplinary backgrounds typically define deepfakes differently based on their disciplinary traditions and primary concerns. Even so, it is necessary to examine the conceptual elements on which most researchers in the field agree to promote a consistent and coherent understanding of deepfakes.
Naturally, research on deepfakes has been characterized by diverse theoretical perspectives and methodological approaches. Theories from different disciplines, such as psychology, communication, and sociology, have been adopted to study the phenomena and issues related to deepfakes. Likewise, a variety of methods, both qualitative and quantitative, have been used to empirically examine deepfakes. Some studies have utilized classic methods of social sciences, such as surveys, experiments, interviews, and focus groups, to collect data from various groups of people, including deepfake creators, audiences, and fact-checkers (Ahmed, 2022; Vaccari & Chadwick, 2020; Weikmann & Lecheler, 2024). Others have investigated deepfakes via digital artifacts using techniques such as discourse analysis, thematic analysis, and topic modeling to extract data from online resources (Brooks, 2021; Tang et al., 2023; Wahl-Jorgensen & Carlson, 2021). The complex patterns of differing theoretical frameworks and methodological approaches highlight the importance of systematically reviewing and summarizing the existing literature on deepfakes across disciplines to identify research gaps and help guide future research.
While previous systematic reviews have provided valuable insights into trends and patterns in deepfake research (Godulla et al., 2021; Vasist & Krishnan, 2022, 2023), relatively little effort has been made to synthesize empirical evidence in this area. To the best of our knowledge, only the study by Birrer and Just (2025) has systematically analyzed both quantitative and qualitative research on deepfakes. As empirical studies continue since the cutoff date of February 2024 for the articles included in that review, there is an urgent need to integrate recent findings from the latest literature and develop an up-to-date agenda for future research. Furthermore, to the best of our knowledge, no systematic review has yet mapped out the guiding theoretical frameworks and main research themes of empirical investigations in this area. Understanding the current state of theoretical applications and research interests can lay the groundwork for future advancements in theory and knowledge. Therefore, we aimed to conduct a systematic review to summarize existing empirical research on deepfakes, clarify key conceptual elements, identify dominant theories and methods employed, and examine the primary research areas and topics of interest. Specifically, the following research questions were proposed:
Methods
Database search
Guided by the Preferred Reporting Items for Systematic Review and Meta-Analysis (PRISMA) statement (Page et al., 2021), we conducted an extensive search for relevant published (journal articles and conference papers) and unpublished (theses and dissertations) works across several electronic databases: EBSCO, ProQuest, Scopus, and Web of Science. This search was performed using specific terms (e.g., “deepfake,” “face-swap,” “image manipulation,” “synthetic media,” “video manipulation,” and “voice manipulation”), along with their derivatives, in article titles, abstracts, and keywords. These terms were selected based on prior systematic reviews (e.g., Rana et al., 2022; Stroebel et al., 2023; Vasist & Krishnan, 2022). The search was limited to articles written in English and available online before 26 July 2025. The initial database search generated 5327 articles.
To locate additional articles not included in the databases listed above, we performed a search with Google Scholar using the same keywords and scrutinized the reference lists of previous review articles on deepfakes (e.g., Godulla et al., 2021; Vasist & Krishnan, 2022, 2023). This process yielded an additional 85 articles, bringing the overall total to 5412 articles.
After removing 898 duplicate articles, 4514 articles remained for initial screening. We examined the titles and abstracts to exclude articles that were clearly irrelevant to our research questions and retained those that appeared to be eligible. A total of 1243 articles were excluded at this stage, and the remaining 3271 articles were included. The full-text versions of these articles were then retrieved, examined, and screened based on the following inclusion and exclusion criteria. Figure 1 summarizes the search process and selection steps.

Flow chart of study retrieval and selection procedures.
Inclusion and exclusion criteria
Articles were included in the current systematic review if they: (a) focused on topics related to deepfakes, (b) presented empirical studies involving the collection and analysis of primary data, (c) were available in full text to the researchers, and (d) were written in English. We excluded articles that focused solely on technical aspects of deepfake production or detection, such as developing algorithms, models, or systems to identify deepfakes (e.g., Ilyas et al., 2023; Islam, 2022), as these were not the focus of the current research. Articles that included nonempirical research, such as review papers, conceptual work, or editorials, were also excluded.
After applying these criteria, the final sample consisted of 219 articles (159 journal articles, 44 conference papers, and 16 theses/dissertations) that documented 261 empirical studies.
Coding
The articles used for analysis were coded by two independent coders. First, basic study information, including article title, year of publication, publication venue, and country of study, was recorded. Additionally, article type (i.e., whether it was a journal article, conference paper, or thesis/dissertation) and journal discipline (i.e., journal categories as classified in the Web of Science Journal Citation Report) were coded.
Second, variables related to the way the concept of a deepfake was conceptualized were coded, such as whether a definition of “deepfake” was provided, and if so, what conceptual elements (e.g., modality, technique, and purpose) were used in the definition. Moreover, whether cited references were used to define deepfakes was coded; if so, those specific references were also coded.
Third, variables related to the theoretical framework were coded. Specifically, we focused on whether theories were used in the study to guide the empirical investigation, and if so, which theories.
Fourth, variables related to methodological approaches, including research design (e.g., quantitative or qualitative), method (e.g., survey, experiment, or interview), and sampling strategies (e.g., convenience, quota, or purposive), were coded.
Finally, variables related to the research area (e.g., politics or entertainment) and research topic (e.g., production, sharing, or identification of deepfakes) were coded.
Intercoder reliability was assessed based on a random subsample of 123 articles (approximately 56% of the total sample) and calculated using Krippendorff's alpha. The reliability scores were satisfactory, with agreement ranging from .82 to one for all variables. Disagreements were resolved through discussion.
Results
Development of the research field
RQ1 asked about the progress of deepfake research in terms of years, disciplines, and national contexts. In our sample, the earliest article was published in 2018. From 2019 to 2020, only a few articles were published on this topic, with 0 in 2019 and 10 in 2020. However, since the beginning of 2021, this area has received increasingly scholarly attention, as evidenced by the rapidly growing number of studies published each year that met the criteria for being included in our review (34 in 2021, 35 in 2022, 55 in 2023, 41 in 2024, and 43 in 2025).
The research on deepfakes that we examined spanned multiple disciplines. Among the 219 journal articles in our sample, nearly one-fifth were published in communication journals (
Regarding the national context in which the data were collected, our results show a pattern that focused primarily on Western countries, with 24.5% of empirical studies coming from North America (
Conceptualization
RQ2 asked how deepfakes have been conceptualized in existing research. The results indicate that, among the 219 articles in our sample, the majority (
Additionally, the majority of definitions (
Of the 163 studies that provided a definition of deepfakes, 62.6% used definitions introduced in the previous literature. The most frequently cited definition in our sample (12.9%) was that of Westerlund (2019: 39), who proposed that deepfakes were “hyper-realistic videos that apply artificial intelligence (AI) to depict someone say and do things that never happened.” This definition was largely in line with the most commonly used conceptual elements observed in the current analysis.
Theoretical perspectives
RQ3 asked about the most commonly used theories in existing research on deepfakes. Of the 219 articles in our sample, 35.6% (
The second most frequently used theory was the modality–agency–interactivity–navigability (MAIN) model (
In addition to these two theories, media richness theory (
Methodological approaches
RQ4 concerns the methodological approaches used in existing research on deepfakes. The results reveal that quantitative design (
Among the qualitative methods used by articles in the sample, they relied mainly on interviews (
The number of samples, whether human or nonhuman, varied widely across the studies. For example, in experiments involving human participants, the sample size ranged from as few as 10 to as many as 9492. Similarly, in content analysis studies, the sample size of online content ranged from 10 videos to 86,425 comments. The populations most frequently observed in the empirical research included adults and college students. Some studies focused on more specialized samples that were less prevalent but involved the purposive sampling of individuals with specific characteristics, such as fact-checking experts, filmmakers, tourism industry employees, and patients experiencing moral injury due to sexual violence.
Research areas and topics
Regarding RQ5, the empirical studies in our sample covered a diverse array of areas. Politics (
In terms of research topics in the sampled articles, the identification and detection of deepfakes was represented in the most empirical studies (
The second most frequent category of research topics focused on the effects of exposure to deepfakes on audience responses, including perception, emotion, physiology, and behavior (
The third most frequent category of studies explored public opinion and understanding of deepfakes (
The fourth most frequent category of research topics involved the prevention and intervention of deepfakes (
The fifth most frequent category included studies examining the sharing and dissemination of deepfakes (
Studies in the sixth most frequent category focused on the creation and generation of deepfakes (
The final category of studies concerned media representations of deepfakes, namely, how deepfake-related issues were portrayed in media outlets (
Discussion
As one of the initial attempts in the field to synthesize both quantitative and qualitative research on deepfakes, the current systematic review provides an overview of the field's development, conceptual frameworks, theoretical perspectives, methodological approaches, and research areas and topics under investigation. Our results provide valuable insights for scholars and practitioners to theoretically and empirically explore the evolving field of deepfakes. In this section, we elaborate on our key findings and provide recommendations for future research.
The results showed that empirical research on deepfakes has grown considerably over the past four years, spanning diverse disciplines ranging from communication and psychology to computer science and information science. Despite the recent proliferation of studies in this area, most have been conducted in Western countries, with relatively little attention paid to Asian countries and even less to African countries. Given the high illiteracy rates in South Asia and sub-Saharan Africa (UNESCO, 2017), it is not far-fetched to imagine that people living in these areas could be more vulnerable to the negative impact of deepfakes. For instance, in 2019, the Gabonese government released a video of President Ali Bongo to dispel public rumors about his health. However, many critics and political opponents claimed that the video was a deepfake. This allegation led to a military coup in the country, which, although ultimately unsuccessful, still posed a considerable threat to domestic stability (Breland, 2019). Our findings highlight the importance of expanding the scope of current deepfake research to more diverse geographic regions, particularly those with relatively low literacy rates.
Moreover, although the growth of deepfakes around the world has become an emerging global issue, previous empirical studies have primarily adopted a single-country approach and overlooked cross-country comparisons. To this point, comparative studies across countries have primarily focused on investigating the effects of exposure to deepfakes on audience responses (e.g., Ahmed et al., 2025; Hameleers et al., 2024) and examining the antecedents of sharing deepfakes across nations (e.g., Ahmed, 2021, 2022). However, there remains a notable lack of empirical research on cross-country differences in the capability to identify deepfakes. Recent studies have documented variations across nations in the ability to detect AI-generated media—specifically text, images, and audio. For example, Frank et al. (2024) found that German participants exhibited greater proficiency in detecting AI-generated audio compared to participants from the United States and China, whereas Chinese participants were more adept at identifying AI-generated images than their counterparts in the United States and Germany. Future investigations should expand this research trajectory into the domain of deepfakes to examine whether and how detection capabilities differ across various cultural and societal contexts. Furthermore, more cross-national comparative studies are needed to evaluate the effectiveness of intervention strategies aimed at combating deepfakes. Rather than treating intervention strategies as one-size-fits-all solutions, scholars should carefully consider context-specific differences and systematically examine how economic, cultural, political, and social factors influence the effectiveness of these strategies.
Our evidence suggests that the majority of empirical studies attempt to define the concept of a deepfake, but the definitions diverge in some respects. While there is no general consensus on how to define deepfakes, our results revealed several core conceptual elements of existing definitions, including video modality, advanced AI-driven technique, inauthentic content, and realistic appearance. Based on these elements, deepfakes can generally be defined as
Although there has been a burgeoning body of studies on deepfakes in recent years, the extant research is still at a nascent stage of theory building and testing. Our results indicated that most studies did not adopt or develop any theories in their empirical investigation. As Kerlinger (1986: 8) pointed out, “The basic aim of science is theory. Perhaps less cryptically, the basic aim of science is to explain natural phenomena. Such explanations are called theories.” According to Craig (2013: 45), “an important function of theories is to explain the regularity of empirical phenomena with reference to the functional or causal processes that produce them.” Given the pivotal role of theory in generating scientific knowledge, more theory-driven research is warranted to develop a theoretical understanding of deepfake phenomena and contribute to theory advancement and synthesis. Future research on deepfakes should devote more effort to the refinement, extension, and critique of existing theories. For example, as the current media and technological environments have changed dramatically from those that gave rise to dual-process theories, it is worthwhile to explore how the cognitive processing of deepfakes confirms, extends, or challenges the foundational assumptions of these theories. To illustrate, Appel and Prietzel (2022) developed a deepfake detection model grounded in two system models of information processing, investigating how individual differences in analytic thinking and political interest affect cognitive processes and the ability to identify deepfakes in a political context. In a similar vein, Lee's (2020) authenticity model of (mass-oriented) computer-mediated communication presents another promising theoretical lens. This model provides an integrative conceptual framework that illustrates the antecedents and consequences of authenticity judgments in digitally mediated communication. Such a framework is particularly relevant to deepfake research, as deepfake technology increasingly blurs the boundaries between genuine and fabricated content. Future research should consider leveraging and extending these frameworks to examine how the message, source, contextual, and technological factors—independently and in combination—shape users’ responses to deepfakes across various contexts.
Additionally, given the interdisciplinary nature of deepfake research (Godulla et al., 2021), it offers numerous opportunities to expand the boundaries of theories across various disciplines and to provide rich theoretical perspectives to guide future work. Thus, we advocate for closer collaboration among deepfake researchers with diverse disciplinary backgrounds—including communication and media studies, computer science, education, ethics, law, political science, and psychology—to facilitate the integration of theories and the development of new theoretical models and frameworks. Each of these disciplines contributes unique theoretical perspectives that, when integrated, have the potential to deepen the current understanding of deepfakes. For instance, social psychological theories, such as motivated reasoning theory and reactance theory, provide valuable insights into the cognitive and psychological mechanisms through which individuals process and respond to deepfake content. Political science draws upon conceptual frameworks centered on governance, power dynamics, and deliberative democracy to examine the political and societal impact of deepfakes. Law enforcement and criminology utilize concepts from digital forensics and cybercrime victimization to understand the detection, investigation, and prevention of the malicious use of deepfakes. Collaborative engagement with these theories and concepts should encompass both empirical and nonempirical approaches; for example, nonempirical research from disciplines such as law (e.g., Chesney & Citron, 2019) and technology ethics (e.g., Diakopoulos & Johnson, 2021) provides crucial insights into the legal and ethical implications of deepfakes.
Some notable methodological gaps may impede a comprehensive understanding of deepfakes. Although an a priori power analysis is generally considered an optimal method for controlling Type I and II errors to validate hypotheses (Kang, 2021), it was largely lacking in the surveys and experiments in our sample. This absence may limit our ability to determine whether nonsignificant results for a relationship are due to insufficient statistical power to detect it or because such a relationship truly does not exist in the population. Furthermore, the scarcity of longitudinal studies poses considerable challenges in identifying patterns and trends over time. This shortcoming is particularly concerning in studies assessing the efficacy of intervention strategies against deepfakes, as the primary purpose of interventions is not merely to produce immediate effects but to equip individuals with the necessary tools and knowledge for their long-term application in daily life. In addition, there was a predominance of quantitative studies using nonprobability samples, suggesting that the conclusions drawn from these studies may lack generalizability to a wider population. In light of these gaps, future quantitative research should not only conduct a priori power analyses before data collection to justify the sample sizes used but also employ longitudinal designs and more representative samples.
Another important trend in the deepfake research in our sample was a predominant focus on the political domain. While deepfakes were not originally created for political purposes, political deepfakes seemed to attract greater academic attention than their use in other areas. This may be because the malicious use of deepfakes in political contexts poses serious democratic threats by manipulating public opinions and interfering with electoral processes (Chesney & Citron, 2019; Diakopoulos & Johnson, 2021). Political deepfakes are often created to purposely spread false information about candidates, discredit opponents and rival parties, and disrupt political campaigns. Previous research has shown that exposure to political deepfakes can heighten uncertainty about information content, reduce trust in online news (Vaccari & Chadwick, 2020), undermine attitudes toward the depicted politicians (Dobber et al., 2021), and increase sharing intentions (Ahmed & Chua, 2023). Such potential consequences, along with an intense focus on political deepfakes in news media (Gosse & Burkell, 2020), may amplify public concerns about the misuse of deepfake technology in politics and receive increasingly scholarly attention. However, it is worth noting that in reality, the majority of deepfake videos circulating online are nonconsensual pornographic content targeting specific women as subjects (Ajder et al., 2019). Deepfakes have emerged as a relatively new form of gender-based violence, leveraging AI technology to humiliate, demean, and objectify women. Targets of sexual deepfakes often suffer reputational damage and experience emotional and psychological distress, such as anxiety, depression, and even suicidal ideation (Kugler & Pace, 2021). The prevalence of sexual deepfakes and the harm they cause highlight the urgent need for future research to devote more effort to investigating and addressing this issue.
When it comes to research topics, our results identified several knowledge gaps and opportunities for future research. First, existing research has focused more on the identification and detection of deepfakes than on the prevention and intervention of deepfakes. Further work is needed to explore various intervention strategies and assess their efficacy in improving the human detection of deepfakes. Second, there is a limited understanding of the psychological processes underlying the impact of deepfakes on audience responses. Most studies have focused on direct effects while overlooking potential psychological mechanisms. Additional research is thus necessary to elucidate these mechanisms and test their effectiveness in explaining deepfake effects. Third, the consequences of sharing deepfakes remain understudied. According to Barasch (2020), sharing information with others leads to two types of consequences: intrapersonal outcomes and interpersonal outcomes. The former focuses on how sharing influences the sharer's internal psychological states, while the latter concerns how sharing affects the sharer's relationship with others. These consequences largely depend on whether the information shared is positive or negative in valence. In light of this framework, future research could examine the intrapersonal and interpersonal consequences of sharing deepfakes and how such consequences vary with the valence of the shared deepfake content (e.g., entertainment vs. disinformation).
Limitations
The current results should be interpreted in light of several limitations. First, although we attempted to account for a potential publication bias by including unpublished theses and dissertations in our sample, other types of unpublished work (e.g., preprint articles) were excluded. To provide a more complete picture of deepfake research, future systematic reviews should take these unpublished materials into account. Additionally, due to the qualitative nature of this systematic review, we were unable to quantify the risk of bias in the included studies. Lastly, our sample was limited to articles published in English, so the results may not be representative of empirical research on deepfakes published in other languages. Future work should expand the inclusion criteria to include articles published in multiple languages.
Conclusion
Powered by recent advances in AI technology, deepfakes have made it easier than ever to generate highly realistic and believable media content that blurs the line between authenticity and fabrication. This study conducted a systematic review of 261 empirical studies on deepfakes to identify research trends and gaps and to point out directions for future research. In conclusion, we call for more research in non-Western and nonpolitical contexts, along with comparative approaches, to provide nuanced insights into deepfakes. We hope this review will spur more interdisciplinary research that integrates theoretical and methodological insights from multiple fields. For instance, collaboration among communication scholars, computer scientists, and psychologists could help assess how advanced deepfake detection tools affect users’ trust in detection warnings and their engagement with media content. By systematically mapping the current state of research and categorizing empirical studies into seven topics—including the identification of deepfakes, the impact of deepfake exposure, and public perceptions of deepfakes—this review provides a navigational resource for scholars. Empirical researchers can leverage this analysis to identify underexplored areas and topics, develop more robust study designs, and situate their work within a broader interdisciplinary context. For scholars dedicated to advancing theoretical frameworks for deepfakes, this review uncovers patterns and gaps that inform more thorough theorizing.
Footnotes
Funding
The author disclosed receipt of the following financial support for the research, authorship, and/or publication of this article: This work was partially supported by the General Research Fund (GRF) of the Research Grants Council (RGC) of the Hong Kong SAR (Project No. 12612624).
Declaration of conflicting interests
The author declared no potential conflicts of interest with respect to the research, authorship, and/or publication of this article.
