Abstract
Introduction
Acne vulgaris is a chronic inflammatory skin disease of the pilosebaceous unit, characterized by a range of lesions including comedones, papules, pustules, nodules, and cysts. 1 It often leads to post-inflammatory hyperpigmentation, permanent scarring, and psychological comorbidities such as anxiety and depression, contributing to a substantial cosmetic and psychosocial burden.1,2 Epidemiological studies have identified acne as the eighth most prevalent disease globally, with a particularly high burden among adolescents in East Asia. 3 Recent advances have expanded the understanding of acne pathophysiology and treatment strategies. 4 However, a persistent gap remains between evidence-based recommendations and real-world treatment decisions. 5
Currently, the global acne treatment market has expanded steadily, driven by increasing healthcare demand and the emergence of new therapeutic strategies. At the same time, short-video platforms have become influential channels for health information dissemination, 6 particularly among younger users. 7 In China, platforms such as Bilibili and TikTok attract millions of daily active users, 8 where dermatologic topics—including acne—are widely discussed.
However, concerns remain regarding the accuracy and credibility of health-related content on these platforms. Videos created by non-medical individuals often receive high engagement despite containing unverified or misleading claims,9–11 highlighting the fragmented nature of online health communication. Although similar quality concerns and patterns of dermatologic misinformation have been reported across multiple social media platforms,11–13 evidence specifically focusing on acne-related information on Chinese short-video platforms remains limited. Moreover, few studies have systematically assessed content quality using validated instruments in conjunction with audience engagement metrics. 13
Given the growing role of short-form videos in public health communication, a clearer understanding of how video characteristics influence both informational quality and user engagement is urgently needed. This study systematically evaluates acne-related videos on Chinese short-video platforms (Bilibili and TikTok) according to uploader characteristics, disease-related topics, and presentation formats. Content quality is assessed using four validated tools: the
Methods
Figure 1 provides an overview of the study workflow and analytical framework.

Overview of the study workflow and analytical framework. JAMA:
Ethical considerations
This study was reviewed and approved by the Ethics Committee of the First Affiliated Hospital of Xi’an Jiaotong University (Approval No. KYLLSL-2026-076-01).
All data analyzed in this study were obtained from publicly accessible short-video platforms, Bilibili and TikTok, in compliance with the respective platforms’ terms of service. No personal identifiers (e.g., user IDs, usernames, or other sensitive information) were collected, stored, or disclosed in the manuscript or supplementary materials. In accordance with institutional review board guidance and the Committee on Publication Ethics (COPE) guidelines, individual informed consent was not required.
Search strategy
From May 1 to 5, 2025, systematic searches were conducted using the Chinese keywords “痤疮” (acne vulgaris), “粉刺” (comedones), and “青春痘” (pubertal acne) on Bilibili and TikTok. To minimize algorithm-driven personalization and potential selection bias, searches were performed using newly registered accounts with no prior browsing history or user interaction.
For each platform, the top 150 videos returned by the default search algorithm were collected, resulting in a total sample of 300 videos.11,14 A ranking-based sampling strategy was adopted to enhance ecological validity, as the sampling frame consisted of content prioritized by platform search and recommendation systems. Although this approach is inherently influenced by algorithmic ranking mechanisms, it more closely reflects the acne-related information that users are likely to encounter during routine browsing. 15 Videos uploaded within the 7 days prior to data collection were excluded to allow engagement metrics to stabilize and to reduce temporal variability associated with early-stage algorithmic promotion,16,17 thereby improving the reliability of analyses examining the relationship between informational quality and user engagement.
Inclusion criteria were as follows: (1) Chinese-language content; and (2) a primary focus on acne-related health information. Exclusion criteria included: (1) Duplicate videos across platforms; (2) content that was primarily commercial or promotional; and (3) videos unrelated to acne or dermatologic health.
Data collection
For each eligible video, a standardized set of variables was extracted. Video characteristics included length (in seconds), duration (days from upload to data collection), and user engagement metrics: number of likes, comments, collections, and shares. Daily engagement metrics were calculated by dividing total engagement counts by the number of days from video upload to data collection. Uploader characteristics comprised verification status, and professional background. Uploaders were categorized as professionals—comprising Western medicine doctor, other medical workers or medical students, hospital, official media, and traditional Chinese medicine (TCM) practitioner—or non-professionals, including self-media and for-profit company. Each video was classified by content category, based on disease-related topics: Anatomy/physiology, causes/prevention, pathology/mechanism, epidemiology, symptoms, diagnosis/examination, treatment, prognosis, and TCM. Presentation format was documented independently, encompassing: Doctor monologue, patient video blog (video log (VLOG)), interactive question-and-answer (Q&A) session, PowerPoint (PPT) or lecture explanation, animation, television (TV) program/documentary, and other formats. Detailed definitions are provided in the Supplemental Methods.
Video quality assessment
Video quality was evaluated using four validated instruments that have been widely applied and demonstrated acceptable validity and reliability in assessing audiovisual health information on video-based social media platforms.14,18,19
The JAMA benchmark criteria, originally developed for text-based medical information, assess transparency and reliability across four domains—authorship, attribution, disclosure, and currency—with a total score ranging from 0 to 4. 20 The mDISCERN tool, originally designed for written health information, evaluates the reliability of digital health content using five equally weighted items (total score: 0–5).21,22 Together, these instruments assess core aspects of informational quality that are applicable regardless of content length or presentation format. The GQS, also rated on a 5-point Likert scale, offers a subjective overall assessment of video quality from a user-oriented perspective, considering clarity, educational value, and ease of understanding.22–24 The VIQI measures four aspects—information flow, accuracy, production quality, and use of visual aids—using a 5-point Likert scale, providing insights into both audiovisual and presentation-related quality.19,22 Detailed definitions and scoring criteria are provided in Tables S1–S4.
All videos were independently evaluated by two trained raters. Any discrepancies were resolved through discussion with a third investigator until consensus was reached. Inter-rater reliability was assessed using an overall weighted Cohen's kappa coefficient (κ), with agreement interpreted as almost perfect (κ > 0.80), substantial (κ = 0.60–0.80), or moderate (κ = 0.40–0.60). 25
Statistical analysis
Data distribution was assessed using the Shapiro–Wilk test. Continuous variables were reported as mean ± standard deviation (SD) or median (interquartile range, IQR), as appropriate. Group comparisons were conducted using the Mann–Whitney
Spearman rank correlation was used to explore associations between video quality scores and engagement metrics, with correlation strength categorized as negligible (<0.20), weak (0.20–0.40), moderate (0.40–0.60), strong (0.60–0.80), or very strong (>0.80).
To identify predictors of high-quality videos, defined as a GQS score ≥ 4, 26 univariable logistic regression models were applied to individual video attributes, including quality assessment scores and engagement metrics, followed by a multivariable model incorporating multiple predictors. Discriminative performance was evaluated using receiver operating characteristic (ROC) curve analyses with the area under the curve (AUC) and 95% confidence intervals (CIs).
All statistical analyses were performed using SPSS version 18.0 (IBM Corp.), R version 4.3.2, and GraphPad Prism version 9.0. A two-tailed
Results
Characteristics of acne-related videos
A total of 272 acne-related videos were included in the final analysis, with 122 videos from Bilibili and 150 from TikTok. Significant differences in video characteristics were observed between the two platforms.
The length of videos on Bilibili was substantially longer than those on TikTok (median [IQR]: 409 [204–637] seconds vs. 51 [38–93] seconds;
Regarding uploader characteristics, the proportion of verified accounts was significantly lower on Bilibili than on TikTok (33.33% vs. 68.10%), and professional accounts contributed a smaller share of content on Bilibili compared to TikTok (25.64% vs. 65.52%) (Figure S1, Table S5). In terms of uploader identity, videos on both platforms were primarily produced by Western medicine doctors and self-media (Figure 2A). Among the 78 Bilibili uploaders, self-media accounted for the largest share (42.31%), whereas among the 116 TikTok uploaders, Western medicine doctors were most common (56.03%), followed by self-media (22.41%) and for-profit companies (11.21%) (Figure 2B, Table S5).

Distribution of acne-related videos on Bilibili and TikTok. (A and B) Distribution by uploader types: Overall proportions across both platforms (A) and platform-specific comparison (B). (C and D) Distribution by disease-related topics: Overall proportions across both platforms (C) and platform-specific comparison (D). (E and F) Distribution by presentation formats: Overall proportions across both platforms (E) and platform-specific comparison (F). VLOG: video log; PPT: PowerPoint; Q&A: question and answer; TV: television.
With respect to disease-related topics, since a single video could address multiple categories, the most frequently covered themes were treatment and symptoms, while epidemiology and prognosis received less attention (Figure 2C). Anatomy/physiology was more often featured on Bilibili (7.38%), whereas diagnosis and examination content was more prevalent on TikTok (17.33%). Content related to TCM appeared in about 10% of videos on both platforms (Figure 2D, Table S6).
As for presentation formats, doctor monologues were the most common overall (Figure 2E). Additionally, patient VLOGs were more frequently observed on TikTok (22.00%) (Figure 2F, Table S7).
Quality and engagement across subgroups
Inter-rater agreement between the two reviewers was excellent (Cohen's κ = 0.889).
By platform
Overall, both platforms exhibited suboptimal video quality, but significant inter-platform differences were observed (Table 1). In terms of the JAMA benchmark criteria, TikTok videos achieved a significantly higher total score than Bilibili videos (1.28 ± 0.45 vs. 0.92 ± 0.98;

Item-level video quality scores and engagement metrics on Bilibili and TikTok. (A–C) Radar plots showing item-level scores for the JAMA (A), the mDISCERN (B), and the VIQI (C). (D–G) Violin plots with embedded boxplots showing the distributions of log-transformed engagement metrics [log10(x + 1)] for likes (D), comments (E), collections (F), and shares (G). JAMA:
Quality assessment of acne-related videos on Bilibili and TikTok.
JAMA:
Mann–Whitney
Chi-squared test.
Continuity correction.
Fisher's exact test.
Regarding engagement metrics, although the absolute numbers of likes, collections, and shares were generally higher on TikTok compared to Bilibili, only the difference in shares reached statistical significance (
By uploader characteristics
Videos from verified uploaders exhibited significantly higher quality scores across all metrics (JAMA, mDISCERN, GQS, and VIQI; all
Regarding different types of uploaders, video quality varied notably across groups (Figure 4A-D, Table S9). Videos uploaded by other medical workers or medical students received the highest scores in JAMA (1.46 ± 0.78), GQS (3.69 ± 0.86), and VIQI (13.31 ± 1.70), followed by those from official media and Western medicine doctors. By contrast, videos uploaded by for-profit companies, self-media creators, and TCM practitioners tended to receive lower quality scores across most indicators. Significant differences were observed across uploader types in GQS (

Comparison of video quality and engagement metrics across uploader types. (A–D) Mean scores of four quality evaluation tools across uploader categories:
Regarding engagement metrics, videos uploaded by official media consistently received the highest number of likes, comments, collections, and shares, significantly outperforming other uploader types, particularly in terms of comments and shares (
By disease-related topics
Among the different types of disease-related topics, videos focusing on anatomy/physiology received the highest quality scores, particularly in GQS (4.11 ± 0.78) and VIQI (14.78 ± 2.28), followed by those addressing epidemiology (mDISCERN: 3.00 ± 1.73; GQS: 4.00 ± 1.00) and pathology/mechanism (GQS: 3.51 ± 0.75; VIQI: 13.11 ± 2.27). In contrast, videos on prognosis exhibited the lowest scores across most indicators, especially GQS (2.50 ± 0.55) and VIQI (9.17 ± 2.79), indicating limited usefulness and presentation quality. Statistical analyses revealed significant differences in mDISCERN, GQS, and VIQI scores among topic categories (all

Comparison of video quality and engagement metrics by disease-related topics. (A–D) Mean scores of four quality evaluation tools across topic categories:
With respect to engagement, videos focusing on epidemiology, although not statistically significant, achieved the highest levels of engagement, especially in comments and shares. Other topic types generally showed lower engagement performance (Figure 5E–H, Table S12).
By presentation formats
Across different presentation formats, videos featuring doctor monologues demonstrated the highest overall quality, with strong scores in GQS (3.45 ± 0.86), and VIQI (12.46 ± 2.29). Both interactive Q&A and PPT or lecture explanations performed well in comparable mDISCERN scores (2.44 ± 0.88 and 2.47 ± 0.92, respectively) and favorable GQS ratings (3.11 ± 0.60 and 3.27 ± 0.59) though the latter showed a relatively low JAMA score (0.80 ± 0.78), suggesting potential limitations in credibility or authorship transparency. Videos presented in animation style achieved the highest VIQI score (12.67 ± 1.75), reflecting relatively strong production quality. In contrast, patient VLOGs showed the lowest scores across most quality indicators, particularly GQS (2.49 ± 0.64) and VIQI (9.97 ± 2.25), reflecting reduced educational value and presentation quality. Statistical analyses revealed significant differences in JAMA, GQS, and VIQI scores across different formats (all

Comparison of video quality and engagement metrics by presentation formats. (A–D) Mean scores of four quality evaluation tools across different presentation formats:
With respect to engagement, TV programs/documentaries showed the highest overall user interaction, particularly in shares (
Correlation and predictive analysis
Spearman correlation analysis was conducted to explore relationships between video quality indicators and engagement metrics (Figure 7). Correlations among quality measures revealed a strong association between GQS and VIQI (ρ = 0.755), whereas JAMA showed only weak correlations with both VIQI and GQS (ρ = 0.309; ρ = 0.383, respectively). Very strong intercorrelations were observed among engagement metrics (ρ = 0.816–0.944, all

Correlations among video quality scores and engagement metrics. Spearman correlation heatmap showing the associations between engagement metrics (likes, comments, collections, and shares) and video quality assessment scores, including the JAMA, mDISCERN, GQS, and VIQI. Color intensity represents the strength of correlation, with numerical values indicating Spearman correlation coefficients. Asterisks denote statistical significance. JAMA:
To further assess the predictive validity, ROC curve analyses was performed using a GQS threshold of ≥ 4 to define high-quality videos. Among quality metrics, VIQI demonstrated the strongest discriminative ability in univariable analyses (AUC = 0.922, 95% CI: 0.892–0.952), while in multivariable analysis, a logistic regression model incorporating all three quality scores further improved performance (AUC = 0.936, 95% CI: 0.908–0.963) (Figure 8A). For engagement metrics, collections had the strongest predictive value (AUC = 0.901, 95% CI: 0.863–0.940), whereas the combined engagement model yielded an AUC of 0.879 (95% CI: 0.836–0.921) (Figure 8B). These results indicate that structured quality assessments provide higher discriminative precision, whereas user interaction patterns can only moderately reflect content quality.

Predictive performance of video quality and engagement metrics for identifying high-quality videos. Receiver operating characteristic curves evaluating the discriminative performance of video quality assessment tools (JAMA, mDISCERN, and VIQI) (A) and engagement metrics (likes, comments, collections, and shares) (B) for identifying high-quality videos (defined as GQS ≥ 4) based on univariable and multivariable logistic regression models. JAMA:
Discussion
Acne vulgaris is a highly prevalent dermatologic condition with substantial psychosocial impact, making its management a subject of widespread public concern. In recent years, short-video platforms have emerged as influential tools for health communication, especially among younger populations. In this context, this study systematically evaluated acne-related videos on two major Chinese platforms, Bilibili and TikTok, by integrating four validated quality assessment tools (JAMA, mDISCERN, GQS, and VIQI) with four engagement metrics (likes, comments, collections, and shares).
A total of 272 videos were analyzed. Overall, video quality on both platforms was suboptimal. However, the observed inter-platform differences reflected complementary strengths rather than consistent superiority. Bilibili videos were generally longer, less interactive, and demonstrated slightly better structural coherence and informational accuracy, as indicated by higher VIQI scores. In contrast, TikTok videos were shorter, attracted significantly greater daily engagement, and were more frequently uploaded by verified and professional accounts, with modestly higher transparency, as indicated by higher JAMA and mDISCERN scores. Beyond platform-level differences, uploader characteristics, disease-related topics, and presentation formats were also associated with variation in quality and engagement. Videos from other medical workers or medical students achieved the highest quality scores, while those uploaded by official media attracted the greatest engagement. Content related to anatomy/physiology received the highest quality ratings, whereas epidemiology-themed videos elicited the most interaction. Among presentation styles, doctor monologues demonstrated superior quality, while TV programs/documentaries generated the highest levels of engagement. Correlation analysis demonstrated a strong association between GQS and VIQI, while engagement metrics were highly intercorrelated. Among them, collections showed notable correlations with both GQS and VIQI. Predictive modeling confirmed that structured quality assessments outperformed engagement metrics in identifying high-quality content. Together, these findings provide an empirical basis for interpreting the observed patterns of quality and engagement in acne-related short-video content.
Platform-specific differences
Given the overall suboptimal quality and cross-platform differences, item-level findings offered detailed insights. TikTok videos universally fulfilled the JAMA authorship criterion (item 1), compared to less than half on Bilibili—aligning with previous research that underscores TikTok's stricter credentialing and emphasis on source transparency. 27 However, TikTok videos scored poorly on JAMA attribution and disclosure (items 2 and 3), underscoring that authorship transparency does not necessarily translate into reliable referencing or conflict-of-interest disclosure. This pattern is consistent with previous observations that social media platforms often prioritize identity verification over citation practices.10,28 TikTok also outperformed Bilibili on mDISCERN item 1, suggesting more explicitly framed content. In contrast, Bilibili scored higher on VIQI domains such as information flow (item 3) and accuracy (item 4), reflecting stronger structural coherence and content reliability.
These patterns may be partly explained by differences in platform design and algorithmic priorities. 29 Bilibili's longer video duration and greater content persistence may facilitate more systematic, knowledge-oriented communication and long-term accessibility of educational materials. The platform is structured around a professional user-generated video model, with interest-based communities, durable content archives, and participatory features such as real-time “danmu” bullet comments, which are commonly associated with sustained attention and communal knowledge sharing. 30 In contrast, TikTok's short-form, algorithm-driven interface emphasizes immediacy and user responsiveness. Its “For You” feed ranks videos based on user engagement signals—such as likes, shares, watch time, and replays—combined with video content and user profile information, thereby promoting the rapid diffusion of highly engaging content. 31 This structural logic may favor concise and visually stimulating formats, which can pose challenges for delivering in-depth and comprehensive medical information. Consistent with this, previous studies have noted TikTok's advantage in visibility and engagement32,33 but relatively lower citation completeness and content comprehensiveness.34,35
From a public health perspective, these observations highlight a fundamental trade-off. TikTok enables rapid dissemination and high audience engagement, supported by greater transparency of authorship, whereas Bilibili offers more coherent and accurate educational content but struggles to achieve comparable reach. These complementary characteristics suggest that platform-specific communication strategies may be warranted, such as concise, high-impact content for TikTok, and comprehensive, structured resources for Bilibili—to improve both accuracy and accessibility in dermatologic communication.
Influence of uploader characteristics
TikTok featured a higher proportion of verified and professionally affiliated contributors, reflecting its stricter credentialing system, which typically requires physicians to hold senior clinical titles or affiliations with professional societies. In contrast, Bilibili applies looser verification standards and promotes creative diversity, resulting in a broader range of contributors, particularly self-media and non-clinical accounts. 23 This structural contrast between a credential-based model and a community-driven framework may partly explain the observed differences in video quality and user engagement across the two platforms.
Quantitative analysis revealed a significant association between uploader credentials and video quality. Across all four evaluation tools (JAMA, mDISCERN, GQS, and VIQI), videos uploaded by verified or professional contributors consistently outperformed those from uncertified or non-professional sources, aligning with recent findings in the field of digital health communication.32,34,36
Engagement metrics, however, revealed a more nuanced pattern. Non-professional creators, particularly self-media, tended to receive more immediate interaction in the form of likes and comments.34,37 In contrast, professional and verified accounts performed better in terms of collections and shares, suggesting that their content may be perceived as more credible, valuable, or worth revisiting and redistributing, 10 although direct evidence linking professional credentials to specific engagement behaviors remains limited. 34 This reflects a familiar tension between popularity and informational reliability: Emotionally appealing or entertaining content may garner more attention, even when its scientific rigor is limited.10,37,38 Notably, institutional accounts, such as official media, stood out as a distinctive subgroup. Although representing a small share of the dataset, their videos consistently achieved the highest engagement across all metrics, indicating that institutional authority remains a strong signal of trustworthiness and audience engagement. 39
In summary, platform governance plays a pivotal role in shaping both the quality and dissemination of medical information. TikTok could promote professionally verified content through algorithmic adjustments, while Bilibili might improve overall quality by strengthening credentialing requirements for health-related uploads.
Impact of disease-related topics and presentation formats
As for disease-related topics, videos covering anatomy and physiology received the highest quality scores, likely due to their structured content and suitability for expert-led explanations. In contrast, videos on prognosis tended to score lowest across most dimensions. Although overall engagement was similar across topics, content focusing on epidemiology showed relatively higher interaction, suggesting audience interest in broader contextual understanding of acne.
Presentation styles significantly influenced video performance.32,40 Doctor monologues, PPT or lecture explanations, and animations consistently demonstrated higher quality, while TV programs/documentaries—though less frequent—achieved the strongest engagement levels, reflecting the communicative strength of institutionally produced, narrative-driven formats. 41
Taken together, these findings suggest that professionally led content and formats are associated with higher informational quality, while diverse narrative and documentary styles are linked to greater audience engagement. Rather than representing competing approaches, these complementary strengths highlight the value of integrating professional expertise with engaging presentation formats to balance credibility and communication impact.
TCM-related content accounted for approximately 10% of acne-related videos across both platforms, whereas TCM uploaders made up only about 3% of all contributors. In TCM theory, acne is typically attributed to internal imbalances, such as “overheating of the lung or stomach” or “blood stasis,” and is treated through herbal medicine, acupuncture, or dietary regulation. 42 Despite its cultural relevance and long-standing presence in dermatologic care, TCM videos consistently received lower quality scores, particularly in the JAMA and mDISCERN domains. This may reflect a predominance of uncertified contributors and limited adherence to standardized, evidence-based communication frameworks. Similar shortcomings have been reported in TCM-related content in other medical fields, particularly regarding source transparency and educational rigor.43,44 To improve the reliability of TCM content on digital platforms, efforts should be made to encourage participation from certified practitioners and to promote collaboration with professional health communicators.
Association between content quality and user engagement
Correlation and predictive analyses revealed a nuanced relationship between content quality and user engagement. Among quality measures, VIQI and GQS showed strong internal correlation, indicating that presentation quality and perceived usefulness often co-occur. Both also demonstrated strong associations with engagement metrics, particularly collections and likes, suggesting that users may be more responsive to visually coherent and educationally valuable content. In contrast, JAMA and mDISCERN, which emphasize source attribution and reliability, exhibited only weak correlations with both other quality tools and engagement indicators, highlighting a disconnect between source transparency and audience attention. Multivariable logistic regression confirmed that structured quality metrics more accurately identified high-quality videos compared to engagement indicators. However, the strong univariable predictive value of collections suggests that certain forms of user behavior may still reflect perceived trustworthiness and informational value. These findings underscore a persistent challenge in short-video health communication: popular content is not always reliable, and reliable content is not always popular. 45 To bridge this gap, platforms may need to move beyond engagement-based algorithms and adopt quality-aware strategies, such as expert labeling, credential verification, or algorithmic weighting based on validated tools to ensure that dermatologic information is not only broadly disseminated, but also accurate and trustworthy.
The divergence between popularity and informational rigor warrants particular attention in adolescent acne communication, as adolescents are both heavily affected by acne46,47 and constitute a major user group of short-video platforms. 48 In our study, self-media videos, typically demonstrating moderate rather than high informational quality, were observed to achieve disproportionately high engagement, suggesting that in adolescent-oriented digital environments, medically rigorous information should not be disadvantaged in terms of accessibility or dissemination competitiveness.
This study has several strengths. First, it provides a structured, cross-platform evaluation of acne-related videos on two widely used Chinese short-video platforms, offering timely insight into dermatologic health communication within a fast-evolving media environment. Second, by incorporating uploader characteristics, disease-related topics, and presentation formats into the analytical framework, the combined use of four validated quality tools (JAMA, mDISCERN, GQS, VIQI) and four user engagement metrics enabled a comprehensive analysis linking informational quality with audience response. All videos were independently rated by trained dermatology physicians, with high inter-rater agreement, and analytic robustness was ensured through correlation analysis, ROC modeling, and multivariable prediction. Third, algorithmic and sampling bias were minimized through predefined keyword queries, newly registered accounts, and standardized inclusion criteria. Collectively, these strengths provide a reproducible methodological framework for future investigations and contribute to the development of platform-aware strategies to improve the quality and impact of digital dermatologic content.
Several limitations should also be acknowledged. First, the cross-sectional design limits causal inference and reflects only a snapshot of content within a rapidly evolving media landscape. Second, by focusing on Chinese-language content and top-ranked videos, the study may not fully represent lower-ranked or non-Chinese content, limiting generalizability across linguistic and regional contexts. Third, although four validated instruments were employed, some (e.g., JAMA, mDISCERN) were originally developed for static, text-based health information and primarily focus on informational transparency and reliability. As a result, certain platform-specific audiovisual and narrative features of short-form videos may not be fully captured. In addition, some categories, including TCM uploaders and epidemiology-related videos, were represented by relatively small sample sizes, which may limit the robustness of subgroup comparisons. Accordingly, related findings should be interpreted with caution. Despite excellent inter-rater reliability, subjective scoring and potential misclassification of uploader identity cannot be fully excluded. Additionally, key engagement indicators such as view counts, dislikes, and watch duration were not accessible due to platform restrictions, which may have constrained the comprehensiveness of user interaction analysis. Lastly, algorithmic amplification may skew engagement metrics, complicating the interpretation of user preferences. 49 These limitations underscore the need for future longitudinal, multi-platform studies incorporating user behavior data and algorithmic transparency to better understand how dermatologic content is encountered, interpreted, and trusted.
Conclusion
This study demonstrated that the overall quality of acne-related videos on both Bilibili and TikTok was suboptimal, with each platform exhibiting distinct strengths. Bilibili videos showed slightly greater structural coherence and informational accuracy, whereas TikTok videos were more frequently uploaded by verified professionals, displayed higher transparency, and achieved substantially higher levels of daily engagement. Videos from professional contributors received higher quality scores, while content from official media and self-media accounts garnered more user interaction. Both topic selection and presentation format significantly influenced video quality and engagement metrics. Although user engagement showed only moderate correlation with content quality, predictive modeling confirmed the value of structured quality assessments and collections among engagement metrics in identifying high-quality content. These findings underscore the need for platform-specific strategies that account for differences in content structure, dissemination mechanisms, and user engagement patterns, in order to enhance the visibility and reliability of dermatologic information.
Supplemental Material
sj-docx-1-dhj-10.1177_20552076261430231 - Supplemental material for Quality, engagement, and predictive validity of acne-related short videos on Chinese platforms Bilibili and TikTok: A cross-sectional content analysis
Supplemental material, sj-docx-1-dhj-10.1177_20552076261430231 for Quality, engagement, and predictive validity of acne-related short videos on Chinese platforms Bilibili and TikTok: A cross-sectional content analysis by Yuhan Xie, Qinxiao Li, Wenmin Deng, Yuxin Yan, Longmei Duan, Yuting Chen, Yusheng Wan, Kainian Han, Heni Ma and Yan Zheng in DIGITAL HEALTH
Supplemental Material
sj-docx-2-dhj-10.1177_20552076261430231 - Supplemental material for Quality, engagement, and predictive validity of acne-related short videos on Chinese platforms Bilibili and TikTok: A cross-sectional content analysis
Supplemental material, sj-docx-2-dhj-10.1177_20552076261430231 for Quality, engagement, and predictive validity of acne-related short videos on Chinese platforms Bilibili and TikTok: A cross-sectional content analysis by Yuhan Xie, Qinxiao Li, Wenmin Deng, Yuxin Yan, Longmei Duan, Yuting Chen, Yusheng Wan, Kainian Han, Heni Ma and Yan Zheng in DIGITAL HEALTH
Footnotes
Acknowledgements
The authors thank the video uploaders for their contributions to public health.
Ethics approval
This study was reviewed and approved by the Ethics Committee of the First Affiliated Hospital of Xi’an Jiaotong University (Approval No. KYLLSL-2026-076-01).
Consent to participate
All data were obtained from publicly accessible short-video platforms (Bilibili and TikTok) in compliance with the respective platforms’ terms of service. No personal identifiers or private user information were collected or reported; therefore, individual informed consent was waived in accordance with Institutional Review Board Guidance and COPE guidelines.
Authors’ contributions
Yuhan Xie and Qinxiao Li conceived the study, contributed to study design and data interpretation, and drafted the manuscript. Wenmin Deng and Yuxin Yan collected and organized the data. Longmei Duan and Yuting Chen performed statistical analyses. Yusheng Wan and Kainian Han contributed to data validation and manuscript revision. Heni Ma assisted with manuscript preparation and formatting. Yan Zheng supervised the study, provided critical revisions, and served as the corresponding author. All authors read and approved the final manuscript.
Funding
The authors received no financial support for the research, authorship, and/or publication of this article.
Declaration of conflicting interests
The authors declared no potential conflicts of interest with respect to the research, authorship, and/or publication of this article.
Availability of data and materials
The datasets generated and analyzed during the current study are available from the corresponding author upon reasonable request.
Supplemental material
Supplemental material for this article is available online.
References
Supplementary Material
Please find the following supplemental material available below.
For Open Access articles published under a Creative Commons License, all supplemental material carries the same license as the article it is associated with.
For non-Open Access articles published, all supplemental material carries a non-exclusive license, and permission requests for re-use of supplemental material or any part of supplemental material shall be sent directly to the copyright owner as specified in the copyright notice associated with the article.
