Abstract
Background
South Africa had an HIV prevalence of 12.7% in 2022, which translates to 7.8 million people living with the disease (Human Sciences Research Council, 2023). The country has the largest HIV programme in the world (Unitaid, 2019), with 91% of people over 15 years old on treatment in 2022 (Human Sciences Research Council, 2023). Treatment of HIV is informed by national guidelines, which are regularly updated. Ongoing training, or continuing professional development (CPD), of healthcare workers (HCWs) is vital for optimal patient care.
In South Africa there are many logistical challenges to ongoing training: the country covers an area of 1.2 million km2 with 4184 health facilities in 2018 (Health Systems Trust, 2020), many in hard-to-reach, rural settings.
Rural clinics are often run almost single-handedly by nurses, supported by community health workers (CHWs), with weekly, fortnightly or monthly visits from a doctor. In South Africa, HIV training has traditionally been face-to-face at centralised points (3–5 days), with a more recent move to online CPD. Distance and a lack of resources – human, financial and infrastructural/technological – reduce uptake of both of these methods (Feldacker, Jacob, et al., 2017; Feldacker, Pintye, et al., 2017).
Going Mobile
There is a need for innovative solutions to meet training and support needs. Social media is recognised as a way to promote research, collaboration, and academic advancement (Shah et al., 2023). WhatsApp, a free and accessible platform, has been shown to be the preferred social media platform by nurses (Ajuwon et al., 2018; Torrejon et al., 2020). With 100% cell phone penetration in 2022 (ICASA, 2022) and 93.2% of South African 16–64 year olds using WhatsApp in 2021 (Staff Writer, 2021), it is an ideal platform for training, even in the most remote clinics.
Keeping it Short, Sweet, and Interactive
Time is a precious commodity in busy clinics. Microlearning is defined as an educational experience that is focused, short and effective, usually driven by mobile devices, social connectedness and time scarcity (Torgerson, 2021). Essentially, microlearning encompasses teaching using ‘bite-sized’ information chunks: each lesson clearly explaining one concept (Kohnke, 2021). Its efficacy lies in interactivity and simplicity of design (Baumgartner, 2013) and it’s been endorsed within the health professions for student learning, training and CPD (De Gagne et al., 2019).
Microlearning has shown efficacy and satisfaction for learners across a range of health professions and social media platforms, including diagnostic evaluations for nurses using Instagram images (Tennyson et al., 2022); quiz-based microlearning on COVID for physicians, nurses, and allied health professionals (Triana et al., 2021); 3-min videos on paediatric emergency procedures for medical students (Cheng et al., 2017); and case-based mobile messaging for pharmacy students (Wang et al., 2018). Haghighat et al. (2023) used microlearning via WhatsApp and an online learning management system to improve final year nursing students’ trauma care knowledge. They found that the intervention improved knowledge and 1-month retention of knowledge, with high levels of satisfaction from participants. In their multi-country, multi-health profession study of microlearning for pharmacovigilance, Hegerius et al. (2020), also reported high levels of satisfaction with their 3- to 5-min video format, but participants requested practical exercises.
One of the regularly cited requests in studies of digital learning is for interaction between trainers and participants, and participants among themselves (Hegerius et al., 2020; Lee, 2021; Reeves et al., 2017; Tudor Car et al., 2019). The new generation of digital learners seek a humanistic approach with an online community, feedback and collaboration (Han et al., 2019).
Post-COVID, the scope of digital learning changed rapidly as educators across levels of education sought creative ways to have active learning on digital platforms (Chen et al., 2022). Tennyson et al. (2022), recommend digital microlearning as “an excellent innovative strategy to facilitate a sense of interconnectedness among nurse practitioner students engaged in distance-based learning.” Tudor Car et al. (2019), in their systematic review of digital education on clinical practice guidelines, proposed that more interactivity and engagement may improve educational gains. Developing platforms that encourage interaction in a ‘live’ online setting is key.
Using Cases to Teach
Case-based learning is a well-established teaching method and, while its effect on learning compared to other methods is inconclusive, it has been shown to be engaging, motivating and enjoyed by teachers and students (Thistlethwaite et al., 2012). Shah et al. (2023) used case-based, asynchronous microlearning via an app to train nurses on antimicrobial stewardship in a small pilot study and found improvements in knowledge and participants’ confidence, with 89% saying they thought it an effective teaching tool.
Lack of Thorough Evaluation of Innovative Training Styles
In their scoping review of 17 studies on using microlearning in health professions education, De Gagne et al. (2019) concluded that it has the potential to change education delivery and improve learning but found that there is a paucity of data on higher level learning outcomes and studies including adult learners.
The Kirkpatrick Model is an extensively used tool to evaluate training at four levels: (1) reaction – is the training favourable, engaging and relevant; (2) learning – does the training improve knowledge, skills, confidence and attitude; (3) behaviour – is what is learnt applied in the workplace; and (4) results – does training result in the desired outcome (Brooks et al., 2016; De Gagne et al., 2019).
Many previous WhatsApp-based and microlearning interventions have only been evaluated at level 1 and 2, leaving data on level 3 and level 4 scarce (Coleman & O'Connor, 2019; De Gagne et al., 2019; Tennyson et al., 2022).
Similarly, numerous other reviews of using social media in medicine have found benefits for communication and training, but an overarching need for rigorous research on its effectiveness with strong theoretical foundations, to explore its full potential (Chan & Leung, 2018; Coleman & O'Connor, 2019; Katz & Nandi, 2021).
The Gap
The call for recognition, research and development in the field of microlearning as a training tool for HCWs was amplified by the COVID pandemic (Dahiya & Bernard, 2021; Mak et al., 2021). The need for rigorous research on innovative, accessible, time-conscious, and effective training is clear.
Aims and Objectives
The primary aim of the study is to design, test and evaluate the effect of WhatsApp group-based HIV training on nurses’ and community health workers’ (CHW) knowledge in a remote area of South Africa. The primary outcome will be a change in knowledge compared to a control group (who do not receive training).
The secondary aims are to assess uptake, acceptability, and feasibility of the intervention; and to explore and describe the changes in prescribing, comparing the intervention group to the control group.
Specific Objectives.
Theoretical Framework
For the purposes of the qualitative analysis of this study, the critical realist stance that there is a reality, but our understanding of that reality is influenced and limited by context, human practices and social forces was followed (Braun & Clarke, 2022; King & Brooks, 2017c).
Two minor theories were used to guide the design of the study – the all-encompassing Unified Theory of Acceptance and Use of Technology (UTAUT) (Marikyan & Papagiannidis, 2023) and the more pragmatic framework used by Asiimwe et al. (2012), based on Jeng’s Usability Assessment of Academic Digital Libraries: Effectiveness, Efficiency, Satisfaction, and Learnability (Jeng, 2005).
The Unified Theory of Acceptance and Use of Technology (UTAUT) has been used widely, across sectors including technological tools for nurses (Barchielli et al., 2021; Su & Chao, 2022; Winckler, 2022) however, many studies cited it, but didn’t use the model (Dwivedi et al., 2019), as is the case here. The modelling itself was not used but its driving factors were used to inform the questionnaire questions and focus group semi-structured interview guide.
While Jeng’s (2005) model was aimed at digital libraries, its suggested factors affecting usability were adapted by Asiimwe et al. (2012) to inform the framework to assess the feasibility, accessibility and use of point-of-care malaria tests in Uganda. This adaption – and further adaptions thereof – have been used widely to guide the qualitative testing of usability across a range of point-of-care interventions in LMICs including South Africa (de Vos et al., 2023), Zambia (Ansbro et al., 2015), and Thailand (Bancone et al., 2022).
Usability Factors and Study Definitions Based on Asiimwe et al.’s (2012) Model.
Methods
Design
A pragmatic, mixed-methods, parallel-group cluster-randomised study (Figure 1). Quantitative analysis will be used to determine the effectiveness of WhatsApp-based training on increasing knowledge and altering prescribing behaviour. Both quantitative and qualitative analysis will be used to describe uptake, acceptability and feasibility of the intervention as a training tool. Study flowchart.
Study Setting
The study will be conducted in four predominantly rural districts in the Eastern Cape province of South Africa (Figure 2). Adult HIV prevalence in the province was 18.8% in 2022 (Human Sciences Research Council, 2023). Map of study district and towns, in South Africa.
Cluster randomisation (Hayes & Moulton, 2017) by town will be used to minimise inter-HCW contamination. Stratification across clusters will be done by number of HCWs and patients attending facilities, to obtain two equally sized groups and minimise variance between groups (Rutterford et al., 2015). Clusters will be randomly allocated within each stratum by the study statistician. Due to the nature of the intervention, blinding is not possible.
Study Population, Sampling and Recruitment
Healthcare Workers for Training
Clinics in this area are staffed by nurses, CHWs and counsellors, with weekly, fortnightly or monthly visits from a shared doctor. Few have an on-site pharmacist. Doctors and pharmacists are excluded from the study to minimise contamination across clinics.
The study is powered to 90% to detect a 2-point difference in mean HCW knowledge, accounting for the cluster design of the study (ICC = 0.01) with a sample size of 24 clusters (12 in each arm, with an average of three HCWs in each cluster) to give 5% significance. Fifty primary care clinics in 30 towns (clusters) will be included.
Sampling for the training intervention is purposive, with all CHWs, counsellors and nurses in the facilities invited to participate during recruitment visits to the clinics and flyers left for those not there on the day, with the option to join the study via WhatsApp.
Healthcare Workers for Focus Groups
Sampling for the focus groups, post-intervention, is purposive, to get rich data relevant to the training (Vasileiou et al., 2018). All HCWs who complete the training will be asked if they are willing to participate in a focus group during the online post-intervention questionnaire, and then convenience sampling will be used to pick the clinics that will be included, considering distance, number of agreeable participants and time.
While methodologies like grounded theory require sampling to ‘saturation’ or ‘data adequacy’ (Morse & Clark, 2019), there has been much debate about the concept within qualitative study (Vasileiou et al., 2018), especially when using thematic analysis (TA), which relies largely on reflexivity: each researcher may capture different nuances from the data collection and analysis (Braun & Clarke, 2019). More recent sample size discussions have recommended that sample size should be guided by the study characteristics e.g., epistemological/theoretical approach, nature of the investigated phenomenon, aims and scope, data quality and researcher skills/experience (Vasileiou et al., 2018). For the purposes of this study, Dey’s definition of ‘theoretical sufficiency’ – data collection stops when the researcher has reached a sufficient understanding – will be used (Braun & Clarke, 2019).
Patients for Folder Reviews
Study Population for Folder Reviews.
Previous prescription reviews have shown rates of incorrect prescriptions due to drug-drug interaction rates of 18.7-43.0%, adjusted mean 20.0% (Agu et al., 2014; Evans-Jones et al., 2010; Kigen et al., 2011; Kuemmerle et al., 2021; Schlaeppi et al., 2020; Seden et al., 2015). Crude population calculations were used to establish population sizes for measurable outcomes (Chiwandire et al., 2021; Epilepsy South Africa, 2020; Human Sciences Research Council, 2019; South African National Department of Health, 2019). Using an expected 10% effect size for sampling calculations, at 80% power, 5% significance, with an ICC of 0.03, between 30 and 40 patient folders per cluster (if 24-26 clusters), will be needed.
Every effort will be made to review as many folders as possible, across all clinics, with the available resources (cost, time and human). If resource constraints force it, clinics for folder review will be randomised to include a smaller number of clusters in the folder reviews. This may mean it will be under-powered but as this is a proof-of-concept, secondary objective, this is considered to be reasonable.
Inclusion and Exclusion Criteria
Inclusion Criteria.
Ethics
Ethical approval has been granted for the project through the University of Cape Town’s Human Research Ethics Committee (HREC 491/2022) and the Eastern Cape Health Research Committee (EC_202209_003).
Patient-level consent for folder reviews is not feasible and there is no risk to patients, as no personal identifiers will be included in the reporting. This is a common challenge with cluster randomised studies which resulted in the development of the Ottawa Statement in 2012 (Taljaard et al., 2013). In accordance with recommendations 6 and 7 of the Ottawa Statement, consent was obtained at provincial level.
Training Intervention
Design of the lessons – six for nurses, four for CHWs – was guided by principles outlined in Corbeil, Khan & Corbeil’s
WhatsApp groups will be made for each of the control and intervention groups, divided by profession – CHWs and counsellors together in one group; nurses in the other. Initially, all four groups will receive cell phone data to facilitate access to online questionnaires and WhatsApp groups. Information messages with links to the baseline online questionnaire will be sent and the online questionnaires will remain open for three weeks. Groups will then be told whether they are in the intervention group (receive training immediately) or control group (receive training at the end of the study).
Each lesson consists of cases and questions, sent in divided messages to encourage interaction, followed by the correct answer in both text and voice note. The training will be conducted ‘live’ in WhatsApp groups: 10–15-min, case-based microlearning lessons within the routine lunch break between 13:00 and 14:00. An introduction session will be given, followed by the lessons twice a week over three weeks (nurses) or twice a week over two weeks (CHWs).
Fifteen minutes before the beginning of each session, a ‘housekeeping’ page (Figure 3) will be sent to the WhatsApp groups. The eight points on conduct within the group were guided by the draft social media ethical guidelines of the Health Professions Council of South Africa (2019); barriers reported in previous studies (Mars & Scott, 2016; Woods et al., 2019) and review by an ethics expert. Housekeeping rules for training intervention on WhatsApp.
Outcomes
Primary and Secondary Outcomes.
Data Collection
Data will be collected in four ways: 1. Questionnaires, pre- and post-intervention: quantitative and qualitative 2. WhatsApp group interaction: quantitative and qualitative 3. Focus group/interviews: qualitative 4. Folder reviews: quantitative
Questionnaires
The questionnaires were designed, completed and stored on LimeSurvey (Community Edition; Version 5.6.10 + 230313). The first section is basic demographic information including town, profession, age, gender and years of experience.
The second section tests knowledge – ten multiple choice questions for nurses; seven for CHWs. The knowledge questions were based on those from the questionnaire disseminated in a previous study testing HCW knowledge of DTG’s drug-drug interactions (Chisholm et al., 2022), which was reviewed by HIV researchers within the Division of Clinical Pharmacology, and the AIDS and Society Research Unit at the Centre for Social Science Research, both at the University of Cape Town, to ensure questions were relevant, phrased to minimise bias and to assess reliability and validity.
Questionnaire Questions to Measure Usability of the Intervention. Answer Options: Yes/No.
The URL link to the online questionnaires will be sent in the WhatsApp groups, and will remain open for three to four weeks, with regular reminders sent to complete them. They will be administered at baseline (intervention and control groups); immediately after the training intervention (intervention group); and three months after the training (intervention and control groups). Multiple choice answers are set to change order at random, to minimise the risk of sharing of answers and/or memorisation.
WhatsApp Group Interaction
Engagement of participants in sessions will be measured using WhatsApp functionalities. WhatsApp Business allows the admin to see to whom messages have been delivered, and who has seen them.
Due to potential disruptions of loadshedding – South Africa has an energy crisis which has resulted in regular, scheduled periods of the electricity being cut off each day – this will not necessarily have to be during the live session. Each participant’s having ‘seen’ the messages in each lesson will be defined as them having received the lesson. This will be measured during the lesson, one hour after, 24 hours after, at the end of the week, and two weeks after the training.
All participant interaction will also be collected. Each day’s messages will be exported from WhatsApp and saved in a password-protected folder.
Focus Groups
Traditionally, in health service and medical education research, focus groups were used to guide the development of research questionnaires, but there has been a move toward their use to get the ‘voice’ on the design, application and evaluation of curricula (Barbour, 2005; Jaschinski & De Villiers, 2008; Lam et al., 2001).
Barbour (2010) describes focus groups as “coming into their own” when the aim is to generate research data relating to new developments or procedures, that requires contributions from everyone to answer ‘why?’ (and ‘why not?’), and to get different perspectives from the same group. The nature of the training intervention – essentially, a group activity – makes focus groups the ideal way to gain insight into the views and experiences of the healthcare workers of the WhatsApp-based training of the study.
Focus groups – with a maximum of ten participants – will be conducted in a convenient, quiet and private location, either at the clinic or in the town where the clinics are. The researcher will conduct the focus groups and an assistant will be present in the meetings. With participant permission, they will be recorded and the recordings will be saved in a password-protected folder. Questions are semi-structured (Annexure 1).
Folder Reviews
To measure changes in prescribing behaviour, retrospective folder reviews will be done in both intervention and control groups at the end of the study. The 48 weeks before the intervention will be compared to the 48 weeks after the intervention. The period of clinic visits for recruitment, prior to the intervention, will be excluded, to minimise confounding.
In the case of drug-drug interaction knowledge, proportion of incorrect prescriptions will be measured in three groups: • Patients on metformin (correct: maximum 500 mg bd) • Patients on rifampicin (correct: 50 mg dolutegravir bd) • Patients on antiepileptics (correct: no phenytoin/phenobarbitone; if carbamazepine, 50 mg dolutegravir bd)
For measuring changes in prescription on dolutegravir use in women of childbearing potential, the use of GeneXpert™, and TB preventive therapy, we will review folders of: • Women aged 18–45 years started on ART (dolutegravir -based regimens) • Pregnant women booking for the first time (GeneXpert™ performed) • All adult patients started on ART (TB preventive therapy)
Data collected will include prescriber (doctor/nurse); patient data (age, gender, pregnancy); clinical data (HIV diagnosis date, ART initiation date; co-morbidities); record of Gene Xpert™; and drug data (dolutegravir start date, NRTI backbone, dosing/dispensed number of dolutegravir, TB treatment, metformin and antiepileptic drugs, other co-prescribed drugs).
Data Analysis
Quantitative Data
At individual HCW level, descriptive analyses will be performed using Excel 365 (Microsoft Version 2309). Proportions and frequencies will be used to describe participants who get correct answers at each follow-up timepoint.
Inferential data analysis will be done using Stata Statistical Software: Release 15.1 (College Station, TX: StataCorp LLC) using expanded linear mixed-effects regression model adjusted for age, gender, years of experience, stratification, and clustering, with the primary analysis being of HCW total knowledge score. Mean differences will be estimated with their respective 95% confidence intervals.
For the patient level analyses, the same methods will be applied, while adjusting to account for the clustering effect. This inflation factor is often referred to as the ‘design effect’ (Campbell et al., 2000). T-test values will be divided by the square root of the design effect ((1 + (
Descriptive Analysis of WhatsApp Data
Text and emoji interaction within the training sessions – that was not direct answering of case-based learning point questions – will be extracted for descriptive quantitative analysis using NVivo (Version 14, released 2023).
Uptake, participation, and interaction will be described using proportions. For uptake, the proportion who participated will be calculated using the formula:
For participation in the training, the proportion will be calculated at each timepoint – during the lesson, one hour after, 24 hours after, at the end of the week, and two weeks after the training – using the formula:
For interaction within the training sessions, counts of participant’s interaction will be counted through line-by-line coding.
All emoticons will be counted in NVivo and used to create a descriptive word cloud on https://www.worditout.com/. Studies, both in non-clinical and clinical settings have shown that emojis are increasingly used to improve communication, add emotion, modulate intensity, contextualise and upgrade/downgrade text, and signal closing or opening of interactions (Halverson et al., 2023; Sampietro, 2019).
Descriptive Analysis of Questionnaire Data
Proportions will be reported for the survey questions assessing acceptability and feasibility (Table 6).
Qualitative Data
Qualitative data, drawn from the online questionnaires, WhatsApp group discussions and focus groups will be coded using NVivo (QSR International Pty Ltd. Version 12, 2018) and analysed thematically.
WhatsApp group discussions will be exported and saved in Excel spreadsheets with personal identifiers removed. Focus groups will be transcribed verbatim by BC. Themes will be drawn out using template analysis and reported descriptively using NVivo.
Template analysis is a ‘generic’ TA approach that allows the researcher to choose the methodology and philosophy that best suits the needs of the study (King & Brooks, 2017b). King and Brooke’s method of template analysis is described as a midpoint between coding reliability approaches (e.g., that commonly used in the grounded theory paradigm) and the reflexive approach, sharing the methods of theme creation of coding reliability but using the qualitative philosophy of the reflexive approach (Byrne, 2022).
The method has been used widely, across a diverse range of research areas (King & Brooks, 2017b), including psychology research (Brooks et al., 2015), sports science (Schneider et al., 2023), primary care guideline implementation (McKillop et al., 2012), and postgraduate medical education (Fokkema et al., 2013) and is suitable within the critical realist framework.
King and Brooks (2017b) situate template analysis within (limited) realist paradigms, where causal explanations are sought, by recommending three points to consider: 1. Template analysis can make use of a priori themes, drawing on existing theory or previous evaluation: in this case, data from the online questionnaires. 2. Reflexivity is vital to develop themes that are not just due to researcher subjectivity. 3. The reflexivity – and realist paradigm – does not rely on technical checks such as inter-rater reliability, but rather relies on researchers constantly reflecting on possible threats to interpretation quality and how to minimise them.
Template analysis has a seven-step process (Figure 4) but is flexible in that it is an iterative process, and emphasises the need to repeatedly go back to the data – quality checks – to support and organise constructed themes (King & Brooks, 2017a). It uses hierarchical coding with the creation of a flexible template that is added to, redefined, and deleted from, throughout the process (Brooks et al., 2015; Keidser et al., 2019). The seven steps of template analysis (King & Brooks, 2017a).
The use of a priori themes can be particularly useful in mixed methods studies like this one, where data from the quantitative portion – in this case, the post-intervention questionnaire – can be used to inform those themes (King & Brooks, 2017a). Template analysis allows researchers to take assumptions from existing data without them restricting the analytic process (Fokkema et al., 2013). The flexibility of the method fits within the pragmatic design and research questions of the study. NVivo will be used throughout the analysis process, to ensure a comprehensive audit trail.
Ensuring Qualitative Rigor
Qualitative analysis exists at the intersection of the researcher, data and, importantly, contextual interpretation (Braun & Clarke, 2022) and, as such, it has very different quality evaluation requirements to quantitative analysis, a subject of ongoing debate (King & Brooks, 2017a; Yadav, 2022) that is beyond the scope of this paper. By virtue of the diversity of paradigms in qualitative research, a single set of quality criteria is impossible and they need to be reflexive (Yadav, 2022).
The most important aspect of ensuring quality is to acknowledge the diversity of qualitative research and to be explicit and transparent about methods used, engaging with the theoretical and philosophical assumptions underlying the research procedures (Braun & Clarke, 2021; Kahlke, 2014; Liu, 2022), a concept often left out of research practice (King & Brooks, 2017c; Lipscomb, 2008; Prestwich et al., 2014).
Numerous checklists for ensuring qualitative rigor have been drawn up over the past 50 years, intersecting each other in some cases, and contradicting in others (Yadav, 2022). The development of a specific list of criteria for ensuring good quality qualitative research is controversial, but they are, invariably, helpful, especially to researchers new to the field.
Critical realist research needs to, instead of relying on standardised checks, be aware of the interpretation of quality and establish a plan to address it (Maxwell, 2011). King and Brooks (2017a) back Symon and Cassell’s recommendation that researcher’s should identify relevant criteria and explicitly justify them, which is echoed by Yadav (2022), who recommends that novice researchers use a combination of the available checklists guided by the particular study paradigm.
Strategies to Enrich Trustworthiness in the Context of the Study (Yadav, 2022).
aTriangulation: analysing a research question from more than one perspective (Amin et al., 2020).
Reporting Results
Reporting will be guided by two sets of guidelines. Firstly, the CONSORT 2010 updated guidelines for reporting parallel group randomised trials (Schulz et al., 2010), with its two relevant extensions: cluster randomised trials (Campbell et al., 2012) and randomised trials of nonpharmacologic treatment (Boutron et al., 2008). Secondly, the Standards for Reporting Qualitative Research (SRQR), a 21-point guide collated by a group with experience in health professions education will be consulted (O’Brien et al., 2014).
