Abstract
Keywords
Literacy interventions that use technology to personalize content and activities are an increasingly common evidence-based approach for improving student outcomes (Almirall, Kasari, McCaffrey, & Nahum-Shani, 2018; Connor, 2019). For example, using information on children’s prior knowledge and performance to adapt print and digital content is a popular intervention strategy for improving learning (U.S. Department of Education, 2016; Walkington & Bernacki, 2018). Adaptive treatment strategies, or adaptive interventions, are tailored to meet the individual needs of students and include decision rules to guide whether, when, and how best to offer treatment components over time (Almirall, Nahum-Shani, Sherwood, & Murphy, 2014). In particular, adaptive literacy interventions closely mirror typical practice conditions in which complex bundles of treatments are frequently modified to support individual learners. In typical practice conditions, educators are likely to confront two critical questions as they implement an adaptive intervention: Which intervention components should be implemented first? And what kinds of follow-up are needed to support children who do not adequately respond to the initial intervention components?
To date, researchers have employed the standard randomized controlled trial (RCT) design to test the efficacy of adaptive interventions relative to a control group (Hedges, 2018). As a result, it is difficult to identify what works for whom and how best to support individual learners who inevitably differ in how they respond to various intervention components. This article describes the development of a novel adaptive literacy intervention that aims to improve reading comprehension among elementary grade children from low-income households during the summer months (Alexander, Entwisle, & Olson, 2007). We illustrate how a sequential multiple assignment randomized trial (SMART) design can be used to develop an adaptive literacy intervention with personalized print texts and app-based activities for kindergarten to Grade 2 children.
Matching the Method to the Complexities of an Adaptive Literacy Intervention
The SMART design is ideally suited to help researchers and practitioners develop multicomponent interventions that adapt treatments based on initial or intermediate student outcomes. A central aim of the SMART design is to identify components that best meet the needs of individual learners. Thus, participants are randomly assigned to intervention components at multiple stages of the intervention; these assignments are determined by a sequence of decision rules, often based on proximal outcomes, that determine the modality, dosage, or delivery of a specific treatment at a given point in time (Almirall et al., 2014). Typically, participants who meet some benchmark on this proximal outcome are deemed “responders” to the treatment and continue receiving the same intervention with no adaptations. Participants who do not meet the benchmark established by the decision rule are deemed “nonresponders” and subsequently receive modification to their intervention (Nahum-Shani et al., 2012). It is common for these treatment modifications to be assigned at the time of nonresponder designation; however, they can also be preassigned as a potential modification to be implemented in the case that the participant is identified as a nonresponder.
The SMART design enables researchers and practitioners to explore critical questions regarding the sequence and dosage of intervention activities in real-world contexts like schools and homes. The resulting combination of initial interventions and later modifications or supports are referred to as “embedded interventions,” and head-to-head comparisons based on the multistage randomization can yield information about the most promising version to proceed with in future research studies (Nahum-Shani et al., 2012). Figure 1 illustrates how we used a SMART design with preassigned potential modifications to develop the components of our adaptive literacy intervention and to determine the different pathways of activities students can receive. In Stage 1, students are randomly assigned to intervention conditions (conceptually coherent texts [CCT] + App or leveled text [LT] + App). In Stage 2, nonresponding students in the Stage 1 intervention conditions are then randomly assigned to modifications that “augment only” (potential for gamification) or “augment and intensify” supports (potential for gamification and text messages).

Design components of the SMART design.
Using a SMART Design to Develop Two Approaches to Context Personalization
Personalized learning in education and literacy refers to efforts by researchers and practitioners to leverage technology (e.g., formative assessments, educational apps, text messages) to adapt content and instruction to meet an individual learner’s needs (Burke & Psaty, 2007; Connor & Morrison, 2016; Doss, Fahle, Loeb, & York, 2018; Greene, 2018; Hirsh-Pasek et al., 2015; Sabatine, 2018). More recently, a style of personalized learning called
To operationalize the first approach—context personalization with an emphasis on
To operationalize the second approach—context personalization with an emphasis on
We were interested in understanding whether an adaptive intervention should provide children with CCTs on a single science topic or LTs on a range of topics and how best to leverage print and digital activities to personalize learning at home during the summer months. Given that the Text Gradient System is incorporated into many formative assessments that provide regular information to teachers about each student’s progress in literacy (Hassrick et al., 2017; Hoffman, 2017; York, Loeb, & Doss, 2017), comparing the CCT condition with the LT condition has both theoretical and practical implications.
Pairing Print Text and App-based Digital Activities to Enhance Context Personalization
Both print text and app-based digital activities may play a critical role in enhancing the benefits of context personalization. First, the medium matters and reading on paper may benefit learners’ recall and memory of expository context (Carr, 2011; Singer & Alexander, 2017a, 2017b). Recent meta-analytic evidence indicates that students perform better on comprehension tasks when they read expository texts on paper than on screens (Clinton, 2019). Furthermore, providing choice of print text can further support intrinsic motivation to read (Deci & Ryan, 2000; Guthrie, McRae, & Klauda, 2007; Patall, Cooper, & Wynn, 2010). Second, accumulating research indicates that “educational” apps should foster meaningful learning activities and social interactions (Hirsh-Pasek et al., 2015; Mayer, 2008; Vaala, Ly, & Levine, 2015). Meaningful learning takes place when children learn with a purpose, learn material that is interesting and personally relevant, and connect new learning to prior knowledge. In other words, meaningful learning enables children to “hook” newly learned content from text into existing information and to grow their domain knowledge (Ausubel, 1968; P. C. Brown, Roediger, & McDaniel, 2014; Fitzgerald, Elmore, Kung, & Stenner, 2017). In addition to meaningful learning, apps should promote social interactions between children and parents (and other family members) to foster language development (Kremar, Grela, & Lin, 2007; Roseberry, Hirsh-Pasek, Parish-Morris, & Golinkoff, 2009; Sawyer, 2006).
In this study, children in both the CCT and LT conditions chose print versions of 10 books and personalized app-based digital activities to foster summer reading activities at home. Based on recommendations from evidence reviews for improving K–2 reading (Foorman et al., 2016; Shanahan et al., 2010), we developed app-based personalized activities that included a variety of text-dependent questions (i.e., questions about the books) and discussion questions (e.g., questions to stimulate parent–child conversations about the book). Questions were both literal (answers could be found directly in the text) and inferential (required children to use context and prior knowledge to answer questions, for example, what is the main idea?).
To further enhance the potential benefits of pairing print texts with app-based digital content, it is critical to develop strategies for helping families access the resources that foster formal (e.g., word reading activities) and informal (e.g., shared book reading) home literacy activities (Sénéchal, 2006). Thus, the CCT and LT conditions included an afterschool family literacy event that focused on helping parents learn how to access the app-based digital activities.
Developing Two Strategies for Addressing Differential Responsiveness
A novel feature of our intervention was the development of a literacy app called MORE@Home that included personalized digital activities. Given prior research indicating that many students would likely not complete the app-based personalized literacy activities paired with each of the 10 print texts (Augustine et al., 2016; Kim et al., 2016), we developed a set of Stage 2 intervention activities that included follow-up protocols to motivate nonresponders. SMART designs enable researchers to compare the effectiveness of different strategies designed to address differential responsiveness to the first stage of an intervention. In this study, because the MORE@Home app is installed on parents’ cell phones, parents must be motivated to download and use the app with their children. Self-determination theory posits that motivation derives from both intrinsic and extrinsic sources (Deci & Ryan, 2000). The version of the MORE@Home app initially provided to students relied primarily on their (or their parents) intrinsic motivation. Subsequently, two strategies were used to motivate nonresponders: (1) white hat gamification features, which was directed primarily at students, and (2) text messaging, which was directed primarily at their parents. In Figure 1, we refer to white gamification as supports that “augment only” and text messaging as supports that “augment and intensify” follow-up with nonresponders.
All students identified as nonresponders based on their app usage at either 2 or 4 weeks following implementation of the Stage 1 components received white hat gamification of the MORE@Home app. Gamification refers to the goal of embedding elements of a game within an activity, and “white hat” gamification, is characterized by positive motivators, such as giving the player a purpose or rewarding the player for demonstrating creativity or skill mastery. This approach contrasts with “black hat” gamification, which is characterized by motivators of avoidance, such as instilling uncertainty or fear of loss in the player (Chou, 2014). When the white hat gamification features were turned on for nonresponders in this study, the MORE@Home framework changed. Instead of seeing a list of activities to complete for each book, reading activities were shown as stepping stones on a path toward creating your own virtual zoo. Each completed set of activities meant a student could choose a new animal for their zoo. These features emphasize meaning, development, empowerment, and ownership—key values of white hat gamification (Chou, 2014)—and they rely on external factors to motivate students to use the app.
Gamification features are designed to motivate the player (in this case, the student) to move through the game (in this case, complete literacy activities in the app). However, even if students are interested in the books and activities, parents must provide them with opportunities to engage, including access to a smartphone with the MORE@Home app installed. Furthermore, some activities encouraged parents to talk with their child about the books. We anticipated that even parents who care about education might fail to download and use the app. For example, parents may have less information than teachers to personalize activities for their children, have logistical difficulties downloading the app, or simply forget to use the app. Thus, in addition to white hat gamification, we employed a second strategy for intensify supports for nonresponders: text messaging. Text messages were sent to parents of nonresponding students twice per week over six weeks, starting a month into summer vacation.
Text messages have become a popular approach to addressing differential responsiveness in part because they are both scalable and cost-effective. Recent studies have demonstrated promising effects of text messages to parents on student outcomes and specifically texts about their children that include guidance for supporting reading habits (Doss et al., 2018; Kraft & Monti-Nussbaum, 2017; York et al., 2017). The text messages used in this study were designed to accomplish three goals. First, to remove barriers to parental access, text messages reminded parents to use the MORE@Home app and included direct links to the iOS/Android apps (Thaler & Sunstein, 2008). Second, to highlight interesting features of the MORE@Home app, text messages mentioned the personalized activities, the gamification features, and the availability of parental discussion questions. Third, to encourage parents to read with their children, text messages shared tips on general reading behaviors, such as leveraging weekends as opportunities to read together (York et al., 2017). Finally, to correct potential misconceptions, text messages presented aggregate usage statistics, such as the number of downloads or the number of books accessed, as way of showing parents that other parents were finding the app to be a useful and fun tool. This final text message type draws on the behavioral theory that correcting parents’ beliefs can lead to desirable behaviors (e.g., Robinson, Lee, Dearing, & Rogers, 2018).
Research Goals and Questions
To improve reading comprehension outcomes for children in kindergarten to Grade 2 during the summer months (Alexander et al., 2007; Kim et al., 2016; Zvoch & Stevens, 2015), we used a SMART design to develop an adaptive literacy intervention with two stages of intervention activities. We asked two research questions2:
In Stage 1 of an adaptive literacy intervention, is it better for K–2 students’ reading comprehension outcomes to receive pretraining in-school lessons, personalized print texts, and app-based digital activities based on (a) 10 conceptually coherent texts (CCT) or (b) 10 leveled texts (LT) on a range of topics?
Is it better for students’ reading comprehension outcomes (a) to augment the Stage 1 treatment with white-hat gamification of the app alone (Gamification Only) or (b) to augment the Stage 1 treatment with gamification of the app and to intensify supports with text message reminders to parents (Gamfication + Texting) if students did not respond to the initial intervention?
Method
Participants
In the spring of 2018, we recruited 16 kindergarten, first-, and second-grade teachers from a southeastern Title-I elementary school to participate in this study. All students in those classrooms were eligible to participate (
Demographic Characteristics of Participants
Design and Procedures
The primary aim of the SMART design is to develop an adaptive intervention using the CCT or LT approach to context personalization. Thus, the SMART design in Figure 1 is characterized by sequential random assignment, with at least one round of this assignment being determined based on a decision rule (i.e., the intermediate tailoring variable indicating whether students are responding to the Stage 1 components). Figure 2 provides a detailed overview of the Stage 1 components, the intermediate tailoring variable used to identify responders and nonresponders, and the Stage 2 components.

Overview of the MORE Stage 1 and Stage 2 intervention components.
Stage 1: CCT Core Components
The CCT core components were designed to instantiate our first approach to context personalization which provides contextual grounding for reading activities at school and home. As shown in the left column of Figure 2, the CCT condition included the three components. In the first component, the CCT pretraining lessons implemented during the last 2 weeks of school, children participated in content literacy instruction and learned about the topic of Arctic animal survival. Children learned to use concept mapping to visualize semantically related words that appeared across conceptually coherent informational texts, and participated in integrated reading, writing, and listening activities to extend their thinking (Connor et al., 2017; Graham et al., 2017; Guthrie et al., 2004; Vitale & Romance, 2012). Concept mapping was used to illustrate the relationship among taught words, including
In the second component, personalized print texts, children chose 10 print books (i.e., paper books) that cohered around the topic of animals and animal survival; students selected from 34 possible titles (see Appendix A in the online supplemental material for a list of books available in both conditions). Based on the Lexile approach to measuring text difficulty and student reading ability designed by MetaMetrics®, the informational books were quite challenging (most books ranged in Lexile level from 375 to 625) for the majority of students given the age and reading levels of the sample (student Lexile levels typically ranged from 25 to 475).
In the third component, app-based digital activities, children and their parents could access the researcher-developed MORE@Home app, which included personalized reading activities. To personalize activities, we used middle of year data from the Measure of Academic Progress (MAP), Primary Grade Reading, which yields Rasch Unit (RIT) scaled scores for children’s performance on four subtests (foundational literacy, comprehension of text, vocabulary, language and writing). Using the four subtest scores on the MAP, we created activities at three levels of difficulty (low, medium, high), and gave children an appropriately leveled activity that included a range of word-level, sentence, and text-level activities, and a discussion question to stimulate conversations with family members at home about the book (see Appendix B in the online supplemental material for example app activities).
Stage 1: LT Core Components
The right column of Figure 2 displays the LT components. Although the LT condition was composed of the same parallel components as the CCT condition, we developed the LT components using typical instructional practices in wide use across many U.S. urban districts. For example, in the context where we conducted this study, the district enacted a balanced literacy curriculum that provides each classroom teacher with LT based on a holistic measure of Text Gradient Level (Fountas & Pinnell, 2010). Teachers are required to teach a daily literacy block involving word study activities (e.g., phonics, spelling), independent reading, small group guided reading and writing, and an emphasis on Tier 2 academic vocabulary words. Finally, the school district was located in a state where all teachers are required to use a formative assessment screener (MCLASS 3D and TRC) at the beginning, middle, and end-of school year.
The primary technology used in component two of the LT condition—that is, the use of LT and formative assessment data based on a holistic measure text difficulty—is widely used in many U.S. school district contexts and thus, represents a face valid typical practice condition (Hassrick et al., 2017; Hoffman, 2017). For example, the Text Gradient Levels measure the level of difficulty for the narrative texts (i.e., fiction or nonfiction text typically written in narrative form) and informational texts (i.e., nonfiction text typically written in expository form) used in the LT condition. The levels range from A (beginning readers) to Z (advanced readers), and the gradient measure is holistic and based on 10 text characteristics (i.e., genre, text structure, content, themes and ideas, language and literacy features, sentence complexity, vocabulary, word difficulty, graphics, and print features). In essence, the materials and technology that already exist in many U.S. school districts afforded us an opportunity to develop a second approach to context personalization.
As shown in the right column of Figure 2, the LT condition included pretraining school lessons, personalized print texts, and app-based digital activities to stimulate students’ interest in reading LT on a range of topics during the summer. In the first component, the LT pretraining in-school lessons, teachers instructed children how to use a word map to build their understanding of target vocabulary (i.e., synonyms, antonyms, examples, and nonexamples of the target word). Word maps are used in the balanced literacy block to teach general purpose academic words that are useful across content areas. In addition to the in-school reading lessons, LT parents were also invited to an end-of-school family literacy event that provided information on the app to parents and adult caregiver. Overall, 40% of LT parents (54 of 134) attended the family literacy event.
In the second component, children chose personalized print texts including five informational texts and five narrative texts. We used the child’s TRC level to find LT on a range of topics that had text gradient levels (D to R), which corresponded to the end of kindergarten to the end of grade 2 (see Appendix A in the online supplemental material). The third component involved app-based digital activities that were tied to each LT and focused on activities included in the TRC; the discussion question was not personalized to the book but included general active reading strategies. In sum, the LT components were designed to model typical instructional practices, to help children learn a general vocabulary strategy for independent reading of personalized and leveled print texts, and to include activities focused on text dependent comprehension and vocabulary questions.
Stage 2: Intermediate Tailoring Variable
In June, we used app usage data to determine whether students were responding or not responding to the Stage 1 intervention components. At this point, the decision rule was implemented to identify the “responder” students and the “nonresponder” students. Using a measure of print exposure as a proxy for later reading achievement (Kim et al., 2016; Mol & Bus, 2011), we defined “responders” as students who had completed all the MORE@Home app activities for at least one book. These students continued using the application with no additional supports. Students who had not yet completed all the app activities for at least one book were deemed “nonresponders” and subsequently received additional supports.
Supporting Nonresponders With Gamification Only or Gamification With Text Message Reminders
A second stage of random assignment defined the type of supports we provided these nonresponders. We blocked students within CCT and LT conditions, noted the date they assessed for nonresponder status, and then randomly assigned to receive gamification of the app only or gamification of the app plus text messaging.
Simplified SMART Design
Initially, this study leveraged three different stages of random assignment, with the middle random assignment assigning different dates for the assessment of nonresponder status (either the middle or the end of June). These two dates, 2 weeks and 4 weeks after the family literacy event marking the end of pretraining, were chosen to represent and “early” and “typical” follow-up period. The typical follow-up period was chosen to represent a period of time over which parents would have had time to settle into their summer routines. The “early” period, 2 weeks after the end of pretraining and one week after the end of the school year, provided an earlier indicator of initial summer vacation behaviors. All randomizations occurred at the beginning of the study and the design and analysis plan were preregistered prior to all program implementation activities (Kim, 2018). However, we decided to simplify our analysis and collapse across the second wave of random assignment for two key reasons. First, while we maintained fidelity of identifying participants as nonresponders at the appropriate time point, there was little difference in the treatment experience for nonresponders in the two conditions. Text messaging only began in July after both groups had been evaluated on the tailoring variable, and while technically, gamification of the app was implemented immediately on identification as a nonresponder, almost all nonresponders had failed to access the app at all. Thus, they would not have learned of the gamification until the first text message was sent. The lack of treatment contrast between the two groups is the primary reason for collapsing across them, but this decision was justified as there was also no significant difference in the tailoring variable between the two groups. This justified the simplification of our analyses to the research design, which is presented in Figure 1.
Stratified Sample of Surveyed Parents
To provide context for our primary research questions and to further explore how families engaged with reading and the MORE@Home app over the summer, we selected a sample of parents for a phone survey in Fall 2018. We randomly identified 2 app users and 2 non–app users (as well as backups) from each of the four embedded interventions, resulting in a total of 16 parents. Both English- and Spanish-speaking parents were called and asked a series of questions about their use of the MORE@Home app over the summer and other ways they engaged with their child about books. We were able to conduct the survey for 12 of the 16 targeted parents. The overall response rate was 75%, and the response rate was higher among app users (88%) than among non–app users (63%).
Measures
Data collection took place immediately after the in-school lessons in Grades K–2, throughout the summer, and following program summer implementation in the fall of Grades 1 to 3. First, we administered three researcher-developed measures immediately after the spring lessons to assess impact on spring science vocabulary, listening comprehension, and word reading skills in Grades 1 and 2. Copies of these can be found in Appendix C in the online supplemental material. Second, we collected usage statistics from the MORE@Home app over the course of the summer, including the books accessed and the activities completed. Finally, we used administrative data to assess impact on two standardized measures of reading in the fall of Grades 1, 2, and 3 and administered a brief parent survey over the phone.
Measure of Science Vocabulary Knowledge
This assessment is a researcher-developed semantic association task designed to measure vocabulary depth and knowledge of taught science concepts. Students were provided with a set of 10 target words, and for each, were asked to identify which of the four-word option was semantically linked to the target word. Each of the 40 word-options was scored as either incorrect (0) or correct (1). In this sample, the Science Vocabulary Knowledge test demonstrated good reliability, with a Cronbach’s alpha score of .94.
Listening Comprehension
This researcher-developed measure is designed to assess whether the student can extract meaning from text that was not taught in any classroom. It includes a short passage about rainforests that was adapted from the Magic Tree House Fact Tracker series (Will Osbourne and Mary Pope Osbourne) and is accompanied by four multiple-choice questions. Students had a copy of the passage in front of them, and the teacher read it aloud. Student scores represent the percent correct among these inferential questions. The assessment had a Cronbach’s alpha of .47 in this sample.
Word Study
This measure is a researcher-developed task that assessed students’ decoding skills that were addressed in the word study component of the MORE intervention. Student scores represent the percent correct among the six multiple-choice questions that asked students to identify similar phonetic components in the word options as in the target word. The assessment had Cronbach’s alpha of .54 in this sample.
Measure of Academic Progress
We used the MAP as the primary pretest and posttest to evaluate impacts on student reading comprehension overall and informational text comprehension in particular. The MAP is a computer-based, adaptive, and vertically scaled test that is administered three times per year and measures a variety of reading skills. It consists of an RIT score (the primary measure) as well as subscores that distinguish between specific types of reading skills. In early grades, these subscores measure both foundation and code-focused skills as well as meaning-focused skills. Once students reach second grade, the measure assesses meaning-focused skills but continues to distinguish between narrative and informational text comprehension ability. The MAP has a reported test–retest reliability ranging from .89 to .96 (R. S. Brown & Coughlin, 2007). For our Grades 1 and 2 baseline sample (also the rising Grades 2 and 3 sample), we report results for the informational text comprehension outcomes. We used RIT scores from the Spring 2018 and Fall 2018 as our primary student outcome.
MCLASS Reading 3D: Dynamic Indicators of Basic Early Literacy Skills (DIBELS) and Text Reading Comprehension
The MCLASS DIBELS (Dynamic Indicators of Basic Early Literacy Skills) battery is a series of teacher-administered assessments that provides systematic information on early grade literacy indicators. From kindergarten to Grade 3, DIBELS subtests include initial sound fluency, phoneme segmentation fluency, letter naming fluency, nonsense word fluency, oral reading fluency, and retell abilities (Good & Kaminski, 2002). We used the DIBELS total composite score since reliability of subtests range from .88 to .98 across tested grades. In addition to the DIBELS composite score, we used the TRC score, which is an individually administered assessment in which teachers use a running record to measure children’s reading accuracy, fluency, comprehension, and oral and written comprehension. Validity evidence indicates that the TRC is correlated (
Analysis
Main Effects
SMART designs use common analytic approaches for traditional RCTs, as random assignment occurs at each phase of the design (Almirall et al., 2014). Our preferred specification is an intent-to-treat (ITT) analysis that leverages the preimplementation random assignments to each potential condition. One reason for preferring an ITT analysis is that not all parents provided a valid phone number. These students would still be able to receive gamification, but were unable to receive any text messages, rendering them as noncompliers in a typical Rubin Causal Model framework (Rubin, 1974). There is a second type of endogenous noncomplier that we include in the analysis by using the ITT approach—these are the responders, those who used the app from the beginning and according to the study design were thus not eligible to receive additional supports. Both these noncompliers were randomly assigned to their respective conditions (including potential additional supports), but due to their own endogenous choices, never received those supports. Using the ITT analytic approach provides a conservative estimate of the effects of additional supports by including individuals who never actually received the intended services.
Figure 1 provides a roadmap for our analyses. For our first research question, our goal was to examine the main effects of the Stage 1 treatments comparing the CCT and LT conditions. Thus, we compared outcomes for all students who received CCT (Groups A + B + C + D) to outcomes for all students who received LT (Groups E + F + G + H). For our second research question, our goal was to examine the main effects of the Stage 2 treatments. Here, we compared outcomes for all students who had the potential to receive gamification and text messaging (Groups C + D + G + H) to students who only had the potential to receive gamification (Groups A + B + E + F). Modeling these treatment effects follows the specifications outlined in the study preregistration: multilevel models nesting students within classrooms.
In all models, Level 1 includes student-level covariate adjustments including student pretest reading scores and demographic characteristics such as gender, race/ethnicity, English language learner status, and special education status and an indicator for the random assignment block. We used the spring assessment of the outcome for the MAP and DIBELS outcomes and include the MAP spring assessment as a pretest control in all other regressions. Level 2 includes a teacher-level random intercept to account for the distribution of student achievement around classroom averages. Treatment indicators are included at the appropriate level based on the specific research question (i.e., the classroom-level treatment is included at Level 2 and the individual treatment assignments are included at Level 1).
Results
Use of the MORE@Home App
Prior to looking at the modeled effects, we investigated app take-up and usage over the summer. Table 2 displays wide variability in parents use of the MORE app. Overall, the MORE@Home app was downloaded 134 times during the study period, but because the app was openly accessible on iOS and Android app stores, this meant that nonparticipants may account for some of this volume. Among participants, 71 unique students (26% of the sample) accessed the app. Access rates were similar in kindergarten (32%) and first grade (29%), and slightly lower for second graders (17%). They were also slightly higher among girls (29%) than boys (22%), but this difference was not statistically significant. On average, students who did access the app read 4.4 books and completed 47 activities (see Table 2). A small number of students completed the same activities multiple times on different devices, so we also measured the number of unique activities that students completed, which was slightly lower at 45 activities per student. We also found that students tended to access the app in concentrated chunks of time; the average app user accessed the app over a period of 3.2 days.
Summary App Usage Statistics
Unique activities removes any activities that were completed more than one time by a student.
Analyses comparing demographic characteristics of users and nonusers can shed light on potential sources of selection bias. The results in Table 3 suggest that there were very few statistically significant differences between app users and nonusers on measured baseline demographic and reading skill measures. Grade 1 app users scored significantly higher on spring MAP and TRC, suggesting that Grade 1 children who used the app were better readers than non–app users.
Demographic Characteristics of App Users and Nonusers
Spring TRC scores converted to numbers ranging from −1 (PR) to 26 (Z). Average kindergarten level is between C and D for kindergarten, between I and J for first grade, and between J and K for second grade.
To address the first research question, we report experimental comparisons in Table 4. The results in the last column display the effect size (ES), the covariate-adjusted standardized mean differences, for three sets of outcome measures. First, we find that the CCT intervention had a slight positive but statistically insignificant effect on app usage statistics relative to LT. Students in this group were 6 percentage points more likely to have accessed the app and completed about 2 more activities. Second, students did score slightly higher on Science Vocabulary Knowledge (ES = 0.10,
Effects of CCT Versus LT
The proximal student outcomes and MAP informational text subscores were only assessed as outcomes for the first and second grade sample.
TRC scores converted to numbers ranging from −1 (PR) to 26 (Z).
Results for Research Question 2 compare the effects of the Gamification + Texting and Gamification Only conditions and are presented in Table 5. Students assigned to the Gamification + Texting support condition were not significantly more likely to use the MORE@Home app by any of the outcomes presented. However, in our sample, these students were 5 percentage points more likely to use the app and completed 3 more activities than students who received just gamification. In the fall, Gamification + Texting students scored marginally significantly higher on the MAP (ES = 0.10,
Effects of Gamification and Texting (G + T) Versus Gamification Alone (G)
The MAP informational text subscores were only assessed as outcomes for the first and second grade sample. bTRC scores converted to numbers ranging from −1 (PR) to 26 (Z).
Using the Parent Surveys to Understand Results
In Table 6, we present description results from a small parent survey. The items were designed to assess whether parents and their children use the personalized print texts and which app features were used. Importantly, the results indicate that parents read an average of 7.7 MORE books to their children and reported their children independently read 6.6 MORE books. Over 50% of parents reported that they “very often” engaged in language activities like reading to their children and talking to their children, and 86% of parents used the “guided questions” in the app, questions designed to foster social interaction and discussion. Across all conditions and combining students who did and did not use the app, parents reported that they read on average 7 of the 10 books that students were provided. While this survey is limited by sample size, the important story here is that parents and children accessed the print texts and the guided questions in the app were the most frequently accessed parent activity.
Parent Survey Results (n = 12)
Discussion
Using a SMART design involving 273 kindergarten to Grade 2 children, we developed an adaptive literacy intervention with multiple components that was designed to improve students’ reading comprehension outcomes. The adaptive intervention components included pretraining in-school lessons, personalized print texts, and app-based digital reading activities, and follow-up procedures to support students’ literacy development. In sum, the results indicate that (a) children who received CCTs performed similarly to children who received LT on reading comprehension outcomes and (b) nonresponders who received both gamification in their app and parent text messages performed better than those nonresponders who received only gamification. Implementation data revealed that one fourth of participants (
Our first research question was to examine whether an adaptive intervention with personalized print texts and app-based activities should provide kindergarten to Grade 2 children with CCTs or LTs. In Stage 1 of our SMART design, our aim was to develop and then compare two approaches to context personalization. Although children in the CCT and LT conditions performed equally well on posttest reading outcomes, there are stronger theoretical and practical reasons for testing the CCT approach to context personalization in a confirmatory RCT. From a theoretical perspective, there is growing evidence that opportunities to read CCTs are critical to building domain knowledge and fostering reading comprehension (Cromley & Azevedo, 2007; Hirsch, 2016; Kintsch, 2009). Our results provide suggestive evidence that children in the CCT condition learned more science vocabulary than children in the LT condition. Some scholars have also suggested that “just as it’s impossible to build muscle without weight or resistance, it’s impossible to build robust reading skills without reading challenging text” (Shanahan, Fisher, & Frey, 2012, p. 58). In other words, the
From a practical perspective, increased exposure to informational texts may be especially vital to supporting informational text comprehension for low-income children. That is, using CCTs rather than LTs in an adaptive literacy intervention may increase low-income children’s access to informational texts at school and home (Duke, 2004; Kim et al., 2016). Hoffman (2017) has argued that the widely used classroom practice of using “just right” LTs in elementary literacy instruction may have unintended negative consequences for low-income and struggling readers by limiting access to informational texts that young children want to read. Future experimental research should examine whether the CCT approach to context personalization developed for this study can enhance children’s reading comprehension outcomes relative to an untreated control group.
Our second research aim was to compare two strategies for nudging nonresponders to participate in the intervention activities. In a standard RCT, the nonresponders—that is, children whose parents did not download and use the app—would not receive any additional follow-up supports. In many out-of-school literacy interventions, particularly those that target the summer months, there is substantial evidence that individual student heterogeneity in response can attenuate the efficacy of the treatment (Augustine et al., 2016; Cooper, Charlton, Valentine, & Muhlenbruck, 2000). To deal with treatment heterogeneity in how students responded to the Stage 1 components, we included Stage 2 procedure to augment or augment and intensify follow-up procedures for nonresponders. Specifically, we developed text messages to nudge parents to complete the app-based activities with their children, as well as a gamification plan in which children could collect animals for a virtual zoo on completion of the app activities. We found that being assigned to receive parental text messages as well as gamification as supports for nonresponsiveness resulted in better reading outcomes for students than being assigned to only gamification. However, there were no significant differences in app usage between these groups. These results suggest that not all reading activity related to MORE@Home was captured by the app statistic. In addition, the results indicate that the text-message supports encouraged parents to engage in other, non–app-based reading activities with their children, underscoring the importance of pairing print and digital content as a strategy to support at-home reading activities.
Finally, how do we explain the positive effects of gamification and texting compared with gamification alone on the reading comprehension outcomes of nonresponders? In our SMART design, children in the CCTs chose 10 informational print books and LT children chose 5 informational print books (and 5 narrative print books). Thus, children in both the CCT and LT conditions, including both responders and nonresponders, enjoyed more opportunities to interact with informational texts in print and to complete app-based digital activities. Our results are consistent with intervention studies that have experimentally increased the number of informational texts in children’s classrooms and demonstrated positive impact on text comprehension outcomes (Duke, 2004). It may be that the text message reminders nudged nonresponders to interact more with informational than narrative texts in ways that supported informational text comprehension outcomes.
Study Limitations and Future Directions
This study underscores several implementation and analytic challenges that are likely to confront researchers seeking to develop and test app-based literacy tools for parents and families to use in out-of-school contexts. For example, only 26% of students accessed the MORE@Home app, a key element of the intervention. Not only is this concerning from an implementation perspective but it also limits the effectiveness of the SMART design, because the decision rule and tailoring variable relied on the MORE@Home app. It would be particularly concerning if app usage was associated with baseline reading achievement; however, there was no significant difference between app users and non–app users on their baseline MAP scores. In this study, we implemented a family literacy event prior to summer vacation to demonstrate the app, and while it was highly attended, individual students could not access the app that evening, and many attendees did not attempt to use the app after that evening. Future research could work to tie the family engagement night to app access or to integrate app usage into the pretraining school lesson component of the CCT intervention. Both of these approaches would ensure that students and their families are familiar with the technology before summer vacation begins. Additionally, from a measurement perspective, it is important to note that researcher-developed measures of listening comprehension and word study had low reliability (α = .47 and α = .54, respectively), which may attenuate the estimated impacts. Future work would benefit from refining these measures to more reliably capture specific dimensions of reading skill. Finally, although we found few observed differences between app users and nonusers, it is critical to explore for whom and under what conditions education apps are most effective in fostering meaningful learning and student engagement (Hirsh-Pasek et al., 2015).
Conclusion
Matching the method to the complexities of developing adaptive interventions is a critical need in education and literacy. To our knowledge, this study represents one of the first efforts to deploy a SMART design to develop an adaptive K–2 literacy intervention with personalized print texts and app-based digital activities. Had we conducted a standard RCT including all the Stage 1 and Stage 2 components in a fixed treatment, we would have limited knowledge of how best to begin an adaptive intervention or how best to help nonresponders participate more fully in the intervention. A SMART design is ideally suited to develop adaptive interventions and to identify potential causal levers of change. Tightly coupling SMART designs that guide the development of adaptive interventions with RCT designs that subsequently evaluate intervention effectiveness should help build usable knowledge to improve young children’s reading comprehension outcomes.
Supplemental Material
DS_10.1177_2332858419872701 – Supplemental material for Using a Sequential Multiple Assignment Randomized Trial (SMART) to Develop an Adaptive K–2 Literacy Intervention With Personalized Print Texts and App-Based Digital Activities
Supplemental material, DS_10.1177_2332858419872701 for Using a Sequential Multiple Assignment Randomized Trial (SMART) to Develop an Adaptive K–2 Literacy Intervention With Personalized Print Texts and App-Based Digital Activities by James S. Kim, Catherine A. Asher, Mary Burkhauser, Laura Mesite and Diana Leyva in AERA Open
Footnotes
Authors
References
Supplementary Material
Please find the following supplemental material available below.
For Open Access articles published under a Creative Commons License, all supplemental material carries the same license as the article it is associated with.
For non-Open Access articles published, all supplemental material carries a non-exclusive license, and permission requests for re-use of supplemental material or any part of supplemental material shall be sent directly to the copyright owner as specified in the copyright notice associated with the article.
