Abstract
Introduction
Vignettes are hypothetical scenarios meant to prompt respondents and elicit reactions without exposing respondents to actual scenarios. Vignette studies have been used widely in the fields of political science, human computer interaction, and social science, as a branch of survey methodology. Hypothetical scenarios can be particularly useful when used as a means to probe perspectives regarding emerging hypotheses or to test hypotheses regarding the reactions of individuals to specific scenarios.
In this paper, we motivate and present the methodology of vignette studies. The primary contribution of this paper is the introduction of a novel vignette study method, SMART vignettes, an adaptation of the sequential multiple assignment randomization trial (SMART) design developed by Murphy (2004), applied to online survey designs. While this method is not a clinical trial, it borrows the critical element of sequential randomization from the methodology. We use a primary motivating example of clinical decision making in AI to demonstrate the utility of the method in allowing inferences on physician reactions toward AI-embedded clinical decision making systems. We argue that our resultant novel method is well poised to probe decision making contexts occurring over discrete time points. Our proposed vignette approach leverages two new design characteristics, namely sequential randomization and adaptive allocation. While conventional vignette studies have been widely deployed in various fields, our newly proposed methodology offers several benefits not yet afforded by vignette studies, by virtue of the novel design features. These features offer three new advantages which are distinct from traditional vignettes: (1) valid causal inferences on the conditional distributions of the primary outcome of interest, given previous outcomes and other contextual factors, (2) balanced allocations across groups, and (3) a greater degree of interactivity for the survey respondent.
Proposed method
Background
Vignettes are narrative stimuli describing hypothetical situations and are intended to elicit reactions from survey respondents. A vignette study may feature a number of key factors of importance, each of which may vary by a fixed number of levels. In a full factorial design, a vignette population is considered to be the complete set of vignettes consisting of all possible combinations of levels for each factor. A single vignette is thus a specific combination of levels from each factor. As the size of the vignette population grows exponentially with each factor, randomization can be used to assign a respondent a single vignette rather than the whole vignette population (Atzmüller and Steiner, 2010). For example, a survey respondent can be randomly assigned to either one of two levels, agreement or no agreement (Table 1).
Example vignette for testing whether level of information impacts response.
In conventional vignette designs, vignette assignment occurs at baseline prior to survey distribution. The vignette is presented in its entirety to the respondent, followed by the question set. With such a design, inferences can be made about the joint distribution of responses. For example, Figure 1 shows a traditional vignette study with two factors

Setup of a traditional vignette study. Randomization is performed by sampling from the vignette population with equal probability at baseline. The path outlined in bold represents one possible vignette that is assigned to the participant. The vignette is presented to the survey respondent followed by a set of questions.
An online SMART vignette survey
We propose an vignette survey study design that is meant to be delivered online and in a dynamic fashion.
We propose a novel vignette method with two features. The novelty is twofold:

Setup of study with sequential randomization. Randomization occurs at multiple time points throughout the study. Respondents are randomly assigned to the first subvignette that characterizes one factor of the vignette then are prompted with a set of questions; the process repeats for other factors. The full vignette is delivered in a sequential fashion, that is as a sequence of subvignettes. The path outlined in bold shows one possible vignette comprising first (solid red line) and second (dotted magenta line) subvignettes.
Sequential randomization
In the clinical trial setting, the sequential multiple assignment randomized trial (SMART) design randomly assigns a patient to an initial line of treatment, and then re-randomizes the patient at a later stage based on the patients intermediate responses to the treatment (Thall et al., 2000). When applied to surveys, SMART vignettes can be viewed as adaptive surveys in the framework of sequential randomization (Montgomery and Cutler, 2013). Instead of presenting a full vignette with a pre-determined combination of characteristics at once, we present a portion of the vignette reflecting a particular characteristic of the vignette at a time. Hence multiple randomization time points are needed, one for each characteristic of the vignette (i.e. subvignette). At each time point, the respondent is randomized to a single subvignette, which is followed by a set of questions. After the response, the participant gets re-randomized to the next subvignette.
Sequential randomization is especially useful when the order of outcomes carries significance, or when subsequent outcomes depend upon the previous outcome. The repeated randomization allows valid causal inferences concerning the relative effectiveness of the intervention (or narrative stimuli) options (Rubin, 1974).
Adaptive allocation
While randomization enables causal inference, it does not guarantee balance across subgroups defined by the vignette population. Complete randomization often results in imbalance. To address this issue of imbalanced groups, Efron (1971) introduced the biased coin design as an adaptive allocation strategy. The concept of this allocation strategy is that it assigns an incoming respondent with a fixed probability
The overall role of an adaptive randomization strategy in this setting is to encourage sufficient cell counts for all vignette and response combinations. Biased coin methods promote balance by encouraging allocations to the under-represented group. Various extensions of the biased coin design have been since proposed and can be used, such as the urn design (Wei, 1978) and the generalized biased coin design proposed by Smith (1984) as a large class of designs covering the biased coin design as a special case (Atkinson, 2014).
Advantages
SMART vignettes allow the researcher to answer questions that are distinct from those that can be answered via traditional vignettes. Our proposed approach is suitable for a research problem for which it is helpful to infer the conditional distribution of the primary outcome, given all other contextual variables. Traditional vignettes with randomization allow for valid causal inference on the joint distribution of multiple outcomes conditioned on the factors (i.e.
Moreover, conventional designs can burden the participant by administering a full set of vignettes. In studies with many factors, only a sample of the set is presented to reduce response burden, but doing so comes at the cost of imbalance. SMART vignettes administer a sample of the full set without sacrificing balance between groups. In addition, while interactivity is not unique to the SMART design, web-based surveys can allow for interactivity between the survey and the participant.
In the next section, we describe the setting of a health care scenario as a motivating example in which the conditional expectation (distribution) of the primary outcome (i.e. decision making) is of interest, instead of the joint distribution of outcomes. This type of outcome may be more salient in settings featuring a decision making process in which a provider or patient receives information across different time points.
Case example: Application to AI and healthcare
We present a case example of an online survey for physicians, which implements the SMART vignette design to probe physicians’ reactions in decision making scenarios featuring an AI-embedded clinical system. This example serves to demonstrate the utility of the proposed method in allowing salient inferences of stakeholders reactions toward AI-embedded clinical decision making systems. In this example, a SMART vignette design was used to randomize hypothetical scenarios to gain a better understanding of the causal impact of physician attitudes, given emerging evidence that a range of factors, including physician understanding, transparency of AI, and trust in AI, among others, may play a role in influencing clinical decisions.
Figure 3 depicts the flow of an online survey by Kim et al. (2024) utilizing the SMART vignette design. This survey for physicians is currently in recruitment. The survey features three randomization time points and characteristics (or factors) of interest

Setup of our SMART vignette study. The above design shows the setup of our SMART vignette study. There are three factors in the vignette: patient risk, information on AI, and certainty of AI output, respectively.
A unique advantage of the sequential delivery of vignettes is that it affords the researcher the ability to infer the conditional distribution of the outcome, which in the example of our physician survey, is the clinical decision based on prior outcomes and variables. This is distinct from the type of inference afforded by traditional vignettes, which is restricted to merely the joint distribution of the outcomes and not the conditional distribution. The second feature of adaptive allocation serves to encourage balanced and sufficient cell counts for inference.
For example, inferences on the
Implementation
The general steps for implementing the SMART survey are as follows.
Randomly allocate the first subvignette
Observe response to the first subvignette
Adaptively allocate the second subvignette
Observe response to the second subvignette
Adaptively allocate the third subvignette
Dropouts, if any, do not impact adaptive allocation because each allocation requires counting the number of people with the same assignments and responses (i.e. referred to as their history). If a given participant drops out and is therefore missing a response
For the example online survey, a web application was developed and is hosted by Stanford University under the domain
Simulation study
We simulated multiple hypothetical studies using Python to compare two designs: (1) SMART and (2) a conventional study using baseline randomization. We assumed each hypothetical vignette study to have two levels for each factor, and we varied the number of factors (to up to 4) for each study. Participant responses to any given subvignette were assumed to be binary and were generated by sampling from a Bernoulli distribution. The probability of the Bernoulli distribution was assumed to be the inverse logit of prior responses and subvignette allocations. We evaluated the two study designs on their imbalance of groups.
Conventional implementation
To simulate the conventional approach to vignette studies, we generated all possible vignettes (i.e. all
SMART implementation
At the first randomization point of the SMART study, participants were assigned to a subvignette at random, with a probability of 0.5. For each subsequent time point, adaptive randomization was used to allocate the subvignettes. This means that all prior responses and allocations (i.e. history) were compared across subvignettes (Figure 4). In the case of unequal allocation across two subvignettes, Efrons biased coin was used to allocate the current participant (Efron, 1971). The biased coin would favor the subvignette with fewer participants with pre-specified probability

Example of adaptive allocation. Subvignette allocation realizations
Simulation results
We adjusted the parameters for the weights and biases of the logit function, as well as the probability of Efrons biased coin. We also defined relative loss in order to quantitatively compare the imbalance between groups for each method (refer to Supplementa Material). For each combination of parameters, the relative loss was calculated under both SMART and conventional study designs at each randomization time point. The conventional design only involves randomization at baseline, whereas the SMART design involves multiple points of randomization. Figure 5 shows the loss as a function of number of respondents

Relative loss as a function of number of respondents
The thick line indicates the average relative loss across all iterations and the shaded region indicates relative loss within 1 standard deviation. Different line styles denote each study varying in number of factors (2–4). When
Discussion
In this paper we presented a novel method, SMART vignettes, as a means to generate stakeholder perspectives to address a critical gap in the current AI healthcare field. A unique aspect of our proposed design of SMART vignettes is the sequential delivery of questions that is adaptive by nature. As demonstrated by the relative loss in our simulations, our proposed approach has an advantage over conventional design.
Strengths of our proposed approach include its potential utility in generating data for decision making scenarios in discrete time settings, and in particular scenarios that are of high risk or are in need of testing before deployment. In the example of AI-embedded clinical care, hypothetical scenarios help probe physician perspectives and generate an understanding of when and under what conditions physicians might accept or ignore AI-augmented output in advance of deployment.
This paper has a few limitations. One is that the novel design features we presented are not yet widely available in existing survey platforms (e.g. REDCap and Qualtrics) and therefore has not yet been widely tested outside simulations. While there is a physician survey currently in progress, making this type of SMART design widely deployable to any research team and domain scientists who wish to construct their own SMART vignette survey is an area of future work. Much like fixed design vignette surveys, responder fatigue may also still pose a challenge. However, we argue that the dynamic nature of the survey allows for balanced statistical analysis and for a greater degree of interactivity which may be favorable for some survey responders. Future studies comparing traditional vignettes with SMART vignettes can be used to evaluate how the methodology impacts the perceived burden experienced by the participants as well as dropout rates.
Studies utilizing SMART vignettes stand to generate data on stakeholder reactions when informed by mixed-methods qualitative work, such as semi-structured interviews, focus groups, and/or surveys to articulate questions of interest and hypotheses. Once such questions are articulated, the researcher can probe further using scenario based testing. We posit that this method holds promise in advancing knowledge about how patients and physicians may respond to novel intelligent systems embedded in their health care.
Supplemental Material
sj-pdf-1-mio-10.1177_20597991241240081 – Supplemental material for A novel experimental vignette methodology: SMART vignettes
Supplemental material, sj-pdf-1-mio-10.1177_20597991241240081 for A novel experimental vignette methodology: SMART vignettes by Jane Paik Kim and Hyun-Joon Yang in Methodological Innovations
Footnotes
Declaration of conflicting interests
Funding
Supplemental material
Author biographies
References
Supplementary Material
Please find the following supplemental material available below.
For Open Access articles published under a Creative Commons License, all supplemental material carries the same license as the article it is associated with.
For non-Open Access articles published, all supplemental material carries a non-exclusive license, and permission requests for re-use of supplemental material or any part of supplemental material shall be sent directly to the copyright owner as specified in the copyright notice associated with the article.
