Abstract
Keywords
The treatment of uncertainty is a key feature of sensitivity analysis in cost-effectiveness analysis (CEA). 1 One source of uncertainty in estimating quality-adjusted life-years (QALYs) and incremental cost-effectiveness ratios (ICERs) is the use of health state values (HSVs) attached to descriptions of health states. i HSVs are not constants but variables that have a distribution as well as a point estimate. However, the main focus of valuation studies, research that identifies HSVs for particular health states, has been to generate robust point estimates. Methods for understanding, identifying, quantifying, and reporting their distributions are less developed, and no comprehensive account of their sources and nature exists.
HSVs used in CEA models rarely have one source and undergo much processing before and during modelling. As a result, there are multiple sources of uncertainty that accumulate in the final values used. This is important because it may affect a CEA’s results and decisions that follow from it. For example, a study comparing estimates of QALY gains from diabetes-related interventions in different simulation models found that differences in estimates were mainly due to different modeling methods, but “varying utility values had an impact on incremental QALYs within each of the models.” 2
Current guidance on how to capture and use information on uncertainty does not cover all aspectsii. For example, an ISPOR Good Practice Report provides guidance on sensitivity analysis where alternative value sets are available but does not analyze other HSV relevant uncertainty sources. 3 The Decision Support Unit (DSU) of England’s National Institute for Health and Care Excellence (NICE) provides uncertainty information for dimension and level parameters for UK EQ-5D-3L values, but CEA models require uncertainty estimated for health states, which combine dimensions and levels.4,5 Some papers identify and discuss different sources and types of uncertainty6–8 but lack empirical evidence on their relative importance. A guide to users of EQ-5D-5L values advises that “the analyst should treat the values in economic appraisal as uncertain parameters and subject them to sensitivity analysis, as with other non-stochastic uncertain variables” but notes that “currently this is not common practice.” 9 Similarly, the NICE DSU notes that although NICE submissions require economic model parameter uncertainty to be captured, typically “the uncertainty in the estimation of preference-based valuation weights has been ignored.” 5
A recent scoping review found few studies of methods for identifying and reporting uncertainty around HSVs used in CEA. 10 Those found mostly focused on methods for capturing one type of uncertainty, parameter uncertainty.11–13
This article addresses this gap by providing a comprehensive account of sources of uncertainty affecting HSVs used to calculate QALYs within CEAs and how they interact. It focuses on values elicitation and modeling studies rather than analyzing in detail all possible uncertainty sources. Our aim is to encourage research to measure and report uncertainty around HSVs for better inclusion in CEAs.
What Do We Mean by Uncertainty?
Inconsistent terminology complicates a comprehensive description of uncertainty affecting HSVs. Different definitions appear in different disciplines and fields, for example, exploratory and inferential statistics, econometrics, meta-analysis, mapping, and cost-effectiveness modeling. The following are closest to those in a key health care economic evaluation text. 14
The term “uncertainty,” in relation to HSVs, refers to the imperfect knowledge that we have about them. Uncertainty has different sources that can be categorized into different types, which we call “variability,”“heterogeneity,”“statistical uncertainty,” and “methodological variation.”
“Variability” denotes intrinsic differences between individual cases, for example, in values that different people attach to health states or responses given when self-reporting their health for the same underlying health state. This “natural” variability is not in itself a problem, although it is one source of statistical uncertainty and is not amenable to change by researchers. It is, however, important to ensure that variability is accounted for and summarized in, for example, the choice of descriptive statistics to summarize an HSV distribution, such as mean and variance or median and quantiles.
“Heterogeneity” is usefully defined as variables relating to different subgroups of a population or sample having different statistical properties, such as a different mean or variance. One source of this may be a characteristic that differs between subgroups. Such characteristics may be known, for example, population mean values differing by sociodemographic characteristics or distributions of HSVs in valuation studies differing because they use different valuation methods. Subgroup characteristics may also be identifiable but unknown, such as differences in responses to the state “dead” in valuation tasks, which can be attributed to unobserved latent classes. A slightly different definition is implied by “unobserved heterogeneity,” in which differences between observations are affected by an unmeasured variable (i.e., an omitted variable).
Heterogeneity takes different forms depending on context, and different implications result. For example, HSV distributions for different sociodemographic groups may have different means, which is uncontroversial where they simply describe each group or where they are used for making decisions about each group’s own health care. There are more difficult issues if decision making relates to the population as a whole without regard for groups’ different HSVs; however, solutions to this are in the realm of policy, not science.15,16
“Statistical uncertainty” arises from using samples to infer population variables, for example, sample mean HSVs as an estimator of a population mean, with a standard error (SE) quantifying the sample mean variation around the true population mean. Methods for assessing and handling statistical uncertainty are well developed. Essentially, these assess estimators’ performance in bias and precision. A biased estimator has an expected value over repeated samples that is not equal to the true population value. This cannot be corrected by more sampling, although if the estimator is consistent, a large sample may better approximate the true value. “Precision” concerns random sampling error, whose size can be quantified and used in, for example, statistically based sensitivity analysis. However, measuring precision can be problematic for the typical use of HSVs in CEA, where multiple estimation procedures are used cumulatively. For example, such data may result from value set modeling within valuation studies, subsequent inclusion in meta-analysis, and cost-effectiveness modeling. The cumulative effect on the final results’ precision may be difficult to estimate.
Uncertainty may also be caused by “methodological variation.” One source of this is the different methods and instruments used to elicit values within studies that generate HSVs and how the study data are collected. Participants in such a study must give precise answers to questions to which they may have only an approximate answer. Moreover, their response may be affected by how they are asked, such as the valuation task that they are given and how this was administered. Such issues are addressed by consideration of the reliability and validity of questionnaire instruments and their application. Similar issues arise with methods that analyze the resulting values data, such as population value modeling, meta-analysis, and cost-effectiveness modeling. Such variability cannot be corrected by statistical methods but can be assessed by reliability, plausibility, and acceptability criteria; nonstatistically based sensitivity analysis may also be used.
How Uncertainty in HSVs Combines with Other Sources of Uncertainty from Health State Descriptive Systems
In principle, a CEA study could elicit HSVs from clinical trial patients, capturing all internal sources of uncertainty such as patient variability in disease severity and treatment effects as well as study design, including valuation methods. In practice, patients’ health states are defined by clinical categories, for example, a diagnosis or disability state, or a health state descriptive system (HDS), which uses generic descriptors such as mobility, pain, and effect on activities of daily living. Externally generated HSVs are then attached to these.
An important type of HDS is one that has a value set created using that system to elicit HSVs from a sample of people, often the general public, such as the EQ-5D, SF-6D, or HUI. We will call this an HDS-V. iii A value set is a list of HSVs that are attached to each possible health state described by an HDS-V, often reported as an algorithm that generates those HSVs. Research that elicits values is a valuation study. A CEA model may incorporate data from an HDS-V and use its associated HSVs. However, the data may sometimes describe patients by clinical categories or an HDS without associated HSVs and use mapping to an HDS-V with associated HSVs. 17 Uncertainty at each stage accumulates, and there are even more complex possibilities, which are described below.
Health technology assessment guidance issued by relevant bodies in, inter alia, Denmark, England, and the Netherlands requires HSVs to be representative of their national populations’ preferences,18–20 and the development of most HDS-Vs, such as EQ-5D-5L 21 and SF-6D, 22 reflects this. A valuation study involves collecting and analyzing data; the use of the resulting HSVs in CEA inherits the uncertainty within that study. Sometimes, the sources of HSVs use studies that themselves use data from mapping and meta-analysis studies, each of which contributes further uncertainty. The HSV uncertainty in a CEA study may therefore accumulate over several studies.
Figure 1 illustrates this as a flow diagram. The boxes represent different types of research. Box 1 is the original source of the HSVs, and Box 7 is the end use of the HSVs. The arrows show where outputs from one study type input to other types. The output of any study includes not only uncertainty generated within the study but also uncertainty inherited from studies that input to it. Box 1 is a valuation study (defined earlier) in which stated preference data are collected from a sample of people. Some studies elicit HSVs for disease-related states described in the form of “vignettes” or “scenarios” that describe disease symptoms and their impact on quality of life.
23
Others elicit HSVs for states defined by an HDS-V, called “profiles” (see

Sources of uncertainty in health state values.
Valuation Studies
Valuation studies estimate a population mean HSV for each health state described by a descriptive system or other classification, using a sample representative of a wider population. Typically, this summarizes over a sample the distribution of different individuals’ subjective estimates of values for described health states if they experienced them. It assumes that individuals’ different expressed values result from different underlying preferences, rather than differing precision in value estimation.
Methodological variation arises from the different empirical methods used. For example, in methods that require respondents to choose between different health outcome scenarios, important features include the framing of choices offered, how different scenarios are presented, prompts given, and the choice of health outcomes included. Each study interviewer may conduct the interview and record data differently.
Statistical uncertainty arises from sampling uncertainty and modeling uncertainty. An important aspect of dealing with sampling uncertainty is testing that the sample is representative of a population. This is normally based on externally observable sociodemographic and personal characteristics, for example, age, sex, and experience of ill health. However, what is required is a sample that is representative of the population’s preferences, but the distribution of this over a population is unobserved and therefore unknown. The assumption is therefore that the tested characteristics are the dominant source of variations in preferences.
Assuming a representative sample, sampling uncertainty concerns the precision with which each health state’s population mean HSV is estimated, measured by the SE of the mean. This is normally estimated using the sample standard deviation and the sample size; however, because HSVs are usually estimated using statistical modeling (see below), this is not straightforward.
Modeling uncertainty arises when the mean population HSVs for health states defined by an HDS-V are estimated using regression analysis, for example, estimating coefficients for the severity level within a health domain or dimension. iv This generates a mean population estimate for each coefficient and an associated SE. However, this does not generate SEs for the mean population HSVs for the health states themselves, although they can be calculated for health states included in the estimation data set. The SEs for other health states can also be estimated, but calculations will depend on both the form of the model and an assumed joint distribution of regression coefficients errors (see the “Discussion” section for an example). Sources of variation arise from both parameter uncertainty and model specification.
HDS-V Profile Data
A profile is a health state described according to an HDS-V. For example, an EQ-5D-5L profile might be: No problems in walking about Slight problems washing or dressing Moderate problems doing usual activities No pain or discomfort Slightly anxious or depressed
Profile data collection studies aim to estimate a population distribution of observed profiles using a sample representative of a wider population.
Methodological uncertainty again arises from the reliability of the instrument used to capture health states and how data are collected. Studies that do not rely on a person’s estimate of their health state, such as proxy completion, have additional complexities. Proxy and own health state data reliability may differ, an issue where data sets include both.24,25
Statistical uncertainty arises from sampling, the precision with which the population distribution for severity levels within each HDS-V dimension is measured. Similar to valuation studies, sampling uncertainty is normally dealt with by testing or assuming that the sample is representative of a population based on sociodemographic and biological characteristics, for example age, sex, and clinical ill-health measures. However, the distributions over people of their subjective perception of severity levels are unobserved and therefore unknown.
Mapping
Mapping studies generate HSVs for health states defined by an HDS using values attached to equivalent states from an HDS-V.17,26 It may also be used to convert values between HDS-Vs. Data are usually collected from people who complete questionnaires that identify them on both the HDS and HDS-V. There are two methods for generating HSVs for the HDS. Response mapping generates a correspondence between HDS and the HDS-V health states. An HDS-V value set is applied to the modeled distribution of the HDS. Direct mapping is used where the HDS also has a summary score calculated by applying an algorithm to questionnaire responses. Regression modeling generates HSVs for particular summary scores, which are applied to the HDS’s health states. A special case is where two or more HDS-Vs are mapped. For example, there are mappings between 6 different HDS-Vs 27 and between different versions of the EQ-5D.28,29
As Figure 1 shows, mapping uses profile data, other HDS data, and value sets from valuation studies. It inherits all of the uncertainty from these, plus 3 additional sources. First, statistical uncertainty relating to the precision with which the joint distribution of the responses to the HDS and HDS-V that would be observed in the population is measured by that observed in the sample. How representative the sample is of the relevant population can be tested using sociodemographic, biological, and clinical characteristics, but again, the population distribution of severity responses is unknown.
Second, response mapping may have methodological uncertainty. In practice, it does not simply apply the joint distribution of the HDS and the HDS-V but uses rules to ensure consistency. For instance, responses indicating “full health” in one HDS might always be converted to “full health” in another, regardless of the actual response in the latter. How the resulting uncertainty can be quantified is not obvious.
Third, direct mapping will inherit uncertainty from whatever methods or assumptions are used to create the HDS’s summary score algorithm. Similarly, where different HDS-Vs’ value sets are mapped, this inherits uncertainty from both valuation studies used to create them.
Other HDS Data
All of the uncertainty sources described in 3.2 also apply to HDS self-reported health. If the HDS has a summary score, similar issues to those described in
Disease State Studies
These studies aim to estimate the value of clinically defined health states according to a particular value set. The usual method is to survey people with specified disease states using an HDS-V and use its value set to calculate the mean value for each state. Some studies use data from patients with one particular condition, such as stroke, coronary artery disease, osteoarthritis, and specific cancers 30 ; others use data from patients with different conditions collected from sources such as hospital records or health surveys.31,32 There are catalogues of different studies’ results for, for example, chronic diseases measured by the EQ-5D.33,34
As Figure 1 shows, disease state studies may use profile data and value sets from valuation studies or mapping. Normal sampling variance reflects the selection of data as a sample from a wider population. These HSVs also inherit all uncertainty sources from the profile data collection and from their source (valuation studies, response mapping or other mapping).
A further complication is that the variance of a distribution of HSVs depends not only on the different profile responses that respondents give but also the value set. This variance is a complex weighted sum of both the variances of each response level within every domain or dimension that the HDS-V has and all of the covariances between them, where the weights are the coefficients from the value set. 35 The covariance element introduces interactions between coefficients that are not in the value sets themselves. Moreover, the greater the differences between coefficients, the higher the variance.
Meta-Analysis
In this context, meta-analysis estimates disease state HSVs using data from different disease state studies. 36 This inherits all of the uncertainty from the disease state studies plus uncertainty from how different studies’ data are combined.
This is important because many such meta-analyses use aggregate data, in which the weights assigned to different studies’ estimates of the mean are the inverse of their variance. If the different studies’ HSVs are calculated using the same value set, the variances will not be truly independent. As described, variances are determined by the value sets’ coefficients and variance attributable to differences between study patients’ responses. They therefore have a fixed element that may mean that observed differences are smaller than true differences.
Cost-Effectiveness Modeling
Cost-effectiveness modeling uses HSV data to estimate QALY gains, comparing 1 or more alternative treatments or other health interventions or policies. As Figure 1 shows, there are multiple possible uncertainty sources depending on how the model defines health states and their attached HSVs. Modeling data may be from a clinical trial that collects HDS or HDS-V data or a clinical disease state measures, with HSV data from a valuation study, mapping, a disease state study, or meta-analysis. There are other uncertainty sources in estimating QALY gains, such as illness duration and survival, but we focus here on HSVs.
Where SEs or information on HSV distributions are missing, functional form assumptions are commonly made, for example, that HSVs follow a beta distribution.
In summary, HSVs used in cost-effectiveness modeling are the cumulation of a wide range of sources that multiply and combine uncertainty.
Sources of Uncertainty within Valuation Studies
This section focuses on the HSV uncertainty generated by valuation studies. The other study types in Figure 1 add to uncertainty around HSVs used in CEA modeling, but valuation studies are especially important as the HSV uncertainty they produce is always present in such models. Table 1 lists these uncertainty sources. The presentation order broadly reflects the process of a valuation study: study design, eliciting stated preferences, and modeling value sets. In each case, we identify uncertainty sources and how these are currently addressed in methods that generate and report value sets. This is not exhaustive, and the presentation order for uncertainty sources does not necessarily reflect their importance order.
Sources of Uncertainty in Values Generated within Valuation Studies
DCE, discrete choice experiment; HDS-V, health state descriptive system with a value set; HSV, health state values; HTA, health technology assessment.
According to the earlier classification of uncertainty sources, only 1 section (1b) refers to variability and heterogeneity, 4 refer to statistical uncertainty (2b, 2c, 3b, 3c), and 8 refer to methodological uncertainty (1a, 1c, 1d, 2a, 2d, 3a, 3b, 3c). This reflects how this list was devised rather than suggesting greater importance or prevalence of different uncertainty sources. However, it suggests that methodological uncertainty and how to deal with it is less understood than other types and deserves more attention. By contrast, there is a large body of knowledge from general statistical and econometric theory and practice to handle statistical uncertainty.
The list illustrates the multiple uncertainty sources in each step of a valuation study. Specific design features of valuation tasks, for example, whether HSVs < 0 are sought and what minimum value is possible, will affect the characteristics of the resulting HSVs. The designs of valuation studies include linked statistical choices concerning which HDS-V states to include, how many choices each participant can reasonably undertake, and sample size—all of which generate statistical uncertainty. There are also normative considerations regarding whose HSVs are considered relevant. For example, in the valuation of adults’ health states, there is ongoing debate about whether to use general population or patient HSVs and, in the valuation of pediatric health states, whether to use adult or children’s HSVs. Different populations may have different HSVs,39–43 so where such issues are unresolved, for example, where decision makers lack a clear view, researchers’ sample choices will determine the HSVs’ characteristics.
A fundamental issue of both study design and data collection is how stated preferences are elicited. There is no health economics consensus about which stated preference method is “best” or about which criteria should determine choice of methods, although preferred methods are stated in some countries’ health technology assessment agency guidelines. 44 Different methods produce HSVs with different characteristics.45–49 Valuation studies usually select and report HSVs from one method selected ex ante, rather than reporting how much HSVs might differ by using different methods.
Data analysis methods vary between studies. For example, HSVs cannot take a value greater than 1, meaning “full health,” but they may take negative values, with no theoretically defined lowest value, which may lead to problems of interpretability of the resulting data. This issue is commonly dealt with by transforming or truncating data; how this is done varies. An important uncertainty source is that, except for a small number of studies using a “saturation” design, value set generation entails estimating HSVs for all HDS-V states from directly observed HSVs for a subset. Uncertainty exists around both the parameters within the selected model and its specification.
Table 1 describes efforts made to address some uncertainty sources by, for example, improved study designs, standardized protocols, processes to identify poor-quality data, and modeling methods. However, the way value sets are currently reported to users reflects little of the associated uncertainty.
Discussion
The issue of HSVs uncertainty is important and more complex than usually acknowledged. The question is how we can improve methods for measuring and reporting it.
Some aspects are easy to deal with. For example, as discussed earlier, the SEs of level and dimension coefficients from regression modeling used to generate EQ-5D HSVs are widely reported, but what would be required for use in CEA are the SEs for HSVs attached to profiles. These can be calculated using the regression model’s variance/covariance matrix (VCM), but this is rarely if ever reported by valuation studies. We showed how to do this 5 using data from the influential Measurement and Valuation of Health study 50 by generating a variance/covariance matrix and calculating SEs for every possible EQ-5D-3L profile’s HSV. The SEs range from 0.008171 to 0.012203, which are small compared with the HSVs’ range, from −0.594 to 1.
This relatively small range is evidence of how robust this study’s HSVs are, resulting from a strong research design and a large sample. In studies with greater variances, for example with a small sample and a less constrained range of possible HSVs, the relative size of the SEs may be significantly larger. We therefore recommend that where relevant either the SEs attached to HSVs for health states should be reported or the VCM so that those HSVs can be derived.
Other uncertainty aspects may be harder to conceptualize and measure. But without a full account of uncertainty and the availability of methods for identifying and reporting it, value sets for HDS-Vs published in journals and books will continue to provide incomplete information about it. It will remain difficult for users to incorporate HSV uncertainty into both cost-effectiveness models and other applications of HSVs, for example, summarizing population health survey data.
Jurisdictions with well-established HTA processes that routinely use cost-effectiveness evidence may have available a local value set for each HDS-V and a preferred HDS-V. Where there are alternative acceptable value sets and it is unclear which to use, a deterministic sensitivity analysis of HSVs may suffice. However, as value sets reporting does not adequately cover HSVs uncertainty, a proper sensitivity analysis of them is precluded. Nevertheless, the processes for generating average HSVs yield valuable information on uncertainty, which may be underreported. HDS-V developers, such as the EuroQol Group, have a responsibility to better support value set users by providing more complete information about them.
Despite much work reporting methods for estimating HSVs, limited information is available about the associated uncertainty. For example, no studies reporting value sets for the EQ-5D-5L 21 or -3L 51 report uncertainty information for the estimated health state profile HSVs. To our knowledge, this also applies to other reported HDS-V value sets.22,52 Value set users lack relevant evidence and have limited ways to test QALY and ICER estimation implications.
Clinical, ethical, social, economic, and political perspectives support medical decision making that takes account of people’s preferences by using HSVs. Our analysis aims to strengthen that by improving the confidence that we can place in estimates of them. We therefore recommend the following:
Clinical trial researchers, decision modelers, and health care decision makers should recognize and account for uncertainty more extensively than is currently practiced.
HSV studies should fully report information on uncertainty; best practice guidelines should include this issue.
Researchers should find solutions for unresolved issues of handling uncertainty.
Conclusions
Despite more than 50 years of research on HSVs, limited attention has been given to identifying all sources of uncertainty around them and developing methods to quantify and report them. Consequently, HSV uncertainty is not adequately reflected in QALY estimates or CEAs based on them. We described multiple sources of uncertainty that affect the elicitation and modeling of value sets and showed that HSV uncertainty affects CEA in complex ways, especially where HSVs are inputs to mapping exercises and meta-analyses, where uncertainty combines at each step. Health state valuation studies should advance beyond providing point estimates and include detailed accounts of HSV uncertainty to better serve their users.
Footnotes
Acknowledgements
We thank the members of a study advisory group for their detailed input: Nigel Rice, Ben van Hout, Bas Janssen, Eleanor Pullenayegum, Bram Roudijk, and Mark Sculpher. We thank participants in a discussion of a EuroQol Group conference paper on this topic, in particular our discussant Michał Jakubczyk. David Parkin, Nancy Devlin, and Andrew Lloyd are members of the EuroQol group, but the views expressed in this article are those of the authors and not necessarily those of the EuroQol Research Foundation, who funded this study.
The authors declared no potential conflicts of interest with respect to the research, authorship, and/or publication of this article. The authors disclosed receipt of the following financial support for the research, authorship, and/or publication of this article: Financial support for this study was provided by a grant from the EuroQol Research Foundation, grant EQ-399-RA. The funding agreement ensured the authors’ independence in designing the study, interpreting the data, writing, and publishing the report.
Ethical Considerations
Not applicable.
Consent to Participate
Not applicable.
Consent for Publication
Not applicable.
Data Availability
Not applicable.
