This article presents a Bayesian approach to estimation in multistage experiments based on the reference prior theory. The idea of deriving design-dependent priors was first introduced using Jeffreys’ criterion. A theoretical framework was then established by showing that explicit reference to the design is fully Bayesian justified and Bayesian objectivity cannot ignore such information. Extending the work to multi-parameter problems, a general form of priors was derived from the reference prior theory. In this article, I evidence the good frequentist properties of the reference posterior estimators with normally distributed data. As a notable advance, I address the issue of the point and the interval estimations upon experiment termination. The approach is applied to a data set collected in a clinical trial in schizophrenia with the possibility to stop the trial early if interim results provide sufficient evidence of efficacy or futility. Finally, I discuss the idea of using the reference posterior estimators as a default choice for objective estimation in multistage experiment.
For ethical and economical reasons there is an increasing interest in experiments which allow for early stopping as soon as interim results are conclusive. Beyond the statistical significance of the test, the estimation of parameters is a key element in decision makings. However, it is known that data-dependent stopping rules affect the sampling distributions of estimators. The effect on the maximum likelihood estimator (MLE) of the mean of normally distributed data has been reported for a long time1 along with the deficiencies of the coverage probability of the Wald confidence interval.2 The issue of the bias induced by the stopping rules has given rise to various researches to propose estimators adapted to multistage experiments.
The frequentist solutions require ordering the observation space. The reason is that the sufficient statistic is the couple of variables formed by the stopping stage, noted in what follows, and the sample sum upon stopping noted . An estimator of the mean based on the MLE ordering was studied in terms of frequentist characteristics3 whereas an approach to estimate confidence intervals based on the stage-wise ordering was described.4 This latter ordering assumes that the results corresponding to earlier termination are more extreme than those which terminate later. However, (pre-)ordering the observation space introduces some subjectivity in the inference.5 Other criticisms are that the solutions are not unique and may depend on the information levels for future (unobserved) stopping stages.
An uniformly minimum variance unbiased estimator (UMVUE) can be found from Rao-Blackwell’s theorem which states that, given an unbiased estimator and a sufficient and complete statistic, the conditional expectation of the first to the second is UMVUE. Although (, ) is not complete in the normal case, the expectation of the sample mean at stage 1 conditional to (, ) is still UMVU.6 A computationally quite extensive formula for calculating the UMVUE of the mean was derived3 but the lack of bias is accompanied by a relatively high variance. Another estimator was proposed which consists of adjusting the MLE by subtracting the estimate of its bias.7 The bias is calculated at the adjusted ML estimate using recursive method.
This article describes an objective Bayesian approach to the point and the interval estimations. A prior is objective if it has minimal impact on the posterior distribution. Any candidate objective prior should also results to posterior estimators with good frequentist properties. The idea of deriving design-dependent priors was first introduced using Jeffreys’ criterion.8 Based on numerical results, it was noted that the coverage of the credible intervals are improved in the negative binomial model.9 A theoretical framework was established by showing that explicit reference to the design is fully Bayesian justified and Bayesian objectivity cannot ignore such information.10 The property of correction for the stopping rule bias was then demonstrated in the Bernoulli sequential design.11 On this basis, a class of point estimators in multistage binomial design was developed12 and the work was extended to hypothesis testing.13 The Jeffreys prior enjoys many optimality properties for regular models where asymptotic normality holds, but in the presence of nuisance parameters this prior suffers from many deficiencies. The reference prior theory developed by Bernardo14 allows to overcome these deficiencies. The theory is based on the Kullback-Leibler divergence between the prior and the posterior. More recently, Sun and Berger15 derived a general form of the reference prior in analysis under sequential experimentation.
In this article, I describe the derivation of reference posteriors in multistage experiments with normally distributed data. The approach is implemented using basic programming. The frequentist properties of the posterior estimators are studied using simulations. The approach is then applied to a data set collected in the double-blind clinical trial FAST16 which compares the efficacy of a new compound versus placebo in the treatment of acute exacerbation of schizophrenia. The treatment effect was originally estimated using a model of covariance analysis. To study the properties of estimators, I introduce fictitious interim analyses with possibility to stop the trial early for futility or efficacy based on frequentist test results.
The key aspects of the reference prior theory are presented in the next section. The frequentist properties of the posterior estimators are investigated in Section “Frequentist properties”. The approach is then applied to the FAST trial in Section “Application of approach in clinical trial” in which I provide the results of the original analysis before turning to the multistage design cases. In the conclusion, I discuss the idea of using the reference posterior estimators as a default choice for objective estimation in multistage experiment. The R scripts to produce the results reported in this article along with computational details are provided in Supplemental Material.
Derivation of the reference prior
Reference prior theory
The idea behind reference prior is to maximize a distance between the prior and the posterior distributions as data are collected. Formally, the data have maximum influence on the posterior if the Kullback-Leibler (K-L) divergence is maximum. By considering the expectation of the K-L divergence, the reference prior can be defined based on virtual data before experiment. In this section, I expose the key aspects of the reference prior theory for the one-parameter problem in the fixed sample case. Many supports describing the theory can be consulted including an overall description,17 a formal definition,18 and a didactic tutorial.19 Let us consider the model , wherein is the outcome variable and the parameter, together with a prior specification . The procedure described below applies for any sufficient statistic for for which is in one-to-one correspondence with . Let us consider the -dimensional vector with density and consider the inferential scenario in which the components of are realizations of from independent experiments. The mutual information average is the average of the K-L divergence between the prior and the posterior with respect to the marginal distribution . This takes the form:A non-informative prior is obtained by maximizing . Formally, we have:The reference prior is then derived by taking where the limit allows for improper priors.
Asymptotic theory makes it possible to obtain a convenient form of by rewriting . The problem is reduced to computing:An analytical solution for the limit as can be found using the Bernstein Von Mises theorem, sometimes called the Bayesian central limit theorem. Therefore, is asymptotically Gaussian and concentrated at a value which is the ‘true value’ of in the frequentist sens of the term, i.e., the value of such that the components of are i.i.d. following . Another consequence is that the variance is given by the Fisher information of the selected model. It comes that, in the one-parameter problem, the reference prior is identical to the Jeffreys prior, noted , obtained using Jeffreys’ criterion. If we denote by the expected Fisher information, we have:
Extension to multistage design
We now assume that the data are collected in a -stage experiment whose the design is noted . In the general case, the experiment outcome is obtained from a sequence of outcome values observed at the interim analyses. The analysis times are predefined according to the statistical information available at each analysis.
We consider a sequence of independent outcomes which is observed until experiment termination and we assume that the density of is known. The model is now wherein is a -dimensional subset of the observation space restriction .
The likelihood function in the design takes the form:The introduction of the design information implies rewriting the relation (1) as a function of the expected design-dependent Fisher information such as:Based on (2), can easily be expressed as a function of the expected naive (i.e., non design-dependent) Fisher information:Jeffreys’ criterion applied to the likelihood (2) yields a design-dependent prior which depends on the naive reference prior and the expected stopping time so that:Based on (3), it is evident that is proper if both the naive reference prior is proper and the stopping rule is proper as well (i.e., the stopping time is finite almost surely). The prior reflects the degree of certainty associated with the projected design by over-weighing the probabilities about values more likely leading to late termination. The greater the certainty about values, the higher their prior probabilities. By so counterbalancing the expected effect of early stopping, the reference posterior estimators benefit from a correction for the stopping rule bias.
Extension to multi-parameter problems
In multi-parameter problems, it is often the case that only one or a subset of parameters is of interest. For the sake of clarity, we first consider the fixed sample case. The model is wherein is the parameter of interest and is a nuisance parameter. We would like to find the joint -reference prior, noted , that captures our unequal interest in both parameters. To handle such situation, the reference prior theory allows a simplification which comes down to sequentially computing Jeffreys prior in one-parameter problem. This is described in the following procedure :
Determine Jeffreys prior of conditional on (i.e., assuming is a constant) and derive the conditional reference prior .
If is proper, integrate out to find:
Determine using the standard procedure for reference priors which yields Jeffreys prior of when is one-dimensional.
The joint -reference prior is . Whereas the derivation of the reference prior can be a tricky technical issue, a major simplification appears under posterior asymptotic normality when the parameter space does not depend on .17 The asymptotic variance matrix of ()depends on the expected Fisher information matrix such that . Let us now define for the following statements.
Given and respectively functions of and to be determined, both and can factorize so thatThen, the so-called -reference prior takes the simple form:This approach to nuisance parameter is based on an implicit ordering according which the first parameter is and the second is . If the parameter is now a -dimensional vector, says , the principle can be extended by considering the ordering (,…,). The reference prior relative to this ordering is obtained after successive conditioning such that:The derivation of the reference prior is greatly simplified if is block diagonal, particularly, if, for , the term can be factored into a product of a function of and a function not depending on , such thatwhere does not depend on . Then we have:Based on (4), the introduction of a stopping rule which only depends on the first component is straightforward. The design-dependent reference prior takes the simple form:By this way, we find the results previously obtained in a more general setting.15
Frequentist properties
In this section, I study the characteristics of reference posterior estimators of the mean of normally distributed data with known variance. Let be the parameter of the mean. To this end, I generate a large number of -stage experiments by drawing sequences of outcomes (), , from the random variables , . The generation process is governed by a stopping rule which is defined based on the mean of the ’s at stage , i.e., . Let denote the interval for the observed values to pursue the experiment to stage . The stopping stage is the first such that so that the stopping rule can be expressed by:For the sake of readability, we denote by the mean of the ’s observed until the experiment termination, i.e., . Formally, is the experiment outcome.
As the naive reference prior of follows a normal distribution with infinite variance, the density of the reference prior as expressed in (3) is proportional to , which is sometimes called the corrective term in what follows. Upon experiment termination, the reference posterior takes the form:A method for simulating from this distribution is the following acceptance-reject algorithm:
Step 1. Sample values of from the naive posterior distribution .
Step 2. Numerically estimate by repeatedly simulating the experiment for each value.
Step 3. Sample values of and accept values if , reject otherwise.
In two-stage design (), step 2 is simplified since can be derived analytically. Let us define the continuation interval for at stage and its counterpart for , the corrective term is defined using the cumulative function for normal density as follows:To figure out the influence of the design information on the reference posteriors, I now consider the case where the experiment stops early as soon as the mean of the ’s is negative. This rule is a special case of (6). It can be expressed by:To ease the presentation, I focus on the two-stage design () where the experiment stops at stage if and the five-stage design () where the experiment stops at stage if or at stage as soon as , . Figure 1 displays the curves of the reference posterior densities of in the two designs if the experiment stops at stage 1 and is observed. The stopping rule expressed in (7) implies that the negative values of are associated with an early stopping. In the design-dependent posteriors, this effect is counterbalanced by over-weighing the probabilities of the positive values of more likely leading to a late termination. This correction effect is more accentuated in the five-stage design wherein the stopping rule effect is stronger.
Naive (- - -) and design-dependent (—) reference posterior densities of if the experiment stops at stage 1 and is observed in the designs and .
Table 1 shows the influence of the design information on the reference posterior estimates. The mean, the median, and the mode of the posteriors are provided in the two designs if the experiment stops at stage and is observed or stops at stage 2 and is observed. These estimates are also given in the five-stage design if the experiment continues to stage and is observed. To complete the description, I give the -credible intervals based on the and the posterior quantiles. The raw shift of the point estimates from is a straightforward indicator to measure the influence of the design information. The influence is greater in the five-stage design since the shift of the posterior mean varies from if the experiment stops at stage (which is of the standard deviation of ) to if the experiment continues to stage (which is 5% of the standard deviation of ). The effect of the prior remains however important in the two-stage design since the shift of the posterior mean represents 9% of the standard deviation of if the experiment stops at stage and 5% of the standard deviation of if the experiment stops at stage .
Mean, median, and mode of the reference posteriors and -credible intervals if the experiment stops at stage and is observed or stops at stage k and is observed for in the design or for in the design .
Naive
Design-dependent posterior
posterior
Stop at stage 1
mean,med,mod
-CI
Stop at stage 2
mean,med,mod
-CI
Stop at stage 5
mean,med,mod
-CI
The bias is an important frequentist property of point estimators. Another obvious characteristic is precision, often quantified in terms of the mean squared error (MSE). The bias may be small while precision can be large. In the case that the bias reduction substantially increases the MSE compared to the MLE, the use of a bias-reduced estimator is more than questionable. Figure 2 shows the bias and the MSE of the reference posterior mean estimator in the two-stage and the five-stage designs. The curves are displayed for an interval of ‘true values’ (in the frequentist sens of the term) of the parameter of the mean, which is noted not to confuse with its Bayesian counterpart. As element of comparison, the curves of Whitehead’s bias-adjusted estimator which is known to have good MSE properties are also shown.
Bias and MSE of the naive (- - -) and the design-dependent (—) reference posterior mean estimator and Whitehead’s bias-adjusted estimator ( - ) in the designs and .
The reference posterior mean estimator based on the design-dependent prior exhibits a lower bias magnitude compared to its naive counterpart with ranges decreasing from [] to [] in the two-stage design and from [] to [] in the five-stage design. Whereas the MSEs of the three estimators are comparable, Whitehead’s bias-adjusted estimator keeps an advantage in term of bias with the range [] in the two-stage design and [] in the five-stage design. It is important to recall here that this estimator has been specifically developed for bias reduction. In return, a major criticism is that the adjustment for the bias only depends on the ML estimate without consideration for the statistical information. On the contrary, the bias correction in the reference posterior estimators depends on the stopping stage: the greater the statistical information, the lower the correction. In line with the reference prior theory principle, the design information is appropriately and objectively used at each interim analysis.
An important frequentist characteristic for credible intervals is the coverage rate, i.e., the proportion of the time that the interval contains the ‘true value’ of . Figure 3 shows the one-sided coverage rates of the -credible intervals based on the reference posteriors in the two-stage and the five-stage designs. For the upper and the lower CI limits, the design-dependent approach allows a lower departure of the coverage rates from the nominal level . For the upper limit, the coverage range improves from [] using the naive approach to [] using the design-dependent approach in the two-stage design and from [] to [] in the five-stage design. For the lower limit, the coverage range improves from [] to [] in the two-stage design and from [] to [] in the five-stage design.
One-sided coverage rates of the -credible intervals based on the naive (- - -) and the design-dependent (—) posteriors in the designs and .
For many statisticians this presentation is of minimal value since the population mean can not be summarized by a fixed value of . Consider now that the values of interest are summarized by a random variable which is normally distributed with a standard deviation of 1. This value is chosen to ease the interpretation as it represents the (known) standard deviation of the outcomes by stage. In this way, the quantity of interest is not the bias for a specific value of but the bias averaged over the values of with respect to its distribution. Table 2 shows that the design-dependent approach allows substantial reductions in terms of so-called ‘average bias’. If the distribution of is centered at 0 such that , this quantity decreases from to in the two-stage design and from to in the five-stage design. Important improvements are also observed in terms of so-called ‘average CI coverage’. If the distribution of is shifted to the critical value such that , the average coverage of the upper limit improves from to in the two-stage design and from to in the five-stage design. These corrections toward the nominal value are also observed for the lower limit if since the average coverage improves from to in the two-stage design and from to in the five-stage design.
Average bias of the posterior mean estimator and average coverage rate of the -credible interval based on the reference posteriors in the designs and .
naive
design-dep.
naive
design-dep.
posterior
posterior
posterior
posterior
Average bias for
Upper limit coverage for
Lower limit coverage for
These results obtained with normally distributed data confirm the good frequentist properties that were evidenced in the Bernouilli and the binomial models.12,13
Application of approach in clinical trial
Description of the case study
A clinical trial design is called ‘adaptive’ if an element can be modified at an interim analysis according to prospectively planned specifications with full control of the frequentist type 1 error.20 However, the simple rejection of a null hypothesis is not sufficient to establish a convincing evidence of the efficacy of a treatment and full interpretation of the results should be based on point and interval estimates. In group sequential design, the trial can stop early if the results reveal sufficiently conclusive evidence to support an hypothesis but parameter estimation is subject to a bias toward the more extreme values of hypothesis. To remedy this, I apply the design-dependent Bayesian approach to estimate the treatment effect in a clinical trial.
The motivating dataset comes from the double-blind clinical trial FAST (NCT number : NCT02151656) to compare the efficacy of a new compound versus placebo in the treatment of acute exacerbation of schizophrenia.16 The primary outcome was the change of the positive and negative syndrome scale from the randomization to 6 weeks (noted PANSS). The analysis was based on an analysis of covariance model (ANCOVA) incorporating the PANSS baseline (noted BPANSS) as a continuous covariate and a stratum factor to adjust for possible imbalance across the four European countries where the trial was conducted. The ANCOVA model can be expressed by:where is the PANSS change value for the th subject () receiving the treatment () in the country () and is his/her baseline value. The parameter is the intercept and represents the treatment effect. A positive value of indicates a greater improvement in the test group than under placebo. Next, is the fixed effect of the country. To avoid model over-parametrization, the parameter of the last level is set to 0. Last, is the experimental error assumed to be distributed according to a .
The trial was powered for a one-sided hypothesis test to detect a treatment effect of with and a significance level set to . The ratio of the treatment effect to the standard deviation yields an effect size of . In the original protocol, the one-sided test was a gatekeeper for the two-sided test version in a hierarchical manner. In other words, the treatment effect is first tested at a one-sided significance level and, if this test is statistically significant, the treatment effect is then tested at a two-sided significance level. Totally, 142 subjects were planned to be randomized in a ratio.
Another important information is that an independent data monitoring committee (IDMC) was set-up to decide whether the trial should stop for futility or continue until its expected end based on interim results. The interim analysis was planned to be conducted on half of the patients having completed their 6-week follow-up. The IDMC statistician was in charge to run the programs of analysis using the actual randomization list. There were no formal rules to stop the trial for futility but to help them in their decision-making process the IDMC members were provided with information like the treatment-effect estimate and the frequentist conditional power. The IDMC members were also informed that the treatment effect estimated upon trial termination should be at least 4 to pursue the clinical development of the compound further. Of note, this treatment-effect magnitude corresponds to half of the expectation considered under the alternative testing hypothesis.
Actually, the interim analysis was conducted on patients using an ANCOVA model with the PANSS baseline as the only covariate. The interim estimate of the treatment effect was . Although this magnitude was under the clinical objective, the IDMC members decided to continue the trial until its expected end based on information combining global efficacy and tolerance. The final analysis was conducted on patients, patients in each treatment group, and the treatment effect estimated via model (8) was () (one-sided). Incidentally, this promising result obtained on the trial population revealed a discrepancy between the patients included before and after the interim analysis although there were no modifications in the inclusion criteria.
Estimation upon frequentist test termination
The objective Bayesian strategy to analyzing treatment effect via model (8) is extremely simple and goes as follows. To derive the reference prior, the parameters need to be grouped and set in descending order of interest. Let us consider () with where is the parameter of interest and () are the nuisance parameters. The expected Fisher information for this ordering takes the simple form of a three-bloc diagonal matrix with in the bottom component and elsewhere. Since only the term appears in the matrix, formula (4) yields a reference prior which is:The marginal reference posterior of was derived a long time ago.21 The authors used a prior proportional to , in place of the Jeffreys prior which is proportional to , based on intuitive arguments. The marginal reference posterior follows a Student distribution such thatwhere is the difference between the treatment groups, and are the sample sizes, is the residual mean sum of squares, and is the associated degree of freedom.
We now turn our attention to the estimation in multistage designs. According to (5), the reference posterior of given the difference and the design takes the form:To this end, the FAST dataset is reanalyzed by introducing fictitious interim analyses with possibility to stop the trial early for futility or efficacy. In what follows, we consider two stopping rules which are defined as:
Stop for futility if the difference between the treatment groups is below 4 (): Interestingly enough, this rule corresponds to the context of the IDMC meeting during the trial. This rule was implicitly suggested as it was a condition to stop the development of the compound if such a difference was observed upon trial termination and this information was known by the IDMC members.
Stop and conclude for efficacy if the frequentist ‘-value’ is lower than the Pocock boundaries for testing the no-effect hypothesis: Pocock’s method is used to preserve the frequentist type I error when several analyses to test the same hypothesis are planned. The method results to equal significance levels across analyses.
After completing the trial, the reference posterior (10) can be derived using the acceptance-rejection algorithm described in Section “Frequentist properties”. In step 2 of algorithm, the corrective term is estimated by repeatedly simulating trials for each value generated from the naive posterior (9). For the futility stopping rule, the estimates of are based on simulated sequences of differences , whereas sequences of -values are simulated for the efficacy stopping rule. It is important to note that a Bayesian strategy could also be considered in either stopping rule. For example, the futility stopping rule could be based on the posterior probability that . In this case, is estimated from simulated sequences of posterior probabilities.
For the reanalysis of the FAST data set, we consider two-stage designs () with one interim analysis on half of the trial population and three-stage designs () with two interim analyses on one-third and two-third of the trial population. The selection of the patients in the interim analyses only depends on their inclusion dates. The implementation of the stopping rules results to the futility and the efficacy designs described in Table 3. To match with the first inference of the original analysis, the Pocock boundaries are given for one-sided tests to preserve an overall -risk of . In addition to the designs shown in Table 3, I introduce the ‘combined design’ which combines both stopping rules: the trial can stop early for futility if or for efficacy if the -value is lower than the Pocock boundary.
Multistage designs based on the rules to stop the trial early for futility or efficacy.
Stopping rules
Futility
Efficacy
Two-stage
Analysis 1 ()
stop if
stop and reject if
designs
Analysis 2 (n=134)
reject if
reject if
Three-stage
Analysis 1 (n=45)
stop if
stop and reject if
designs
Analysis 2 (n=90)
stop if
stop and reject if
Analysis 3 (n=134)
reject if
reject if
Table 4 allows measuring the influence of the design information on estimation. The results are shown for all analyses at any stage regardless the earlier decision to stop or continue the trial. The reference posterior means are used to estimate the treatment effect. The interval estimation is based on the 90%-credible intervals whose the limits are given by the 5% and the 95% posterior quantiles. The coverage probability of 90% does not match with the Pocock boundaries. However, this unique coverage for all the analyses whatever the design eases the appraisal of the influence of the design information on the credible intervals.
Reference posterior means and 90%-credible intervals in the naive and the design-dependent approaches for each multistage design.
Naive
Design-dependent approach
approach
Futility
Efficacy
Combined
Two-stage
Analysis 1
designs
Analysis 2
Three-stage
Analysis 1
designs
Analysis 2
Analysis 3
As observed in the previous section, the densities of the design-dependent reference posteriors are moved away from the stopping boundaries. Relative to the naive approach, the treatment-effect estimates are increased in the futility designs and decreased in the efficacy designs. In the combined designs, the posterior densities are moved away from both the futility and the efficacy boundaries and makes the distribution more concentrated on its central value. Consequently, the posterior means are close to those obtained in the naive approach whereas the lengths of the credible intervals are reduced. The fluctuation of the results is maximum in the three-stage designs at the first interim analysis with treatment-effect estimates varying from to . This variation is caused by both the influence of the design information which is stronger in the three-stage designs and the prior weight which is greater at the first interim analysis in presence of a limited number of patients. For the inverse reason, the results at the last planned analyses are more stable with estimates varying from to whatever the design configuration.
Whatever the design, no trials stop early for efficacy since none of the -values pass the Pocock boundaries. The significance levels are at analysis 1 in the two-stage designs and at analysis 1 and at analysis 2 in the three-stage designs. However, the treatment effect estimated using model (8) is at analysis 1 in the two-stage designs so that this analysis is the final one in the futility and the combined designs. Table 5 shows the statistical decisions and the estimates obtained upon trial termination. For the efficacy and the combined designs, the CI limits are based on the and the quantiles of the reference posterior densities, whereas the 90%-credible intervals are shown for the futility designs.
Statistical decisions and estimates obtained upon trial termination using the design-dependent Bayesian approach for each multistage design.
Stopping rules
Futility
Efficacy
Combined
Two-stage
stop at analysis 1
reject at analysis 2
stop at analysis 1
designs
Posterior mean=
Posterior mean=
Posterior mean=
*
*
Three-stage
reject at analysis 3
reject at analysis 3
reject at analysis 3
designs
Posterior mean=
Posterior mean=
Posterior mean=
* The -credible interval is given for descriptive purpose as the trial stops early for futility.
Concluding remarks
The use of data-dependent stopping rules in experiments has long been a source of controversies among theoretical statisticians. Some are reluctant to transgress the stopping rule principle according to which once the data have been obtained, the reasons for stopping the experiment should have no bearing on the evidence reported about the parameter. The stopping rule principle is the main consequence of the likelihood principle which states that all of the information about the parameter provided by an experimental outcome is expressed in the likelihood function. In turn, the likelihood principle is considered as a direct implication of Bayes’ theorem. However, any applied statistician considers that the design information cannot be ignored because of the bias induced by the stopping rule.
An important breakthrough was made by showing that the likelihood principle is no longer a direct implication of Bayes’ rule.10 Bayes’ rule was expressed with an explicit reference to the experimental design . Using the formalism of Section “Extension to multistage design”, is a sequence of independent outcomes observed in the design . We assume that has a known density function which satisfies minimum conditions of regularity. Bayes’ rule can be expressed as:Formulation (11) holds for any multistage design governed by a proper stopping rule. It becomes now evident that a state of prior ignorance cannot be characterized without reference to the experimental design. This formulation also shows that the approach developed in this article is fully Bayesian justified.
The idea behind the reference prior theory is to maximize a distance between the prior and the posterior distributions, as data are collected. In return, the ‘data collected’ have maximum influence on the posterior estimates. In this article, I extend this interpretation to the ‘data collected in a given experimental design’. I also show that the reference prior theory applied to multistage experiments yields a class of Bayesian estimators with good frequentist properties while allowing a unified approach to the point and the interval estimations. The reference posterior estimators make it possible to avoid the various problems of the alternatives such as the pre-ordering in the observation space in the frequentist approach or the bias-correction regardless the level of information as in Whitehead’s bias-adjusted point estimator. These arguments justify the use of the reference posterior estimators as a default choice for objective estimation in multistage experiments.
Footnotes
ORCID iD
Pierre Bunouf
Supplemental material
Supplementary material for this article is available online.
References
1.
ArmitageP. Numerical studies in the sequential estimation of a binomial parameter. Biometrika1958; 45: 1–15.
2.
TsiatisAARosnerGLMehtaCR. Exact confidence interval following group sequential test. Biometrics1984; 40: 797–803.
3.
EmersonSSFlemingTR. Parameter estimation following group sequential hypothesis testing. Biometrika1990; 77: 875–892.
4.
JennisonCTurnbullBW. Group Sequential Methods with Applications to Clinical Trials. New York: Chapman & Hall, 2000.
5.
WhiteheadJ. The case for frequentism in clinical trials. Stat Med1993; 12: 1405–1413.
6.
LiuAHallWJ. Unbiased estimation following a group sequential test. Biometrika1999; 86: 71–78.
7.
WhiteheadJ. On the bias of maximum likelihood estimation following a sequential test. Biometrika1986; 73: 573–581.
8.
GovindarajuluZ. The Statistical Analysis of Hypothesis Testing, Point and Interval Estimation, and Decision Theory. Columbus, OH: American Sciences Press, 1981.
9.
YeK. Reference priors when the stopping rule depends on the parameter of interest. J Am Stat Assoc1993; 88: 360–363.
10.
de CristofaroR. On the foundations of likelihood principle. J Stat Plan Inference2004; 126: 401–411.
11.
BunoufPLecoutreB. Bayesian priors in sequential binomial design. C R Acad Sci Paris, Ser I2006; 343: 339–344.
12.
BunoufPLecoutreB. On Bayesian estimators in multistage binomial designs. J Stat Plan Inference2008; 138: 3915–3926.
13.
BunoufPLecoutreB. An objective Bayesian approach to multistage hypothesis testing. Seq Anal2010; 29: 88–101.
14.
BernardoJM. Reference posterior distributions for Bayesian inference. J Roy Statist Soc B1979; 41: 113–147. (with discussion).
15.
SunDBergerJ. Objective Bayesian analysis under sequential experimentation. IMS Collections, Pushing The Limits of Contemporary Statistics: Contributions in Honour of Jayanta K Ghosh2008; 3: 19–32.
16.
BitterIIstvanJA. Randomized, double-blind, placebo-controlled study of F17464, a preferential antagonist, in the treatment of acute exacerbation of schizophrenia. Neuropsychopharmacology2019; 44: 1917–1924.
FDA. Center for Drug Evaluation and Research (2019), Adaptive Design Clinical Trials for Drugs and Biologics Guidance for Industry.
21.
BoxGETiaoGC. Bayesian Inference in Statistical Analysis. New York: Wiley, 1992.
Supplementary Material
Please find the following supplemental material available below.
For Open Access articles published under a Creative Commons License, all supplemental material carries the same license as the article it is associated with.
For non-Open Access articles published, all supplemental material carries a non-exclusive license, and permission requests for re-use of supplemental material or any part of supplemental material shall be sent directly to the copyright owner as specified in the copyright notice associated with the article.