Concise Definitive Review Series Editor, Jonathan E. Sevransky, MD, MHS

Bayesian Methodology for the Design and Interpretation of Clinical Trials in Critical Care Medicine: A Primer for Clinicians What we want (from clinical trials) is not what we are getting. —Robert Matthews (1) Andre C. Kalil, MD, MPH1; Junfeng Sun, PhD2

Objectives: To review Bayesian methodology and its utility to clinical decision making and research in the critical care field. Data Source and Study Selection: Clinical, epidemiological, and biostatistical studies on Bayesian methods in PubMed and Embase from their inception to December 2013. Data Synthesis: Bayesian methods have been extensively used by a wide range of scientific fields, including astronomy, engineering, chemistry, genetics, physics, geology, paleontology, climatology, cryptography, linguistics, ecology, and computational sciences. The application of medical knowledge in clinical research is analogous to the application of medical knowledge in clinical practice. Bedside physicians have to make most diagnostic and treatment decisions on critically ill patients every day without clear-cut evidence-based medicine (more subjective than objective evidence). Similarly, clinical researchers have to make most decisions about trial design with limited available data. Bayesian methodology allows both subjective and objective aspects of knowledge to be formally measured and transparently incorporated into the design, execution, and interpretation of clinical trials. In addition, various degrees of knowledge and several hypotheses can be tested at the same time in a single clinical trial without the risk of multiplicity. Notably, the Bayesian technology is naturally suited for the interpretation of clinical trial findings for the individualized care of critically ill patients and for the optimization of public health policies.

Division of Infectious Diseases, Department of Internal Medicine, University of Nebraska Medical Center, Omaha, NE. 2 Department of Critical Care, National Institutes of Health, Bethesda, MD. Dr. Sun has disclosed government work. Dr. Kalil has disclosed that he does not have any potential conflicts of interest. For information regarding this article, E-mail: [email protected] 1

Copyright © 2014 by the Society of Critical Care Medicine and Lippincott Williams & Wilkins DOI: 10.1097/CCM.0000000000000576

Critical Care Medicine

Conclusions: We propose that the application of the versatile Bayesian methodology in conjunction with the conventional statistical methods is not only ripe for actual use in critical care clinical research but it is also a necessary step to maximize the performance of clinical trials and its translation to the practice of critical care medicine. (Crit Care Med 2014; 42:2267–2277) Key Words: Bayes; clinical trial; critical care

T

he choice for the quote above was not a random choice. The question we clinicians want to ask every time we have in our hands the publication of a new clinical trial is: What is the probability that the new treatment is better, given the trial result? It sounds very logical, doesn’t it? However, this question cannot be answered by the way most clinical trials have been performed, that is, by the conventional (i.e., frequentist) statistical methodology. Fortunately, this important question for everyone who works at the bedside of critically ill patients can be properly answered if we add the Bayesian methodology to the design and interpretation of clinical trials. Bayes’ rule combines mathematical simplicity and cleverness. It states that if we multiply the likelihood of the data with the known knowledge or beliefs (prior probability) regarding a parameter of interest, we can produce the posterior probability of this parameter; this new probability will became part of future new prior probabilities as new data arise. In other words, in science, as in real life, we should learn from our experiences and readjust our knowledge according to the accumulation of growing information on a given subject. Bayes’ theorem can be described mathematically as follows: Probability (Hypothesis |Data) =

Probability (Data |Hypothesis) × Probability (Hypothesis) Probability (Data)

Many scientific fields have fully adopted and applied Bayes’ technology to their research methodology, for example, www.ccmjournal.org

2267

Kalil and Sun

astronomy, engineering, chemistry, genetics, physics, geology, paleontology, climatology, cryptography, linguistics, ecology, and computational sciences, just to cite a few. The uptake of the Bayesian methodology by the medical sciences has been slow but steady. In 1994, a PubMed search for “Bayesian Clinical Trials” generated a total of only 23 articles, while 20 years later, the same search generated 1,852 articles—of which 32 were associated with critical or intensive care. Furthermore, the Food and Drug Administration has released an official guidance for the use of Bayesian statistics in medical device clinical trials in 2010 and the agency has already used Bayesian trials and analyses for its advisory committee panel meetings (2). Bayesian guidelines from the European Medicines Agency have also been released (3). It is not the scope of this article to discuss this disparity between Medicine and other scientific fields, but it suffices to say that this impressive 80-fold increase in medical publications using Bayes methods in just 20 years is the tip of the iceberg. In addition, the rapid progress made in computational technology in the last few decades has allowed the application of Bayesian methodology in ways never before imagined. We strongly believe that by being more proactive in both understanding and applying this technology to critical care research, we will be closer to our ultimate goal as physicians and scientists: to improve the survival outcomes of our critically ill patients. In this article, we aim to introduce the reader of critical care medicine to Bayesian methodology and demonstrate the utility of this methodology for critical care research. We Table 1.

will neither address the historical/philosophical controversies between Bayesian and conventional statistics nor the methodology’s mathematical principles; however, we submit the readers to a set of didactic articles geared to the medical field (4–16). We provide a table highlighting the major differences between both methods regarding clinical trial research (Table 1) adapted from five studies (4–7, 14).

WHAT IS THE BAYES’ THEOREM? A clinical example is used to facilitate the understanding of Bayes’ theorem in the ICU setting. The patient is admitted with shock of unclear etiology. The knowledge of the shock status is not enough for the clinician to make inferences about the etiology, and without a clearer picture about the etiology, the treatment of shock will likely be compromised and potentially harm the patient. The physician will search for more data through both history and physical examination and, if available, through radiologic and laboratory tests. Now the physician finds out that the patient is a 20-year-old college student with recent onset of fevers and headaches and a physical examination with petechial rash; induction reasoning leads the physician to the diagnostic hypothesis of meningococcal meningitis. Thus, aggressive fluid replacement and immediate administration of antibiotics would be lifesaving at this time. However, if the history was different, that is, the patient was 60 years old and presented with shock and new-onset severe chest pain, and the physical

Clinical Trials: Comparison of Conventional and Bayesian Designs

Area of Contrast

Conventional

Bayesian

Critical question

What is the probability of the trial result (observed data), given the treatment effect (study hypothesis)?

What is the probability of the treatment effect (study hypothesis), given the trial result (observed data)?

Reasoning

Deduction

Induction

Prior information

Informally used with design but not with interpretation

Formally used with design and interpretation

α and β errors

Prespecified

Unnecessary

Test hypothesis

Prespecified

Unrestricted

Sample size

Prespecified

Unrestricted

Interim monitoring plan

Prespecified

Unnecessary

Data source

Random

Random

Variables of interest (e.g., risk ratio)

Fixed

Random

Summary measure

p value, estimates, and CIs

Posterior probability

Inference basis

Observed and unobserved trial data

Observed trial data and prior knowledge

Uncertainty

Unquantified

Quantified

Subgroup analysis

Multiplicity is present

Multiplicity is absent

Trial interpretation

Confusing

Straightforward

Clinical decision making

Difficult to translate to bedside

Easy to translate to bedside

Practice guidelines

Rigid and less intuitive

Flexible and more intuitive

2268

www.ccmjournal.org

October 2014 • Volume 42 • Number 10

Concise Definitive Review

examination suggested cardiac tamponade, the use of aggressive fluids and antibiotics would not save the life of the patient; in fact, would lead to his/her demise. This type of induction reasoning we use every day in our clinical decision making, that is, we update our diagnostic hypothesis with the gathering of the new data is the core of the Bayes theorem. In other words, the probability that our hypothesis is true is appropriately influenced (i.e., conditional) by a new piece of evidence. Hence, the classical Bayes theorem formula: Probability (Meningitis |Physical) =

Probability (Physical |Meningitis) ´ Probability (Meningitis) Probability (Physical)

Translation: The probability of meningococcal meningitis (hypothesis) given the history and physical examination (data), also called posterior probability, is equal to the probability of these data given the hypothesis divided by the probability of the history and physical examination findings (new evidence) and multiplied by the probability of meningitis (also called prior probability). This is analogous to what we do daily in our clinical practice for positive predicted values, which has been well captured by a previous article by Gill et al (17). For a comparison, here is the Bayesian approach for the positive predictive value: Probability (Disease |Test Positive) =

Probability (Test Positive |Disease) ´ Probability (Disease) Probability (Test Positive)

As sensitivity and specificity alone are poorly informative (and even misleading) without knowing the prevalence (prior) of the disease to determine the positive predictive value, so are the p value and power of a clinical trial without knowing the prior probability of the treatment effect (15). As Matthews (1) described, “Bayes’ theorem is ultimately an expression of an unavoidable mathematical fact: if we want to answer the question we are interested in—what is the probability that the treatment is efficacious, given the trial result—we must include any prior insights we have.”

WHERE DO THE PRIORS COME FROM? In the meningitis example above, the prior came from the epidemiological data captured by the physician’s history taking. The prevalence determines the prior for the Bayes’ rule. This is classically depicted by the diagnostic positive predictive values for a specific disease, which will increase with the increasing prevalence of the disease; this disease prevalence is exactly the definition of the prior information, which we use as part of our diagnostic tools when we see new patients. How can we use priors to update our view of the benefits of a given treatment? Similar to diagnostic decisions, if we have data derived from randomized trials, meta-analyses, observational studies, or institutional experiences with a specific treatment, then all these data can be formally entered into the Bayes’ rule to Critical Care Medicine

compose single or multiple priors (1, 5, 18–24). The availability of multiple priors is essential to understand the flexibility and advantage of Bayes for treatment approaches. Clinicians can have different interpretations of the very same data from the same trial; this may (or may not) modify their current view of a given treatment according to their prior information. This constant data gathering and accumulation (prior) gives us the de facto opportunity to update our current knowledge status with new posteriors in order to make the best treatment decisions. Of note, the definition of a prior does not require a temporal order, that is, data from previous, concurrent, or selected trials can be incorporated into the prior (7, 22, 25, 26). What about if we are going to design a phase II or III clinical trial for a new therapy, where do we get the priors? The way it has been done for decades is by finding the best “educated guess” as a single prior to design a new phase II or III trial with a single hypothesis. This may explain in part the failure of sepsis trials in the last 30 years (27). Here, all possible evidence, be it small or large, has to be evaluated and this includes animal studies, human pharmacokinetic and pharmacodynamics studies, drug biological activity, drug effect on biomarkers, activity of similar drugs or similar classes of drugs used for the same or other disease processes, and last, expert opinion. How do we get the priors in the worst case scenario, in which no external information is available? One Bayesian approach is by assuming minimal or no external information through the use of noninformative priors. This means that the posteriors are calculated with the assumption that no other related data is available to inform the value of the prior. The other approach is by methodically constructing priors through the evaluation of different degrees of belief among many experts and formally translating these various degrees into mathematical calculations to evaluate them in the context of both study design and interpretation. It has been suggested by Moyé (16) that the active incorporation of counterintuitive prior information should be performed in this worst case scenario as “the more passionate the investigator, the greater the protection the priors require from their strongly held opinion.” Hence, the reader should be aware that for small sample size scenarios for priors, the expert opinion can be overly influential. The good news is that this extreme scenario of no external information is very rare before the execution of phase III efficacy trials since they are usually preceded by preclinical studies and phase I, IIa, and IIb trials. Once the new data accumulate, the precise source and accuracy of the original prior information will naturally become less relevant because the posteriors will progressively become less influenced by the initial priors (1). In other words, the constant updating process naturally allowed by the Bayesian approach will build more robust posterior predictions (i.e., decrease uncertainty), which is exactly what we search for to best interpret and translate the trial results to our patient care. Thus, the major benefit of the Bayesian approach is that all priors are explicitly derived and reported in the clinical trial with several treatment hypotheses, whereas the conventional approach describes only implicitly the single prior for the single treatment hypothesis. www.ccmjournal.org

2269

Kalil and Sun

BAYESIAN APPLICATION TO THE INTERPRETATION OF CLINICAL TRIALS Bayes’ rule touches one of the most important issues faced by Medicine as a scientific field and as a clinical practice in the 21st century: Why do practicing physicians commonly interpret the results of the very same trial so differently? If we ask two critical care physicians about how they would approach a patient with septic shock—considering that both trained in the same fellowship program and now practice in the same hospital and both read the last version of the Surviving Sepsis Campaign (28)—they could come up with very different treatment regimens. In reality, this can be seen with the wide range of compliance with the sepsis bundles that has been documented in many different studies from many countries (29–36). If the evidence and current information they have available is the same, why would their treatment approach be different? We conjecture that conventional statistics is poorly suited for the trial interpretation and individualization of patient care. We demonstrated in a previous comprehensive Bayesian analyses of sepsis trials that the misinterpretation of the evidence is a major barrier to improve the outcome of patients with severe sepsis (37). Table 2 shows the research hypothesis (15, 38–40) to provide a didactic example: for the same clinical trial with

type I error of 5% and 80% power, the probability of a positive finding (i.e., p < 0.05) being true is dramatically different for a different prior probability that a new vasopressor is better than the control. Would you accept to use the vasopressor results from the first interpretation and apply it to your patient care? We hope not! The accurate interpretation of this clinical trial is entirely dependent on the prior probabilities, which are not used at all with conventional methodology. Bayesian analysis can improve the trial interpretation and at the same time can allow the clinician to translate this information into bedside clinical practice, that is, shifting the focus from statistical to clinical significance. Here is a real-life example. The trial on early goal-directed therapy (EGDT) for patients with septic shock (41) showed one of the largest survival benefits in the history of critical care medicine: 16% absolute risk reduction for mortality or a 42% relative risk reduction with the 95% CI ranging from 13% to 62% and a p = 0.009. By the conventional statistics, this is a very strong result. However, several other studies (42–51) have shown inconsistent results regarding survival benefits with this approach. A straightforward Bayesian analysis which incorporate information from these additional studies is shown in Table 3: if you believe that the none of these other studies should change your

Interpretation of a Clinical Trial With Positive Results (p < 0.05) Replicated 1,000 Times With Different Priors Table 2. Variable

New Drug Is Better Than Control

New Drug Is Not Better Than Control

Total

First interpretation  Prior probability that new vasopressor works = 5%   Positive trial result (p < 0.05)

40 (true-positive trial [power] = 80%)

48 (false-positive trial [p] = 5%)

88

  Negative trial result (p ≥ 0.05)

10 (false-negative trial = 20%)

902 (true-negative trial = 95%)

912

   Total

50 (prior = 5%)

950

1,000

Conclusion: New vasopressor is better in 46% (40/88) of all trials with p < 0.05, or 54% of all positive trials will have a significant result despite the fact the new vasopressor is not better (54% false-positive). Second interpretation  Prior probability that new vasopressor works = 15%   Positive trial result (p < 0.05)

120 (true-positive trial [power] = 80%)

43 (false-positive trial [p] = 5%)

163

  Negative trial result (p ≥ 0.05)

30 (false-negative trial = 20%)

807 (true-negative trial = 95%)

837

   Total

150 (prior = 15%)

850

1,000

Conclusion: New vasopressor is better in 74% (120/163) of all trials with p < 0.05, or 26% of all positive trials will have a significant result despite the fact the new vasopressor is not better (26% false-positive). Third interpretation  Prior probability that new vasopressor works = 60%   Positive trial result (p < 0.05)

480 (true-positive trial [power] = 80%)

20 (false-positive trial [p] = 5%)

500

  Negative trial result (p ≥ 0.05)

120 (false-negative trial = 20%)

380 (true-negative trial = 95%)

500

   Total

600 (prior = 60%)

400

1,000

Conclusion: New vasopressor is better in 96% (480/500) of all trials with p < 0.05, or 4% of all positive trials will have a significant result despite the fact the new vasopressor is not better (4% false-positive). 2270

www.ccmjournal.org

October 2014 • Volume 42 • Number 10

Concise Definitive Review

Posterior Probabilities With Early Goal-Directed Therapy Table 3.

Published p (Frequentist) Prior Probability (Bayesian)

0.009 Posterior Probability of New Treatment Being No Better Than Control

Enthusiastic

0.01

Mild skeptic

0.05

Mild-moderate skeptic

0.14

Moderate skeptic

0.25

Severe skeptic

0.41

Adapted with permission from Kalil and Sun (37). Adaptations are themselves works protected by copyright. So in order to publish this adaptation, authorization must be obtained both from the owner of the copyright in the original work and from the owner of copyright in the translation or adaptation.

interpretation of the study by Rivers et al (41), then you are part of the enthusiastic interpretation, that is, there is only 1% chance (posterior probability) that the EGDT would be no better than controls. On the other side, if you believe that some of the other studies should cast doubts on the very large survival benefits seen with the pivotal trial, that is, you are mild to moderate skeptical, then your interpretation translates into a 14% chance of EGDT being no better than controls. Based on the current conventional view that a 5% false-positive rate would be the most we would tolerate in order to call it a significant result, a 14% chance of not working would likely prevent you to use it in a standard ICU practice. If these posterior probabilities are helpful to you but not enough to make your final decision regarding EGDT, then Bayesian analysis can be used to ask an additional question regarding a meaningful clinical threshold based on your knowledge and clinical experience: What is the (posterior) probability of decreasing the relative risk of death of my patient with septic shock by at least 15% with EGDT? Table 4 shows that if you are enthusiastic about this therapy, your patient will have a 94% chance of having a 15% reduction in death, but if you are only mildly skeptic, then the chances of your patient reaching this clinical threshold are

reduced to 62%, which would be too low to be introduced into standard clinical practice. Hence, instead of putting too much or too little trust in the opinions of an editorialist or an expert on the field, the ICU clinician can feel empowered to make his/ her best clinical decisions by the use of the Bayesian interpretation in conjunction with the conventional statistical reports. Because of the controversies surrounding EGDT (52), there are currently three phase III clinical trials being executed in the United States, United Kingdom, and Australia. The Bayesian interpretation provided here should not only help the ICU physician to direct his/her own clinical practice for patients with severe sepsis while this trials’ results are not available but also provide the prior to put the findings of these new clinical trials in the context of the entire available evidence. During the peer-review process of this paper, the ProCESS trial, which evaluated the efficacy of EGDT in the United States was published (53); the results showed no significant mortality differences between EGDT, protocol-based therapy, and usual care arms. If we update the original report by Rivers et al (41) with the ProCESS trial results and utilize the probabilities we provide in tables 3 and 4 of this manuscript, the Bayesian posterior probability that EGDT is no better than control is 0.25 (i.e., probability of a false-positive result with EGDT = 25%), and the posterior probability that EGDT reduces the relative risk of death by at least 15% is 6% (i.e., a 94% chance that EGDT does not reduce mortality by a clinically meaningful threshold).

BAYESIAN APPLICATION TO THE DESIGN OF CLINICAL TRIALS The implementation of Bayesian design in clinical trials has grown in the last two decades (12, 54–56). Two thirds of the articles on this subject have been published on biostatistical journals (54). Here, we provide a summary of the strengths and limitations of the Bayesian methods on phase III clinical trials, meta-analysis, and comparative-effectiveness research. Phase I and II clinical trials are described elsewhere (57–64) due to space limitation. Bayesian Phase III Trials Phase III trials are characteristically multicenter, randomized, comparative studies aimed at confirming the efficacy of a new

Posterior Probabilities of Relative Risk Reduction (RRR) With Early Goal-Directed Therapy Table 4.

Posterior Probability of Mortality Relative Risk Reduction Prior Probability (%)

RRR > 15%

RRR > 20%

RRR > 25%

Enthusiastic

94

87

78

Mild skeptic

62

41

22

Mild-moderate skeptic

27

9

2

Moderate skeptic

6

1

0

Severe skeptic

0.4

0

0

Adapted with permission from Kalil and Sun (37). Adaptations are themselves works protected by copyright. So in order to publish this adaptation, authorization must be obtained both from the owner of the copyright in the original work and from the owner of copyright in the translation or adaptation.

Critical Care Medicine

www.ccmjournal.org

2271

Kalil and Sun

treatment. One attractive feature (and point of controversy) of Bayesian designs is the choice of a prior probability distribution that summarizes the prior knowledge (or the lack of). Jennison and Turnball (23) divided prior distributions into five categories: 1) clinical priors elicited from experts; 2) skeptical priors centered at zero with a small probability of having a meaningful benefit; 3) enthusiastic priors with a small probability of no beneficial effect; 4) noninformative priors that convey as little prior information as possible; and 5) “pragmatic Bayes” proposed by Grossman et al (24) with a skeptical prior that is equivalent to an existing study with zero difference between treatments and a sample size 25%, the size of the planned study. Studies based on Markov Chain Monte Carlo simulations have shown that with five or fewer interim analyses, the type I error rate is controlled at 0.05 (24). One area of important application for Bayesian designs is the use of adult data to guide pediatric trials. Children are routinely excluded from adult trials, so when pediatric trials are proposed, there are usually substantial amount of adult data available. Therefore, it is very attractive to be able to utilize adult data for the design and analyses of pediatric trials, especially for rare diseases for which a standalone trial is unfeasible. Schoenfeld et al (65) proposed a Bayesian hierarchical modeling approach that allows pediatric trial data to be pooled with adult trial data and showed a substantial increase in power. Goodman and Sladky (66) proposed a Bayesian noninferiority trial for pediatric Guillain-Barré syndrome. They used a meta-analysis of adult data to generate a prior for the difference between the new treatment and the positive control. They generated the priors from the meta-analysis results by disregarding a small beneficial effect of the new treatment and then by reducing the information from the adult trials by half. The predominance of frequentist designs in drug trials motivated the development of “hybrid” methods (67). These methods use prior information in the design of trials, but the analysis remains frequentist. For example, frequentist power calculations assume the true variable (e.g., risk ratio) to be a fixed number, whereas Bayesians assumes it to be random with a (prior) distribution. For large phase III trials, interim analysis is important in case convincing evidence that has been accumulated before the planned termination of the trial. For example, investigators may be ethically obligated to terminate a trial early if the new treatment is shown to be worse than the control treatment so that no more participants would be harmed. On the other hand, if the new treatment is shown to be superior, the trial could also be terminated early so that the control patients would not need to continue on suboptimal therapy. Under the frequentist framework, with each interim analysis, the rate of false-positive findings increases due to the repeated evaluation of the ongoing data gathering. To allow interim analyses and control the probability of false-positive findings, group sequential (GS) methods are widely used (23) with the frequentist approach. One criticism of the GS methods is that by simply looking at the data in interim, it could affect the final finding of the whole study, especially when the final p value is near the nominal 0.05. Under the Bayesian framework, there 2272

www.ccmjournal.org

is no need to adjust for interim analysis, since each analysis straightforwardly update the posterior distribution, hence, eliminating the risk for multiplicity. Berry (67) suggested that decision should be also made based on the benefits and costs of the different outcomes. Another way of interim monitoring is the stochastic curtailment (23). Conditioning on the data at the current evidence and fixed variable values for the distribution of the remaining data, a decision will be made regarding whether the trial will be stopped early. For example, the trial can be stopped early if the probability of failing to reject the null hypothesis is low under the alternative hypothesis since continuing the study is “futile” or if the probability of rejecting the null hypothesis is high under the null hypothesis since rejection is mostly assured. Note here the variables are assumed fixed, without accounting for the uncertainty. A hybrid approach is to use the posterior distribution of these values to calculate the probabilities of positive findings in the final frequentist analysis, which accounts for the uncertainty and incorporates all available information, also called the “Predictive Probability of Success” (69). Adaptive design has been a very intense area of research in recent years (54, 70–75). The major difference between traditional and adaptive trials is regarding the accumulating data while the trial is ongoing, which can be used to modify the trials’ course in the adaptive, but not in the traditional design. This approach has already been used in clinical research, but uncommonly through the Bayesian approach, which allows for several types of adaptation at the same time, for example, more efficient drug-dosing evaluation, the need for dropping or adding drug doses or study arms, seamless transition between drug development phases within the same trial, more efficient distribution of patients to each study arm, and modifications of accrual rate (70). The Pharmaceutical Research and Manufacturers of America working group defined adaptive design as “a clinical study design that uses accumulating data to decide how to modify aspects of the study as it continues, without undermining the validity and integrity of the trial” (74). They emphasize the “by design” nature of adaptive designs. Adaptive designs include a broad range of designs: Kairalla et al (73) categorized adaptive designs into three phases: 1) Learning phase is composed of adaptive dose response studies based on toxicity and/or efficacy. 2) Combined (seamless) phase includes seamless phase I–IIa and phase IIb–III designs. Adaptive seamless designs combine more than one phase into a single study to speed up the drug development, and all data are used in the analysis to be more efficient. 3) Confirmatory phase includes adaptive randomization, GS, sample size reestimation (SSE), and combination of GS and SSE. Adaptive randomization methods include covariate adaptive randomization, which attempts to make treatment groups more balanced with respect to key covariates, and response adaptive randomization, which tries to assign more patients to the superior treatment group. SSE allows sample size to be updated using the accumulating data during the trial. The frequentist properties and the interpretability of adaptive designs may be rigid and confusing (73). In comparison, Bayesian methods are October 2014 • Volume 42 • Number 10

Concise Definitive Review

very attractive for adaptive designs due to their flexibility. It has been argued that Bayesian methods allow data to be analyzed any time during the ongoing trial without the well-known limitations of the conventional statistics (26). Importantly, there are limitations pertinent to conventional statistical methods that also apply to Bayesian methods, for example, the presence of bias and/or confounding can make the results of clinical trials less relevant or even invalidate them. These problems cannot be resolved alone by either conventional or Bayesian inferences. This is essential to recognize because no currently available statistical methodology is fool proof against bias and confounding (76). Bayesian Analysis to Evaluate the Need for Confirmatory Phase III Trials If a phase III trial shows strong and convincing efficacy results, it is unlikely that its new findings would be much influenced by skeptic priors. However, if the results are not strong, then they can be possibly changed by priors. Because the definition of strong or weak efficacy results necessarily changes according to the specific disease process, the clinical relevance of the results of a single phase III trial can be more properly evaluated by different priors to test the strength of the new evidence in the context of the clinical needs expressed by different priors. For example, if after a positive phase III trial a major efficacy difference is observed between noninformative and skeptic priors, this may suggest that a confirmatory trial is warranted (i.e., potential false-positive trial). On the other side, if after a negative phase III trial a major efficacy difference is observed between noninformative and enthusiastic priors, this may suggest that another trial is warranted (i.e., potential false-negative trial). This is an example of how even noninformative priors can be highly relevant to clarify the need for confirmatory trials. This type of sensitivity analysis is an excellent tool to assess the need or not for a confirmatory phase III trial (37, 77, 78). In addition, the needed sample size for the next (confirmatory) phase III trial can be estimated, optimized, and most importantly, designed with a substantially smaller number of patients by straightforward Bayesian tools compared with conventional methods. This was recently demonstrated in a severe sepsis trial in which 600 patients were needed for a confirmatory trial with Bayesian design, whereas 2,000 patients were needed for the same trial by conventional design (78). Bayesian Reporting for Clinical Trials There are two guidelines (79, 80) and three comprehensive analyses (37, 77, 81) regarding the items that should be reported in Bayesian studies. Table 5 summarizes the main points to be reported.

BAYESIAN APPLICATION TO META-ANALYSIS Meta-analytic methodology has expanded substantially in the last decade and the limitations of the conventional methods are increasingly being resolved by the application of Bayesian technology (82, 83). The advantages of including the Bayesian approach into the performance of meta-analyses include the following (79, Critical Care Medicine

Table 5. Optimal Reporting of Bayesian Clinical Trials Introduction  Background  Intervention  Objectives Methods  Design of study  Statistical model    Prior distribution     Rationale for choice     Sensitivity analysis   Likelihood of observed data    Posterior distribution     Variable estimates     Posterior densities    Joint probabilities if multiple comparisons     Bayes’ factors    Computation    Checks for convergence if Markov Chain Monte Carlo used    Methods for generating posteriors    Software used, how validated if not publically available Results  Description of all posterior distributions  Description of all sensitivity analyses if alternative priors used Limitations  Description of all design and methodological limitations

82–85): 1) the full spectrum of uncertainty is accounted for and measured in all analyses based on the entire state of knowledge; 2) it enables evidence from different sources to be formally incorporated in the probability calculations; 3) studies with different designs can be evaluated with appropriate weights to combine comparable studies through Bayesian hierarchical modeling, also called cross-design synthesis; 4) the borrowing strength concept can be applied through Bayesian hierarchical models; this concept allows a fully Bayesian implementation of random effect models, which accounts for the uncertainty in the estimated random effect distribution; 5) no asymptotic normal approximation is necessary, thus making the occurrence of zero events in an arm of a study mathematically treatable; 6) multiple testing issues, a frequent problem with conventional statistics, is eliminated with Bayesian analysis; 7) the final estimates of the Bayesian meta-analysis are reported as posterior probabilities, which are easily translated to individual-patient application; 8) the results from the Bayesian meta-analysis may be used for the design of future studies; and 9) the Bayesian framework naturally allows www.ccmjournal.org

2273

Kalil and Sun

same parameter estimates as the three previously published meta-analyses (i.e., risk ratio or Study name Events / Total Risk ratio and 95% CI odds ratio) (Fig. 1A), and then Risk Lower Upper we applied a Bayesian analytic ratio limit limit Steroids Control approach. Of note, we started Bollaert 1998 0.50 0.25 1.02 7 / 22 12 / 19 the analysis without any skeptiChawla 1999 0.55 0.24 1.25 6 / 23 10 / 21 cism by using a noninformative Briegel 1999 0.75 0.19 2.93 3 / 20 4 / 20 Yildiz 2002 0.67 0.35 1.27 8 / 20 12 / 20 prior probability. Our findAnnane 2002 0.89 0.73 1.08 82 / 151 91 / 149 ings of the three meta-analyses Confalonieri 2005 0.08 0.00 1.29 0 / 23 6 / 23 consistently showed that in the Tandan 2005 0.85 0.62 1.15 11 / 14 13 / 14 best case scenario, low-dose Oppert 2005 0.99 0.52 1.88 10 / 23 11 / 25 steroids had a 57% probability Rinaldi 2006 0.86 0.33 2.21 6 / 26 7 / 26 Cicarelli 2006 0.63 0.35 1.12 7 / 14 12 / 15 of reducing the relative risk of Meduri 2007 0.57 0.27 1.20 10 / 42 8 / 19 death by 15% in patients with Sprung 2008 1.09 0.85 1.40 86 / 251 78 / 248 septic shock. Figure 1B shows 0.84 0.72 0.97 236 / 629 264 / 599 in a didactic form the Bayesian 0.1 0.2 0.5 1 2 5 10 results parallel to the forest Favors Steroids Favors Control plot previously published by Annane et al (87). The Bayesian graph is highly transparent by B showing the changes in survival probability after the addition of each trial until the report of the most updated posterior probability (57%) after the last performed trial by Sprung et al (90). This calculated probability (57%) of benefits with steroids was not far from flipping a coin, and this could never be appreciated if we only used the conventional methodology. The estimated probability would be even lower if we had used any skeptic prior probabilities. In addition, we found that the probability of increasing the risk of superinfections, hyperglycemia, and gastroinFigure 1. A, Conventional analysis of low-dose steroids mortality relative risk reduction. B, Bayesian analysis testinal bleeding were all eleof low-dose steroids. RRR = relative risk reduction. Adapted with permission from Kalil and Sun (86). Adaptavated with low-dose steroids, tions are themselves works protected by copyright. So in order to publish this adaptation, authorization must 81%, 99%, and 73%, respecbe obtained both from the owner of the copyright in the original work and from the owner of copyright in the translation or adaptation. tively. Based on the disagreements the design differences for accounting cost and utilities, which is ideal for guidelines and and potential reasons for the negative results found in the last public health policy decisions. phase III trial by Sprung et al (89), we redid the entire Bayesian Here is a practical example of the utility of the Bayesian meta-analyses with and without the trial by Sprung et al (89): meta-analytic approach in critical care medicine. The use of the overall results remained similar. In addition, two real-life low-dose steroids in septic shock remains controversial due large ICU cohorts by Ferrer et al (91) (n = 2,786) and Casserly et to contradictory findings from different randomized trials al (91) (n = 8,992) did not find significant survival benefits with (86). Our research group hypothesized that the re-evaluation low-dose steroids in patients with septic shock, which further of three meta-analysis (87–89) done with the conventional confirmed our Bayesian results. Taken together, the efficacy and methods through the Bayesian lens would provide further safety findings from our combined Bayesian and conventional insight into this controversy (86). We replicated the methods analyses brought scientific clarity to the field and were highly (i.e., fixed-effect or random-effects modeling) and used the applicable to the bedside practice of critical care. Notably, the

A

2274

Low-dose Steroids for Severe Sepsis and Septic Shock (Annane et al)

www.ccmjournal.org

October 2014 • Volume 42 • Number 10

Concise Definitive Review

Bayesian methods discussed above are also an excellent fit for the performance of individual-patient meta-analyses (52), as well as for network meta-analyses (93, 94).

BAYESIAN APPLICATION TO CLINICAL PRACTICE GUIDELINES AND COMPARATIVEEFFECTIVENESS RESEARCH There is a growing body of literature that is already applying Bayesian methodology to guideline development (95–102). In addition, Diamond and Kaul (103) recently proposed a new methodology to report clinical practice guidelines based on Bayesian principles; they suggest that the use of this Bayesian approach could potentially reduce practice variation and increase the adherence of practice guidelines, which in turn would culminate in better clinical outcomes and lower healthcare cost. The growing application of Bayes in both metaanalyses and practice guidelines should propel this method to be used in comparative-effectiveness research. Berry et al (104) recently provided the basis and rationale for why Bayes technology is perfectly suited for comparative-effectiveness research. As a real-world example on comparative-effectiveness research, a recently published study by the American Heart Association Task Force on Practice Guidelines (99) demonstrated that percutaneous coronary intervention, similar to coronary artery bypass surgery, improved survival for patients with unprotected left main coronary artery disease compared with medical therapy. The application of Bayesian decision making to evidence-based medicine and public health has been comprehensively reviewed by Ashby and Smith (105) and Lilford and Braunholtz (106).

CONCLUSIONS The addition of the Bayesian approach to clinical trials in critical care medicine can realistically and timely lead to the better performance of trials of all phases, including the concrete benefits of faster trial completion with the reduction of unnecessary placebo exposure (positive trial) or treatment exposure (negative trial), fewer false-positive trials, and fewer false-positive findings in subgroup analyses. In addition, the Bayesian application can lead to more rapid pace from drug development to drug approval and improved translation of trial results into public health policies. Further, the bedside utilization of Bayesians posteriors can provide the ICU clinician with a more individualized application of trial results to determine the best course of action to treat critically ill patients. Both conventional and Bayesian approaches have their own strengths and weaknesses. Both approaches carry a substantial degree of subjectivity, implicitly with the conventional approach, and explicitly with the Bayesian approach. If either approach were fully objective as one would dream of, all medical statistics would be done by automated processes without the need for statisticians. This is not different from clinical medicine, which also cannot be fully objective; otherwise, we would not need physicians anymore, just computers to make perfect diagnoses and deliver the ideal treatments. Critical Care Medicine

What we propose in this article is that today we already have the clinical, mathematical, and computational skills to apply the Bayesian technology in conjunction with the conventional methodology to maximize the design, execution, and interpretation of clinical trials in the field of critical care. The failure to recognize the fact that we could be doing better clinical research in the critical care field is the same as the physician who provides a less than optimal treatment to a critically ill patient because he/she does not know enough about a known optimal treatment. This is inadmissible in today’s clinical medicine, so the same high standards should hold for clinical research. We hope that our comprehensive review and suggestions to improve clinical research in our field of critical care will spark the interest and excitement of clinical researchers with this versatile Bayesian world.

REFERENCES

1. Matthews RA: Bayesian statistical methods: What, why–and when. J Altern Complement Med 1998; 4:361–363 2. FDA. 2010. Available at: http://www.fda.gov/MedicalDevices/ DeviceRegulationandGuidance/GuidanceDocuments/ucm071072. htm. Accessed January 8, 2014 3. EMA. 2009. Available at: http://www.ema.europa.eu/docs/en_GB/ document_library/Scientific_guideline/2009/09/WC500003615. pdf. Accessed January 8, 2014 4. Adamina M, Tomlinson G, Guller U: Bayesian statistics in oncology: A guide for the clinical investigator. Cancer 2009; 115:5371–5381 5. Diamond GA, Kaul S: Prior convictions: Bayesian approaches to the analysis and interpretation of clinical megatrials. J Am Coll Cardiol 2004; 43:1929–1939 6. Sterne JA, Davey Smith G: Sifting the evidence-what’s wrong with significance tests? BMJ 2001; 322:226–231 7. Spiegelhalter DJ, Myles JP, Jones DR, et al: Methods in health service research. An introduction to Bayesian methods in health technology assessment. BMJ 1999; 319:508–512 8. Lewis RJ, Wears RL: An introduction to the Bayesian analysis of clinical trials. Ann Emerg Med 1993; 22:1328–1336 9. Goodman SN: Introduction to Bayesian methods I: Measuring the strength of evidence. Clin Trials 2005; 2:282–290; discussion 301–304, 364–378 10. Efron B: Mathematics. Bayes’ theorem in the 21st century. Science 2013; 340:1177–1178 11. Diamond GA, Forrester JS: Clinical trials and statistical verdicts: Probable grounds for appeal. Ann Intern Med 1983; 98:385–394 12. Ashby D: Bayesian statistics in medicine: A 25 year review. Stat Med 2006; 25:3589–3631 13. Goodman SN: Toward evidence-based medical statistics. 1: The P value fallacy. Ann Intern Med 1999; 130:995–1004 14. Austin PC, Brunner LJ, Hux JE: Bayeswatch: An overview of Bayesian statistics. J Eval Clin Pract 2002; 8:277–286 15. Browner WS, Newman TB: Are all significant P values created equal? The analogy between diagnostic tests and clinical research. JAMA 1987; 257:2459–2463 16. Moyé LA: Bayesians in clinical trials: Asleep at the switch. Stat Med 2008; 27:469–482; discussion 483 17. Gill CJ, Sabin L, Schmid CH: Why clinicians are natural Bayesians. BMJ 2005; 330:1080–1083 18. Johnson SR, Tomlinson GA, Hawker GA, et al: A valid and reliable belief elicitation method for Bayesian priors. J Clin Epidemiol 2010; 63:370–383 19. Thomas DC, Witte JS, Greenland S: Dissecting effects of complex mixtures: Who’s afraid of informative priors? Epidemiology 2007; 18:186–190 www.ccmjournal.org

2275

Kalil and Sun 20. Greenland S: Prior data for non-normal priors. Stat Med 2007; 26:3578–3590 21. Aickin M: Bayes without priors. J Clin Epidemiol 2004; 57:4–13 22. Greenland S: Bayesian perspectives for epidemiological research: I. Foundations and basic methods. Int J Epidemiol 2006; 35:765–775 23. Jennison C, Turnball BW: Group Sequential Methods With Applications to Clinical Trials. Boca Raton, FL, Chapman & Hall/CRC, 2000 24. Grossman J, Parmar MK, Spiegelhalter DJ, et al: A unified method for monitoring and analysing controlled trials. Stat Med 1994; 13:1815–1826 25. Burton PR, Gurrin LC, Campbell MJ: Clinical significance not statistical significance: A simple Bayesian alternative to p values. J Epidemiol Community Health 1998; 52:318–323 26. Berry DA: Bayesian clinical trials. Nat Rev Drug Discov 2006; 5:27–36 27. Kalil AC: Deciphering the sepsis riddle: We can learn from Star Trek. Crit Care Med 2013; 41:2458–2460 28. Dellinger RP, Levy MM, Rhodes A, et al; Surviving Sepsis Campaign Guidelines Committee including the Pediatric Subgroup: Surviving sepsis campaign: International guidelines for management of severe sepsis and septic shock: 2012. Crit Care Med 2013; 41:580–637 29. Ferrer R, Artigas A, Levy MM, et al; Edusepsis Study Group: Improvement in process of care and outcome after a multicenter severe sepsis educational program in Spain. JAMA 2008; 299:2294–2303 30. Fujishima S, Gando S, Saitoh D, et al; Japanese Association for Acute Medicine Sepsis Registry (JAAM SR) Study Group: A multicenter, prospective evaluation of quality of care and mortality in Japan based on the Surviving Sepsis Campaign guidelines. J Infect Chemother 2014; 20:115–120 31. Li ZQ, Xi XM, Luo X, et al: Implementing surviving sepsis campaign bundles in China: A prospective cohort study. Chin Med J (Engl) 2013; 126:1819–1825 32. Phua J, Ho BC, Tee A, et al: The impact of clinical protocols in the management of severe sepsis: A prospective cohort study. Anaesth Intensive Care 2012; 40:663–674 33. Shiramizo SC, Marra AR, Durão MS, et al: Decreasing mortality in severe sepsis and septic shock patients by implementing a sepsis bundle in a hospital setting. PLoS One 2011; 6:e26790 34. Vesteinsdottir E, Karason S, Sigurdsson SE, et al: Severe sepsis and septic shock: A prospective population-based study in Icelandic intensive care units. Acta Anaesthesiol Scand 2011; 55:722–731 35. Levy MM, Dellinger RP, Townsend SR, et al; Surviving Sepsis Campaign: The Surviving Sepsis Campaign: Results of an international guideline-based performance improvement program targeting severe sepsis. Crit Care Med 2010; 38:367–374 36. Noritomi DT, Ranzani OT, Monteiro MB, et al: Implementation of a multifaceted sepsis education program in an emerging country setting: Clinical outcomes and cost-effectiveness in a long-term followup study. Intensive Care Med 2014; 40:182–191 37. Kalil AC, Sun J: Why are clinicians not embracing the results from pivotal clinical trials in severe sepsis? A Bayesian analysis. PLoS One 2008; 3:e2291 38. Motulsky H: Intuitive Biostatistics: A Nonmathematical Guide to Statistical Thinking. Completely rev. Second Edition. New York, Oxford University Press, 2010 39. Kirkwood BR, Sterne JAC, Kirkwood BR: Essential Medical Statistics. Second Edition. Malden, MA, Blackwell Science, 2003 40. Rothman KJ, Greenland S, Lash TL: Modern Epidemiology. Third Edition. Philadelphia, PA, Wolters Kluwer Health/Lippincott Williams & Wilkins, 2008 41. Rivers E, Nguyen B, Havstad S, et al; Early Goal-Directed Therapy Collaborative Group: Early goal-directed therapy in the treatment of severe sepsis and septic shock. N Engl J Med 2001; 345:1368–1377 42. Lin SM, Huang CD, Lin HC, et al: A modified goal-directed protocol improves clinical outcomes in intensive care unit patients with septic shock: A randomized controlled trial. Shock 2006; 26:551–557 43. Benes J, Chytra I, Altmann P, et al: Intraoperative fluid optimization using stroke volume variation in high risk surgical patients: Results of prospective randomized study. Crit Care 2010; 14:R118 44. Cecconi M, Fasano N, Langiano N, et al: Goal-directed haemodynamic therapy during elective total hip arthroplasty under regional anaesthesia. Crit Care 2011; 15:R132

2276

www.ccmjournal.org

45. Donati A, Loggi S, Preiser JC, et al: Goal-directed intraoperative therapy reduces morbidity and length of hospital stay in high-risk surgical patients. Chest 2007; 132:1817–1824 46. Gan TJ, Soppitt A, Maroof M, et al: Goal-directed intraoperative fluid administration reduces length of hospital stay after major surgery. Anesthesiology 2002; 97:820–826 47. Jhanji S, Vivian-Smith A, Lucena-Amaro S, et al: Haemodynamic optimisation improves tissue microvascular flow and oxygenation after major surgery: A randomised controlled trial. Crit Care 2010; 14:R151 48. Lobo SM, Lobo FR, Polachini CA, et al: Prospective, randomized trial comparing fluids and dobutamine optimization of oxygen delivery in highrisk surgical patients [ISRCTN42445141]. Crit Care 2006; 10:R72 49. Hayes MA, Timmins AC, Yau EH, et al: Elevation of systemic oxygen delivery in the treatment of critically ill patients. N Engl J Med 1994; 330:1717–1722 50. Yu M, Levy MM, Smith P, et al: Effect of maximizing oxygen delivery on morbidity and mortality rates in critically ill patients: A prospective, randomized, controlled study. Crit Care Med 1993; 21:830–838 51. Alía I, Esteban A, Gordo F, et al: A randomized and controlled trial of the effect of treatment aimed at maximizing oxygen delivery in patients with severe sepsis or septic shock. Chest 1999; 115:453–461 52. Kalil AC: Wanted: Early goal-directed therapy for septic shock–dead or alive, but not critically ill! Intensive Care Med 2010; 36:1–3 53. ProCESS Investigators, Yealy DM, Kellum JA, Huang DT, et al: A randomized trial of protocol-based care for early septic shock. N Engl J Med 2014; 370:1683–1693 54. Chevret S: Bayesian adaptive clinical trials: A dream for statisticians only? Stat Med 2012; 31:1002–1013 55. Lee JJ, Chu CT: Bayesian clinical trials in action. Stat Med 2012; 31:2955–2972 56. Gamalo MA, Tiwari RC, LaVange LM: Bayesian approach to the design and analysis of non-inferiority trials for anti-infective products. Pharm Stat 2014; 13:25–40 57. O’Quigley J, Pepe M, Fisher L: Continual reassessment method: A practical design for phase 1 clinical trials in cancer. Biometrics 1990; 46:33–48 58. Cheung YK, Chappell R: Sequential designs for phase I clinical trials with late-onset toxicities. Biometrics 2000; 56:1177–1182 59. Babb J, Rogatko A, Zacks S: Cancer phase I clinical trials: Efficient dose escalation with overdose control. Stat Med 1998; 17:1103–1120 60. Desfrere L, Zohar S, Morville P, et al: Dose-finding study of ibuprofen in patent ductus arteriosus using the continual reassessment method. J Clin Pharm Ther 2005; 30:121–132 61. Krams M, Lees KR, Hacke W, et al; ASTIN Study Investigators: Acute Stroke Therapy by Inhibition of Neutrophils (ASTIN): An adaptive dose-response study of UK-279,276 in acute ischemic stroke. Stroke 2003; 34:2543–2548 62. Thall PF, Simon R: Incorporating historical control data in planning phase II clinical trials. Stat Med 1990; 9:215–228 63. Thall PF, Wathen JK, Bekele BN, et al: Hierarchical Bayesian approaches to phase II trials in diseases with multiple subtypes. Stat Med 2003; 22:763–780 64. Thall PF, Simon R: Practical Bayesian guidelines for phase IIB clinical trials. Biometrics 1994; 50:337–349 65. Schoenfeld DA, Hui Zheng, Finkelstein DM: Bayesian design using adult data to augment pediatric trials. Clin Trials 2009; 6:297–304 66. Goodman SN, Sladky JT: A Bayesian approach to randomized controlled trials in children utilizing information from adults: The case of Guillain-Barré syndrome. Clin Trials 2005; 2:305–310; discussion 364 67. Spiegelhalter D, Abrams, KR, Myles, JP: Bayesian Approaches to Clinical Trials and Health-Care Evaluation. Chichester, England, John Wiley & Sons, 2004 68. Berry DA: A case for Bayesianism in clinical trials. Stat Med 1993; 12:1377–1393 69. Trzaskoma B, Sashegyi A: Predictive probability of success and the assessment of futility in large outcomes trials. J Biopharm Stat 2007; 17:45–63 70. Berry DA: Adaptive clinical trials in oncology. Nat Rev Clin Oncol 2012; 9:199–207 October 2014 • Volume 42 • Number 10

Concise Definitive Review 71. Meurer WJ, Lewis RJ, Tagle D, et al: An overview of the adaptive designs accelerating promising trials into treatments (ADAPT-IT) project. Ann Emerg Med 2012; 60:451–457 72. Chow SC, Chang M: Adaptive design methods in clinical trials—A review. Orphanet J Rare Dis 2008; 3:11 73. Kairalla JA, Coffey CS, Thomann MA, et al: Adaptive trial designs: A review of barriers and opportunities. Trials 2012; 13:145 74. Gallo P, Chuang-Stein C, Dragalin V, et al: Adaptive designs in clinical drug development—An executive summary of the PhRMA Working Group. J Biopharm Stat 2006; 16:275–283; discussion 285–291, 293–298, 311–312 75. Simon N, Simon R: Adaptive enrichment designs for clinical trials. Biostatistics 2013; 14:613–625 76. Brennan P, Croft P: Interpreting the results of observational research: Chance is not such a fine thing. BMJ 1994; 309:727–730 77. Wijeysundera DN, Austin PC, Hux JE, et al: Bayesian statistical inference enhances the interpretation of contemporary randomized controlled trials. J Clin Epidemiol 2009; 62:13–21.e5 78. Kalil AC, Sun J: How many patients with severe sepsis are needed to confirm the efficacy of drotrecogin alfa activated? A Bayesian design. Intensive Care Med 2008; 34:1804–1811 79. Spiegelhalter DJ, Myles JP, Jones DR, et al: Bayesian methods in health technology assessment: A review. Health Technol Assess 2000; 4:1–130 80. BaSis. 2001. Available at: http://lib.stat.cmu.edu/bayesworkshop/2001/BaSis.html. Accessed January 8, 2014 81. Sung L, Hayden J, Greenberg ML, et al: Seven items were identified for inclusion when reporting a Bayesian analysis of a clinical study. J Clin Epidemiol 2005; 58:261–268 82. Copetti M, Fontana A, Graziano G, et al: Advances in meta-analysis: Examples from internal medicine to neurology. Neuroepidemiology 2014; 42:59–67 83. Sutton AJ, Higgins JP: Recent developments in meta-analysis. Stat Med 2008; 27:625–650 84. Kwok H, Lewis RJ: Bayesian hierarchical modeling and the integration of heterogeneous information on the effectiveness of cardiovascular therapies. Circ Cardiovasc Qual Outcomes 2011; 4:657–666 85. Sutton AJ, Abrams KR: Bayesian methods in meta-analysis and evidence synthesis. Stat Methods Med Res 2001; 10:277–303 86. Kalil AC, Sun J: Low-dose steroids for septic shock and severe sepsis: The use of Bayesian statistics to resolve clinical trial controversies. Intensive Care Med 2011; 37:420–429 87. Annane D, Bellissant E, Bollaert PE, et al: Corticosteroids in the treatment of severe sepsis and septic shock in adults: A systematic review. JAMA 2009; 301:2362–2375 88. Minneci PC, Deans KJ, Eichacker PQ, et al: The effects of steroids during sepsis depend on dose and severity of illness: An updated meta-analysis. Clin Microbiol Infect 2009; 15:308–318 8 9. Marik PE, Pastores SM, Annane D, et al; American College of Critical Care Medicine: Recommendations for the diagnosis and management of corticosteroid insufficiency in critically ill adult patients: Consensus statements from an international task force by the American College of Critical Care Medicine. Crit Care Med 2008; 36:1937–1949 90. Sprung CL, Annane D, Keh D, et al; CORTICUS Study Group: Hydrocortisone therapy for patients with septic shock. N Engl J Med 2008; 358:111–124

Critical Care Medicine

91. Ferrer R, Artigas A, Suarez D, et al; Edusepsis Study Group: Effectiveness of treatments for severe sepsis: A prospective, multicenter, observational study. Am J Respir Crit Care Med 2009; 180:861–866 92. Casserly B, Gerlach H, Phillips GS, et al: Low-dose steroids in adult septic shock: Results of the Surviving Sepsis Campaign. Intensive Care Med 2012; 38:1946–1954 93. Cipriani A, Higgins JP, Geddes JR, et al: Conceptual and technical challenges in network meta-analysis. Ann Intern Med 2013; 159:130–137 94. Mills EJ, Thorlund K, Ioannidis JP: Demystifying trial networks and network meta-analysis. BMJ 2013; 346:f2914 95. Zhang W, Doherty M, Peat G, et al: EULAR evidence-based recommendations for the diagnosis of knee osteoarthritis. Ann Rheum Dis 2010; 69:483–489 96. van Hulst LT, Kievit W, van Bommel R, et al: Rheumatoid arthritis patients and rheumatologists approach the decision to escalate care differently: Results of a maximum difference scaling experiment. Arthritis Care Res (Hoboken) 2011; 63:1407–1414 97. Lackner KJ, Peetz D, von Landenberg P: Revision of the Sapporo criteria for the antiphospholipid syndrome—Coming to grips with evidence and Thomas Bayes? Thromb Haemost 2006; 95:917–919 98. Bangalore S, Kumar S, Volodarskiy A, et al: Blood pressure targets in patients with coronary artery disease: Observations from traditional and Bayesian random effects meta-analysis of randomised trials. Heart 2013; 99:601–613 99. Bittl JA, He Y, Jacobs AK, et al; American College of Cardiology Foundation/American Heart Association Task Force on Practice Guidelines: Bayesian methods affirm the use of percutaneous coronary intervention to improve survival in patients with unprotected left main coronary artery disease. Circulation 2013; 127:2177–2185 100. Khoury MJ, Bertram L, Boffetta P, et al: Genome-wide association studies, field synopses, and the development of the knowledge base on genetic variation and human diseases. Am J Epidemiol 2009; 170:269–279 101. Ankerst DP, Till C, Boeck A, et al: The impact of prostate volume, number of biopsy cores and American Urological Association symptom score on the sensitivity of cancer detection using the Prostate Cancer Prevention Trial risk calculator. J Urol 2013; 190:70–76 102. Tappenden P, Chilcott J, Brennan A, et al: Using whole disease modeling to inform resource allocation decisions: Economic evaluation of a clinical guideline for colorectal cancer using a single model. Value Health 2013; 16:542–553 103. Diamond GA, Kaul S: Bayesian classification of clinical practice guidelines. Arch Intern Med 2009; 169:1431–1435 104. Berry SM, Ishak KJ, Luce BR, et al: Bayesian meta-analyses for comparative effectiveness and informing coverage decisions. Med Care 2010; 48:S137–S144 105. Ashby D, Smith AF: Evidence-based medicine as Bayesian decisionmaking. Stat Med 2000; 19:3291–3305 106. Lilford RJ, Braunholtz D: The statistical basis of public policy: A paradigm shift is overdue. BMJ 1996; 313:603–607

www.ccmjournal.org

2277

Bayesian methodology for the design and interpretation of clinical trials in critical care medicine: a primer for clinicians.

To review Bayesian methodology and its utility to clinical decision making and research in the critical care field...
586KB Sizes 4 Downloads 9 Views