1. Please check and comment entries here.
Table of Contents

    Topic review

    Adaptive Clinical Trials' Bayesian Design

    Subjects: Others
    View times: 29
    Submitted by:

    Definition

    Adaptive designs are attracting a keen interest in several disciplines, from a theoretical viewpoint and also—potentially—from a practical one, and Bayesian adaptive designs, in particular, have raised high expectations in clinical trials.

    1. Introduction

    This entry is a bird’s eye view of the recent literature on adaptive designs of clinical trials from a Bayesian perspective. Statistics plays a prominent role in the design as well as the analysis of the results of a clinical study and Bayesian ideas are well received by clinicians. In their book, Spiegelhalter and his coauthors [1] make a strong case in favour of Bayesian methods in health care, and in the last two decades Bayesian statistics has had a large impact in the medical field (see the superb review by Ashby [2]), the more so as its implementation gets easier thanks to better computational facilities. “Bayesian clinical trials: no more excuses” is the title of an editorial in Vol 6(3) of Clinical Trials [3]. The Bayesian approach has a good reputation at producing scientific openness and honesty.

    The Bayesian paradigm is especially appropriate at the planning stage of a clinical trial, when external information, such as historical data, findings from previous studies, and expert opinions, is often available and awaiting to be made the most of. As Donald Berry and his colleagues state in [4], we are all Bayesian at the design stage! Health authorities have issued important statements on the statistical, clinical and regulatory aspects of Bayesian clinical trials ([5][6]), recently allowing and even advocating the use of innovative methods, in particular adaptive design; as the editors of this Special Issue point out, most statistical and biomedical journals have recently hosted proposals of trial designs with a Bayesian slant, in some cases virtual re-executions of published trials. A search carried out in PubMed in August 2020 has returned nearly 300 publications (half of them published in the last decade) which either propose or use Bayesian adaptive methods in the design of a clinical trial. This may be also thanks to the popularization by Donald Berry [7][8][9][10][11] and the efforts made by statisticians working in the pharmaceutical industry, one of the main players in the design of clinical trials, to incorporate Bayesian methods. This is shown in leading journals in clinical trial methodology, like Pharmaceutical Statistics, The Journal of Biopharmaceutical Statistics or Biometrical Journal.

    Some confusion occasionally arises between the concepts of “Bayesian” and of “adaptive” design, because of similarities in the outlook: in the Bayesian paradigm, accrued data are used to update the prior distribution on the parameters, via Bayes’ Theorem, and in response-adaptive experiments the accrued data are used at each step, namely after each observation or predefined group of observations, to update the next planning decision. Either approach (Bayesian or adaptive) can stand on its own, and has developed independently of the other: we clarify this point later.

    We are interested in trial designs that are both Bayesian and adaptive. The data are recursively evaluated during the experiment: the posterior parameter distribution is recursively updated and used to modify the execution of the trial according to a previously established rule. The textbook by Berry, Carlin, Lee, and Muller [12] successfully illustrates Bayesian adaptive methods in clinical research and deals with design issues too. It goes almost without saying that randomization is a must in a clinical trial (for Bayesians too), to counteract several types of bias, for instance selection bias.

    2. Bayesian Adaptive Designs in Registered Trials

    Adaptive designs are mathematically sophisticated instruments. Their development is fairly recent, and the split that can be observed between theory and practice is not at all surprising. There are several obstacles—both technical and practical—to launching an adaptive trial, beyond the significant time and effort required by any clinical trial. Among other things, adaptive design requires updating information on accrued data, the speed of acquisition may be highly variable so there is the need to identify short-term endpoints that can be used to accurately predict treatment responses such as long-term mortality in terms of a gold-standard endpoint. The steps required to establish this type of design in a novel context are indeed fairly complex, as some case studies show (see for instance Mason et al. [13]. As to the Bayesian approach, this may include specialized software programs to run the study design, only made possible by recent advancements in computational algorithms and computer hardware ([14]).

    Nevertheless, it is worth remarking that the philosophy of Bayesian adaptive designs has already made its way into the clinic. They are now fairly well established in cancer research ([10]), and to a lesser extent, in other clinical areas. As well as single study designs, Bayesian adaptive methods are being employed to build “platform” designs (Adaptive Platform Trials). These are trials for simultaneous testing of multiple treatment strategies in separate groups, with plans to discontinue any group that is definitively inferior at planned interim analyses. Trial patients are enrolled in a continuous manner via a common master protocol, with interventions entering and leaving the platform on the basis of a predefined decision algorithm. Several Adaptive Platform Trials are now funded in various disease areas (see Angus et al. [15], Brown et al. [16] and Talisa et al. [17] for a discussion).

    The following is a non-exhaustive list of recent or still on-going clinical trials that incorporate Bayesian adaptive design features:

    • The Randomized Embedded Multifactorial Adaptive Platform Trial in Community Acquired Pneumonia (REMAP-CAP): see [18]. It has set-up a sub-platform called “REMAP−COVID” on which the evaluation of specific treatments for COVID-19 is run.
    • Anti-Thrombotic Therapy to Ameliorate Complications of COVID-19 (ATTACC) (see [19]), similar in purpose to RECAP-COVID.
    • GBM AGILE, an adaptive clinical trial to deliver improved treatments for glioblastoma, now open and enrolling patients ([20]).
    • STURDY, a randomized clinical trial of Vitamin D supplement doses for the prevention of falls in older adults ([21]).
    • The SPRINT trial on safety and efficacy of neublastin in painful lumbosacral radiculopathy ([22]).
    • SARC009: A Phase II study in patients with previously treated, high-grade, advanced sarcoma ([23]).
    • The SHINE clinical trial for hyperglycaemia in stroke patients ([24][25]).
    • The EPAD project in neurology ([26]).
    • The BATTLE and BATTLE-2 trials for lung cancer ([27][28]).
    • The I-SPY 2 platform for breast cancer chemotherapy ([29]; see also [30][31][32]).
    • A study on Lemborexant, for the treatment of insomnia disorder ([33]).
    • A Phase I non-randomized trial of a combination therapy in patients with pancreatic adenocarcinoma ([34]).
    • A first-in-human study of RG7342 for the treatment of schizophrenia in healthy male subjects ([35]).
    • A newly started Phase II trial in Japan for sarcoma ([36]) also shows the utility of a Bayesian adaptive design.
    • A Bayesian response-adaptive trial in tuberculosis is the endTB trial ([37]).
    • Acute Stroke Therapy by Inhibition of Neutrophils (ASTIN) was a Bayesian adaptive phase 2 dose-response study to establish whether UK-279,276 improves recovery in acute ischemic stroke. The adaptive design facilitated early termination for futility ([38]).

    3. Conclusions

    “Bayesian adaptive clinical trials: a dream for statisticians only?” asks Chevret [39]. Clearly, Bayesian adaptive experiments are not easy to design, let alone to implement. For a start, elicitation of a prior is not a simple matter. In clinical trials it is generally assumed to be based on historical data. In their book ([1]) Spiegelhalter, Abrams and Myles recommend attempting both an “enthusiastic” and a “skeptical” prior. On the other hand, Bayesian statistics exercises greater appeal than frequentist on most applied researchers, and the same can be said of adaptive design rules. This explains why the presence of Bayesian and adaptive design methods combined together has become massive in the biostatistical literature, notwithstanding the fact that adaptive algorithms are more complex than non-adaptive.

    It is this author’s opinion that although there is a widespread consensus that the Bayesian and the adaptive approaches to design go very well together, the field is still rather fragmented. The development has taken place in a relatively short time and Bayesian adaptive designs are still awaiting in-depth investigation. It is a sad state of affairs that in general there is no sounder way to evaluate the performance of Bayesian (and non-Bayesian) designs other than by computer simulations. Often the simulation scenarios are chosen on the basis of the researchers’ personal preferences, so the conclusions may be debatable.

    The book by Yin [40] is a thorough presentation of both Bayesian and frequentist adaptive methods in clinical trial design, but the two approaches are based on fundamentally different paradigms and a comparison of Bayesian and non-Bayesian designs is possible only in restricted cases. As an example, when several experimental treatments are available for testing, Wason and Trippa [41] compare Bayesian adaptive randomization, which allocates a greater proportion of future patients to treatments that have performed well, to multi-arm multi-stage designs, which use pre-specified stopping boundaries to determine whether experimental treatments should be dropped. The authors show that in this case both are efficient, but neither is superior: it depends on the true state of nature.

    In conclusion, it is worth quoting the words of Stallard et al. [42]: “Bayesian adaptive methods are often more bespoke than frequentist approaches… They require more design work than the use of a more standard frequentist method but can be advantageous in that design choices and their consequences are considered carefully”.

    This entry is adapted from 10.3390/ijerph18020530

    References

    1. Spiegelhalter, D.J.; Abrams, K.R.; Myles, J.P. Bayesian Approaches to Clinical Trials and Health-Care Evaluation; John Wiley & Sons: Hoboken, NJ, USA, 2004.
    2. Ashby, D. Bayesian statistics in medicine: A 25 year review. Stat. Med. 2006, 25, 3589–3631.
    3. EDITORIAL Bayesian clinical trials: No more excuses. Clin. Trials 2009, 6, 203–204.
    4. Biswas, S.; Liu, D.D.; Lee, J.J.; Berry, D.A. Bayesian clinical trials at the University of Texas M. D. Anderson Cancer Center. Clin. Trials 2009, 6, 205–216.
    5. European Medicines Agency. Reflection Paper on Methodological Issues in Confirmatory Clinical Trials Planned with an Adaptive Design; European Medicines Agency: Amsterdam, The Netherlands, 2007; CHMP/EWP/2459/02.
    6. Food and Drug Administration, USA. Adaptive Designs for Clinical Trials of Drugs and Biologics Guidance for Industry; Docket Number: FDA-2018-D-3124; U.S. Food and Drug Administration: Silver Spring, MD, USA, 2019.
    7. Berry, D.A.; Müller, P.; Grieve, A.; Smith, M.; Parke, T.; Blazek, R.; Mitchard, N.; Krams, M. Adaptive Bayesian designs for dose-ranging drug trials. In Case Studies in Bayesian Statistics; Gatsonis, C., Kass, R.E., Carlin, B., Carriquiry, A., Gelman, A., Verdinelli, I., West, M., Eds.; Springer: New York, NY, USA, 2002; Volume 162, pp. 99–181.
    8. Berry, D.A. Statistical innovations in cancer research. In Cancer Medicine, 6th ed.; Holland, J., Frei, T., Eds.; BC Decker: London, UK, 2003; pp. 465–478.
    9. Berry, D.A. Bayesian clinical trials. Nat. Rev. Drug Discov. 2006, 5, 27–36.
    10. Berry, D.A. Adaptive clinical trials in oncology. Nat. Rev. Clin. Oncol. 2012, 9, 199–207.
    11. Stangl, D.; Lurdes, Y.T.I.; Telba, Z.I. Celebrating 70: An interview with Don Berry. Stat. Sci. 2012, 27, 144–159.
    12. Berry, S.M.; Carlin, B.P.; Lee, J.J.; Muller, P. Bayesian Adaptive Methods for Clinical Trials; Chapman & Hall/CRC Biostatistics Series; Chapman and Hall/CRC: Boca Raton, FL, USA, 2011.
    13. Mason, A.J.; Gonzalez-Maffe, J.; Quinn, K.; Doyle, N.; Legg, K.; Norsworthy, P.; Trevelion, R.; Winston, A.; Ashby, D. Developing a Bayesian adaptive design for a Phase I clinical trial: A case study for a novel HIV treatment. Stat. Med. 2017, 36, 754–771.
    14. Lee, J.J.; Chu, C.T. Bayesian clinical trials in action. Stat. Med. 2012, 31, 2955–2972.
    15. Angus, D.C.; Alexander, B.M.; Berry, S.; Buxton, M.; Lewis, R.; Paoloni, M. Adaptive platform trials, definition, design, conduct and reporting considerations. Nat. Rev. Drug Discov. 2019, 18, 797–807.
    16. Brown, A.R.; Gajewski, B.J.; Aaronson, L.S.; Mudaranthakam, D.P.; Hunt, S.L.; Berry, S.M.; Quintana, M.; Pasnoor, M.; Dimachkie, M.M.; Jawdat, O.; et al. A Bayesian comparative effectiveness trial in action: Developing a platform for multisite study adaptive randomization. Trials 2016, 17, 428.
    17. Talisa, V.B.; Yende, S.; Seymour, C.W.; Angus, D.C. Arguing for adaptive clinical trials in sepsis. Front. Immunol. 2018, 9, 1502.
    18. Angus, D.C.; Berry, S.; Lewis, R.J.; Al-Beidh, F.; Arabi, Y.; van Bentum-Puijk, W.; Bhimani, Z.; Bonten, M.; Broglio, K.; Brunkhorst, F.; et al. The Randomized Embedded Multifactorial Adaptive Platform for Community-Acquired Pneumonia (REMAP-CAP) study: Rationale and design. Ann. Am. Thorac. Soc. 2020.
    19. Houston, B.L.; Lawler, P.R.; Goligher, E.C.; Farkouh, M.E.; Bradbury, C.; Carrier, M.; Zarychanski, R. Anti-Thrombotic Therapy to Ameliorate Complications of COVID-19 (ATTACC): Study design and methodology for an international, adaptive Bayesian randomized controlled trial. Clin. Trials 2020, 17, 491–500.
    20. Alexander, B.M.; Ba, S.; Berger, M.S.; Berry, D.A.; Cavenee, W.K.; Chang, S.M.; Cloughesy, T.F.; Jiang, T.; Khasraw, M.; Li, W.; et al. Adaptive global innovative learning environment for glioblastoma. Clin. Cancer Res. 2018, 24, 737–743.
    21. Michos, E.D.; Mitchell, C.M.; Miller, E.R., 3rd; Sternberg, A.L.; Juraschek, S.P.; Schrack, J.A.; Szanton, S.L.; Walston, J.D.; Kalyani, R.R.; Plante, T.B.; et al. Rationale and design of the Study to Understand Fall Reduction and Vitamin D in You (STURDY): A randomized clinical trial of Vitamin D supplement doses for the prevention of falls in older adults. Contemp. Clin. Trials. 2018, 73, 111–122.
    22. Backonja, M.; Williams, L.; Miao, X.; Katz, N.; Chen, C. Safety and efficacy of neublastin in painful lumbosacral radiculopathy: A randomized, double-blinded, placebo-controlled phase 2 trial using Bayesian adaptive design (the SPRINT trial). Pain 2017, 158, 1802–1812.
    23. Schuetze, S.M.; Wathen, J.K.; Lucas, D.R.; Choy, E.; Samuels, B.L.; Staddon, A.P.; Ganjoo, K.N.; von Mehren, M.; Chow, W.A.; Loeb, D.M.; et al. SARC009: Phase 2 study of Dasatinib in patients with previously treated, high-grade, advanced sarcoma. Cancer 2016, 122, 868–874.
    24. Connor, J.T.; Broglio, K.R.; Durkalski, V.; Meurer, W.J.; Johnston, K.C. The Stroke Hyperglycemia Insulin Network Effort (SHINE) trial: An adaptive trial design case study. Trials 2015, 16, 72.
    25. Johnston, K.; Bruno, A.; Pauls, Q.; Hall, C.E.; Barrett, K.M.; Barsan, W.; Fansler, A.; Van de Bruinhorst, K.; Janis, S.; Durkalski-Mauldin, V.L. for the Neurological Emergencies Treatment Trials Network and the SHINE Trial Investigators. Intensive vs standard treatment of hyperglycemia and functional outcome in patients with acute ischemic stroke: The SHINE randomized clinical trial. J. Am. Med. Assoc. 2019, 322, 326–335.
    26. Ritchie, C.W.; Molinuevo, J.L.; Truyen, L.; Satlin, A.; van der Geyten, S.; Lovestone, S. Development of interventions for the secondary prevention of Alzheimer’s dementia: The European Prevention of Alzheimer’s Dementia. Lancet Psych. 2015, 3, 179–186.
    27. Kim, E.S.; Herbst, R.S.; Wistuba, I.I.; Lee, J.J.; Blumenschein, G.R.; Tsao, A.; Stewart, D.J.; Hicks, M.E.; Erasmus, J., Jr.; Gupta, S.; et al. The BATTLE trial: Personalizing therapy for lung cancer. Cancer Discov. 2011, 1, 44–53.
    28. Papadimitrakopoulou, V.; Lee, J.J.; Wistuba, I.; Tsao, A.; Fossella, F.; Kalhor, N.; Gupta, S.; Byers, L.A.; Izzo, J.; Gettinger, S.; et al. The BATTLE-2 Study: A biomarker-integrated targeted therapy study in previously treated patients with advanced non-small-cell lung cancer. J. Clin. Oncol. 2016, 34, 3638–3647.
    29. Barker, A.D.; Sigman, C.C.; Kelloff, G.J.; Hylton, N.M.; Berry, D.A.; Esserman, L.J. I-SPY 2: An adaptive breast cancer trial design in the setting of neoadjuvant chemotherapy. Clin. Pharmacol. Ther. 2009, 86, 97–100.
    30. Carey, L.A.; Winer, E.P. I-SPY 2: Toward more rapid progress in breast cancer treatment. N. Eng. J. Med. 2016, 375, 83–84.
    31. Park, J.W.; Liu, M.C.; Yee, D.; Yau, C.; van’t Veer, L.J.; Symmans, W.F.; Paoloni, M.; Perlmutter, J.; Hylton, N.M.; Hogarth, M.; et al. for the I-SPY 2 Investigators. Adaptive randomization of Neratinib in early breast cancer. N. Eng. J. Med. 2016, 375, 11–22.
    32. Rugo, H.S.; Olopade, O.I.; DeMichele, A.; Yau, C.; van’t Veer, L.J.; Buxton, M.B.; Hogarth, M.; Hylton, N.M.; Paoloni, M.; Perlmutter, J.; et al. Adaptive randomization of Veliparib-Carboplatin treatment in breast cancer. N. Engl. J. Med. 2016, 375, 23–34.
    33. Murphy, P.; Moline, M.; Mayleben, D.; Rosenberg, R.; Zammit, G.; Pinner, K.; Dhadda, S.; Hong, Q.; Giorgi, L.; Satlin, A.; et al. Dual Orexin Receptor Antagonist (DORA) for the treatment of insomnia disorder: Results from a Bayesian, adaptive, randomized, double-blind, placebo-controlled study. J. Clin. Sleep Med. 2017, 13, 1289–1299.
    34. Cook, N.; Basu, B.; Smith, D.M.; Gopinathan, A.; Evans, J.; Steward, W.P.; Palmer, D.; Propper, D.; Venugopal, B.; Hategan, M.; et al. A Phase I trial of the γ-secretase inhibitor MK-0752 in combination with Gemcitabine in patients with pancreatic ductal adenocarcinoma. Br. J. Cancer 2018, 118, 793–801.
    35. Sturm, S.; Delporte, M.L.; Hadi, S.; Schobel, S.; Lindemann, L.; Weikert, R.; Jaeschke, G.; Derks, M.; Palermo, G. Results and evaluation of a first-in-human study of RG7342, an mGlu5 positive allosteric modulator, utilizing Bayesian adaptive methods. Br. J. Clin. Pharmacol. 2018, 84, 445–455.
    36. Hirakawa, A.; Nishikawa, T.; Yonemori, K.; Shibata, T.; Nakamura, K.; Ando, M.; Ueda, T.; Ozaki, T.; Tamura, K.; Kawai, A.; et al. Utility of Bayesian single-arm design in new drug application for rare cancers in Japan: A case study of Phase 2 trial for sarcoma. Ther. Innov. Regul. Sci. 2018, 52, 334–338.
    37. Cellamare, M.; Ventz, S.; Baudin, E.; Mitnick, C.D.; Trippa, L. A Bayesian response-adaptive trial in tuberculosis: The end TB trial. Clin. Trials 2017, 14, 17–28.
    38. Krams, M.; Lees, K.R.; Hacke, W.; Grieve, A.P.; Orgogozo, J.M.; Ford, G.A. Acute Stroke Therapy by Inhibition of Neutrophils (ASTIN). An adaptive dose-response study of UK-279,276 in acute ischemic stroke. Stroke 2003, 34, 2543–2548.
    39. Chevret, S. Bayesian adaptive clinical trials: A dream for statisticians only? Stat. Med. 2012, 31, 1002–1013.
    40. Yin, G. Clinical Trial Design: Bayesian and Frequentist Adaptive Methods; Wiley & Sons: Hoboken, NJ, USA, 2012; 368p.
    41. Wason, J.M.S.; Trippa, L. A comparison of Bayesian adaptive randomization and multi-stage designs for multi-arm clinical trials. Stat. Med. 2014, 33, 2206–2221.
    42. Stallard, N.; Todd, S.; Ryan, E.G.; Gates, S. Comparison of Bayesian and frequentist group-sequential clinical trial designs. BMC Med. Res. Methodol. 2020, 20, 4.
    More