On the Existence of Constant Accrual Rates in Clinical Trials and Direction for Future Research

Similar documents
Monitoring Accrual and Events in a Time-to-Event Endpoint Trial. BASS November 2, 2015 Jeff Palmer

**BEGINNING OF EXAMINATION** A random sample of five observations from a population is:

Extracting Information from the Markets: A Bayesian Approach

Exam 2 Spring 2015 Statistics for Applications 4/9/2015

An Improved Skewness Measure

Appendix A: Introduction to Queueing Theory

Omitted Variables Bias in Regime-Switching Models with Slope-Constrained Estimators: Evidence from Monte Carlo Simulations

Equity, Vacancy, and Time to Sale in Real Estate.

Application of Conditional Autoregressive Value at Risk Model to Kenyan Stocks: A Comparative Study

Does Calendar Time Portfolio Approach Really Lack Power?

Section 3.1: Discrete Event Simulation

Bootstrap Inference for Multiple Imputation Under Uncongeniality

Technical Appendices to Extracting Summary Piles from Sorting Task Data

Valuation of Exit Strategy under Decaying Abandonment Value

CS 361: Probability & Statistics

List of Examples. Chapter 1

Superiority by a Margin Tests for the Ratio of Two Proportions

Sample Size for Assessing Agreement between Two Methods of Measurement by Bland Altman Method

ME3620. Theory of Engineering Experimentation. Spring Chapter III. Random Variables and Probability Distributions.

Monitoring Processes with Highly Censored Data

1.010 Uncertainty in Engineering Fall 2008

Volatility Forecasting and Interpolation

The mathematical model of portfolio optimal size (Tehran exchange market)

Real-Options Analysis: A Luxury-Condo Building in Old-Montreal

BAYESIAN MAINTENANCE POLICIES DURING A WARRANTY PERIOD

Term Par Swap Rate Term Par Swap Rate 2Y 2.70% 15Y 4.80% 5Y 3.60% 20Y 4.80% 10Y 4.60% 25Y 4.75%

Obsolescence Risk and the Systematic Destruction of Wealth

Improving the accuracy of estimates for complex sampling in auditing 1.

[D7] PROBABILITY DISTRIBUTION OF OUTSTANDING LIABILITY FROM INDIVIDUAL PAYMENTS DATA Contributed by T S Wright

GENERATION OF STANDARD NORMAL RANDOM NUMBERS. Naveen Kumar Boiroju and M. Krishna Reddy

Luke and Jen Smith. MONTE CARLO ANALYSIS November 24, 2014

A Bayesian Implementation of the Standard Optimal Hedging Model: Parameter Estimation Risk and Subjective Views

A Model to Quantify the Return On Information Assurance

Time Observations Time Period, t

Window Width Selection for L 2 Adjusted Quantile Regression

Journal of Insurance and Financial Management, Vol. 1, Issue 4 (2016)

To be two or not be two, that is a LOGISTIC question

EXAMINATION OF THE THRESHOLD FOR THE TO COMPLETE INDEXES By Walt Lipke, PMI Oklahoma City Chapter

Recovery Risk: Application of the Latent Competing Risks Model to Non-performing Loans

M249 Diagnostic Quiz

Survival Analysis Employed in Predicting Corporate Failure: A Forecasting Model Proposal

Fixed-Income Securities Lecture 5: Tools from Option Pricing

Review for Final Exam Spring 2014 Jeremy Orloff and Jonathan Bloom

Optimal rebalancing of portfolios with transaction costs assuming constant risk aversion

Technical Appendix: Policy Uncertainty and Aggregate Fluctuations.

This item is the archived peer-reviewed author-version of:

Corporate Investment and Portfolio Returns in Japan: A Markov Switching Approach

Computational Statistics Handbook with MATLAB

(11) Case Studies: Adaptive clinical trials. ST440/540: Applied Bayesian Analysis

PARAMETRIC AND NON-PARAMETRIC BOOTSTRAP: A SIMULATION STUDY FOR A LINEAR REGRESSION WITH RESIDUALS FROM A MIXTURE OF LAPLACE DISTRIBUTIONS

Machine Learning for Quantitative Finance

Stochastic Analysis Of Long Term Multiple-Decrement Contracts

APPROXIMATING FREE EXERCISE BOUNDARIES FOR AMERICAN-STYLE OPTIONS USING SIMULATION AND OPTIMIZATION. Barry R. Cobb John M. Charnes

MAINTAINABILITY DATA DECISION METHODOLOGY (MDDM)

An Approximation Algorithm for Capacity Allocation over a Single Flight Leg with Fare-Locking

Ipsos Poll Conducted for Reuters Core Political Approval

Estimation after Model Selection

Measuring the Amount of Asymmetric Information in the Foreign Exchange Market

Properly Assessing Diagnostic Credit in Safety Instrumented Functions Operating in High Demand Mode

Probability Models.S2 Discrete Random Variables

An Improved Saddlepoint Approximation Based on the Negative Binomial Distribution for the General Birth Process

FURTHER ASPECTS OF GAMBLING WITH THE KELLY CRITERION. We consider two aspects of gambling with the Kelly criterion. First, we show that for

The Fundamentals of Reserve Variability: From Methods to Models Central States Actuarial Forum August 26-27, 2010

Two-Sample Z-Tests Assuming Equal Variance

Assembly systems with non-exponential machines: Throughput and bottlenecks

WC-5 Just How Credible Is That Employer? Exploring GLMs and Multilevel Modeling for NCCI s Excess Loss Factor Methodology

Appendix A. Selecting and Using Probability Distributions. In this appendix

THE PENNSYLVANIA STATE UNIVERSITY SCHREYER HONORS COLLEGE DEPARTMENT OF FINANCE

Dealing with forecast uncertainty in inventory models

An Introduction to Bayesian Inference and MCMC Methods for Capture-Recapture

3.4 Copula approach for modeling default dependency. Two aspects of modeling the default times of several obligors

yuimagui: A graphical user interface for the yuima package. User Guide yuimagui v1.0

Artificially Intelligent Forecasting of Stock Market Indexes

PROJECT 73 TRACK D: EXPECTED USEFUL LIFE (EUL) ESTIMATION FOR AIR-CONDITIONING EQUIPMENT FROM CURRENT AGE DISTRIBUTION, RESULTS TO DATE

TABLE OF CONTENTS - VOLUME 2

The Multinomial Logit Model Revisited: A Semiparametric Approach in Discrete Choice Analysis

Simulating the Need of Working Capital for Decision Making in Investments

A Comparative Study of Various Forecasting Techniques in Predicting. BSE S&P Sensex

FISHER TOTAL FACTOR PRODUCTIVITY INDEX FOR TIME SERIES DATA WITH UNKNOWN PRICES. Thanh Ngo ψ School of Aviation, Massey University, New Zealand

An Analysis of a Dynamic Application of Black-Scholes in Option Trading

Breakeven holding periods for tax advantaged savings accounts with early withdrawal penalties

Subject CS1 Actuarial Statistics 1 Core Principles. Syllabus. for the 2019 exams. 1 June 2018

Probability Theory and Simulation Methods. April 9th, Lecture 20: Special distributions

Performance risk evaluation of long term infrastructure projects (PPP-BOT projects) using probabilistic methods

,,, be any other strategy for selling items. It yields no more revenue than, based on the

THE CHANGING SIZE DISTRIBUTION OF U.S. TRADE UNIONS AND ITS DESCRIPTION BY PARETO S DISTRIBUTION. John Pencavel. Mainz, June 2012

Assessing volatility and credibility of experience a comparison of approaches

Posterior Inference. , where should we start? Consider the following computational procedure: 1. draw samples. 2. convert. 3. compute properties

Department of Statistics, University of Regensburg, Germany

Probability is the tool used for anticipating what the distribution of data should look like under a given model.

Estimating A Smooth Term Structure of Interest Rates

On the investment}uncertainty relationship in a real options model

Statistical and Machine Learning Approach in Forex Prediction Based on Empirical Data

WORKING PAPER. The Option Value of Delay in Health Technology Assessment (2006/06) CENTRE FOR APPLIED ECONOMIC RESEARCH. By S. Eckermann and A.

On modelling of electricity spot price

Objective calibration of the Bayesian CRM. Ken Cheung Department of Biostatistics, Columbia University

INSTITUTE AND FACULTY OF ACTUARIES. Curriculum 2019 SPECIMEN EXAMINATION

Probability Distributions II

I Bonds versus TIPS: should individual investors prefer one to the other?

Forecasting Life Expectancy in an International Context

Transcription:

University of Kansas From the SelectedWorks of Byron J Gajewski Summer June 15, 2012 On the Existence of Constant Accrual Rates in Clinical Trials and Direction for Future Research Byron J Gajewski, University of Kansas Medical Center Stephen D Simon Susan E Carslon, University of Kansas Medical Center Available at: https://works.bepress.com/byron_gajewski/21/

International Journal of Statistics and Probability; Vol. 1, No. 2; 2012 ISSN 1927-7032 E-ISSN 1927-7040 Published by Canadian Center of Science and Education On the Existence of Constant Accrual Rates in Clinical Trials and Direction for Future Research Byron J. Gajewski 1, Stephen D. Simon 2 & Susan E. Carlson 3 1 Department of Biostatistics, The University of Kansas Medical Center, Kansas City, USA 2 Department of Biomedical and Health Informatics, University of Missouri-Kansas City, Kansas City, USA 3 Department of Dietetics and Nutrition, The University of Kansas Medical Center, Kansas City, USA Correspondence: Byron J. Gajewski, Department of Biostatistics, The University of Kansas Medical Center, 3901 Rainbow Blvd., Kansas City, KS 66160, USA. Tel: 1-913-588-1603. E-mail: bgajewski@kumc.edu Received: April 6, 2012 Accepted: April 20, 2012 Online Published: June 15, 2012 doi:10.5539/ijsp.v1n2p43 URL: http://dx.doi.org/10.5539/ijsp.v1n2p43 This work was supported in part by DHA Supplementation and Pregnancy Outcomes 1R01 HD047315 (BJG & SEC) and Kansas Frontiers: The Heartland Institute for Clinical and Translational Research CTSA UL1RR033179 (BJG). The contents are solely the responsibility of the authors and do not necessarily represent the official views of the NIH Abstract Many clinical trials fall short of their accrual goals. This can be avoided with accurate accrual prediction tools. Past researchers provide important methodological alternative models for predicting accrual in clinical trials. One model allows for slow accrual at the start of the study, which eventually reaches a threshold. A simpler model assumes a constant rate of accrual. A comparison has been attempted but we wish to point out some important considerations when comparing these two models. In fact, we can examine the reasonableness of a constant accrual assumption (simpler model) which had data 239 days into a three- year study. We can now update that and report accumulated from the full three years of accrual data and we can demonstrate that constant accrual rate assumption was met in this particular study. We will use this report to frame future research in the area of accrual prediction. Keywords: prior elicitation, exponential, inverse gamma, Bayesian, sample size 1. Introduction Zhang and Long (2010) provide an important methodological contribution to the literature for predicting accrual in clinical trials. They accurately describe their effort as an extension of Gajewski, Simon and Carlson (2008). An important parallel result was provided by Anisimov and Fedorov (2007), and was derived and published independently. The model developed by Zhang and Long allows for slow accrual at the start of the study, which eventually reaches a threshold. The Gajewski et al. model is simpler in that it assumes a constant rate of accrual. Zhang and Long compare their methodology to Gajewski et al., but we wish to extend some important considerations when comparing these two models. Zhang and Long assert that in most real trial situations, the constant accrual rate assumption does not hold. We have found evidence to the contrary. In fact, we can examine the reasonableness of a constant accrual assumption using data cited in Gajewski et al. which had data 239 days into a three-year study. We can now update that and report accumulated from the full three years of accrual data and we can demonstrate that constant accrual rate assumption was met in this particular study. We will use this report to frame future simulations in the area of accrual prediction. 2. Review of Gajewski et al. Model & Bayesian Runs Test Before reporting the prediction results, we will review the model in Gajewski et al. (constant accrual). We also report here a new Bayesian runs test that we claim is an important diagnostic that should be computed for any accrual problem. 43

2.1 Review of Model We wish to predict accrual after accruing m patients. Let w 1, w 2,...,w m represent the gap in time that each new patient is accrued. The goal of the accrual monitoring process is to develop a model for the yet to be observed waiting times W m+1, W m+2,...,w n, where n is the actual patient accrual at the end of the trial. We assume that w i θ exp(1/θ) where exp( ) is the exponential distribution and E(w i ) = θ. In Gajewski et al. two priors were proposed: a flat prior and an informative prior. These were respectively θ IG(k = 0, V = 0) and θ IG(k = 175, V = 1.5), where IG( ) is the inverse gamma distribution. The 175 and 1.5 comes from answering two questions: (1) How long will it take to accrue n subjects? (2) On a scale of 1-10, how confident are in your answer to (1)? The answer to (1) provides T and the answer to (2)/10 provides P. In Gajewski et al. we have T=3 years and P=0.5. We arrive at our informative prior since k = np and V = TP(the flat sets P = 0). This conjugate prior results in alternative posteriors (flat- and informative-based) θ w IG(m, t m ) and θ w IG(175 + m, 1.5 + t m ), where t m = m w i represents the time the last patient was accrued. i=1 2.2. Review of Prediction Algorithm The overall goal is to predict n with m gap times. First we predict the n m data W m+1,...,w n. To achieve this, first we randomly select θ 1 from the posterior distribution and then randomly select waiting time n m random variables from W m+1,1,...,w n,1 from an exponential distribution with parameter θ 1. This process is repeated for θ 2,θ 3,...,θ b. The sum of observed and simulated waiting times, S b(n) = w 1 + w 2 + + w m + W m+1,b + + W n,b represents b estimates of the total duration of the clinical trial of size n. However, n is the unknown, so we use this process to obtain a posterior predictive sample size (n p ). Let T represent the time point at which the study ends (for the purposes here T = 3 years). We then compute partial sums S b(m+1), S b(m+2),... until the partial sum exceeds T. The values nb P which represent the largest values where the partial sums do not exceed T, provides a realization of the predictive distribution of sample sizes. Replication of this process provides the posterior distribution n P.In this paper we will use observations in 1/12 year increments to explore the cross validated prediction of the true accrual (n = 265) for T = 3 years of accrual. 2.3. New Bayesian Runs Test A Bayesian runs test, motivated by (Gelman, 2004, Chapter 6), tests the assumption of independence and identical distribution. This test is performed using all n = 265 gap data points. First, the number of runs of the observed gap data (w 1, w 2,..., w m ) relative to posterior mean (θ b ) is calculated. This is repeated for posterior predictive gap data (W 1,b,..., W n,b ) and posterior mean (θ b ). 3. Results of Prediction The probability of observed runs larger than predictive runs is 0.3986, suggesting independent and identically distributed gap data. A graphical examination of the accrual data (Figure 1) supports the use of exponential waiting Figure 1. Probability plot for Exponential distribution fit of the gap data after three years 44

times rather than a more complex waiting time distribution. We evaluate the prediction accuracy using the expected absolute deviation from the true accrual (n = 265), E( n p 265 ). Figure 2 displays the monthly prediction across 36 months using a non-informative prediction and an informative prediction. The first column displays the true three-year accrual (n = 265) and the point estimate with 95% prediction intervals using only the data up to that point. We can see that the informative prior does much better than the flat prior early on. Past the two-year point the flat and informative versions essentially agree. The second column displays the error across time as measured by E( n p 265 ). This can be described in terms of error %= E( n p 265 )/265. Early in the process (first year) the error for the flat prior is above 20% (20-60%) whereas during that same timeframe the informative prior is always less than 20%. The true a prior defined informative simple prediction model (Exponential) was extremely useful for prediction in this clinical trial. Figure 2. Monthly prediction across 36 months using a non-informative prediction with 95% intervals 4. Direction for Future Research Our experience is that a constant rate of accrual seems quite reasonable. One difference, perhaps, between our experience and the experience of Zhang and Long is that we work in an academic setting with smaller trials, typically at a single location. We do not know if our experience, or the experience of Zhang and Long hold for most other researchers and suggest that data be collected in a systematic fashion to better understand accrual patterns in most clinical trials. It is clear that a more complex model can be superior to a simpler model. We are in favor of more complex models in some settings, but a further assessment would note the drawbacks of a more complex model. First, specifying a prior distribution is far more difficult. Important elements in a complex model, such as the number of knots in the cubic spline (Zhang & Long, 2010) are not incorporated at all into the prior distribution, and those elements which are incorporated are too complex for the average researcher to fathom. Second, a more complex model is frequently inefficient with limited data. Limited data, of course, occurs early in the study. We believe that accurate early predictions are very important because small changes to the study at an early stage to improve a sagging accrual rate are easier and more efficient than changes made later in the trial. Third, a simple model of accrual has a closed form solution for the posterior predictive distribution that is intuitively plausible. The mean of the posterior predictive distribution, for example, is simply a weighted average of the data and the prior mean. A 45

closed form solution also means that tracking accrual throughout a clinical trial could be conducted directly by the researcher on a daily or weekly basis, perhaps even on a simple spreadsheet. Perhaps a compromise between complexity and simplicity is most appropriate. In fact, we are looking at a linear piecewise regression model as an alternative to a complex spline and a compromise between the two approaches. The piecewise approach would allow for slow early accrual rates (both a step and elbow). Regardless of using simple, complex, or compromise we would like to propose guidance for evaluating the approaches with simulation studies. While it is impossible to conduct a simulation study that covers every possible research scenario, we believe a broad number of conditions need consideration to show scenarios where a simple model would perform well. Here are some suggested conditions: 1) performance under a constant accrual model. We believe that a simple model will perform well relative to the complex model in settings where a comlex model over fits the data. 2) performance early in the trial. We believe that a simple model will perform well relative to a complex model when only a small fraction of the accrual data is available. For example, in Zhang and Long, the simulation examined the performance of the model only when 30% and 60% of the accrual data was available. It would be very valuable to see the performance when only 5% or 10% of the accrual data was available. 3) performance under slow accrual rates. The average threshold accrual rate in the Zhang and Long simulation was 12 patients per day. While this may be normal in large multi-center trials, our experience with smaller academic center trials is that accrual rates of fewer than one person per day is more common. It would be instructive to test the cubic spline model with data where the Poisson counts are mostly zeros and ones. 4) performance under a weak, but not totally data driven prior. While we suggested an initial approach for getting a prior distribution using a simple question (how confident are you on a scale of 1 to 10), that prior was not intended to be plugged in thoughtlessly. Instead, that initial assessment would be used to examine the behavior of the predictive distribution. Review of that distribution would then lead the researcher to revise the prior accordingly. With a total sample size of 3,000 patients (much larger than the norm in an academic setting), P=0.5 constitutes an extremely strong prior. It says that after accumulating 1,500 patients, the prior and the data should still have equal weight. We would suggest that P=0.1 might be a more reasonable prior with such a large sample size, even when the researchers had strong prior information. In fact, all models need to be testing with a range of informative priors which needs to be balanced between two competing models of different complexity. 5. Conclusion A simpler model (e.g. Gajewski et al.) can and should be used in many other settings. The availability of both a simple and a complex (e.g. Zhang and Long) model of accrual will allow researchers to choose the approach that best fits their needs. Carefully crafted simulation studies designed to better understand the tradeoffs between simplicity and complexity would be most beneficial. References Anisimov, V., & Fedorov, V. D. (2007). Modelling, prediction and adaptive adjustment of recruitment in multicentre trials. Statistics in Medicine, 26(27), 4958-4975. http://dx.doi.org/10.1002/sim.2956 Gelman, A., Carlin, J. B., Stern, H. S., & Rubin, D. B. (2004). Bayesian Data Analysis (2nd ed.). Washington DC, Chapman and Hall/CRC. Gajewski, B., Simon, S., & Carlson, S. (2008). Predicting accrual in clinical trials with Bayesian posterior predictive distributions. Statistics in Medicine, 27(13), 2328-2340. http://dx.doi.org/10.1002/sim.3128 Zhang, X., & Long, Q. (2010). Stochastic modeling and prediction for accrual in clinical trials. Statistics in Medicine, 29(6), 649-658. http://dx.doi.org/10.1002/sim.3847 46