Presentation is loading. Please wait.

Presentation is loading. Please wait.

October 28, 2008 Model qualification and assumption checking 1 To Validate or not to Validate? If all models are wrong but some.

Similar presentations


Presentation on theme: "October 28, 2008 Model qualification and assumption checking 1 To Validate or not to Validate? If all models are wrong but some."— Presentation transcript:

1 Pravin.Jadhav@fda.hhs.gov October 28, 2008 Model qualification and assumption checking 1 To Validate or not to Validate? If all models are wrong but some are useful, why bother? To Qualify or not to Qualify? To Check or not to Check? To Evaluate or not to Evaluate? Pravin Jadhav, Pharmacometrics

2 Pravin.Jadhav@fda.hhs.gov October 28, 2008 2 Model qualification and assumption checking Model Qualification or Checking Qualification/Checking can be used to provide the following –feedback on how to improve the current model (learn), and/or –some reassurance that the model can at the least regenerate the data that were used to build the model (confirm).

3 Pravin.Jadhav@fda.hhs.gov October 28, 2008 3 Model qualification and assumption checking Model Qualification or Checking Before you go to quantitative qualification methods, (Qualitative)- –Use prior knowledge (nothing better than this) Properties of the drug Parameter estimates from previous studies or analyses (e.g. NonCompartmental analysis) Drugs in the same class In vitro data, etc. –Quality of the experimental design and data used in building the model –Specify model qualification objectives

4 Pravin.Jadhav@fda.hhs.gov October 28, 2008 4 Model qualification and assumption checking Precision of parameter estimates Log-likelihood profiling Bootstrap –Parametric –Nonparametric

5 Pravin.Jadhav@fda.hhs.gov October 28, 2008 5 Model qualification and assumption checking Log-Likelihood profiling: Principles Obj Fn CL 0 Maximum Likelihood Estimate Derive confidence interval for parameter estimates –Procedure Using the final model run, fix the value of parameter estimate to different values than maximum likelihood estimate Run estimation Plot difference in OBJ from the final model run against parameter –Automated In WFN, use nmllp runname paraname val1 val2…

6 Pravin.Jadhav@fda.hhs.gov October 28, 2008 6 Model qualification and assumption checking Log-Likelihood profiling Clearance estimate

7 Pravin.Jadhav@fda.hhs.gov October 28, 2008 7 Model qualification and assumption checking Log-Likelihood profiling This method is dependent on –-2*log-likelihood is chi-square distributed with 1 degree of freedom (one reduced parameter compared to full model) For some estimation methods (for example, FO) the assumption might not be accurate –Gobburu and Lawrence Pharm Res. 2002 Jan;19(1):92-8 –Wahlby et. al. J Pharmacokinet Pharmacodyn. 2001 Jun;28(3):231-52 What does this mean to you? How do you derive this?

8 Pravin.Jadhav@fda.hhs.gov October 28, 2008 8 Model qualification and assumption checking What does this mean? For most examples, if you use FO method- change of 3.84 in objective function value for the reduced model versus full model is probably not accurate for 95% significance level? If the exact p-value is needed –Do you need higher or lower change based on the previous graph? ???? How do you derive conditional distribution? If the exact p-value is needed, one will need to generate conditional distribution of the log-lileklihood and not rely on theoretical distribution (using randomization test, which will be not covered today) Generally speaking, the conditional distribution will be different for each combination of full model, reduced model and dataset (dense/sparse etc.) based on the approximations used-- But we don’t do it for every run- why?

9 Pravin.Jadhav@fda.hhs.gov October 28, 2008 9 Model qualification and assumption checking Bootstrap Wikipedia: bootstrapping or booting which began in the 1880s as a leather strap and evolved into a group of metaphors that share a common meaning, a self-sustaining process that proceeds without external help. Wikipedia: Bootstrapping is the practice of estimating properties of an estimator (such as its variance) by measuring those properties when sampling from an approximating distribution. –Smooth bootstrap –Parametric bootstrap –Case resampling (Non-parametric bootstrap) –Resampling residuals –Wild bootstrap http://en.wikipedia.org/wiki/Bootstrapping_(statistics)

10 Pravin.Jadhav@fda.hhs.gov October 28, 2008 10 Model qualification and assumption checking Bootstrap Dataset 1000: Sample 100 subjects Dataset 1: Sample 100 subjects Non-parametric –Sample individuals to create several datasets from the original data Example from homework #3 Run E S T I M A T I O N 1000 Population estimates Original Data ID=9 ID=5 ID=45 ID=67 WFN use: nmbs runname 100 will call nmgo with 100 bootstrap sampled data sets taken from dataset supplied in runname

11 Pravin.Jadhav@fda.hhs.gov October 28, 2008 11 Model qualification and assumption checking Dataset 1000: using 100 CLi, Vi, ERRij and original data structure Dataset 1: using 100 CLi, Vi, ERRij and original data structure Bootstrap Parametric –Monte Carlo Simulations to create several datasets from the final model and model parameters One compartment example from homework #2 CL V Final model ESTIMATEESTIMATE 1000 Population estimates Error

12 Pravin.Jadhav@fda.hhs.gov October 28, 2008 12 Model qualification and assumption checking Bootstrap Parametric or Non-Parametric method will yield N (one for each successfully converged boostrap samples) sets of parameters –For example, N=1000 sets of CL, V,  2 CL,  2 V and  2

13 Pravin.Jadhav@fda.hhs.gov October 28, 2008 13 Model qualification and assumption checking Model Qualification or Checking Diagnostic plots (Slide 7 From Dr. Tornoe’s slides) –Observed and predicted concentration vs. time –Observed vs. predicted concentration –Residuals vs. time –Residuals vs. predictions Did you do this ever? –Homework #1, #2, #3

14 Pravin.Jadhav@fda.hhs.gov October 28, 2008 14 Model qualification and assumption checking Model Qualification or Checking: Observed vs. predicted concentration Homework #3 Student 1: There is no systemic bias (over-prediction or under-prediction) in individual predictions from the model; however, population predictions appear to be slightly under- predicted (biased). Student 2:The model was able to describe the data very well with no systematic bias. Student 3: Individual Predictions: The observed and predicted concentrations appear to be closely distributed around the line of identity suggesting minimal residual error and hence, the validity of the one-compartment model with first order absorption.

15 Pravin.Jadhav@fda.hhs.gov October 28, 2008 15 Model qualification and assumption checking Model Qualification or Checking: Residuals vs. time or predictions Homework #3 Student 1: Weighted residual are homogenously and randomly distributed around the line with zero mean without any trend, suggesting………………………. Student 2: Observation of the weighted residuals vs. time and vs. population predicted shows no heteroscedasticity.

16 Pravin.Jadhav@fda.hhs.gov October 28, 2008 16 Model qualification and assumption checking Model Qualification or Checking: Observed and predicted concentration vs. time Student 1: From a visual point of view, observed concentrations are well described by a one- cmpt body model with oral absorption.

17 Pravin.Jadhav@fda.hhs.gov October 28, 2008 17 Model qualification and assumption checking Model Qualification or Checking: Other plots There are several other plots one could make to test assumptions etc. We will not discuss these or others but the underlying message is to check assumptions that went into the model

18 Pravin.Jadhav@fda.hhs.gov October 28, 2008 18 Model qualification and assumption checking Model Qualification or Checking: Predictive check (Posterior) Predictive check is proposed to check whether the posited model should be excluded, because the model fails to provide a reasonable summary of the data used for modeling. Originally developed for checking fully Bayesian models. –The posterior distribution, a reflection of the uncertainty of a parameter, is influenced by the strength of the prior knowledge. Recollect Session 1 and today’s Q&A –Major question: How close the posterior distribution was to the current data? a summary feature, called a statistic (for example, SSE), calculated from the current data, are compared with the same statistic calculated under the posterior distribution. If this comparison failed to meet a prespecified criterion, the model might be rejected. Makes a lot of sense in Bayesian framework.

19 Pravin.Jadhav@fda.hhs.gov October 28, 2008 19 Model qualification and assumption checking Model Qualification or Checking: Predictive check Why (posterior) in parentheses –ML methods do not use priors –The ML approach yields only the point estimates of the parameters (called the ML estimates) and the asymptotic standard errors, and not a posterior distribution Yano Y, Beal SL, Sheiner LB. J Pharmacokinet Pharmacodyn. 2001;28:171-192 Jadhav, P. R.; Gobburu, J.V.S.; AAPS Journal, Vol. 7 No. 3 (2005)

20 Pravin.Jadhav@fda.hhs.gov October 28, 2008 20 Model qualification and assumption checking Model Qualification or Checking: Predictive check Three steps in PC or PPC –Estimation step y OD : Original data (For example, Drug Concentration)  : Estimated population parameters –What about 1 compartment model for IV administration –Simulation step y 1 rep ….. Y n rep are generated using  –Evaluation step Compare –test statistics T(y OD ) to T(y i rep ): mean concentration at time t and area under the curve (determined empirically) –Discrepancy variable T(y OD,  ) to T(y i rep,  ): sum of squared errors (SSE), determined using the observed and model- predicted variables (eg, concentrations), mean prediction and mean absolute prediction errors

21 Pravin.Jadhav@fda.hhs.gov October 28, 2008 21 Model qualification and assumption checking Evaluation step –graphical assessment of the 95% prediction interval (visual PC) –considerable scatter beyond the 95% prediction interval could indicate a poor model, but the converse may not be valid. Model Qualification or Checking: Predictive check

22 Pravin.Jadhav@fda.hhs.gov October 28, 2008 22 Model qualification and assumption checking Model Qualification or Checking: Predictive check –predictive p-value (Pp) What does Pp =0, 0.5 and 1 mean? –probability of equivalence (p eqv ),

23 Pravin.Jadhav@fda.hhs.gov October 28, 2008 23 Model qualification and assumption checking Predictive check: References Gelman A, Carlin JB, Stern HS, Rubin DB. Model checking and sensitivity analysis. In: Gelman A, ed. Bayesian Data Analysis. London: Chapman & Hall; 1995:161-189

24 Pravin.Jadhav@fda.hhs.gov October 28, 2008 24 Model qualification and assumption checking Other approaches Internal validation External validation Sensitivity analysis


Download ppt "October 28, 2008 Model qualification and assumption checking 1 To Validate or not to Validate? If all models are wrong but some."

Similar presentations


Ads by Google