1 / 24

Model qualification and assumption checking

Model qualification and assumption checking. Pravin Jadhav, Pharmacometrics. To Validate or not to Validate? If all models are wrong but some are useful, why bother? To Qualify or not to Qualify? To Check or not to Check? To Evaluate or not to Evaluate?. Model Qualification or Checking.

lovie
Télécharger la présentation

Model qualification and assumption checking

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Model qualification and assumption checking Pravin Jadhav, Pharmacometrics To Validate or not to Validate? If all models are wrong but some are useful, why bother? To Qualify or not to Qualify? To Check or not to Check? To Evaluate or not to Evaluate? Model qualification and assumption checking

  2. Model Qualification or Checking • Qualification/Checking can be used to provide the following • feedback on how to improve the current model (learn), and/or • some reassurance that the model can at the least regenerate the data that were used to build the model (confirm). Model qualification and assumption checking

  3. Model Qualification or Checking • Before you go to quantitative qualification methods, (Qualitative)- • Use prior knowledge (nothing better than this) • Properties of the drug • Parameter estimates from previous studies or analyses (e.g. NonCompartmental analysis) • Drugs in the same class • In vitro data, etc. • Quality of the experimental design and data used in building the model • Specify model qualification objectives Model qualification and assumption checking

  4. Precision of parameter estimates • Log-likelihood profiling • Bootstrap • Parametric • Nonparametric Model qualification and assumption checking

  5. Log-Likelihood profiling: Principles • Derive confidence interval for parameter estimates • Procedure • Using the final model run, fix the value of parameter estimate to different values than maximum likelihood estimate • Run estimation • Plot difference in OBJ from the final model run against parameter • Automated • In WFN, use • nmllp runname paraname val1 val2… Obj Fn Maximum Likelihood Estimate 0 CL Model qualification and assumption checking

  6. Log-Likelihood profiling Clearance estimate Model qualification and assumption checking

  7. Log-Likelihood profiling • This method is dependent on • -2*log-likelihood is chi-square distributed with 1 degree of freedom (one reduced parameter compared to full model) • For some estimation methods (for example, FO) the assumption might not be accurate • Gobburu and Lawrence Pharm Res. 2002 Jan;19(1):92-8 • Wahlby et. al. J Pharmacokinet Pharmacodyn. 2001 Jun;28(3):231-52 What does this mean to you? How do you derive this? Model qualification and assumption checking

  8. What does this mean? • For most examples, if you use FO method- change of 3.84 in objective function value for the reduced model versus full model is probably not accurate for 95% significance level? • If the exact p-value is needed • Do you need higher or lower change based on the previous graph? • ???? • How do you derive conditional distribution? If the exact p-value is needed, one will need to generate conditional distribution of the log-lileklihood and not rely on theoretical distribution (using randomization test, which will be not covered today) • Generally speaking, the conditional distribution will be different for each combination of full model, reduced model and dataset (dense/sparse etc.) based on the approximations used-- But we don’t do it for every run- why? Model qualification and assumption checking

  9. Bootstrap • Wikipedia: bootstrapping or booting which began in the 1880s as a leather strap and evolved into a group of metaphors that share a common meaning, a self-sustaining process that proceeds without external help. • Wikipedia: Bootstrapping is the practice of estimating properties of an estimator (such as its variance) by measuring those properties when sampling from an approximating distribution. • Smooth bootstrap • Parametric bootstrap • Case resampling (Non-parametric bootstrap) • Resampling residuals • Wild bootstrap http://en.wikipedia.org/wiki/Bootstrapping_(statistics) Model qualification and assumption checking

  10. Bootstrap • Non-parametric • Sample individuals to create several datasets from the original data • Example from homework #3 Original Data Run E S T I M A T I O N ID=9 Dataset 1: Sample 100 subjects ID=5 1000 Population estimates ID=45 Dataset 1000: Sample 100 subjects ID=67 WFN use: nmbs runname 100 will call nmgo with 100 bootstrap sampled data sets taken from dataset supplied in runname Model qualification and assumption checking

  11. CL Error V Bootstrap • Parametric • Monte Carlo Simulations to create several datasets from the final model and model parameters • One compartment example from homework #2 Dataset 1: using 100 CLi, Vi, ERRij and original data structure E S T I M A T E 1000 Population estimates Dataset 1000: using 100 CLi, Vi, ERRij and original data structure Final model Model qualification and assumption checking

  12. Bootstrap • Parametric or Non-Parametric method will yield N (one for each successfully converged boostrap samples) sets of parameters • For example, N=1000 sets of CL, V, 2CL, 2V and 2 Model qualification and assumption checking

  13. Model Qualification or Checking • Diagnostic plots (Slide 7 From Dr. Tornoe’s slides) • Observed and predicted concentration vs. time • Observed vs. predicted concentration • Residuals vs. time • Residuals vs. predictions • Did you do this ever? • Homework #1, #2, #3 Model qualification and assumption checking

  14. Model Qualification or Checking: Observed vs. predicted concentration • Homework #3 • Student 1: There is no systemic bias (over-prediction or under-prediction) in individual predictions from the model; however, population predictions appear to be slightly under-predicted (biased). • Student 2:The model was able to describe the data very well with no systematic bias. • Student 3: Individual Predictions: The observed and predicted concentrations appear to be closely distributed around the line of identity suggesting minimal residual error and hence, the validity of the one-compartment model with first order absorption. Model qualification and assumption checking

  15. Model Qualification or Checking: Residuals vs. time or predictions • Homework #3 • Student 1: Weighted residual are homogenously and randomly distributed around the line with zero mean without any trend, suggesting………………………. • Student 2: Observation of the weighted residuals vs. time and vs. population predicted shows no heteroscedasticity. Model qualification and assumption checking

  16. Model Qualification or Checking: Observed and predicted concentration vs. time • Student 1: From a visual point of view, observed concentrations are well described by a one-cmpt body model with oral absorption. Model qualification and assumption checking

  17. Model Qualification or Checking: Other plots • There are several other plots one could make to test assumptions etc. etc. We will not discuss these or others but the underlying message is to check assumptions that went into the model Model qualification and assumption checking

  18. Model Qualification or Checking: Predictive check • (Posterior) Predictive check is proposed to check whether the posited model should be excluded, because the model fails to provide a reasonable summary of the data used for modeling. • Originally developed for checking fully Bayesian models. • The posterior distribution, a reflection of the uncertainty of a parameter, is influenced by the strength of the prior knowledge. • Recollect Session 1 and today’s Q&A • Major question: How close the posterior distribution was to the current data? • a summary feature, called a statistic (for example, SSE), calculated from the current data, are compared with the same statistic calculated under the posterior distribution. • If this comparison failed to meet a prespecified criterion, the model might be rejected. • Makes a lot of sense in Bayesian framework. Model qualification and assumption checking

  19. Model Qualification or Checking: Predictive check • Why (posterior) in parentheses • ML methods do not use priors • The ML approach yields only the point estimates of the parameters (called the ML estimates) and the asymptotic standard errors, and not a posterior distribution • Yano Y, Beal SL, Sheiner LB. J Pharmacokinet Pharmacodyn. 2001;28:171-192 • Jadhav, P. R.; Gobburu, J.V.S.; AAPS Journal, Vol. 7 No. 3 (2005) Model qualification and assumption checking

  20. Model Qualification or Checking: Predictive check • Three steps in PC or PPC • Estimation step • yOD : Original data (For example, Drug Concentration) •  : Estimated population parameters • What about 1 compartment model for IV administration • Simulation step • y1rep ….. Ynrep are generated using  • Evaluation step • Compare • test statistics T(yOD) to T(yirep): mean concentration at time t and area under the curve (determined empirically) • Discrepancy variable T(yOD,) to T(yirep ,): sum of squared errors (SSE), determined using the observed and model-predicted variables (eg, concentrations), mean prediction and mean absolute prediction errors Model qualification and assumption checking

  21. Model Qualification or Checking: Predictive check • Evaluation step • graphical assessment of the 95% prediction interval (visual PC) • considerable scatter beyond the 95% prediction interval could indicate a poor model, but the converse may not be valid. Model qualification and assumption checking

  22. Model Qualification or Checking: Predictive check • predictive p-value (Pp) • What does Pp =0, 0.5 and 1 mean? • probability of equivalence (peqv), Model qualification and assumption checking

  23. Predictive check: References • Gelman A, Carlin JB, Stern HS, Rubin DB. Model checking and sensitivity analysis. In: Gelman A, ed. Bayesian Data Analysis. London: Chapman & Hall; 1995:161-189 Model qualification and assumption checking

  24. Other approaches • Internal validation • External validation • Sensitivity analysis Model qualification and assumption checking

More Related