Presentation is loading. Please wait.

Presentation is loading. Please wait.

Comparison of several prediction errors for model evaluation K. Brendel 1, E. Comets 1, C. Laveille 2, R. Jochemsen 2, F. Mentré 1 The model (model B)

Similar presentations


Presentation on theme: "Comparison of several prediction errors for model evaluation K. Brendel 1, E. Comets 1, C. Laveille 2, R. Jochemsen 2, F. Mentré 1 The model (model B)"— Presentation transcript:

1 Comparison of several prediction errors for model evaluation K. Brendel 1, E. Comets 1, C. Laveille 2, R. Jochemsen 2, F. Mentré 1 The model (model B) was built from 2 phase II studies of an antidiabetic drug. Model B was a one compartment model with zero order absorption and first order elimination, with exponential random effects on the apparent volume of distribution (V/F) and on the apparent clearance (Cl/F). A proportional error model was selected. Two “validation” datasets were simulated according to the design of a real phase I study (12 subjects with 16 samples): the first (V true ) was simulated with the parameters values estimated previously (B); the second (V false ) was simulated using the same model, but with the mean value for V/F and Cl/F divided by two,corresponding to a bioavailability multiplied by two. We simulated these two datasets to check the ability to the metrics to validate V true and to reject V false. We considered metrics without and with simulations. These last metrics, called posterior predictive check (PPC), evaluate the adequacy between data and model by comparing a given statistic, computed with the data, to its posterior predictive distribution computed under the model. This distribution was estimated using Monte Carlo simulations with model B to obtain K datasets simulated according to the phase I design. In the following, DVsimijk is the jth simulated concentration for the ith subject and the kth simulation. Metrics based on concentrations Standardized Prediction Error for Concentration (SPEC) SPEC ij = WRES ij = (OBS ij - PRED ij )/SD(PRED ij ) obtained on datasets V using model B and their hyperparameters. Standardized Prediction Error for Concentration with Simulation (SPECS) SPECS ij = (DV ij – E(DVsim ij )) / SD(DVsim ij ) where E(DVsim ij ) and SD(DVsim ij ) are the empirical mean and SD of the K simulated DVsim ijk. Normalized Prediction Distribution Errors for Concentrations with Simulation (NPDECS) NPDECS ij are normalized PDECS ij, which are computed as the percentile of DV ij in the whole predicted distribution of the DVsim ijk. More precisely, the K values of DVsim ijk are sorted and the percentile of DV ij is the number of DVsim ijk lower than DV ij divided by K. These PDECS ij were normalized using the inverse of the cumulative density function of a normal distribution (with mean 0 and variance 1). For each of these metrics, we tested whether the mean was significantly different from 0 using a Wilcoxon signed-rank test and whether the variance was significantly different from 1 with a Fisher test. We then tested normality using the Shapiro-Wilks (SW) and the Kolmogorov-Smirnov (KS) tests. Metrics based on random effects Standardized Prediction Error for the post hoc Random effect (SPERHOC) For each random effect, SPERHOC i are computed from the Bayes estimate of the random effects (  i ) obtained on V using model B and their hyperparameters: SPERHOC i =  i /  B where  B is the variance for the parameter esimated with model B with tha initial dataset. Tests for SPERHOC are the same as for metrics on concentrations. Metrics based on hyperparameters Standardized Prediction Error on the Hyperparameters (SPEH) The posterior predictive distribution of the hyperparameters under the model was estimated by fitting the K simulated datasets (Ψ B ). The hyperparameters (Ψ V ) estimated with an independent population analysis on V were then compared to the region of acceptance of this distribution. SPEH are based on the differences of the estimated values t SPEH = (Ψ V – Ψ B )/ [(SE V ) 2 +( SE B ) 2 ] 1/2 where SE are the standard errors of estimation. We tested whether SPEH = 0 using a Wald test. Standardized Prediction Error on the Hyperparameters with simulations For the fitting of the simulations of K datasets, we compared the hyperparameters Ψ V ( estimated with an independent population analysis on V) to their posterior predictive distribution from the hyperparameter estimates in the K simulated datasets. METHODS Data V true 1 INSERM E0357, Department of Epidemiology, Biostatistics and Clinical research, AP-HP, Bichat University Hospital, Paris, France, 2 Servier,Courbevoie, France Data V false SPEC SPECS NPDECS SPERHOC on CL INTRODUCTION External validation refers to a comparison between the validation dataset (V) and the predictions from the model built using the learning dataset (B). The validation dataset is not used for model building and parameters estimation. The aim of this study was to compare criteria for the evaluation of a population pharmacokinetic model. Several types of prediction errors on concentrations, random effects and hyperparameters are proposed and evaluated on simulated validation datasets. RESULTS Even on V true, both SPEC and SPECS were found to differ significantly from a normal distribution. On the other hand, NPDECS and SPERHOC on CL/F and V/F followed a normal distribution, as illustrated on the QQ plots.The mean was not significantly different from 0 for the four metrics. For hyperparameters without simulation, Wald tests were not significant. For hyperparameters with simulation, the p value was not significant for CL/F and OmegaCL/F as illustrated on the histograms (same results for V/F and OmegaV/F). On V false, SPEC, SPECS and NPDECS were not found to follow a normal distribution, but SPERHOC on CL and V did. The mean was significantly different from 0 for the four metrics. For hyperparameters without simulation, Wald tests were significantly different for V/F and CL/F, and with simulation, the test was significant for CL and not OmegaCL/F (same results for V/F and OmegaV/F). CLSimulatedOmegaCLSimulated TESTp Meanns Variancens Normality KS SW 0.03 <0.0001 TESTp Mean<0.0001 Variance0.01 Normality KS SW <0.010 <0.0001 TESTp Mean<0.0001 Variance0.01 Normality KS SW <0.010 0.001 TESTp Meanns Variancens Normality KS SW <0.010 <0.0001 TESTp Meanns Variancens Normality KS SW ns TESTp Mean<0.0001 Variancens Normality KS SW 0.033 0.001 TESTp Meanns Variancens Normality KS SW ns TESTp Mean0.024 Variancens Normality KS SW ns hyperparameters B Estimate (SE) Vtrue Estimate (SE) p Vfalse Estimate (SE) p CL 1.0 (0.042) 0.98 (0.159) ns 0.48 (0.083) <0.01 V40 (2.3) 42 (3.3) ns 20 (1.5) <0.01 D1 6.6 (0.22) 6.5 (0.29) ns 7.0 (0.063) ns (OmegaCL) 2 0.35 (0.057) 0.27 (0.13) ns 0.34 (0.13) ns (OmegaV) 2 0.11 (0.028) 0.086 (0.026) ns 0.062 (0.022) ns (Sigma) 2 0.061 (0.0064) 0.052 (0.0048) ns 0.057 (0.0064) ns Wald tests on estimated hyperparameters With 500 simulations and fittings With 1000 simulations CONCLUSION pVtrue =ns pVfalse=0.002 pVtrue =ns pVfalse=ns For metrics on concentrations NPDECS was able to validate V true and to reject V false, while SPEC and SPECS showed less discrimination and rejected V true with the normality test. Wald tests as well as PPC on hyperparameters were also able in this example to detect model misfit. These metrics will be applied to the real phase I dataset (the design of which was used here for the simulations) and evaluated on repeated simulations of validation datasets. The type I error and the power of the proposed tests will be estimated.


Download ppt "Comparison of several prediction errors for model evaluation K. Brendel 1, E. Comets 1, C. Laveille 2, R. Jochemsen 2, F. Mentré 1 The model (model B)"

Similar presentations


Ads by Google