Presentation is loading. Please wait.

Presentation is loading. Please wait.

An Introduction to Bayesian GLM Methods for Cost-Effectiveness Analysis of Primary Data Dave Vanness, Ph.D. University of Wisconsin School of Medicine.

Similar presentations

Presentation on theme: "An Introduction to Bayesian GLM Methods for Cost-Effectiveness Analysis of Primary Data Dave Vanness, Ph.D. University of Wisconsin School of Medicine."— Presentation transcript:

1 An Introduction to Bayesian GLM Methods for Cost-Effectiveness Analysis of Primary Data Dave Vanness, Ph.D. University of Wisconsin School of Medicine and Public Health

2 Introduction In cost-effectiveness analysis conducted alongside clinical trials, it is common to simply calculate the Incremental Cost-Effectiveness Ratio (ICER) or Net Benefit (NB) directly from average costs and outcomes observed in treatment groups. In a classical (frequentist) inference, we would construct confidence intervals for the ICER or NB. –95% confidence does NOT express the probability that an unknown (e.g., the "true" NB) lies within a specific confidence interval. –Rather, it represents the likelihood that your procedure of calculating such intervals would capture the true parameter value about 95% of the time when an experiment is repeated over and over.

3 Is Inference Irrelevant in CEA? If the objective is to maximise health gain for a given budget, then programmes should be selected based on the posterior mean net benefit irrespective of whether any differences are regarded as statistically significant or fall outside a Bayesian range of equivalence. This is because one of the mutually exclusive alternatives must be chosen and this decision cannot be deferred. The opportunity costs of failing to make the correct decision based on the mean are symmetrical and the historical accident that dictates which of the alternatives is regarded as current practice is irrelevant. -- Karl Claxton, 1999 p. 347-8.

4 Overview/Learning Objectives Introduction to Bayesian Analysis –Probability as Uncertainty –Bayes Rule –Markov Chain Monte Carlo Analytical Example –Hypothetical CEA alongside a trial (simulated dataset) –Generalized Linear Models (GLM) of: Cure and Adverse Events - Bernoulli with Probit Link Cost - Gamma with Log Link QALY - Beta with Logistic Link –Decision-Theoretic Analysis of Results Goal: The probability that a treatment is cost-effective Posterior ICER, expected net benefit and acceptability

5 Introduction to Bayesian Analysis

6 Probability as Uncertainty 1654 – Pascal and Fermat (The Points Problem) – how to divide winnings when a sequence of games is interrupted. stochastic 1662, "pertaining to conjecture," from Gk. stokhastikos "able to guess, conjecturing," from stokhazesthai "guess," from stokhos "a guess, aim, target, mark," lit. "pointed stick set up for archers to shoot at" (see sting). [] 1701?-1761, Rev. Thomas Bayes

7 Philos. Trans. R. Soc. London, 1763, 53: 370-418.

8 PosteriorPriorLikelihood Normalizing Constant Bayes Rule

9 Is proportional to…

10 A Simple Bayesian Hierarchical Model Let Y i = 1 if a treatment is successful for individual i, Y i = 0 otherwise. Y i ~ Bernoulli (θ) θ = P(Y i = 1) for all i. All individuals are exchangeable members of the population with unknown probability of success, θ. As Bayesians, we represent uncertainty about θ with a probability distribution: θ ~ P(θ) (prior: before observing data) θ ~ P(θ|Y) (posterior: after observing data)

11 01 0 1 P(θ) θ Prior beliefs for θ (must integrate to 1) Pr(Y i = 1) = E[Y i ] = θ

12 Suppose we observe one individual… Suppose Y 1 = 0. What is P(Y 2 = 1 | Y 1 = 0)? –Its still θ – but do we now have the same beliefs about θ that we had before? –No! We apply Bayes Rule to update our prior beliefs.

13 01 0 1 L(Y 1 = 0 | θ) θ The Likelihood Function

14 01 0 1 P(θ) θ X 01 0 1 L(Y 1 = 1 | θ) θ Applying Bayes Rule

15 = 01 0 1 θ This doesnt integrate to 1. P(θ|Y 1 = 1)

16 Just a normalizing constant: P(Y)

17 01 0 1 θ L(Y 1 = 1 | θ)

18 = 01 0 1 θθ ÷ Now it integrates to 1 P(θ|Y 1 )

19 Now, we observe one more… Suppose Y 2 = 1. Now, what is P(Y 3 = 1 | Y 2 = 1)? –Its still θ. –This time, we brought some information with us. –Our posterior P(θ|Y 1 ) becomes our new prior, and we apply Bayes Rule again.

20 01 0 2 θθ P(θ) X 01 0 1 L(Y 1 = 1 |θ) θ

21 α P(θ|Y 2 ) 01 0 1 θ

22 Conjugate Analysis It happens to be that the prior and posterior distributions you just saw are Beta distributions. The likelihood function was Bernoulli (binomial if we observed multiple individuals at once). If you multiply a Beta prior distribution by a Bernoulli/Binomial Likelihood function, you get back a Beta posterior distribution. Posterior( θ |Data) = Beta(NS +, NF + ) –NS = number of observed successes (say, 6) –NF = number of observed failures (say, 3) –Where prior is Beta(, ) [ in our example]

23 Prior( θ ) = Beta(1,1) Posterior( θ ) = Beta(7,4)

24 Markov Chain Monte Carlo The Metropolis-Hastings algorithm uses distributions we already know to generate random draws from a Markov Chain whose stationary distribution is P(θ|X). We then collect those draws and analyze them (take their mean, median, etc., or run them through as parameters of a cost- effectiveness simulation, etc.).

25 Gibbs Sampling When θ is multidimensional, it can be useful to break down the joint distribution P(θ|X) into a sequence of full conditional distributions P(θ j |θ -j,X) = L(X|θ j,θ -j ) P(θ j |θ -j ) where -j signifies all elements of θ other than j. We can then specify a starting vector θ -j 0 and, if P(θ j |θ -j,X) is not from a known type of distribution, we can use the Metropolis algorithm to sample from it. –Running from j = 1 to M gives one full sample of θ.

26 θ1θ1 θ02θ02 θ2θ2 1 2 3 4 5

27 Heterogeneity Our first example was a very simple and homogenous model: every individuals outcome is drawn from the same distribution. The extreme in the opposite direction (complete heterogeneity) is also fairly simple: Y i ~ Bernoulli(θ i ) But its pretty difficult to extrapolate (make predictions) when there is no systematic variation.

28 Modeling Heterogeneity Usually, we assume there is a systematic relationship that explains some of the heterogeneity in observed outcomes. The classical normal regression model (which we usually estimate with Ordinary Least Squares) can be thought of as a hierarchical model. Y i = X i β + ε i –X i is a row vector of individual covariates –β is a column vector of parameters –ε i ~ N(0,σ 2 ) We can also write this as: Y i ~ N(µ i,,σ 2 ), µ i = X i β θ = {β, σ 2 } ~ P(θ) –P(θ) summarizes our knowledge about the joint distribution of unknown parameters. –β is probably multidimensional; σ 2 is a variance term – has to be positive; this is probably a weird mixture of distributions. Yikes!


30 Analytical Example: Cost-Effectiveness Analysis

31 Using MCMC to Conduct CEA We can use Markov Chain Monte Carlo in WinBUGS to estimate models of virtually any type. Draws from the posterior distributions can be used to conduct inference, test simple hypotheses – or can become inputs for policy- relevant simulations. Using the flexibility of WinBUGS, we can also explore the relationships among treatments, covariates, costs and health outcomes using regression analysis with Generalized Linear Models (GLM) and perform probabilistic sensitivity analysis at the same time.

32 Simulated CEA Dataset We use a combination of real-world variables and simulated relationships. 800 individuals selected at random from 2,452 individuals who self-reported hypertension in the 2005 MEPS-HC Covariates were: –Age –Sex (Male = 1) –BMI We created a latent class that equals 1 if an individual self-reported diabetes; 0 otherwise. The class variable was excluded from all analysis (assumed to be unobservable)

33 Simulated CEA Dataset Treatment (T = 1) T i ~ Bernoulli (0.5) Adverse events (AE = 1) AE i ~ Bernoulli (P i AE ) P i AE = 0.1 + 0.9*T i *Class i Cure (S = 1) S i ~ Bernoulli (P i S ) P i S = 0.8*T i + 0.1*(1-T i )

34 Simulated CEA Dataset Costs were simulated from Gamma- distributions as follows: C i = C i T *T i + C i X C i T ~ Gamma(25,1/400) C i X ~ Gamma(4,1/exp(X i β C )) where X i is a row vector consisting of: 1~Age i ~Sex i ~AE i ~S i and β C is a column vector of parameters: 7|.03|0|1.5|-.5.

35 Simulated Costs by Treatment Group

36 Simulated CEA Dataset QALYs were simulated from Beta distributions: Q i ~ Beta(α i,β i ) α i = β i exp(X i β Q ) β i = 1.2 + 2*T i where X i is a row vector consisting of: 1~Age i ~Sex i ~AE i ~S i and β Q is a column vector of parameters: 1|-.01|.25|-1|1.5

37 Simulated QALYs by Treatment Group

38 -> t = 0 Variable | Obs Mean Std. Dev. Min Max -------------+-------------------------------------------------------- c | 390 7671.389 7850.924 477.4742 65191 q | 390.6354654.2557378.0053504.9961702 t | 390 0 0 0 0 age | 390 54.7641 5.513782 45 64 male | 390.4871795.5004777 0 1 ae | 390.1230769.3289475 0 1 s | 390.1025641.3037784 0 1 bmi | 390 31.78385 6.994469 16 57.4 class | 390.2512821.4343075 0 1 -> t = 1 Variable | Obs Mean Std. Dev. Min Max -------------+-------------------------------------------------------- c | 410 18385.52 8746.241 6958.342 55654.68 q | 410.7849.163569.1765754.9979415 t | 410 1 0 1 1 age | 410 54.7561 5.482493 45 64 male | 410.4414634.4971683 0 1 ae | 410.297561.4577439 0 1 s | 410.802439.3986455 0 1 bmi | 410 30.90756 6.540621 17.2 56.5 class | 410.2268293.4192929 0 1

39 Sample (n=800) ICER: $74,357/QALY Population (n=2452) ICER: $79,933/QALY Population ICER by (unobserved class): –Class 0 (no diabetes): $43,519/QALY –Class 1 (diabetes): $589,954/QALY

40 BMI by Class

41 GLM Cure and Adverse Event Models (Bernoulli with Probit Link) S ~ Bernoulli(P i s ) P i s = Φ(X i s β s ) X i s = 1~Age i ~BMI i ~Sex i ~T i ~T i *(Age i ~BMI i ~Sex i ) AE ~ Bernoulli(P i AE ) P i AE = Φ(X i AE β AE ) X i AE = 1~Age i ~BMI i ~Sex i ~T i ~T i *(Age i ~BMI i ~Sex i ) Note: we are using non-informative (flat) priors, which give results comparable to Maximum Likelihood. But we could bring outside information into the prior through meta-analysis.

42 What the BUGS Code Looks Like for ( i in 1 : 800 ){ S[i] ~ dbern(p_S[i]) p_S[i] <- phi(arg.S[i]) arg.S[i] <- max(min(bS[1] + bS[2]*st_AGE[i] + bs[3]*MALE[i] + bS[4]*st_BMI[i] + bS[5]*T[i] + bS[6]*T[i]*st_AGE[i] + bS[7]*T[i]*MALE[i] + bS[8]*T[i]*st_BMI[i],5),-5) AE[i] ~ dbern(p_AE[i]) p_AE[i] <- phi(arg.AE[i]) arg.AE[i] <- max(min(bAE[1] + bAE[2]*st_AGE[i] + bAE[3]*MALE[i] + bAE[4]*st_BMI[i] + bAE[5]*T[i] + bAE[6]*T[i]*st_AGE[i] + bAE[7]*T[i]*MALE[i] + bAE[8] *T[i]*st_BMI[i],5),-5) } Note: For complete code, email

43 GLM Cost Model (Gamma with Log Link) We model cost as a mixture of Gammas (separate distributions of cost with and without treatment). C i = T i *C i 1 + (1-T i )*C i 0 C i 1 ~ Gamma(shape C1,scale C1 ) C i 0 ~ Gamma(shape C0,scale C0 ) Using log function to link mean cost to shape and scale parameters. Ln[mean(C)] = Xβ exp(Ln[mean(C)]) = exp(Xβ) mean(C) = exp(Xβ) shape/scale = exp(Xβ) scale = r/exp(Xβ)

44 GLM QALY Model (Beta with Logit Link) Rescale Q to [0,1] interval by dividing by maximum possible Q (follow-up time) Q i = T i *Q i 1 + (1-T i )*Q i 0 Q i 1 ~ Beta(a Q1, b Q1 ) Q i 0 ~ Beta(a Q0, b Q0 ) We use the logit function to link mean QALYs to the Beta parameters. mean(Q) = exp(Xβ)/(1+exp(Xβ)) mean(Q) = a/(a + b) a + a exp(Xβ) = a exp(Xβ) + b exp(Xβ) a = b exp(Xβ)

45 Posterior Credible Interval Node Posterior Mean Posterior Standard Deviation Monte Carlo Error2.50%median97.50% Adverse Events (Bernoulli - Probit) AGE0.00940.07830.0025-0.13630.00900.1647 BMI-0.01190.07830.0029-0.1665-0.00950.1339 CONSTANT-1.16300.11810.0054-1.4030-1.1610-0.9381 MALE-0.01210.16900.0080-0.3376-0.01350.3178 T0.63910.14870.00690.35010.64050.9348 TxAGE0.06080.10230.0034-0.14570.06200.2571 TxBMI0.23380.10430.00370.03420.23530.4395 TxMALE-0.00470.22110.0104-0.4334-0.00140.4152 Cure (Bernoulli - Probit) AGE-0.00220.08850.0031-0.1753-0.00130.1693 BMI-0.00120.08590.0031-0.17640.00020.1633 CONSTANT-1.35000.12140.0055-1.5990-1.3490-1.1180 MALE0.13910.16870.0077-0.17960.13840.4846 T2.24600.15190.00681.96002.24802.5600 TxAGE0.03260.11560.0039-0.19360.03360.2552 TxBMI-0.15570.11310.0040-0.3741-0.15600.0684 TxMALE-0.22740.21890.0097-0.6529-0.22420.2165 Posterior Inference




49 Posterior Credible Interval Node Posterior Mean Posterior Standard Deviation Monte Carlo Error2.50%median97.50% Cost w/o Treatment (Gamma-Log) AE1.51900.07370.00151.37701.51901.6690 AGE0.15520.02400.00040.10980.15500.2037 BMI0.00310.02300.0004-0.04240.00320.0477 CONSTANT-0.86000.03580.0012-0.9289-0.8600-0.7893 MALE0.02770.04900.0015-0.06710.02890.1231 S-0.54400.07750.0016-0.6897-0.5430-0.3867 SHAPE4.55900.31300.00293.97004.55305.1890 Cost w/Treatment (Gamma - Log) AE0.70350.02530.00060.65440.70350.7544 AGE0.07460.01140.00020.05220.07470.0971 BMI0.01470.01210.0002-0.00800.01440.0390 CONSTANT0.20020.03030.00150.13820.20130.2563 MALE0.00160.02290.0006-0.04200.00150.0472 S-0.17250.02990.0014-0.2294-0.1736-0.1133 SHAPE18.35001.28900.011915.900018.330020.9600 Posterior Inference

50 Posterior Credible Interval Node Posterior Mean Posterior Standard Deviation Monte Carlo Error2.50%median97.50% QALY w/o Treatment (Beta - Logit) AE-1.17600.15780.0030-1.4950-1.1740-0.8751 AGE-0.01930.04930.0009-0.1144-0.02050.0795 BMI-0.00150.04790.0010-0.0954-0.00170.0927 CONSTANT0.52940.07330.00230.38490.53000.6753 MALE0.13660.09630.0030-0.05650.13680.3198 S1.31000.16150.00300.99281.31601.6240 β1.16000.07560.00081.01701.15801.3110 QALY w/Treatment (Beta - Logit) AE-0.95420.06770.0015-1.0880-0.9524-0.8262 AGE-0.00920.03000.0006-0.0673-0.00880.0484 BMI0.08650.03050.00060.02980.08620.1478 CONSTANT0.41600.07970.00410.26250.41290.5812 MALE0.28120.05770.00170.16210.28270.3906 S1.47200.07990.00401.31201.47601.6230 β3.49800.23220.00223.05803.49503.9740 Posterior Inference

51 Subgroup Simulations Within WinBUGS, we take the draws from the parameter posteriors and, for a hypothetical individual (X profile): –Assign to No Treatment (T = 0) Simulate cure or no cure Simulate adverse event Simulate cost and QALY, given simulated cure and adverse event status –Assign to Treatment (T = 1) Repeat cure, adverse event, cost and QALY simulations –Repeat, say, 1,000 times and calculate average incremental costs and QALYs. The following ICERs apply to a female (Sex i = 0) of average age (z-transformed Age i = 0) at 5 different levels of z-transformed BMI i = {-2,-1,0,1,2}

52 Posterior ICERs by BMI z-score BMI = -2BMI = -1 BMI = 0BMI = 1BMI = 2

53 Posterior Acceptability (BMI = -2) (WTP from $0 to $200,000)

54 Posterior Acceptability (BMI = 2) (WTP from $0 to $200,000)

55 Posterior Net Benefit (WTP from $0 to $200,000) BMI = -2 BMI = -1 BMI = 0 BMI = 1 BMI = 2

56 References George Woodworths book Biostatistics: A Bayesian Introduction (Wiley-Interscience, 2004 ISBN: 0471468428 9780471468424 has an excellent WinBUGS tutorial (Appendix B), the text of which may be found here: Carlin BP, TA Louis. Bayes and Empirical Bayes Methods for Data Analysis. 2nd Edition, London: Chapman & Hall, 2000. Claxton K. The irrelevance of inference: a decision-making approach to the stochastic evaluation of health care technologies. J.Health Econ. 1999 Jun;18(3):341-364. Fryback DG, NK Stout, MA Rosenberg. An Elementary Introduction to Bayesian Computing Using WinBUGS. International Journal of Technology Assessment in Health Care, 2001;17(1):98-113. Gelman A and J Hill. Data Analysis Using Regression and Multilevel/Hierarchical Models. Cambridge: Cambridge University Press, 2007. Gilks WR, S Richardson, DG Spiegelhalter. Markov Chain Monte Carlo in Practice. London: Chapman & Hall, 1996. Hastie T, R Tibshirani and J Friedman. The Elements of Statistical Learning. Data Mining, Inference and Prediction. Springer, 2002. Luce BR, Claxton K. Redefining the analytical approach to pharmacoeconomics. Health Econ. 1999 May;8(3):187-189. Lunn DJ, Thomas A, Best N, Spiegelhalter D. WinBUGS-A Bayesian modelling framework: Concepts, structure, and extensibility. Statistics and Computing 2000;10(4):325-337. O'Hagan A, Stevens JW, Montmartin J. Bayesian cost-effectiveness analysis from clinical trial data. Stat.Med. 2001 Mar 15;20(5):733-753. Skrepnek GH. The contrast and convergence of Bayesian and frequentist statistical approaches in pharmacoeconomic analysis. Pharmacoeconomics 2007;25(8):649-664. Spiegelhalter DJ, Myles JP, Jones DR, Abrams KR. Bayesian methods in health technology assessment: a review. Health Technol.Assess. 2000;4(38):1-130. Tanner MA. Tools for Statistical Inference. Methods for the Exploration of Posterior Distributions and Likelihood Functions. 3d ed. Springer, 1996. Vanness DJ and Kim WR. Empirical Modeling, Simulation and Uncertainty Analysis Using Markov Chain Monte Carlo: Ganciclovir Prophylaxis in Liver Transplantation, Health Economics, 2002: 11(6), 551-566.

Download ppt "An Introduction to Bayesian GLM Methods for Cost-Effectiveness Analysis of Primary Data Dave Vanness, Ph.D. University of Wisconsin School of Medicine."

Similar presentations

Ads by Google