Presentation on theme: "Nicola Cooper Centre for Biostatistics & Genetic Epidemiology,"— Presentation transcript:
1USE OF EVIDENCE IN DECISION MODELS: An appraisal of health technology assessments in the UK Nicola CooperCentre for Biostatistics & Genetic Epidemiology,Department of Health Sciences, University of Leicester, U.K.Acknowledgements to: Doug Coyle, Keith Abrams,Miranda Mugford & Alex Sutton
2OUTLINE Background to empirical research Methods & Findings from Study Conclusions
3BACKGROUNDIncreasingly decision models developed to inform complex clinical/economic decisions (e.g. NICE technology appraisals)Decision models provide:Explicit quantitative & systematic approach to decision makingCompares at least 2 alternativesUseful way of synthesising evidence from multiple sources (e.g. effectiveness data from trials, adverse event rates from observational studies, etc.)
4BACKGROUND Decision modelling techniques commonly used for: Decision modelling techniques commonly used for:i) Extrapolation of primary data beyond endpoint of a trial,ii) Indirect comparisons when no ‘head-to-head’ trialsiii) Investigation of how cost-effectiveness of clinical strategies/interventions changes with values of key parametersiv) Linking intermediate endpoints to ultimate measures of health gain (e.g. QALYs)v) Incorporation of country specific data relating to disease history and management.
5BACKGROUNDDecision models contain many unknown parameters & evidence may include published data, controlled trial data, observational study data, or expert knowledge.Need to utilise/synthesise available evidenceModel parameters can include:clinical effectiveness,costs,disease progression rates, &utilitiesEvidence-based models – Require systematic methods for identification & synthesis of evidence to estimate model parameters with appropriate levels of uncertaintyIf select only “best” (most relevant) evidence – potentially ignore valuable information from other sources
6ECONOMIC DECISION MODEL RCT1RCT2RCT3OBS1OBS2ROUTINEEXPERTDATASOURCESMeta-analysisEVIDENCESYNTHESISGen. synthesisOpinion poolingBayes theoremIn combinationAdverse EventsClinicalEffectMODELINPUTSUtilityCostDECISIONMODEL
7MRC FELLOWSHIPThe use of evidence synthesis & uncertainty modelling in economic evidence-based health-related decision modelsPart 1) To review and critique use of evidence in decision models developed as part of health technology assessments to datePart 2) Develop practical solutions for synthesising evidence, with appropriate uncertainty, to inform model inputs:For example, combining evidence in different formats (e.g. mean and median), from different sources (e.g. RCT, cohort, registry, etc.), etc.
8NICE GUIDANCENICE methods guidelines to Health Technology Assessment (2004)‘all relevant evidence must be identified’‘evidence must be identified, quality assessed and, where appropriate, pooled using explicit criteria and justifiable and reproducible methods’and‘explicit criteria by which studies are included or excluded’
9USE OF EVIDENCE IN HTA DECISION MODELS (Cooper et al, In press) OBJECTIVE:Review sources & quality of evidence used in the development of economic decision models in health technology assessments in the UKMETHODOLOGY:Review included all economic decision models developed as part of the NHS Research & Development Health Technology Assessment (HTA) Programme between 1997 and 2003 inclusively.Quality of evidence was assessed using a hierarchy of data sources developed for economic analyses (Coyle & Lee 2002) & good practice guidelines (Philips et al 2004).
10GOOD PRACTICE CRITERIA FOR DECISION MODELS (Philips et al 2004) Statement of perspective (e.g. healthcare, societal, etc.)Description of strategies/comparatorsDiagram of model/disease pathwaysDevelopment of model structure and assumptions discussedTable of model input parameters presentedSource of parameters clearly statedModel parameters expressed as distributionsDiscussion of model assumptionsSensitivity analysis performedKey drivers/influential parameters identifiedEvaluation of internal consistency undertaken
11HIERARCHY OF DATA SOURCES Hierarchy of evidence - a list of potential sources of evidence for each data component of interest:Main clinical effectivenessBaseline clinical dataAdverse events and complicationsResource useCostsUtilitiesSources ranked on increasing scale from 1 to 6, most appropriate (best quality) assigned a rank of 1
12HIERARCHY OF DATA SOURCES #Surrogate outcomes = an endpoint measured in lieu of some other so-called true endpoint (including survival at end of clinical trial as predictor of lifetime survival)
14FLOW DIAGRAM 180 HTA published 1997-2003 5 out of 42 (12%) Individual Sampling##One HTA reported both decision & Markov models, one reported both Markov & Individual Patient models, and one model type was unclear.26 out of 42 (62%) Decision Trees#12 out of 42 (29%) Markov Models#147 out of 180 (73%) considered Health Economics48 out of 147 (33%) Developed Decision Models42 out of 48 (88%) Economic Evaluation Models6 out of 48 (15%) Cost Analyses Models22 (out of 42) NICE Appraisals
15GOOD PRACTICE CRITERIA FOR DECISION MODELS (n=42)
16RESULTS FROM APPLYING HIERARCHIES OF EVIDENCE (n=42 decision models)
18CONCLUSIONS Evidence on main clinical effect mostly: identified & quality assessed (76%) as part of companion systematic review for HTAreported in a fairly transparent & reproducible way.For all other model inputs (i.e. adverse events, baseline clinical data, resource use, and utilities)search strategies for identifying relevant evidence rarely made explicitsources of specific evidence not always reported
19CONCLUSIONS Concerns about decision models confirmed by this study: (1) Use of data from diverse sources (e.g. RCTs, observational studies, expert opinion) - may be subject to varying degrees of bias due to confounding variables, patient selection, or methods of analysis(2) Lack of transparency regarding identification of model input data & key assumptions underlying model structure and evaluation(3) Bias introduced by the researcher with regards to choice of model structure & selection of parameter values to input into the model.
20CONCLUSIONSHierarchies of evidence for different data components provide useful tool for assessingi) quality of evidence,ii) promoting transparency, &iii) informing weakest aspects of model for future work.Acknowledged, highly ranked evidence for certain model parameters may not always be available but needs to be made explicit (e.g. expert opinion used as no other data available?).Value of evidence input into decision models, regardless of position in hierarchy, depends on its quality & relevance to question of interest.QUANTITY vs. QUALITY (PRECISION vs. BIAS)
21UNANSWERED QUESTIONS How best to identify the relevant evidence? How much evidence is sufficient and when would there be benefit from identifying additional/supplementary evidence (possibly from lower levels of the hierarchy)?How to appropriately assess, and where possible adjust for, quality of different types of evidence?- Instruments for assessing quality within study designs but across different study designs non-trivial (Downs & Black 1998)How to appropriately combine/synthesis evidence from different study types? For example,meta-analyse all data assuming equal weight,observational data as prior for RCT data, orhierarchical synthesis model
22Copy of slides available at: http://www.hs.le.ac.uk/personal/njc21/ REFERENCESCooper NJ, Coyle D, Abrams KR, Mugford M, Sutton AJ. Use of evidence in decision models: An appraisal of health technology assessments in the UK to date. Journal of Health Services Research and Policy (In press 2005).Coyle D, Lee KM. Evidence-based economic evaluation: how the use of different data sources can impact results. Donaldson C, Mugford M, Vale L. Evidence-based health economics: From effectiveness to efficiency in systematic review. London: BMJ Publishing Group, 2002:Downs SH,.Black N. The feasibility of creating a checklist for the assessment of the methodological quality both of randomised and non-randomised studies of health care interventions. Journal of Epidemiology and Community Health 1998;52:Philips Z, Ginnelly L, Sculpher M et al. Review of guidelines for good practice in decision-analytic modelling in health technology assessment. Health Technology Assessment. 2004; 8(36).National Institute for Clinical Excellence (NICE). Guide to the methods of technology appraisal. London: National Institute of Clinical Excellence, 2004.Copy of slides available at: