Presentation is loading. Please wait.

Presentation is loading. Please wait.

Department of O UTCOMES R ESEARCH. Daniel I. Sessler, M.D. Professor and Chair Department of O UTCOMES R ESEARCH The Cleveland Clinic Clinical Research.

Similar presentations


Presentation on theme: "Department of O UTCOMES R ESEARCH. Daniel I. Sessler, M.D. Professor and Chair Department of O UTCOMES R ESEARCH The Cleveland Clinic Clinical Research."— Presentation transcript:

1 Department of O UTCOMES R ESEARCH

2 Daniel I. Sessler, M.D. Professor and Chair Department of O UTCOMES R ESEARCH The Cleveland Clinic Clinical Research Design Sources of Error Types of Clinical Research Randomized Trials

3 Sources of Error There is no perfect study All are limited by practical and ethical considerations It is impossible to control all potential confounders Multiple studies required to prove a hypothesis Good design limits risk of false results Statistics at best partially compensate for systematic error Major types of error Selection bias Measurement bias Confounding Reverse causation Chance

4 Statistical Association

5 Selection Bias Non-random selection for inclusion / treatment Or selective loss Subtle forms of disease may be missed When treatment is non-random: Newer treatments assigned to patients most likely to benefit “Better” patients seek out latest treatments “Nice” patients may be given the preferred treatment Compliance may vary as a function of treatment Patients drop out for lack of efficacy or because of side effects Largely prevented by randomization

6 Confounding Association between two factors caused by third factor For example: Transfusions are associated with high mortality But larger, longer operations require more blood Increased mortality consequent to larger operations Another example: Mortality greater in Florida than Alaska But average age is much higher in Florida Increased mortality from age, rather than geography of FL Largely prevented by randomization

7 Measurement Bias Quality of measurement varies non-randomly Quality of records generally poor Not necessarily randomly so Patients given new treatments watched more closely Subjects with disease may better remember exposures When treatment is unblinded Benefit may be over-estimated Complications may be under-estimated Largely prevented by blinding

8 Example of Measurement Bias Reported parental historyArthritis (%)No arthritis (%) Neither parent2750 One parent5842 Both parents158 From Schull & Cobb, J Chronic Dis, 1969 P = 0.003

9 Reverse Causation Factor of interest causes or unmasks disease For example: Morphine use is common in patients with gall bladder disease But morphine worsens symptoms which promotes diagnosis Conclusion that morphine causes gall bladder disease incorrect Another example: Patients with cancer have frequent bacterial infections However, cancer is immunosuppressive Conclusion that bacteria cause cancer is incorrect Largely prevented by randomization

10 External Threats to Validity Population of interest Eligible Subjects Subjects enrolled Selection bias Measurement bias Confounding Chance Conclusion Internal validityExternal validity ? ??

11 Types of Clinical Research Observational Case series –Implicit historical control –“The pleural of anecdote is not data” Single cohort (natural history) Retrospective cohort Case-control Retrospective versus prospective Prospective data usually of higher quality Randomized clinical trial Strongest design; gold standard First major example: use of streptomycin for TB in 1948

12 Case-Control Studies Identify cases & matched controls Look back in time and compare on exposure Time Case Group Control Group ExposureExposure

13 Cohort Studies Identify exposed & matched unexposed patients Look forward in time and compare on disease Time Exposed Unexposed DiseaseDisease

14 Timing of Cohort Studies Time Initial exposuresDisease onset or diagnosis PROSPECTIVE COHORT STUDY AMBIDIRECTIONAL COHORT STUDY RETROSPECTIVE COHORT STUDY

15 Randomized Clinical Trials (RCTs) A type of prospective cohort study Best protection again bias and confounding Randomization: reduces selection bias & confounding Blinding: reduces measurement error Not subject to reverse causation RCTs often “correct” observational results Types Parallel group Cross-over Factorial Cluster

16 Parallel Group Randomize participants to treatment groups Intervention AIntervention B Outcome AOutcome B Enrollment Criteria

17 Cross-over Diagram Treatment A ± Washout Treatment B ± Washout Treatment A Randomize individuals To sequential treatment Enrollment Criteria

18 Pros & Cons of Cross-over Design Strategy Sequential treatments in each participant Patients act as their own controls Advantages Paired statistical analysis markedly increases power Good when treatment effect small versus population variability Disadvantages Assumes underlying disease state is static Assumes lack of carry-over effect May require a treatment-free washout period Evaluate markers rather than “hard” outcomes Can not be used for one-time treatments such as surgery

19 Factorial Trials Simultaneously test 2 or more interventions Clonidine +ASAPlacebo + ASA Clonidine + PlaceboPlacebo + Placebo Clonidine +ASAPlacebo + ASA Clonidine + PlaceboPlacebo + Placebo Clonidine vs. Placebo ASA vs. Placebo

20 Pros & Cons Advantages More efficient than separate trials Can test for interactions Disadvantages Complexity, potential for reduced compliance Reduces fraction of eligible subjects and enrollment Rarely powered for interactions –But interactions influence sample size requirements

21 Factorial Outcome Example Apfel, et al. NEJM 2004

22 Subject Selection Tight criteria Reduces variability and sample size Excludes subjects at risk of treatment complications Includes subjects most likely to benefit May restrict to advance disease, compliant patients, etc. Slows enrollment “Best case” results –Compliant low-risk patients with ideal disease stage Loose criteria Includes more “real world” participants Increases variability and sample size Speeds enrollment Enhances generalizability

23 Randomization and Allocation Only reliable protection against Selection bias Confounding Concealed allocation Independent of investigators Unpredictable Methods Computer-controlled Random-block Envelopes, web-accessed, telephone Stratification Rarely necessary

24 Blinding / Masking Only reliable prevention for measurement bias Essential for subjective responses –Use for objective responses whenever possible Careful design required to maintain blinding Potential groups to blind Patients Providers Investigators, including data collection & adjudicators Maintain blinding throughout data analysis Even data-entry errors can be non-random Statisticians are not immune to bias! Placebo effect can be enormous

25 Placebo Effect Kaptchuk, PLoS ONE, 2010

26 Selection of Outcomes Surrogate or intermediate May not actually relate to outcomes of interest –Bone density for fractures –Intraoperative hypotension for stroke Usually continuous: implies smaller sample size Rarely powered for complications Major outcomes Severe events (i.e., myocardial infarction, stroke) Usual dichotomous: implies larger sample size Mortality Cost effectiveness / cost utility Quality-of-life

27 Composite Outcomes Any of ≥2 component outcomes, for example: Cardiac death, myocardial infarction, or non-fatal arrest Wound infection, anastomotic leak, abscess, or sepsis Usually permits a smaller sample size Incidence of each should be comparable Otherwise common outcome(s) dominate composite Severity of each should be comparable Unreasonable to lump minor and major events Death often included to prevent survivor bias Beware of heterogeneous results

28 Outcomes Approaches

29 Trial Management Case-report forms Require careful design and specific field definitions Every field should be completed –Missing data can’t be assumed to be zero or no event Data-management (custom database best) Evaluate quality and completeness in real time Range and statistical checks Trace to source documents Independent monitoring team

30 Multiple “Looks” Type 1 error = 1 – (1 – alpha) k Where k is the number of evaluations Number of “looks”Alpha error 10.05 20.10 30.14 40.19 50.23 100.40 Informal evaluations count

31 Stopping Rules Corresponds to p < 0.05 at each analysis

32 Interim Analyses & Stopping Rules Reasons trials are stopped early Ethics Money Regulatory issues Drug expiration Personnel Other opportunities Pre-defined interim analyses Spend alpha and beta power Avoid “convenience sample” Avoid “looking” between scheduled analyses Pre-defined stopping rules Efficacy versus futility

33 Potential Problems Poor compliance Patients Clinicians Drop-outs Crossovers Insufficient power Greater-than-expected variability Treatment effect smaller than anticipated

34 Fragile Results Consider two identical trials of treatment for infarction N=200 versus n=8,000 Which result do you believe? Which is biologically plausible? What happens if you add two events to each Rx group? Study A p=0.13 Study B p=0.02 TrialN Treatment Events Placebo Events RRP A200190.110.02 B4,0002002500.800.02

35 Four versus Five Rx for CML

36 Problem Solved?

37 How About Now?

38 Small Studies Often Wrong!

39 Multi-center Trials Advantages Necessary when large enrollment required Diverse populations increase generalizability of results Problems in individual center(s) balanced by other centers –Often required by Food and Drug Administration Disadvantages Difficult to enforce protocol –Inevitable subtle protocol differences among centers Expensive! “Multi-center” does not necessarily mean “better”

40 Unsupported Conclusions Beta error Insufficient detection power confused with negative result Conclusions that flat-out contradict presented results “Wishful thinking” — evidence of bias Inappropriate generalization: internal vs. external validity To healthier or sicker patients than studied To alternative care environments Efficacy versus effectiveness Failure to acknowledge substantial limitations Statistical significance ≠ clinical importance And the reverse!

41 Conclusion: Good Clinical Trials… Test a specific a priori hypothesis Evaluate clinically important outcomes Are well designed, with A priori and adequate sample size Defined stopping rules Are randomized and blinded when possible Use appropriate statistical analysis Make conclusions that follow from the data And acknowledged substantive limitations

42 Meta-analysis “Super analysis” of multiple similar studies Often helpful when there are many marginally powered studies Many serious limitations Search and selection bias Publication bias –Authors –Editors –Corporate sponsors Heterogeneity of results Good generalizability Rajagopalan, Anesthesiology 2008

43 Department of O UTCOMES R ESEARCH

44 Design Strategies Life is short; do things that matter! Is the question important? Is it worth years of your life? Concise hypothesis testing of important outcomes Usually only one or two hypotheses per study Beware of studies without a specified hypothesis A priori design Planned comparisons with identified primary outcome Intention-to-treat design General statistical approach Superiority, equivalence, non-inferiority Two-tailed versus one-tailed It’s not brain surgery, but…


Download ppt "Department of O UTCOMES R ESEARCH. Daniel I. Sessler, M.D. Professor and Chair Department of O UTCOMES R ESEARCH The Cleveland Clinic Clinical Research."

Similar presentations


Ads by Google