Expanded CONSORT Figure for Planning and Reporting D & I Research

Slides:



Advertisements
Similar presentations
Template: Making Effective Presentation about Your Evidence-based Health Promotion Program This template is intended for you to adapt to your own program.
Advertisements

Introduction to the unit and mixed methods approaches to research Kerry Hood.
Using RE-AIM as a tool for Program Evaluation From Research to Practice.
A brief overview What is program evaluation? How is an evaluation conducted? When should it be used? When can it be used? Used with Permission of: John.
Evaluation is a professional and ethical responsibility and is a core part of PHN professional practice Commitment to evaluation helps build the PHN intelligence.
Title slide PIPELINE QRA SEMINAR. PIPELINE RISK ASSESSMENT INTRODUCTION TO GENERAL RISK MANAGEMENT 2.
DC Home visiting Implementation and impact evaluation
STUDY PLANNING & DESIGN TO ENHANCE TRANSLATION OF HEALTH BEHAVIOR RESEARCH Lisa Klesges, Russell Glasgow, Paul Estabrooks, David Dzewaltowski, Sheana Bull.
Program Evaluation. Program evaluation Methodological techniques of the social sciences social policy public welfare administration.
The Audit Process Tahera Chaudry March Clinical audit A quality improvement process that seeks to improve patient care and outcomes through systematic.
Systematic Reviews.
Barbara Resnick, PhD, CRNP, FAAN, FAANP
Implementation and process evaluation: developing our approach Ann Lendrum University of Manchester Neil Humphrey University of Manchester Gemma Moss Institute.
Implementation Science: Finding Common Ground and Perspectives Laura Reichenbach, Evidence Project, Population Council International Conference on Family.
[Presentation location] [Presentation date] (Confirm ABT logo) Building Bridges and Bonds (B3): An introduction.
PRAGMATIC Study Designs: Elderly Cancer Trials
The PRECIS-2 tool: Matching Intent with Methods David Hahn, MD, MS, WREN Director Department of Family Medicine & Community Health University.
1 HSRC EPWP SEMINAR 07 JUNE 2016 Presentation by Lulu Mdletshe EPWP Co-ordinator: Department of Transport.
2 What’s in this presentation? We are seeking the board’s approval and advice on beginning a programme of work on culture and leadership across our trust.
Module 8 Guidelines for evaluating the SDGs through an equity focused and gender responsive lens: Overview Technical Assistance on Evaluating SDGs: Leave.
Overview of Intervention Mapping
Translating Research Into Practice: Pragmatic Research Approaches
Russell E. Glasgow, PhD (and Bridget Gaglio, PhD- PCORI )
Incorporating Evaluation into a Clinical Project
The Visual Storyteller:
Performance Improvement Project Validation Process Outcome Focused Scoring Methodology and Critical Analysis Presenter: Christi Melendez, RN, CPHQ Associate.
Evaluation Plan Akm Alamgir, PhD June 30, 2017.
WASL Science: Grade 10 Specific Title Slide for School.

Bringing About Cultural Change Among Providers
Russell Glasgow Marina McCreight Borsika Rabin
Literature review Methods
Right-sized Evaluation
D&I Grant Funding: Tips for Success and Possible Futures
Supplementary Table 1. PRISMA checklist
Welcome slide.
Performance Improvement Project Validation Process Outcome Focused Scoring Methodology and Critical Analysis Presenter: Christi Melendez, RN, CPHQ Associate.
Conducting Efficacy Trials
CENTER FOR COORDINATION OF RESEARCH INTERNATIONAL FEDERATION OF CATHOLIC UNIVERSITIES  (IFCU) Continuing of Education for Disadvantaged Adolescents in.
Chapter 16 Nursing Informatics: Improving Workflow and Meaningful Use
YSGOL GWYDDORAU GOFAL IECHYD / SCHOOL OF HEALTHCARE SCIENCES
Asset Governance – Integrated Strategic Asset Management
Reading Research Papers-A Basic Guide to Critical Analysis
RESEARCH IMPLEMENTATION PRACTICE
What Can Researchers Do to Address Health Inequities
Developing a Health Maintenance Schedule
Assessed for eligibility (N = )
Pilot Studies: What we need to know
Fralin Translational Obesity Research Center
Performance Improvement Poster Title Author(s) Name(s), Degree, License, Certification, Institution and/or Department Affiliations Focus Develop Evaluate.
Dissemination and Implementation Research: Past, Present, and Future (One View) Russell E. Glasgow, PhD Department of Family Medicine, Univ. Colorado.
Describe and Evaluate the Cognitive Treatment for Schizophrenia
12/9/2018 3:38 PM Quality Improvement Plans (QIPs): Aligning the Content to the QIP Guidelines.
Good Practices on Disaster Prevention in Europe
Project Title Subtitle: make sure to specify that project is an improvement project (see SQUIRES elaboration article Presenter(s) Date of presentation.
Presenter: Kate Bell, MA PIP Reviewer
Experience Applying Pragmatic Models for D&I Research: Current and Future Directions for the RE-AIM Planning and Evaluation Model Russell E. Glasgow, PhD.
SUSTAINABLE MICRO-FINANCE for WOMEN’S EMPOWERMENT
What are systematic reviews and why do we need them?
Group work on challenges Learning oriented M&E System
Considering Fidelity as an Element Within Scale Up Initiatives Validation of a Multi-Phase Scale Up Design for a Knowledge-Based Intervention in Science.
Chapter 15 Community As Client: Applying the Nursing Process
STROBE Statement revision
Contemporary Issues of HRM
Program Planning: Models and Theories
Some Further Considerations in Combining Single Case and Group Designs
Sustainability of Organizational Change
Integrating Gender into Rural Development M&E in Projects and Programs
Picking and Applying a Dissemination & Implementation Science Theory
APMP Professional Certification
Presentation transcript:

Expanded CONSORT Figure for Planning and Reporting D & I Research Russell E. Glasgow, PhD Amy Huebschmann, MD, MS and friends ACCORDS Dissemination and Implementation Science Program Departments of Family and General Internal Medicine www.ucdenver.edu/accords/implementation www.evaluationhub.org

D & I Issues in Pragmatic Research External Validity/Pragmatic Criteria, often ignored Participant Representativeness Setting Representativeness Multi-level Context and Setting Adaptation/Change in Intervention and Implementation Strategies Sustainability Reasons for Participation and Drop Out

WHY Are These Important? Understanding Context Generalizability Transparency in Reporting Address Failure to Replicate

The RE-AIM Framework and External Validity Reporting www.re-aim.org Focus on enhancing: Reach – Participation rates and representativeness Effectiveness – Breadth (quality of life), including negative or unintended effects Adoption - Setting and staff participation rate and representativeness Implementation – Consistency, adaptation and costs to deliver the program Maintenance – Extent to which program and effects are sustained References ok, Russ? Gaglio B, et al. The RE-AIM Framework:  A systematic review of use over time.  Am J Public Health. 2013 Jun;103(6):e38-46. Kessler RS, et al. What does it mean to “Employ” the RE-AIM Model?  Eval Health Prof. 2012 Mar; 36(1):44-66.

Key Translation and Pragmatic Questions to Consider for the RE-AIM Dimensions Key Pragmatic Questions to Consider and Answer Reach Who is intended to benefit, and who actually participates or is exposed to the intervention? Effectiveness What is the most important benefits you are trying to achieve and what is the likelihood of negative outcomes? Adoption Where is the program or policy applied, and WHO applied it? Implementation How consistently is the program or policy delivered; How will it be adapted; How much will it cost; and How/Why will the results come about? Maintenance When will the initiative become operational; how long will it be sustained (setting level); and how long are the results sustained (individual level)? I repasted this figure…didn’t capture the first column initially.

Example of Translation of Interventions into Practice Why Is This Important? Consider the DPP Case Study Impact of Loss at Each RE-AIM Concept Example of Translation of Interventions into Practice Dissemination Step RE-AIM Concept % Impact 50% of settings use intervention Adoption 50.0% 50% of staff take part 25.0% 50% of patients identified, accept Reach 12.5% 50% follow regimen correctly Implementation 6.2% 50% benefit from the intervention Effectiveness 3.2% 50% continue to benefit after six months Maintenance 1.6% www.re-aim.org

CONSORT Figure CONSORT Figure and/or checklist required by most leading healthcare journals Use associated with improved reporting and quality Focused predominantly on internal validity AMY- do you have basic CONSORT. REF or 2? LINDA- Need consistency of font size and styles across slides Russ…where possible I have created the titles in 32 Arial and the titles in the body, 28 Arial and the bullets below 24 Arial…not always possible given amount, etc. Also, I removed the ACCORDS ‘bit’ at the bottom of the slides as it was repetitive of what’s above and we needed the space. (Linda)

Assessed for eligibility (n= ) Enrollment Assessed for eligibility (n= ) Excluded (n= ) ¨  Not meeting inclusion criteria (n= ) ¨  Declined to participate (n= ) ¨  Other reasons (n= ) Randomized (n= ) Allocated to intervention (n= ) ¨ Received allocated intervention (n= ) ¨ Did not receive allocated intervention (give reasons) (n= ) ¨ Did not receive allocated intervention (why?) (n= ) Allocation Lost to follow-up (give reasons) (n= ) Discontinued intervention (why?) (n= ) Discontinued intervention (give reasons) (n= ) Follow-Up Analysed (n= ) ¨ Excluded from analysis (give reasons) (n= ) Analysis AMY- do you have? Thought you did one? I cannot find Is this what you had in mind, Russ?

Excluded by Investigator CONSORT Figure - Reach Total Potential Participants (n) Individuals Eligible (n and %) Individuals Decline Not Contacted/Other Excluded by Investigator (n, %, and reasons) Critical Considerations: -Characteristics of Enrollers vs. Decliners

Adoption: Setting and Staff Issues 1. Setting Level Adoption Factor Settings Who is invited/excluded Participation rate Who participates How representative are they Reasons for participating/declining 2. Staff Level Adoption Factors Linda- Messed up- letter under 2 should be a, b, c, etc. NO ‘f”- this is supposed to be just main point 2 – DONE (Linda)

Expanded CONSORT issue – Adoption Total Number Potential Settings Settings Eligible (n and %) Settings and Agents Who Participate Settings and Agents Who Decline Other Excluded by Investigator (n, %, and reasons) Critical Considerations: -Characteristics of Adopting Settings vs. Non -Characteristics of Adopting Staff vs. Non

First Expanded CONSORT Figure Total Number Potential Settings Settings Eligible (n and %) Settings and Agents Who Participate (n and %) Settings and Agents Who Decline Other Excluded by Investigator (n, % and reasons) Total Potential Participants (n) Individuals Eligible (n and %) Individuals Enroll (n and %) Individuals Decline (n, %, and reason) Not Contacted/Other (n and %) Excluded by Investigator (n, %, and reasons) ADOPTION REACH RE-AIM Issue

DPP Example Adoption Data Total Number Potential Settings n=100 Settings Eligible n=75 Settings (and Agents) Who Participate n=40 (and 20) Settings (and Agents) Who Decline and Reasons n=30 (and 20) Other n=5 (and 5) Excluded by Investigator n=25 and reasons Total Potential Participants n=1000 Individuals Eligible n=800 Individuals Enroll n=400 Individuals Decline n=300 and reason Not Contacted/Other n=100 Excluded by Investigator n=200 and reasons ADOPTION REACH RE-AIM Issue

Implementation Factors Overall Level of Delivery of Key Components ‘Fidelity’ or Consistency of Delivery Variability in Implementation and Outcomes Across: Staff Setting Time Patient sub-group Adaptations made, same issues as above Reasons for high and low levels

Expanded CONSORT issues – Implementation & Effectiveness Extent Tx Delivered by Different Agents as in Protocol Adaptations and Reasons Component A = XX% Component B = YY% etc. IMPLEMENTATION Complete Tx (n and %) Amount of Change (by Condition) Drop Out of Tx (n, %, and reasons) Amount of Change EFFECTIVENESS RE-AIM Issue Critical Considerations: -Factors Related to Fidelity (Implementation) and Adaptation -Characteristics of Drop-outs vs Completers (Effectiveness)

Second Expanded CONSORT Figure Total Number Potential Settings Settings Eligible (n and %) Settings and Agents Who Participate (n and %) Settings and Agents Who Decline Other Excluded by Investigator (n, % and reasons) Total Potential Participants (n) Individuals Eligible (n and %) Individuals Enroll (n and %) Individuals Decline (n, %, and reasons) Not Contacted/Other Excluded by Investigator (n, % and reasons) ADOPTION REACH RE-AIM Issue Extent Tx Delivered by Different Agents as in Protocol Adaptations and Reasons Component A = XX% Component B = YY% etc. IMPLEMENTATION Complete Tx (n and %) Amount of Change (by Condition) Drop Out of Tx (n, %, and reasons) Amount of Change (by Condition) EFFECTIVENESS

Maintenance or Sustainability Level or Step Patient Level: longer term follow-up than usual When do you lose patients? Who is lost? Why are they lost or stop benefitting? Setting Level Sustainability or Maintenance Do settings continue after research is completed? Do they discontinue program entirely? Do they modify/adapt the program or policy? Why?? Chambers D and Glasgow RE (2013). The Dynamic Sustainability Framework: Addressing the Paradox of Sustainment amidst Ongoing Change. Implementation Science, 8(1):117.

Expanded CONSORT issue – Maintenance Present at Follow-up (n and %) and Amount of Change or Relapse (By Condition) Lost to Follow-up (n, %, and reasons) Amount of Change or Relapse (By Condition) A) Individual Level Setting in which Program in Continued and/or Modified After Research is Over (n, %, and reasons) Settings in which Program Adapted (n, %, and reasons) B) Setting Level Settings in which Program Discontinued (n, %, and reasons) Critical Considerations: -Characteristics of Drop-outs vs. Completers (Individual) -Characteristics of Settings the Continue vs. Do Not

Expanded CONSORT Figure Total Number Potential Settings Settings Eligible (n and %) Settings and Agents Who Participate(n and %) Settings and Agents Who Decline(n and %) Other (n and %) Excluded by Investigator (n, %, and reasons) Total Potential Participants (n) Individuals Eligible (n and %) Individuals Enroll (n and %) Individuals Decline (n, %, and reasons) Not Contacted/Other (n and %) Excluded by Investigator (n, %, and reasons) ADOPTION REACH RE-AIM Issue Extent Tx Delivered by Different Agents as in Protocol Adaptations and Reasons Component A = XX% Component B = YY% etc. IMPLEMENTATION Complete Tx (n and %) Amount of Change (by Condition) Drop Out of Tx (n, %, and reasons) Amount of Change (by Condition) EFFECTIVENESS Present at Follow-up (n and %) and Amount of Change or Relapse (By Condition) Lost to Follow-up (n, %, and reasons) Amount of Change or Relapse (By Condition) Setting in which Program is Continued and/or Modified After Research is Over (n, %, and reasons) Settings in which Program Adapted (n, %, and reasons) MAINTENANCE Settings in which Program Discontinued (n, %, and reasons)

Questions? Comments thus far? Do we have a graphic. Cartoon..something to indicate questions- someone raising hand; Rodin thinkng statue, interacting students in classroom? Some ideas

Qualitative Reporting Issues To understand how and why results happen (especially reasons for participation and declining, in this case) To provide more in-depth understanding of characteristics of those who participate vs. those who do not Linda- problems with right hand side DONE (Linda)

Challenges to Using the Expanded CONSORT Takes too much space Already have to report CONSORT, figure added little No money or time to collect necessary data No room in text for qualitative reasons ‘why and how’ Do not have data on many of these issues Do not want to know he answers or see these data Requires culture change

Why is it Important to Use the Expanded CONSORT Why is it Important to Use the Expanded CONSORT? (or similar reporting approach) To enhance transparency and indicate generalizability factors To increase the frequency and quality of external validity reporting To address the replication failure crisis To succinctly and visually summarize key reporting issues To align with evolving D&I reporting criteria (StaRI)

Standardized Criteria for Reporting Implementation Research (StaRI) 27 reporting criteria identified by systematic review and international experts, including: both implementation strategies and intervention effects related to, but expansion of, CONSORT criteria reporting of context, including SDOH and policy issues cost and economic outcomes fidelity and adaptations harm and unintended effects representativeness and estimated scalability Pinnock H. Barwick M, Carpenter CR, Eldridge S, Grandes G, et al. Standards for reporting implementation studies (StaRI) statement.  BMJ 2017;356:i6795.

Summary The Expanded CONSORT provides an efficient, transparent way to summarize key contextual and external validity issues Greater use of it, along with StaRI and similar checklists, can enhance the quality of reporting and attention to these key D&I issues Linda- problems with right hand slide. (I do not like this template.  I know what you mean. DONE (Linda)

Questions? Comments? I’m all ears.

References and Resources