Presentation is loading. Please wait.

Presentation is loading. Please wait.

Mountains or Molehills? Reducing the perceived burden of monitoring and evaluating outcomes.

Similar presentations

Presentation on theme: "Mountains or Molehills? Reducing the perceived burden of monitoring and evaluating outcomes."— Presentation transcript:

1 Mountains or Molehills? Reducing the perceived burden of monitoring and evaluating outcomes

2 Rebecca Colclough Dr Mark Llewellyn

3 Workshop Aim To provide participants with useful background information and a practical step -by-step guide to planning your evaluation approach and developing a simple and effective outcome monitoring questionnaire for your project

4 STEP 1: Digging Around Use simple Google searches to do some research into your area of interest What evaluation tools and techniques are similar organisations using? Some common systems are available including Outcome Stars and Quality of Life Indicators: check if these could be appropriate Take stock each time you start a new project Talk to other people doing similar work

5 STEP 2: Set your Outcomes “Outcomes are the difference your project can make to your community or beneficiaries” Big Lottery

6 SMART Outcomes Specific Measurable Achievable Realistic Time-bound –Significant –Meaningful –Attainable –Relevant –Resourced –Results-based –Timely Outcomes need to be SMART before you can begin to work out how to measure whether you have achieved them or not!

7 Example 1 OUTCOME Increased confidence SMART OUTCOME Older people who participate in the programme will evidence an increase in their confidence to speak up for themselves, which will be shown immediately following their participation, and will still be evident 6 months later

8 Example 2 OUTCOME Increased use of recreational facilities SMART OUTCOME The number of people using the recreational facilities will be evidenced to have increased by 10% in the first year of the project and by 20% by the end of the project Or The number of people who report using more than one of the recreational facilities at the Centre will have increased from 15% to 20% in the first year of the project and will reach 25% by the end of the project

9 STEP 3: Evaluation Questions Turn your SMART outcomes into evaluation questions Remember, evaluation is COMPARATIVE by nature  past with present  project 1 with project 2  method a) with method b)  before and after

10 GOOD evaluation questions Easy to understand Written in plain language (no jargon or abbreviations) Not too long Only ask one thing per question Don’t ask ‘leading’ questions Offer mutually exclusive and exhaustive categories Can be closed or open-ended Can use recognised ‘scales’

11 Ordinal Scales (LIKERT) ExcellentVery goodFairPoor Strongly agree AgreeNeither agree nor disagree DisagreeStrongly disagree AlwaysOftenSometimesRarelyNever NoneVery mildMildModerateSevere Completely satisfied Very satisfied Somewhat satisfied Somewhat dissatisfied Very dissatisfied Completely dissatisfied

12 Developing Evaluation Questions SMART Outcome: Older people who participate in the programme will evidence an increase in their confidence to speak up for themselves, which will be shown immediately following their participation, and will still be evident 6 months later

13 …continued Evaluation Metric: What % of beneficiaries report an improvement in their confidence to speak up for themselves?

14 …continued Evaluation Question: To make this evaluation question comparative you must ask it at least twice: BEFORE and AFTER your project intervention (in this case the question must also be asked 6 months later as well)

15 STEP 4: Develop your Evaluation Plan Develop an evaluation plan which details all of your intended project outcomes and evaluation questions, and sets out how you will collect the data to monitor your achievement of these, alongside wider service monitoring and evaluation

16 Evaluation Plan: Contents (i) Data reporting deadlines Service monitoring and evaluation questions (RBA ‘How much are we doing’ and ‘How well are we doing it’?) SMART outcomes (RBA ‘Is anyone better off’?) Outcome evaluation questions

17 Evaluation Plan: Contents (ii) Use of monitoring and evaluation data  To meet funding requirements?  To influence social policy?  To gain further project funding?  To scope a particular issue or problem? Evaluation participants (who you will collect the data from)?  Beneficiaries?  Volunteers?  Stakeholders?

18 Evaluation Plan: Contents (iii) The type of data you will collect The data collection methods you will use

19 Evaluation Plan: Contents (iv) Analysis How you will analyse your results Reporting What kind of reports you will need to produce from your results Dissemination How your findings will be shared to have the greatest impact Unfortunately there won’t be time today to go into these areas in more detail

20 Quantitative / Qualitative Quantitative data is essentially NUMBERS Qualitative data is essentially WORDS

21 Quantitative Learning little from many: can make generalisations Standardised responses Non-deliberative Numerical Based on sample sizes Learning a lot from a few: hard to make generalisations Open responses; conversational and flexible Deliberative Non-numerical; ideas; beliefs; meaning Based on experience Qualitative

22 What kinds of data can be collected? “Everything that can be counted does not necessarily count; everything that counts cannot necessarily be counted” Albert Einstein

23 Data Collection Methods Different types of data require different data collection methods to be used Quantitative Survey Routine monitoring data collection (stats) Qualitative Interview Focus group Digital story


25 Ensuring Quality Data Making sure you collect the right data at the right time and in a standardised way is essential to making sure the data you collect, analyse and share is good quality Different ‘methods’ use different ‘tools’ to do this, such as a questionnaire, a structured interview schedule or a focus group discussion guide

26 What is an Outcome TOOL? An outcome tool is something designed to ensure you can assess or describe change in a consistent and standardised way. This will produce data which is of a better quality Good quality data makes it possible to draw together information on a number of beneficiaries and give an overview of the change achieved for a whole project or group of projects

27 Developing an Outcome Tool for your Project Measuring outcomes for beneficiaries


29 Simple and Effective Make the TOOL as useful as possible Simple design which is easy to follow Gather data which answers your evaluation questions Use recognised SCALES where appropriate to ensure quality Think about when and how the TOOL is used

30 Service Quality Take the opportunity to assess the QUALITY of your service along with beneficiary outcomes

31 Free Text Space It is always useful to include free text space for beneficiaries to say more about something Whilst not everyone will complete free text space, those who do will often have something constructive or positive to say You can get great quotes for future use!

32 Capture project ‘champions’ Make sure there is space for beneficiaries to fill in their contact details if they wish Take the opportunity to ask if they would mind being interviewed or contributing further to your project evaluation User voice is exceptionally powerful so don’t risk losing it!

33 Simple Structure Order the questions well according to any natural timelines Move from familiar to unfamiliar Don’t ask sensitive questions near the beginning Limit the number of questions End with easy questions Make the completion instructions clear

34 Data Collection Points “Timing has a lot to do with the outcome of a rain dance” Cowboy Proverb

35 Data Collection Points Baseline Progress End of Intervention Impact You must ask exactly the same questions each time when trying to evaluate outcomes. Service monitoring and quality questions can be one-off

36 Completion Method Face to face By post By the client alone or with support By a proxy (for people with dementia) By a volunteer Standardising your data collection method helps ensure the data gathered is high quality

37 Useful Resources Link to a document called ‘Explaining the difference your project makes’ Power point module on writing good evaluation questions Slides on good practice in writing evaluation questions %20Tools%20Second%20Edition.pdf Really useful publication describing outcome tools (written for the homelessness sector but more widely applicable)

Download ppt "Mountains or Molehills? Reducing the perceived burden of monitoring and evaluating outcomes."

Similar presentations

Ads by Google