Some Problems evaluating large programs or program going in many different directions 1. the objectives are too general; they are really goals many.

Slides:



Advertisements
Similar presentations
Introduction to the Environment Rating Scales
Advertisements

Your Best Year Yet! Ten Questions for Making the Next Twelve Months Your Most Successful Ever.
Square Peg and Round Hole… As parents and educators, the change in grading systems requires a fundamental switch in our thinking… 4=A 1=F 2=D 3=B.
Goal Attainment Scales Definitions Examples Scoring.
Conceptualization, Operationalization, and Measurement
Measurement Concepts Operational Definition: is the definition of a variable in terms of the actual procedures used by the researcher to measure and/or.
Victorian Curriculum and Assessment Authority
ATTITUDE. Attitude is another Cognitive process like Personality. While personality is usually thought of as the whole person, attitude is said to be.
Reliability for Teachers Kansas State Department of Education ASSESSMENT LITERACY PROJECT1 Reliability = Consistency.
Student Learning Targets (SLT) You Can Do This! Getting Ready for the School Year.
© Cambridge International Examinations 2013 Component/Paper 1.
Algebra Problems… Solutions Algebra Problems… Solutions © 2007 Herbert I. Gross Set 4 By Herb I. Gross and Richard A. Medeiros next.
1 Contextual Value Added and Data For Dummies The mystery explained.
CHAPTER 3 ~~~~~ INFORMAL ASSESSMENT: SELECTING, SCORING, REPORTING.
HRM-755 PERFORMANCE MANAGEMENT OSMAN BIN SAIF Session: Three 1.
Evaluating Hypotheses
Uses of Language Tests.
Common Triggers for Progress Reports and Possible Solutions Bob DeLorme, EdS, RRT-NPS.
Lehrstuhl für Informatik 2 Gabriella Kókai: Maschine Learning 1 Evaluating Hypotheses.
Review Performance Management and Appraisal
1 Psych 5500/6500 Statistics and Parameters Fall, 2008.
© Curriculum Foundation1 Section 2 The nature of the assessment task Section 2 The nature of the assessment task There are three key questions: What are.
Near East University Department of English Language Teaching Advanced Research Techniques Correlational Studies Abdalmonam H. Elkorbow.
If you don’t know where you’re going, any road will take you there.
Lecture 12 Statistical Inference (Estimation) Point and Interval estimation By Aziza Munir.
Chapter 3 Understanding Test Scores Robert J. Drummond and Karyn Dayle Jones Assessment Procedures for Counselors and Helping Professionals, 6 th edition.
© 2006 McGraw-Hill Higher Education. All rights reserved. Numbers Numbers mean different things in different situations. Consider three answers that appear.
Steps In Appraisal Performance The performance appraisal process contains three steps. 1.Define the job. 2.Appraise.
This Outcome Report is based on data from patients who completed a Functional Restoration Programme (FRP) at the RealHealth Treatment Centre in Coventry.
Between groups designs (2) – outline 1.Block randomization 2.Natural groups designs 3.Subject loss 4.Some unsatisfactory alternatives to true experiments.
Intro to Research: How Psychologists Ask and Answer Questions Correlation and Description.
25 June 2009, London Impact significance in air quality assessment Application of EPUK criteria to road schemes?
THE RELATIONSHIP BETWEEN PRE-SERVICE TEACHERS’ PERCEPTIONS TOWARD ACTIVE LEARNING IN STATISTIC 2 COURSE AND THEIR ACADEMIC ACHIEVEMENT Vanny Septia Efendi.
A Solution Oriented Approach In Educational Settings The aim of this series of training sessions is to give an introduction to the principles of solution.
1 Psych 5500/6500 Standard Deviations, Standard Scores, and Areas Under the Normal Curve Fall, 2008.
Student Growth Measures in Teacher Evaluation Using Data to Inform Growth Targets and Submitting Your SLO 1.
Performance-Based Assessment HPHE 3150 Dr. Ayers.
Ms. Paschitti. What is your definition of success? bcitech.org/lpaschitti 2.
Understanding Theory and Research Frameworks. Theories help to: Establish a way to think about nursing phenomena Provide a link between concepts Clarify.
3 Some Key Ingredients for Inferential Statistics.
Assessment at KS4 Bury C of E High School Engaging Parents Information.
Differentiation What is meant by differences between learners?
Psychometrics. Goals of statistics Describe what is happening now –DESCRIPTIVE STATISTICS Determine what is probably happening or what might happen in.
Chapter 7 Measuring of data Reliability of measuring instruments The reliability* of instrument is the consistency with which it measures the target attribute.
SAT’s Information Parent’s Meeting 10 th February February 2016.
DESCRIPTIVE STUDIES Let’s get the facts, fool!. Descriptive Methods Methods that yield descriptions or behavior but not necessarily causal explanations.
What attitude towards assessment did you have as a high school student?
Assessment Assessment is the collection, recording and analysis of data about students as they work over a period of time. This should include, teacher,
In-Service Teacher Training Assessment in IGCSE Biology 0610 Session 2: Question papers and mark schemes.
© 2013 by Nelson Education1 Selection III: Interviewing.
Statistics 19 Confidence Intervals for Proportions.
Assessment Evening March Previously… Children’s attainment was described in terms of levels. 5b 5c 4a 4b 4c 3a 3b 3c 2a 2b 2c 1a 1b 1c Expectation.
ASSESSMENT WITHOUT LEVELS IN THE CULM VALLEY FEDERATION.
A2 unit 4 Clinical Psychology 4) Content Reliability of the diagnosis of mental disorders Validity of the diagnosis of mental disorders Cultural issues.
Inspiring today’s children for tomorrow’s world Early Years Foundation Stage Assessment Procedure 2016.
Evaluation Of and For Learning
A Growth Measure for ALL Students.
Module 5 HAIL Research This module provides an overview of how researchers will collect data with both Independent Living Specialist(ILS) and consumers.
PROGRAM & POLICY EVALUATION ‘Goal Attainment Scaling’
Assessment Theory and Models Part II
St Peter’s Catholic Primary
Overview: Understanding and Building a Schoolwide Assessment Plan
Skills for a Healthy Life
EVAAS Overview.
Introduction to Statistics for the Social Sciences SBS200, COMM200, GEOG200, PA200, POL200, or SOC200 Lecture Section 001, Spring 2016 Room 150 Harvill.
An Introduction to e-Assessment
THE RELATIONSHIP BETWEEN PRE-SERVICE TEACHERS’ PERCEPTIONS TOWARD ACTIVE LEARNING IN STATISTIC 2 COURSE AND THEIR ACADEMIC ACHIEVEMENT Vanny Septia Efendi.
Analyzing test data using Excel Gerard Seinhorst
Peer Support Master Class 2018
How to Design an Experiment: Part II
Presentation transcript:

Some Problems evaluating large programs or program going in many different directions 1. the objectives are too general; they are really goals many large program such as mental health outpatient programs have objectives such as “help clinets live satisfying lives”. Nlw it could be that they have poorly written objectives OR When there are objectives like this, they may have been written this broadly on purpose; because there are many clients and each client has particulalrized objectives that are peculiar to his or her own situation. In other words, each clients objectives are different. 2. there are no objectives only very general goals. The same is true here. It is possible that your job might be to help them write objectives as we discussed earlier OR they have purposefully chosen to NOT write narrow and specific objectives because there are many clients with a broad range of problems and each client has particularized objectives that are peculiar to his or her own situation. 3. even when very large program do have specific objectives, often there are so many clients each with his or her own focus, purpose goals and objectives, that it is impossible to use the goals and objectives of the program to evaluate the clients success or achievement in the program. What do I do then? In program evaluation, if you are evaluating each client with a different set of standard, then there is really no COHESIVE PROGRAM or at least no way to evaluate it as a PROGRAMMATIC WHOLE111

Evaluators must find some way to measure clients a. On either the same factor or b. Using some kind of standardized measure that allows different problems to be measured using the same measure

Solution a: A note on client satisfaction: client satisfaction is often the factor that is used for solution a. We will get more into client satisfaction in the future. However it is important to know that many evaluators consider “client satisfaction” a hollow measure. That means, it really does not measure anything meaningful. It would be like asking you “who is your favorite teacher?”. Although the answer would be nice to know, it is in no way, a measure of your learning! Solution b: Goal Attainment Scaling. GAS is a way to measure very different objectives, using the same measure. It is not a perfect solution!!!

Gas does not measure specific outcomes or objectives, as much as it converts each client’s progress toward their specific objective into a standard measure; that is GAS standardizedes levels of outcomes rather than specific outcomes.. in this way it is able to montor client achievement within a program EVEN WHEN THEY HAVE DIFFERENT KINDS OF PROBLEMS Steps in developing a GAS system 1. using numbers that mean something 2. writing individualized outcome anchors for each number on each goal 3. training staff to use GAS.

1. Using numbers that mean something In GAS measurement, a. each treatment objective for a single client is measured on the degree to which the client has achieved or under-achieved or exceeded that objective. b. each treatment objective is scored using a series of numbers that are indicative of a level of achievement toward that objective c. each number indicates a clients achievement on each of his or her particular objectives and also allows the evaluator to sompile all clients’ degree of achievement together, even to the point of giving a mean achievement score!!

1. Using numbers that mean something-con’t Usually, goal attainment scales use 5 or 7 measurement points. This number allows the researcher to create a ‘mid-point’ ( 3 on a 5 point scale and 5 on a 7 point scale, to establish the expected outcome The mid-point is the most important In goal attainment scaling. It establishes the Center for all the other measures. In other words, there is an equal number of points Below the mid-point and an equal number above. OR

1. Using numbers that mean something-con’t Think of the mid-point as the center Far below below average average above average far above Average average The mid-point is the most important In goal attainment scaling. It establishes the Center for all the other measures. In other words, there is an equal number of points below the mid-point and an equal number above. OR Much less less than meets expected better than much better Than expected expected outcome. expected than expected Outcomeoutcome outcome outcome

1. Using numbers that mean something-con’t This example, with variations, is found on likert type ratings used on many standardized scale or RAIs Far below below average average above average far above Average average In the 1 st example, the mid-point is the average or ‘neutral point’. In the 2 nd Example, the mid-point is the point where the expected outcome is actually achieved. This example is found on many “Goal Attainment Scales” Much less less than meets expected better than much better Than expected expected outcome. expected than expected Outcomeoutcome outcome outcome

1. Using numbers that mean something-con’t As you can see, both examples show a progression from ‘worst’ to ‘best’ But with goal attainment scales, the standard for meeting the objective is the mid-point. Goal attainment scales allow you to a. measure achievement of an objective (midpoint), b. lack of achievement (-1, -2), as well as c. surpassing an objective (+1, +2) Thus you can measure different objective for different clients, using the same scale of measurement

2. writing individualized outcome anchors for each number on each goal Although the numbers now mean something, they must also be meaningful for each individual client objective. An anchor is a description of what each number (level on the GAS) actually looks like for each client objective Anchors connect each client particular objective with the numbers on the scale Thus each client objective is “operationalized” on the same scale of measurement

2. writing individualized outcome anchors for each number on each goal Lets take a client who is depressed. Some of her individual symptoms/problems are: –Sleeps all the time she is not working –Continual thoughts of worthlessness –Trouble concentrating at work 7 finishing tasks Finally her overall score on the BDI upon admission is 37 (37= severe dep is the clinical cutoff for mild dep.)

2. writing individualized outcome anchors for each number on each goal Let’s write “expected outcomes” for those. Normally we would do this with the client. –Sleeps all the time she is not working = goes to bed at 9P.M. and takes no naps –Continual thoughts of worthlessness = no more than 1 thought /day re worthless and dismisses –Trouble concentrating at work & finishing tasks = focused completely on work and produces as as before depression And For the 37 on the BDI we might have scores no higher than 18 on the BDI We might typically expect that if a client achieved these outcomes, she would no longer be terribly affected by depression.

2. writing individualized outcome anchors for each number on each goal Now we would use each of those outcomes as the mid-point anchors on the GAS. Lets use negative thoughts & BDI no more than 1 thought /day re worthless and dismisses it score on BDI

2. writing individualized outcome anchors for each number on each goal As you can see, there is now room for both greater improvement as well as failure. The next step would be to write individualized anchors for the other attainment points; for example no more than 1 thought /day re worthless and dismisses it score on BDI Has occasional + Thoughts about Self along with no – Scores On BDI Has + thoughts About self Much of time & No neg. thought Scores 9 or below on BDI Still has same Thoughts of Worthlessness Score on BDI Remains thoughts of Worthlessness Have increased or Become more Severe BDI score worse 38 or above

As you can see, the GAS has problems –It is not really a true likert scale and it is not designed to be treated as interval level data (although people do). –It is highly subjective and depends on peoples (staff & clients) ability to come up with precise operational definitions of each number. –Its reliability and validity can be questioned –Using the scoring the way it is, small negative scores can be positive! –Te severity of clients individual problems and their ability to solve them, vary widely and this is NOT captured on GAS. It can make all problems look the same. (unless you “weight” each objective) –Training all staff to use it can be a major pain!

Although GAS can be a very useful tool when clients are not focused on program objectives, but individual ones, the key to its effective use is getting staff to use it. Training staff in the use of GAS is difficult and time-consuming.