Presentation is loading. Please wait.

Presentation is loading. Please wait.

SWPBS Forum October 2008 Claudia Vincent and Scott Spaulding University of.

Similar presentations

Presentation on theme: "SWPBS Forum October 2008 Claudia Vincent and Scott Spaulding University of."— Presentation transcript:

1 SWPBS Forum October 2008 Claudia Vincent and Scott Spaulding University of Oregon

2  Provide information about desirable features of SWPBS evaluation tools  Provide an overview of the extent to which SWPBS evaluation tools meet these desirable features

3 PBS Self-Assessment Implement systems to support practices Implement practices Improved student outcomes EVALUATION DATA Use eval data for decision-making Fidelity measures Student outcome measures Action Plan Interpret eval data 1.Drive implementation decisions 2.Provide evidence for SWPBS impact on student outcomes

4  A measure that drives implementation decisions should be: ◦ socially valid ◦ contextually appropriate ◦ sufficiently reliable (reliable enough to make defensible decisions) ◦ easy to use  A measure that builds the evidence base for SWPBS should: ◦ have known reliability ◦ have known validity ◦ clearly link implementation status to student outcomes

5  Measurement scores have two components: ◦ True score, e.g. a school’s true performance on “teaching behavioral expectations” ◦ Error, e.g. features of the measurement process itself  Our goal is to use tools that 1.maximize true score and minimize measurement error, and therefore 2.yield precise and interpretable data, and therefore 3.lead to sound implementation decisions and defensible evidence. True score (relevant to construct) Error (noise)

6  True score is maximized and error minimized if the evaluation tool is technically adequate, i.e. ◦ can be applied consistently (has good reliability) ◦ measures the construct of interest (has good validity)  Sound implementation decisions are made if the evaluation tool is practical, i.e. data ◦ are cost efficient to collect (low impact) ◦ are easy to aggregate across units of analysis (e.g. students, classrooms, schools, districts, states) ◦ are consistently used to make meaningful decisions (have high utility)

7  Consistency across ◦ Items/subscales/total scales (“internal consistency”) ◦ Data collectors (“inter-rater reliability” or “inter-observer agreement”) ◦ Time (“test-retest reliability”)

8  Definition: ◦ Extent to which the items on an instrument adequately and randomly sample a cohesive construct, e.g. “SWPBS implementation”  Assessment: ◦ If the instrument adequately and randomly samples one construct, and if it were divided into two equal parts, both parts should correlate strongly  Metric: ◦ coefficient alpha ( the average split-half correlation based on all possible divisions of an instrument into two parts)  Interpretation: ◦ α ≥.70 (adequate for measures under development) ◦ α ≥.80 (adequate for basic research) ◦ α ≥.90 (adequate for measures on which consequential decisions are based)

9  Definition: ◦ Extent to which the instrument measures the same construct regardless of who collects the data  Assessment: ◦ If the same construct were observed by two data collectors, their ratings should be almost identical  Metric: ◦ Expressed as percentage of agreement between two data collectors  Interpretation: ◦ ≥ 90% good ◦ ≥ 80% acceptable ◦ < 80% problematic

10  Definition: ◦ Extent to which the instrument yields consistent results at two points in time  Assessment: ◦ The measure is administered at two points in time. The time interval is set so that no improvement is expected to occur between first and second administration.  Metric: ◦ Expressed as correlation between pairs of scores from the same schools obtained at the two measurement administrations  Interpretation: ◦ r ≥.6 acceptable

11 How can we interpret this graph?

12  Interpretability of data!  Did these schools truly differ in the extent to which they taught behavioral expectations?  Or…did these schools obtain different scores because ◦ the tool’s items captured only some schools’ approach to teaching expectations? (tool lacked internal consistency) ◦ they had different data collectors? (tool lacked inter-rater agreement) ◦ some collected data in week 1 and some in week 2 of the same month? (tool lacked test-retest reliability)

13  Content validity  Criterion-related validity ◦ Concurrent validity ◦ Predictive validity  Construct validity

14  Definition: ◦ Extent to which the items on an instrument relate to the construct of interest, e.g. “student behavior”  Assessment: ◦ Expert judgment if items measure content theoretically or empirically linked to the construct  Metric: ◦ Expressed as percentage of expert agreement  Interpretation: ◦ ≥ 80% agreement desirable

15  Definition: ◦ Extent to which the instrument correlates with another instrument measuring a similar aspect of the construct of interest and administered concurrently or subsequently  Assessment: ◦ Concurrent validity: compare data from concurrently administered measures for agreement ◦ Predictive validity: compare data from subsequently administered measures for predictive accuracy  Metric: ◦ Expressed as a correlation between two measures  Interpretation: ◦ Moderate to high correlations are desirable ◦ Concurrent validity: Very high correlations might indicate redundancy of measures

16  Definition: ◦ Extent to which the instrument measures what it is supposed to measure (e.g. the theorized construct “student behavior”)  Assessment: ◦ factor analyses yielding information about the instrument’s dimensions (e.g. aspects of “student behavior”) ◦ Correlations between constructs hypothesized to impact each other (e.g. “student behavior” and “student reading achievement”)  Metric: ◦ Statistical model fit indices (e.g. Chi-Square)  Interpretation: ◦ Statistical significance

17 How can we interpret this graph?

18  Interpretability of data!  Can we truly conclude that student behavior is better in school F than school J? ◦ Does the tool truly measure well-defined behaviors? (content validity) ◦ Do student behaviors measured with this tool have any relevance for the school’s overall climate? For the student’s long-term success? (concurrent, predictive validity) ◦ Does the tool actually measure “student behavior”, or does it measure “teacher behavior”, “administrator behavior”, “parent behavior” ? (construct validity)

19  Consider sample size ◦ Psychometric data derived from large samples are better than psychometric data derived from small samples.  Consider sample characteristics ◦ Psychometric data derived from specific samples (e.g. elementary schools) do not automatically generalize to all contexts (e.g. middle schools, high schools).

20  Making implementation decisions based on evaluation data ◦ When has a school reached “full” implementation?  “Criterion” scores on implementation measures should be calibrated based on student outcomes Implementation student outcome goals criterion academic achievement 102030405060708090100 social achievement

21  Evaluation data lead to consequential decisions, e.g. ◦ Additional trainings when data indicate insufficient implementation ◦ Emphasis on specific supports where data indicate greatest student needs  To make sure we arrive at defensible decisions, we need to collect evaluation data with tools that ◦ have documented reliability and validity ◦ clearly link implementation to student outcomes

22 1. Collect evaluation data regularly 2. Collect evaluation data with tools that have good reliability and validity 3. Guide implementation decisions with evaluation data clearly linked to student outcomes

23  Provide information about desirable features of SWPBS evaluation tools  Provide an overview of the extent to which SWPBS evaluation tools meet these desirable features

24  How is my school doing?  My school is “80/80”. Now what?  My school is just beginning SWPBS. Where do I start?  How do we handle the kids still on support plans?  I’ve heard about school climate. What is that?  What about the classroom problems we still have?

25  Measurement within SWPBS  Research or evaluation?  What tools do we have?  What evidence exists for use of these tools?  Guidelines for using the measures

26  Focus on the whole school  School-wide PBS began with a focus on multiple systems  Evaluation of a process  Evaluation of an outcome  Growth beyond initial implementation Non- classroom Classroom Individual Student School-wide Systems Sugai & Horner (2002)

27 Primary Prevention: School-/Classroom- Wide Systems for All Students, Staff, & Settings Continuum of School-wide Positive Behavior Support ~80% of Students ~15% ~5% Secondary Prevention: Specialized Group Systems for Students with At-Risk Behavior Tertiary Prevention: Specialized Individualized Systems for Students with High-Risk Behavior

28 Student Unit of Measurement and Analysis School Nonclassroom Classroom ? Academics Behavior Academics Behavior Level of Prevention and Intervention Dimension of Measurement Academic Achievement Social Behavior ? ? ? ? ? ? ? ? Tertiary Secondary Primary OutcomesProcess

29  Measurement within SWPBS  Research or evaluation?  What tools do we have?  What evidence exists for use of these tools?  Guidelines for using the measures

30 1. Drive implementation decisions 2. Provide evidence for SWPBS impact on student outcomes  Measures have developed to support research-quality assessment of SWPBS  Measures have developed to assist teams in monitoring their progress

31  Measurement within SWPBS  Research or evaluation?  What tools do we have?  What evidence exists for use of these tools?  Guidelines for using the measures

32 Some commonly used measures:  Effective Behavior Supports Survey  Team Implementation Checklist  Benchmarks of Quality  School-wide Evaluation Tool  Implementation Phases Inventory

33 Newer measures:  Individual Student Schoolwide Evaluation Tool  Checklist for Individual Student Systems  Self-assessment and Program Review

34 Whole-SchoolNon-classroomClassroom TertiaryISSET CISS SecondaryISSET CISS UniversalEBS TIC SET BoQ EBS

35  Measurement within SWPBS  Research or evaluation?  What tools do we have?  What evidence exists for use of these tools?  Guidelines for using the measures

36  Is it important, acceptable, and meaningful?  Can we use it in our school?  Is it consistent?  Is it easy to use?  Is it “expensive”?  Does it measure what it’s supposed to?  Does it link implementation to outcome?

37  Effective Behavior Supports Survey (EBS)  School-wide Evaluation Tool (SET)  Benchmarks of Quality (BoQ)

38  Effective Behavior Supports Survey ◦ Sugai, Horner, & Todd (2003) ◦ Hagan-Burke et al. (2005) ◦ Safran (2006) Internal consist. T-RInter-raterContentCriterionConstruct ╳

39  46-item, support team self-assessment  Facilitates initial and annual action planning  Current status and priority for improvement across four systems: ◦ School-wide ◦ Specific Setting ◦ Classroom ◦ Individual Student  Summary by domain, action planning activities  20-30 minutes, conducted at initial assessment, quarterly, and annual intervals

40  Internal consistency ◦ Sample of 3 schools ◦ current status: α =.85 ◦ improvement priority: α =.94 ◦ Subscale α from.60 to.75 for “current status” and.81 to.92 for “improvement priority”  Internal consistency for School-wide ◦ Sample of 37 schools ◦ α =.88 for “current status” ◦ α =.94 for the “improvement priority”

41  School-wide Evaluation Tool ◦ Sugai, Horner & Todd (2000) ◦ Horner et al. (2004) Internal consist. T-RInter-raterContentCriterionConstruct ╳╳╳╳╳

42  28-item, research evaluation of universal implementation  Total implementation score and 7 subscale scores: behavioral expectations behavioral expectations taught 3.acknowledgement system 4.consequences for problem behavior 5.system for monitoring of problem behavior 6.administrative support 7.District support  2-3 hours, external evaluation, annual

43  Internal consistency ◦ Sample of 45 middle and elementary schools ◦ α =.96 for total score ◦ α from.71 (district-level support) to.91 (administrative support)  Test-retest analysis ◦ Sample of 17 schools ◦ Total score, IOA = 97.3% ◦ Individual subscales, IOA = 89.8% (acknowledgement of appropriate behaviors) to 100% (district-level support)

44  Content validity ◦ Collaboration with teachers, staff, and administrators at 150 middle and elementary schools over a 3-year period

45  Construct validity ◦ Sample of 31 schools ◦ SET correlated with EBS Survey ◦ Pearson r =.75, p <.01  Sensitivity to differences in implementation across schools ◦ Sample of 13 schools ◦ Comparison of average scores before and after implementation ◦ t = 7.63; df = 12, p <.001

46  Schoolwide Benchmarks of Quality ◦ Kincaid, Childs, & George (2005) ◦ Cohen, Kincaid, & Childs (2007) Internal consist. T-RInter-raterContentCriterionConstruct ╳╳╳╳╳

47  Used to identify areas of success / improvement  Self-assessment completed by all team members  53-items rating level of implementation  Team coaches create summary form, noting discrepancies in ratings  Areas of strength, needing development, and of discrepancy noted for discussion and planning  1-1.5 hours (1 team member plus coach)  Completed annually in spring

48  Items grouped into 10 subscales: 1.PBS team 2.faculty commitment 3.effective discipline procedures entry 5.expectations and rules 6.reward system 7.lesson plans for teaching behavioral expectations 8.implementation plans 9.crisis plans 10.evaluation

49  Internal consistency ◦ Sample of 105 schools ◦ Florida and Maryland  44 ES, 35 MS, 10 HS, 16 center schools ◦ overall α of.96 ◦ α values for subscales .43 “PBS team” to .87 “lesson plans for teaching expectations”

50  Test-retest reliability ◦ Sample of 28 schools ◦ Coaches scores only ◦ Total score: r =.94, p < 0.01 ◦ r values for subscales:  0.63 “implementation plan” to  0.93 “evaluation”  acceptable test-retest reliability  Inter-observer agreement (IOA) ◦ Sample of 32 schools ◦ IOA = 89%

51  Content validity ◦ Florida PBS training manual & core SWPBS elements ◦ Feedback from 20 SWPBS research and evaluation professionals ◦ Interviewing to identify response error in the items ◦ Pilot efforts with 10 support teams  Concurrent validity ◦ Sample of 42 schools ◦ Correlation between BoQ and SET ◦ Pearson r =.51, p <.05

52  Measurement within SWPBS  Research or evaluation?  What tools do we have?  What evidence exists for use of these tools?  Guidelines for using the measures

53  What measures do I use?  How do I translate a score into “practice”?  What school variables affect measurement choices? ◦ SWPBS implementation status  Evaluation template

54 Fidelity ToolYear 1Year 2Year 3 EBS SurveyX 123412341234 Universal TIC XXXXXXXXX SET / BoQ XXX Secondary / tertiary CISS XXXXXXXXX ISSET XXX Classroom setting Class internal XXXXXXXXX Class external XXX

55  Evaluation of School-wide PBS occurs for implementation and outcomes  Evidence of a “good” measure depends on its intended use  The quality of implementation decisions depends on the quality of evaluation tools  Evaluation occurs throughout the implementation process, with different tools for different purposes at different stages

Download ppt "SWPBS Forum October 2008 Claudia Vincent and Scott Spaulding University of."

Similar presentations

Ads by Google