Presentation is loading. Please wait.

Presentation is loading. Please wait.

1 INFORMAL ASSESSMENT. Usefulness of Norm-Referenced Tests Norm-referenced tests are standardized methods for assessing skills or behaviors. They are.

Similar presentations


Presentation on theme: "1 INFORMAL ASSESSMENT. Usefulness of Norm-Referenced Tests Norm-referenced tests are standardized methods for assessing skills or behaviors. They are."— Presentation transcript:

1 1 INFORMAL ASSESSMENT

2 Usefulness of Norm-Referenced Tests Norm-referenced tests are standardized methods for assessing skills or behaviors. They are useful for making statistically accurate comparisons between the performance of a target student and the performances of other students of the same age or grade level. With the results of norm referenced tests, it is possible to determine whether a target student ’ s performance is typical or atypical, based on national norms. 2

3 Problems with Norm-Referenced Tests Norm-referenced tests are useful in assessing students for eligibility for special education, but they do have significant limitations; such as… May not accurately reflect curriculum being taught Have limited numbers of alternative forms; problems exist with “ test wiseness ” May not be sensitive to small gains in academic growth Problems of bias exist in selection of items 3

4 Informal Assessment Informal assessment means using non-standardized methods of diagnosing learning problems and measuring student progress. 4

5 Advantages of Informal Assessment Informal assessment has the following advantages over norm-referenced testing: Can be more closely related to the curriculum More sensitive to small gains Less cumbersome to administer and score Relates directly to planning instruction and teaching Can identify specific error patterns 5

6 Comprehensive Assessment= Comprehensive Test Battery 6

7 Types of Informal Assessment criterion-related assessment--an assessment which involves comparing a student ’ s performance to a given criteria (rather than to a norm group; mastery testing) curriculum-based assessment--tests which use excerpts from the general education curriculum as the subject matter for testing direct measurement--measuring progress by using the same instructional materials or tasks that are used in the classroom probes--brief tests used for assessment of mastery of specific skill or sub-skill 7

8 Teacher-Made Tests When designing informal instruments research has shown that teachers are prone to make errors that may skew the results. E.g., matching items, followed by completion, essay and true/false. Teachers may write test items using different levels of learning, although, many teachers use items at the knowledge level because they are easier to write. Such items require the student merely to recall, recognize, or match the material. Higher order thinking skills are needed to assess a student ’ s ability to sequence, apply information, analyze, synthesize, infer, or deduct. 8

9 Criterion-Referenced Testing Criterion-referenced tests (CRTs) compare the performance of a student to a given criterion for mastery. Criterion-referenced testing can be used to determine the examinee ’ s position along the continuum from acquisition to mastery. To be accurate, criterion-referenced tests must have “ item density, ” enough items in each domain to make sure that the topic is covered adequately. The advantages of CRTs include: Practical Fair Assists with measuring educational accountability

10 Sources for CRTs Adapt existing norm-referenced instruments Use published criterion-referenced tests (like The Brigance Inventories) Design teacher-made CRTs – Curriculum-based – Direct measurement 10

11 Establishing Criterion With published CRTs, the authors provide a criterion for mastery on their instrument. When a teacher designs a CRT, the teacher must determine an appropriate mastery criterion. Some tasks require 100% mastery (e.g., math facts) and others can tolerate a lower standard like 80 or 90% (e.g., reading comprehension). Typical criterion for mastery are listed below: More than 95% = mastery of objective 90% to 95% = instructional level 76% to 89% = difficult level Less than 76% = failure level (See Activity 5.6, p. 165) 11

12 What Does Mastery Mean? If a student is able to reach a mastery level score just once on a particular criterion-referenced instrument, this does not necessarily mean that the student actually has mastered the skills being tested. To establish mastery with some certainty, the student would need to be tested over multiple trials. 12

13 Beyond Mastery: Other Considerations Does passing the test mean that the student is proficient and will maintain the skills? Is the student ready to progress to the next level in the curriculum? Will the student be able to generalize and apply the skills in other contexts? Would the student pass the mastery test if it were given again at a later date? 13

14 Brigance Inventories The Brigance is standardized assessment system that provides criterion-related assessment of basic academic skills. There are three age levels of Brigance Inventories. Brigance Diagnostic Inventory of Early Development (birth to age 7) Brigance Diagnostic Inventory of Basic Skills (elementary- aged students) Brigance Diagnostic Inventory of Essential Skills (intermediate and secondary students) 14

15 Brigance Diagnostic Inventory of Early Development The Brigance Inventory of Early Development, Revised (IED) is a criterion-referenced scale for young children. It contains ten skill clusters: 1.Pre-ambulatory motor 2.Gross motor 3.Fine motor 4.Self-help 5.Speech and language 6.General knowledge and comprehension 7.Readiness 8.Basic reading 9.Manuscript writing 10.Basic math. The IED is designed for teachers to identify present levels of performance and measure progress and as an instructional guide with written objective for developing intervention programs. 15

16 Brigance Diagnostic Inventory of Basic Skills Author: Albert Brigance (1991) Publisher: Curriculum Associates Description of Test: The test is presented in a plastic ring binder that is designed to be laid open and placed between the examiner and the student. A separate student booklet provided for the student ’ s answers is designed so that the skills range from easy to difficult; thus, the teacher can quickly ascertain the skills level the student has achieved. Administration Time: Specific time limits are listed on many tests; others are untimed. Age/Grade Levels: Grades K through 9. It is also used for academic assessment of older students who are functioning below sixth-grade academic levels. Subtest Information: There are four subtests including 143 pencil and-paper or oral-response tests: readiness, reading, language arts, and mathematics 16

17 Brigance Comprehensive Diagnostic Inventory of Basic Skills-Revised (CIBS-R) The CIBS-R is a criterion-referenced inventory designed for use with elementary and middle school students. The inventory includes subtests in the following domains: readiness, reading (word recognition, passage reading, word analysis, vocabulary), language arts (handwriting, grammar mechanics, spelling, and reference skills), and math (grade level math, numbers, operations, and measurement). For each area, several subskills are assessed. For each of the items assessed within a subskill, objectives are included. If a student fails to show mastery of a particular subskill, the objective for that subskill can be used for educational planning. Some of the tests on the CIBS-R have been normed and results can be expressed as standard scores, percentile ranks, and grade equivalents. A computer program, the CIBS-R Standardized Scoring Conversion Software, is available to assist in the scoring. 17

18 Brigance Inventory of Essential Skills The criterion-referenced individually administered Brigance Diagnostic Inventory of Essential Skills covers academic skill areas and life skills. The former includes reading/language arts, math, and study skills. Life skill subtests include food and clothing, money and finance, travel and transportation, and communication and telephone skills. The Inventory of Essential Skills also includes rating scales for measuring health and attitude, responsibility and self-discipline, job interview preparation, communication, and auto safety. Inventory materials include a student record book that records competency levels and defines instructional objectives and a class record book that provides a matrix of skills assessed, skills mastered, and objectives for a group of up to 15 students. The inventory is widely used to assess secondary level students and adult learners with special needs. 18

19 Strengths of the Brigance The Brigance is considered one of the most comprehensive criterion-referenced instruments. It is also viewed as being well suited to determining mastery of very specific learning objectives. The test manual states that results of the Brigance should be considered in conjunction with the student ’ s classroom performance, classroom observations, and scrutiny of actual curriculum goals. The specific strengths of the Brigance include: Helps to determine what a student has or has not learned Contains suggestions for specific instructional objectives Requires no testing expertise Can help with referral decisions 19

20 Curriculum-Based Assessment Curriculum-based assessment (CBA) means using materials and tasks from the general curriculum to diagnose learning problems or to measure student progress. Curriculum-based assessments are usually given at the end of an instructional period (summative). CBA assesses mastery of specific content or skill taught during an academic period. Students results are compared against a standard of mastery (e.g., student must pass with 80% of items correct). 20

21 What Is Curriculum-Based Measurement? Curriculum-Based Measurement (CBM) is the method of monitoring student progress through direct, continuous assessment of basic skills. CBM is used to assess skills such as reading fluency, comprehension, spelling, mathematics, and written expression. Early literacy skills (phonics and phonological awareness) are similar measures and are downward extensions of CBM. 21

22 CBM Is Formative Assessment With curriculum-based assessment, the student is measured from the beginning of instruction against the ultimate goal for the student ’ s learning. For example, a child in third grade would be measured against a year-end goal in reading (e.g., beginning fourth grade), even though at the beginning of the year, the child would not be expected to have mastered the goal. Throughout the school year, the student would be measured against the year-end goal to see if the student is making reasonable progress. Measuring progress during instruction is called formative assessment. Formative assessment allows the teacher to make changes in instruction based upon the student ’ s academic performance. Thus, the teacher is able to make quick adjustments so the student does not “ get stuck ” and continues to make progress toward the ultimate learning goal. 22

23 How Valid Is CBM? CBM assessment practices are based on 25 years of scientific research at the University of Minnesota and elsewhere (Deno, 1985; Deno, Marston, & Mirkin, 1982; Deno, Marston, Shinn, & Tindal, 1983). These informal tests are time efficient and inexpensive, yet produce accurate charts of student growth over time. 23

24 What Is a CBM Probe Like? CBM probes last from 1 to 5 minutes depending on the skill being measured and student performance is scored for speed and accuracy to determine proficiency. Because CBM probes are quick to administer and simple to score, they can be given frequently to provide continuous progress data. The results are charted and provide for timely evaluation based on hard data. 24

25 What Is the Content of CBM? As the name implies, CBM materials have historically been derived from individual school curricula. Currently, the CBM field is moving towards standard general curriculum probes to increase standardization and make more accurate comparisons. This is especially helpful when curriculum changes over time. 25

26 Rule of Thumb Teachers can design their own curriculum-based assessments using classroom materials. There are some basic guidelines for developing curriculum-based assessments. Below are some “ rules ” for assessment design in various academic areas: In reading, students should read aloud from reading materials for 1 minute. The number of words read correctly per minute (WCPM) constitutes the basic decision-making unit. In spelling, students write words that are dictated at specific intervals (either 5, 7, or 10 seconds) for 2 minutes. The number of correct letter sequences and words spelled correctly are counted. In written expression, students write a story for 3 minutes after viewing a story starter. The number of words written, spelled correctly, and/or correct word sequences are counted. In mathematics, students write answers to computational problems via two minute probes. The numbers of correctly written digits in correct position are counted. 26

27 Baseline and Goal In order to set up a measurement process, the teacher first determines baseline in the skills to be taught. For example, the teacher would do three probes of a skill or set of skills. The scores for the three probes are averaged or the median score can be selected as the baseline score. Once the baseline number has been determined, the teacher can estimate the goal (e.g., number of words read correctly, number of problems solved correctly, number of words spelled correctly) to be reached by the end of the year. The research literature provides guidance for reasonable yearly gains by grade level. For example, a second grader with a baseline of 55 correctly read words per minute can be expected to increase oral reading proficiency by approximately 38 words by the end of the year. The goal then is 93 correctly read words per minute. 27

28 Aimline The aimline is the goal line against which progress is measured in curriculum-based measurement. In order to plot the aimline, the teacher would begin at the baseline score and draw a line to the goal. To monitor the instruction, the data are plotted twice per week. When a student falls below the aimline for three consecutive measures, the instruction should be adjusted. When the student excels above the aimline for three consecutive measures, the instruction should be made more challenging. 28

29 Content for CBA When a teacher is designing a curriculum-based assessment, a good source of information for what to include in the assessment is a scope and sequence. A scope and sequence is a formal listing of the range of skills and the sequence in which those skills must be learned in a particular academic domain (e.g., reading, written expression, mathematics). To see an example of a Scope and Sequence, go to the Document Sharing Section of the course. 29 SCOPE SEQUENCE

30 Task Analysis Sometimes when developing CBA, the test designer will look at a specific skill or sub-skill and try to address it with items that constitute the steps toward completing a task or skill. The name for this process is task analysis. Task analysis simply means analyzing a task by breaking it down into the smallest steps or sub- skills. 30

31 Task Analysis Example Let ’ s assume, for example, that the teacher wants to assess the following skill: Recognizes initial consonant sounds and their association with the consonants in the alphabet A task analysis for recognizing initial consonant sounds might include steps like these: Test single letters (uppercase) for identification in this order– M, T, S, F, D, G, L, H, C, B, N, K, V, W, J, P Test single letters (lowercase) for identification in this order--m, t, s, f, d, r, g, l, h, c, b, n, k,v, w, j, p Test matching of upper- and lowercase consonants Test matching most common sounds with consonants Test initial consonant sound identification in CVC words 31

32 Error Analysis Error analysis is one of the best ways to determine what types of academic problems a student may be having. What the teacher does is look for patterns of errors. Once the error patterns are discovered, the teacher can then develop instructional lessons to correct the errors. ( See Activity 5.9,p. 167 ) 32

33 CBA/CBM Summary Curriculum-based assessment is a relatively simple process, involving a thorough analysis of the requirements of the curriculum in a particular domain, the development of items to cover the domain, and arrangement of those items in order from the simplest (or easiest) to the most complex (or most difficult). Using CBM allows the teacher to keep track of a student ’ s progress in the curriculum and to compare one student ’ s scores to those of other classmates learning the same curriculum. 33

34 Value of Curriculum-Based Assessment Provides more direct feedback to students Supports increases in student achievement Provides accurate screening information for eligibility Provides useful data to determine when students are ready to return to the general education program Is appropriate for assessing medication effects Is useful in designing instructional programs 34

35 Cautions in Using CBM Limited to measuring discrete skills; can ’ t measure global skills like creativity More sensitive to changes in rote learning than in higher level thinking skills 35

36 Issues in Informal Testing Are standards appropriate for student in terms of race, culture and gender? Are test items free from cultural bias? Is the language appropriate for the student? Does the measure bypass the limitations imposed by the disability? Are CBA measures of sufficient technical quality? Does the CBA measure thoroughly cover the skill range? Is the test long enough to provide enough information on the student ’ s performance? 36

37 Informal Assessment of Reading Decoding, word recognition, fluency, and comprehension are the broad areas of reading that teachers typically assess using informal methods. Decoding--the ability to associate sounds and symbols Word Recognition—the ability to read words instantly on sight Fluency--rate and ease with which a student reads orally Comprehension--ability to derive meaning from written language 37

38 DecodingWord Recognition FluencyComprehensionAttitudes Toward Reading Match letters and sounds Read grade level word lists Timed oral reading of letters Answering questionsInterview Read isolated letters, blends, syllables and real words Read Dolch WordsTimed oral reading of word lists ParaphrasingQuestionnaire Sound out nonsense words Timed oral reading of phrases Story retellingStudent history Sound out combinations of vowel sounds and patterns, consonant blends, and digraphs Timed oral reading of sentences Cloze Read sentences that contain new words Timed oral reading of paragraphs Maze Sentence Verification Vocabulary identification 38 Informal Techniques for Assessing Reading

39 Informal Reading Inventories Informal reading inventories are informal assessments that usually contain graded word lists for testing word recognition ability and graded reading passages for evaluating oral reading, silent reading, and comprehension. Teachers may select from among many commercially available informal reading inventories or design one of their own. To learn more about commercial informal reading inventories, conduct a search with the ETS Test File at http://ericae.net/testcol.htm using thehttp://ericae.net/testcol.htm keywords reading inventories. On the next slide is a description of one commercial inventory. 39

40 Informal Reading Inventory—5th Ed IRI-5 The Informal Reading Inventory by Burns and Roe introduces reading passages with a sentence that provides a purpose for reading. The student reads the selection orally and then responds to eight comprehension questions read by the examiner The number and percentage of word recognition and comprehension errors made by the student are recorded to determine whether the selection falls at the student ’ s Independent, Instruction, or Frustration reading level. 40

41 41 Basic Sight Words aka High Frequency Words aka Dolch List

42 Sample Informal Reading Inventory Motivation Statement: Imagine how you would feel if you were up to bat and this was your team ’ s last chance to win the game! Please read this story. Passage: Whiz! The baseball went right by me, and I struck at the air! “ Strike one! ” called the man. I could feel my legs begin to shake! Whiz! The ball went by me again, and I began to feel bad. “ Strike two, ” screamed the man. I held the bat back because this time I would kill the ball! I would hit it right out of the park! I was so scared that I bit down on my lip. My knees shook and my hands grew wet. Swish! The ball came right over the plate. Crack! I hit it a good one! Then I ran like the wind. Everyone was yelling for me because I was now a baseball star! 42

43 Comprehension Questions and Possible Answers 1. What is this story about? (Main idea--a baseball game, someone who gets two strikes and finally gets a hit 2. After the second strike, what did the batter plan to do? (Factual--Hit the ball right out of the park) 3. Who is the “ man ” in this story who called strikes? (Inferential--the umpire) 4. In this story, what was meant when the batter said, “ I would kill the ball ” ? (Terminology--Hit it very hard) 5. Why was the last pitch a good one? (Cause and effect--Because it went right over the plate) 6. What did the batter do after the last pitch? (Case and effect--The batter hit it a good one and ran like the wind.) 43

44 Scoring an IRI Error Count: Omissions _____Aided words _____ Insertions _____Repetitions _____ Substitutions _____Reversals _____ Scoring Guide Word Recognition ErrorsComprehension Errors Independent10 Instructional61-2 Frustration12+3+ 44

45 IRI Reading Levels The results obtained from IRIs are grade level scores. Typically, informal inventories provide three reading levels: Independent Level, Instructional Level, and Frustration Level. A student ’ s Independent Level is the level of graded reading materials that can be read easily with a high degree of comprehension and few errors in decoding. At this level, the student reads independently, without instruction or assistance from the teacher. Reading materials at the student ’ s Instructional Level are somewhat more difficult; this is the level appropriate for reading instruction. Materials at the Frustration Level are too difficult for the student; decoding errors are too frequent and comprehension too poor for instruction to occur. 45

46 Criteria for Reading Levels According to Kirk, Kliebhan, and Lerner (1978), the usual criteria for determining independent, instructional and frustration levels are as shown in the chart below: Independent Reading Level Word Recognition: 98% to 100% Comprehension: 90% to 100% Instructional Reading Level Word Recognition: 95% Comprehension: 75% Frustration Reading Level Word Recognition: less than 90% Comprehension: less than 50% These levels have been criticized for being too stringent. For example, Spache (1972) warned that “ if the teacher employs an Informal Reading Inventory (IRI) for his estimate of instructional level, he may be expecting children to read with a very unrealistic degree of oral accuracy. ” 46

47 Error Analysis in Reading Error analysis is generally used to investigate decoding mistakes in oral reading. The teacher records deviations from the printed text that the student makes while reading orally. Several types of errors can occur when students read connected text. Most systems of error analysis include at least four classes of errors: Additions—the reader adds words or parts of words to the text Substitutions—the reader mispronounces a word or parts of words; this type of error is also called a mispronunciation. (e.g., want for what) Omissions—the reader fails to pronounce words or parts of words. This error occurs when readers skip words, when they hesitate in responding, or when they say the do not know a word. Reversals—The reader changes the order of the words in a phrase or sentence or the order of sounds within a word. 47

48 Miscue Analysis An alternate method of error analysis takes into account the quality of the errors that readers make. This is called miscue analysis. Miscues are analyzed to determine whether they represent a change in meaning from the original test. For example, the substitution of hold for fight in “ fight back the tears ” is semantically correct and does not alter meaning. However, the substitution of ready for right in “ he ’ ll be all right ” does change the sense of the passage. Miscues that produce changes in meaning can be further analyzed. For example, the student ’ s miscue and the original text can be compared in these three ways: Graphic Similarity: How much do the two words look alike? Sound Similarity: How much do the two words sound alike? Grammatical Function: Is the grammatical function of the reader ’ s word the same as the grammatical function of the text word? 48

49 Informal Assessment of Mathematics Math is a relatively easy subject to assess using informal methods. The areas that are usually assessed include: Math facts Computation Math reasoning Math applications The assessment should be combined with both task analysis and error analysis to determine specific problem areas. These problem areas should be further assessed by using probes to determine the specific difficulty. Interviewing the student is also helpful in determining how the student is reasoning through a problem. 49

50 Methods of Informal Math Assessment Informal Inventories—Informal inventories survey a variety of skills to determine where the student ’ s strengths and weaknesses lie. Inventories usually have only one or two examples of each type of math problem so further analysis of errors is necessary in more specific probes. Criterion-Referenced Tests--CRTs are used to assess mastery of specific mathematics skills (e.g. multiplication by 9). Error Analysis—Error analysis is a process of looking at the student ’ s responses to determine why a mistake was made and to see if there is a pattern of repeated types of errors. Error analysis differentiates between systematic computation errors and errors that are random or careless mistakes. Diagnostic Probes—Probes are in-depth assessments of the mastery of a specific skill or sub-skill; typically a probe contains several items focused on the same skill. Clinical Math Interviews—Clinical interviews elicit information about the procedures that students use to arrive at their answers. The student is observed going about the mathematics task and then the student is interviewed to find out the cognitive strategies he or she used to accomplish the task. Portfolio Assessment—A portfolio should contain several examples of the student ’ s work, including classroom quizzes or assignments, group or individual projects, written math reports or math logs, or artwork related to mathematics. Portfolios may also contain results of standardized tests and informal assessments, student self-assessments, and student interest surveys and questions. Teachers might include checklists of student progress, graphs of results from CBA measures and records of clinical math interviews. 50

51 Example of an Informal Inventory Addition 6 3 4 10 8 11173367 +2 + 5 + 0 + 5 + 3 + 4 + 5 +15 +71 Subtraction 6 3 4 17 98 47101427 -4 -3 -0 -3 -4 -32 -3 -6 -24 Multiplication 3 2 2 6 33 22 3 232 204 x2x2x8 x0x 4x 422x3 x 4 51

52 Example of Math CRT Criterion for Mastery: 100% (10/10) correct Directions: Round off each number to the nearest hundred (100). 1.721__________ 2.7,879__________ 3.6,834__________ 4.881__________ 5.8502__________ 6.13,782__________ 7.789,332__________ 8.3,055__________ 9.803__________ 10.419__________ 52

53 Math Error Analysis 53 The teacher examines the student’s work and observes how the student goes about solving the problems. The teacher can then analyze what types of errors the student is making. Common types of errors include: Incorrect operation Incorrect number fact Incorrect algorithm Errors in place value Failure to follow sequence of Placement (working from right to left) Copying or handwriting errors Random errors

54 Error Analysis Practice For each of the following problems, analyze and describe the types of errors the student is making. Note that within the same box, all of the problems display the same error. 83 66 476 753 + 67+29+851+693 1410 815 1481113 67 58 627 861 +31+12-486-489 17 16 261 428 54 AB CD

55 More Error Analysis 175 185 632 523563 - 54- 22-147 -366 -382 11111513 495 167 181 17 461206 divided by 6 = 21 x 4x 8 128648 55 EF GH

56 Teacher-Made Probes Teacher-made probes can be used to identify specific problem areas. Mixed probes are used to locate areas that need further assessment or instruction. In the probe on the following page, each of the following categories has nine items: basic addition facts of sums to 9 (first item and then every fourth item), two-digit numbers plus two-digit numbers with no regrouping (second item and then every fourth item), two-digit number plus one-digit number with no regrouping (third item and then every fourth item), and basic addition facts of sums to 18 (fourth item and then every fourth item). When scoring a probe, the student receives one point for every correct digit in the correct place. On this probe, the student can obtain a maximum score of 63 correct digits with no errors. After three times, a high score of 40 or more correct digits per minute with no errors is a reasonable criterion for diagnostic purposes. 56

57 Example of a Math Probe 4 22 33 9 6 3641 6 8 +3 +41 +6 +7 +2 +62 +3 +5 +0 53 78 5 7 43 82 7 5 61 +44 +1 +8 +2 +36 +5 +4 +3 +37 7 5 82 37 6 4 31 57 7 +6 +2 +13 +2 + 6 +4 +18 +32 +9 Patterns: 0-9 facts, 2D +2D, 2D + 1D, and 0-18 facts Number of Correct Digits: ___________ 57

58 Error Analysis of Probe Results If the student fails to reach the criterion on a mixed probe, it is important to analyze the responses and locate the errors in the items missed. This analysis provides the teacher with information for further assessment with specific skill probes (such as 0-9 facts). Also, specific skill probes can be used to monitor the daily progress of the student. For what grade level do you think this probe was designed? 58

59 Analysis of Math Observation and Clinical Interview What previous knowledge did the student bring to this problem? To what extent are the ideas accurate and complete? What strategy did the student employ to solve the problem? Could the student do the steps of the problem in proper sequence? Were the student ’ s calculations accurate? 59

60 Diagnostic Questions for Analysis of Problem Solving AgreeDisagree ______ ______ Decodes words correctly in story problem ______ ______ Understands the meaning of the situation described in the story problem ______ ______Identifies the relevant and irrelevant information in the problem ______ ______Can illustrate the components of the problem ______ ______Selects the appropriate operation (addition, subtraction, multiplication or division) ______ ______Writes down the computational problem correctly ______ ______Remembers number facts correctly ______ ______Selects the appropriate computational algorithm ______ ______Estimates the correct answer ______ ______Determines is answer “ makes sense ” 60

61 Informal Assessment of Written Expression Written expression includes a complex array of skills which all must be working relatively well in order for the written product to be legible, understandable, and persuasive. Informal probes of writing skills can be done in each of the important areas, including handwriting, writing mechanics, spelling, and composition. On the next slide is a chart showing common methods of informal writing assessment by category. 61

62 Common Informal Methods of Assessing Writing HandwritingWriting MechanicsSpellingComposition Analysis of handwriting sample (copy 100 word passage) CRTs of punctuation, capitalization, and grammar Paper and pencil dictation test Rating scales and checklists of skills Rating handwriting sample according to grade level template (e.g., shape, slant, spacing, size, smoothness) Informal surveys of punctuation, capitalization, and grammar Oral spellingWriting sample analysis (write for 15 minutes on a topic or from a story starter) Inventories and CRTs (e.g., Denver Handwriting Analysis, Brigance) Rating scalesMultiple-choice format (choose the correct spelling) Work samples Writing samplesSpontaneous writing sampleObservation and clinical interview Work samples Inventories of regular words, irregular words and homophones Criterion-referenced test (usually of grade level words 62

63 Writing Sample the peopol of englind didn ’ t the cherch roals. So a group of pepol got to gether and desided to live. So after a lot of comfermising. The king gov them 3 ships and they set sail for a mew land. They sailed a long ways for a to long tine. Then they saw it land it was North amareca. They landid on plymouth rock. There they started to beld the ferst coliny. The firs winter wase the hardes a lot peopl dide from being sick. Afte the winter was over the ingin ’ s becom frinds with them and to them how to hunt and grow food. What kinds of errors do you see in this composition? What is your error analysis? 63

64 Excerpt from a Writing Checklist AgreeDisagree Content ______ ______Does the writing clearly communicate an idea or ideas to the reader? ______ ______Is the content adequately developed? ______ ______Is the content interesting to the potential reader? Vocabulary ______ ______Does the writer select appropriate words to communicate his/her ideas? ______ ______Does the writer use precise/vivid vocabulary? ______ ______Does the writer effectively use verbs, nouns, adjectives and adverbs? ______ ______Does the vocabulary meet acceptable standards for written English (e.g., isn ’ t vs. ain ’ t)? Sentences ______ ______Are the sentences complete (subject and predicate)? ______ ______Are run-on sentences avoided? ______ ______Are exceptionally complex sentences avoided? ______ ______Are the sentences grammatically correct (e.g., word order, subject-verb agreement)? Paragraphs ______ ______Do the sentences in the paragraph relate to one topic? ______ ______Are the sentences organized to reflect the relationships between ideas within the paragraph? ______ ______Does the paragraph include a topical, introductory or transition sentence? 64

65 Sample Writing Interview Questions You ’ ve finished your composition. Tell me about what you ’ ve written. When you finished, did you read over what you had written? Did you make any changes? What did you change? Did you have anyone else read your paper? Did you change your paper on the basis of suggestions that someone else made? While you were writing, what did you think about? Did you consider the ideas you were writing about? What should come first, second, and so on? Choosing the exact words to express your meaning? Spelling the words correctly, using correct punctuation, and following all the rules for correct grammar? Do you think that you ’ ve accomplished your purpose in writing? Why or why not? If not, what do you need to change? Do you think your writing will be understandable for your audience? Is the vocabulary suitable? The tone? If not, what do you need to change? 65

66 Spelling Assessment A short informal spelling test can be designed by selecting grade level words from a frequency-of-use word list. The student is asked to spell on paper words from each grade list until three words in a grade list are missed. The student ’ s spelling level can be estimated as that at which two or fewer words are missed. 66

67 Common Types of Spelling Errors Dysphonetic errors. Spelling errors which reflect inaccurate spellings without regard to phonics. Words may have some correct letters, but the letters are placed in bizarre positions, such as ronaeg for orange. Students with this problem read and spell primarily through visualization. Dyseidetic errors. Spelling errors reflect phonic-equivalent errors (e.g., pese for peace, det for debt). What kinds of spelling errors do you see in the essay on Slide #65? 67

68 Summary of Methods of Informal Assessment Criterion Referenced Tests—tests of one skill with a designated level of accuracy in order for the skill to be considered mastered Curriculum-Based Assessments— informal tests using content from the curriculum Probes—tests of specific skills or sub- skills with multiple examples of the same skill to determine strengths and weaknesses of the student Checklists—lists of academic or behavioral skills Questionnaires—questions about a student ’ s behavior or academic performance that can be answered by the student or by a parent or teacher Work Samples—samples of a student ’ s classroom work Permanent Products—products made by the student that can be analyzed for academic or behavioral performance Performance assessment--assessment that requires the student to create an answer or product to demonstrate knowledge Authentic assessment--assessment that requires the student to apply knowledge in the real world Portfolio assessment--evaluating student progress, strengths, and weaknesses using a collection of different measurements and work samples 68

69 Portfolio Contents daily assignments samples of student writing examples of student problem solving informal inventories and probes criterion-referenced tests results of standardized tests story maps reading log or reading list math log vocabulary journal artwork, project papers, photographs, and other products group work, papers, projects, and products daily journal writing ideas letters to pen pals, letters exchanged with teacher out-of-school writing and artwork unit and lesson tests multi-media products independent research self-evaluations handwriting samples math computation exercises 69

70 Elements of High Quality Assessment Authentic and valid Encompass the whole child Involve repeated observations of patterns of behavior Ongoing and continuous Use a variety of methods Provide a means for feedback Provide an opportunity for students, teachers, and parents to discuss progress 70


Download ppt "1 INFORMAL ASSESSMENT. Usefulness of Norm-Referenced Tests Norm-referenced tests are standardized methods for assessing skills or behaviors. They are."

Similar presentations


Ads by Google