Copyright © 2012 by Educational Testing Service. Computer Based Achievement Assessment of Young Students (Grades 1, 2 and 3) Christine M. Mills ETS, Princeton,

Slides:



Advertisements
Similar presentations
Common Core Standards (What this means in computer class)
Advertisements

Sheltered Instruction Observation Protocol
Performance Tasks for English Language Arts
Primary Measures of Academic Progress (P-MAP) & Measures of Academic Progress (MAP)
“Data Walk” Faculty Orientation
PARCC Read Aloud Decision Making
Stepping Up with PARCC All students deserve access to a world-class education that will prepare them for success in college and careers. To work toward.
Beyond the Classroom: The Use of Essential Skills for Remediation and Extension Christine Koch November 2008.
PARCC Accommodation: Text-to-Speech, Screen Reader Version, ASL Video, Human Reader/Human Signer For the ELA/Literacy Assessment December 2014.
Assessment of Behavior
1 National Reading First Impact Study: Critique in the Context of Oregon Reading First Oregon Reading First Center May 13, 2008 Scott K. Baker, Ph.D. Hank.
MARCH 12, 2015 Testing at Lees Corner ES. Still Online? Online Testing  Grade Level Common Assessments Mostly in grades 3-6  eCart Windows Grades 3-6.
Secondary PARCC Tests Partnership for Assessment of Readiness for College and Careers.
Principles of Assessment
The Five New Multi-State Assessment Systems Under Development April 1, 2012 These illustrations have been approved by the leadership of each Consortium.
Lewiston Porter PEC Standards- Based Report Card Grades K-3 Report Card Committee Members: Heidi Kazulak, Suzanne Hedemann, Lisa Winslow, Kelly Millville,
PARCC Update June 6, PARCC Update Today’s Presentation:  PARCC Field Test  Lessons Learned from the Field Test  PARCC Resources 2.
Identifying the gaps in state assessment systems CCSSO Large-Scale Assessment Conference Nashville June 19, 2007 Sue Bechard Office of Inclusive Educational.
McLendon and Polis1 An Administrator’s Guide to Assessment: A Menu of Assessment Options for MAERS and Instructional Guidance.
ELA SCHOOL TEAM SESSION Welcome to EEA, 2012! 10/2/2015MSDE1.
 Closing the loop: Providing test developers with performance level descriptors so standard setters can do their job Amanda A. Wolkowitz Alpine Testing.
Becoming Familiar with the GRE General Test GRE Test Preparation Workshop for Campus Educators.
Alternate Assessment Transitions in West Virginia Melissa Gholson Office of Assessment.
Let’s Look at... Assessing Group Performance 1. Performance Groups Material for this section largely adapted from: “Assessing group work” © Copyright.
March 11,  Under Tab 2  Look at the Table of Contents  Based on the Table of Contents, what peeks your interest and want to learn more about?
THINK / PUZZLE / EXPLORE Think about what you may have read as part of the pre-reading activity for this workshop as well as what you already know about.
Smarter Balanced Assessment System March 11, 2013.
Evaluation of Shoreline Science Jia Wang & Joan Herman UCLA Graduate School of Education & Information Studies National Center for Research on Evaluation,
Universal Access to Assessments. Project Overview Four Implementing States: New Hampshire, Vermont, Rhode Island, and Maine Eight Partner States: Connecticut,
VIRGINIA Communication and Literacy Assessment –In June 22, 2005, the Board of Education approved the implementation of the Communication and Literacy.
A course is designed to increase mathematical comprehension. In order to evaluate the effectiveness of the course, students are given a test before and.
State Support for Classroom Assessment Fen Chou, Ph.D. Louisiana Department of Education National Conference on Student Assessment June 27, 2012.
Performance Task Overview Introduction This training module answers the following questions: –What is a performance task? –What is a Classroom Activity?
State Board of Education Presentation Jeffrey Hauger, Ed.D. Peggy McDonald, Ed.D. Elizabeth Celentano, M.Ed. January 11, 2016 NEW JERSEY DYNAMIC LEARNING.
The Interactive Strategies Approach to Early Literacy Intervention (ISA) Michelle Eackles RDG 692 Best Practices in Early Literacy Instruction Diane M.
1 NCEXTEND1 Alternate Assessment with Alternate Achievement Standards Conference on Exceptional Children November 17-18, 2008 NCDPI Division of Accountability.
Effectiveness of Selected Supplemental Reading Comprehension Interventions: Impacts on a First Cohort of Fifth-Grade Students June 8, 2009 IES Annual Research.
Evaluation Results MRI’s Evaluation Activities: Surveys Teacher Beliefs and Practices (pre/post) Annual Participant Questionnaire Data Collection.
C R E S S T / U C L A UCLA Graduate School of Education & Information Studies Center for the Study of Evaluation National Center for Research on Evaluation,
Literacy-Based Promotion Act & 3 rd Grade Summative Assessment Parent Information Night September 29, 2015.
Project VIABLE - Direct Behavior Rating: Evaluating Behaviors with Positive and Negative Definitions Rose Jaffery 1, Albee T. Ongusco 3, Amy M. Briesch.
AzMERIT. The AzMERIT what?? AIMS is gone. Bye. AzMERIT is the new AIMS type test Still Writing and Reading sections (and math-but you don’t want me teaching.
BY: HEATHER BEAVER, JOY DAUGHTRY, AND JENNIFER HAMBLEY Digital Books and Primary Literacy Skills.
Evan Jones. A Quick Background First year economics has a historically high failure rate of approximately 50%. Unlike accounting, statistics, mathematics.
Tier III Preparing for First Meeting. Making the Decision  When making the decision to move to Tier III, all those involve with the implementation of.
Research Questions  What is the nature of the distribution of assignment quality dimensions of rigor, knowledge construction, and relevance in Math and.
Unit 3 Learning Styles Learning Styles Study Styles LASSI Discussion Assignment Seminar.
1 IT/Cybersecurity - ICRDCE Conference Day Using Blooms to Write Student Learning Outcomes (SLO’s)
Measuring the Power of Learning.™ 2015–16 CAASPP March, 2016 Laurie Carlson California Assessment of Student Performance and Progress (CAASPP)
LEAP 2025 Practice Test Webinar for Teachers
ACSP Presents FSA Night Welcome!
What is a CAT? What is a CAT?.
Parent Night 2017.
AP Spanish Language and Culture Adapted from a presentation by: Maritza Sloan, Plano West Senior High School, Texas.
Massachusetts Comprehensive Assessment System (MCAS)
GLoCALL & PCBET 2017 Joint Conference, 7-9 September 2017 at Universiti Teknologi Brunei, Brunei Darussalam, Presented at Room 1, 11:00-11:30. Effect of.
3rd-5th Grade FSA Parent Night
PARCC Math Options Stephanie Boyd, PARCC Consultant
3rd-5th Grade FSA Parent Night
Annual Assessment and Accountability Meeting Updates
Usability Research: Lessons Learned For Digital Assessment Delivery Cathy Wendler Educational Testing Service June 21, 2013.
Massachusetts Comprehensive Assessment System (MCAS)
State Board of Education Presentation Jeffrey Hauger, Ed.D.
Information for Parents
Georgia’s Systems of Continuous Improvement Georgia School Assessment of Performance on Systems (GSAPS) Orientation Anita Smith Program Assessment Specialist.
Somerset Academy Silver Palms at Princeton Presents FSA Night Welcome!
FSA Parent Information
3rd-5th Grade FSA Parent Night
Presentation transcript:

Copyright © 2012 by Educational Testing Service. Computer Based Achievement Assessment of Young Students (Grades 1, 2 and 3) Christine M. Mills ETS, Princeton, NJ Paper presented at the National Conference on Student Assessment June , 2013, National Harbor, Maryland Unpublished Work Copyright © 2012 by Educational Testing Service. All Rights Reserved. These materials are an unpublished, proprietary work of ETS. Any limited distribution shall not constitute publication. This work may not be reproduced or distributed to third parties without ETS's prior written consent. Submit all requests through Educational Testing Service, ETS, the ETS logo, and Listening. Learning. Leading. are registered trademarks of Educational Testing Service (ETS). 1

Copyright © 2012 by Educational Testing Service. What problem did we try to solve? The client wanted a battery of diagnostic assessments measuring student achievement in English language arts and mathematics to a computer delivery system. These studies were completed within the scope of the transition to inform the design decisions for the tests assessing students in Grades 1, 2, and 3. 2

Copyright © 2012 by Educational Testing Service. Three Small Studies General computer usability study – informing design and test presentation decisions Modality comparability study – average scores comparable in CBT and PBT Audio delivery usability study – informing decisions regarding audio delivered tests (those typically read aloud by teachers) 3

Copyright © 2012 by Educational Testing Service. Computer Usability Study Inform design and presentation decisions and provide evidence that young students could manage completing the CBT assessment – 15 students (4 - Grade 1, 8 - Grade 2 and 3 - Grade 3) participated in individual interviews – Researchers followed a scripted protocol to walk students through the exercises. – Students recorded their responses using the mouse or keyboard and were interviewed about their experience. 4

Copyright © 2012 by Educational Testing Service. Computer Usability Study Results: – Most students thought both CBT and PBT were easy to use prefer taking the test on a computer – General hardware requirements should replicate what students use daily – Practice makes perfect – Different font was required for CBT – Different assumptions about practice items 5

Copyright © 2012 by Educational Testing Service. Modality Comparability Study Are average scores for schools comparable in CBT and PBT? – Schools were recruited on a voluntary basis and received $20 per student for participation – Provided guidance and asked to randomly assign students within a classroom to take a subset of the assessment battery on PBT group or CBT group – Performed t-test for the comparison of CBT and PBT group means 1 6

Copyright © 2012 by Educational Testing Service. Modality Comparability Study Results: – Across the three grades mean scores tended to be slightly higher for the CBT group. – Observed more variability in the scores for the PBT group with the exception of Grade 1 where we observed the opposite result – Only statistically significant differences in mean scores were observed for Grade 1 Auditory Comprehension and Grade 3 Mathematics – in both cases the CBT group had a slightly higher mean score. – Effect sizes ranged between 0.00 to 0.41 with largest values for Grade 1 Auditory Comprehension (.41) and Mathematics Grade 3 (0.25) – all others were between.0 and

Copyright © 2012 by Educational Testing Service. Audio Delivery Usability Study Inform design decisions for audio delivery – Students were presented with item instructions, stimulus and responses (where applicable) and asked to adjust the audio settings and/or determine how to have the audio stimulus repeated. – One kindergartner and four Grade 1 students responded to the directions, sample items and 12 live items assessing Word Analysis and 5 mathematics items. – Six Grade 2 students were asked to answer sample questions and 6 live items assessing writing mechanics. 8

Copyright © 2012 by Educational Testing Service. Audio Delivery Usability Study – The students followed the audio delivered instructions with ease. – They were able to answer the sample questions and in most cases understood how to follow the directions to answer the questions. – When students had trouble answering questions it did not appear to be due to difficulty in understanding the speech of the narrators who had recorded the stimulus presented. – Students were able to adjust their headsets with ease. Only one student had trouble understanding the tutorial’s description of how to adjust the volume; however had no trouble during the test. – Ten minutes was allotted for each student to complete the tutorial to learn how to adjust the volume of the narrator, his headset, and to learn how to replay the audio presentation of an item and this seemed to be plenty of time to take each student through the tutorial. – More time was needed after practice questions for students to respond to the item. However, most students used “Replay” spontaneously or with veiled coaching, such as “Pretend you wanted to listen to the lady again” if they wanted to hear the item text again. 9

Copyright © 2012 by Educational Testing Service. Conclusions – With deliberate, well thought out development decisions young students can take computer based assessments – With the appropriate item and stimulus presentation the scores form the PBT and CBT versions are comparable for this low stakes test battery. – Continued investigation is warranted as more schools transition to the CBT – Future work should consider the emerging literature on teacher preparation for implementing use of technology in the classrooms and how this affects what students are able to do. 10

Copyright © 2012 by Educational Testing Service. Limitations – Study included only traditional selected response item types – Sample sizes were small so caution should be used and more study is warranted as data become available to substantiate findings and evaluate subpopulations – There was not a comparison of the audio usability on CBT to PBT and performing this information 11