Evaluating Assessments of Novice Programming Environments Paul Gross and Kris Powers Tufts University.

Slides:



Advertisements
Similar presentations
Internal Assessment Your overall IB mark (the one sent to universities after the IB test) in any IB science course is based upon two kinds of assessments.
Advertisements

Critical Reading Strategies: Overview of Research Process
Andrea M. Landis, PhD, RN UW LEAH
Developing an Outcomes Assessment Plan. Part One: Asking a Meaningful Question OA is not hard science as we are doing it. Data that you collect is only.
Summer Internship Program Outline
Qualitative Social Work Research
Designs to Estimate Impacts of MSP Projects with Confidence. Ellen Bobronnikov March 29, 2010.
Using Rubrics for Evaluating Student Learning. Purpose To review the development of rubrics for the purpose of assessment To share an example of how a.
S519: Evaluation of Information Systems Analyzing data: Merit Ch8.
Reading the Dental Literature
Using Rubrics for Evaluating Student Learning Office of Assessment and Accreditation Indiana State University.
Business research methods: data sources
Refining Your Research Question. In this session, we will… Discuss guidelines for creating a ‘good’ research question Provide time to revisit and revise.
Experimental Research
Chapter 2 Understanding the Research Process
Chapter 8 Experimental Research
I want to test a wound treatment or educational program but I have no funding or resources, How do I do it? Implementing & evaluating wound research conducted.
Writing a Research Proposal
Copyright © 2014 by The University of Kansas Choosing Questions and Planning the Evaluation.
The Vocabulary of Research. What is Credibility? A researcher’s ability to demonstrate that the study is accurate based on the way the study was conducted.
 Be familiar with the types of research study designs  Be aware of the advantages, disadvantages, and uses of the various research design types  Recognize.
Classroom Assessments Checklists, Rating Scales, and Rubrics
Nursing Research Prof. Nawal A. Fouad (5) March 2007.
Systematic Review Module 7: Rating the Quality of Individual Studies Meera Viswanathan, PhD RTI-UNC EPC.
Section 2.4: Simple Comparative Experiments. Experiment – A planned intervention undertaken to observe the effects of one or more explanatory variables,
Chapter 2 Research Methods in Social Psychology. Chapter Outline  Characteristics of Empirical Research  Research Methods  Research in Diverse Populations.
Is research in education important?. What is the difference between Qualitative and Quantitative Research Methods?
Research Model Problem and Research Approach  In today’s climate of increasingly large computer science classes, educators face the challenge of delivering.
ScWk 242 Course Overview and Review of ScWk 240 Concepts ScWk 242 Session 1 Slides.
Eloise Forster, Ed.D. Foundation for Educational Administration (FEA)
1 Experimental Research Cause + Effect Manipulation Control.
Construct-Centered Design (CCD) What is CCD? Adaptation of aspects of learning-goals-driven design (Krajcik, McNeill, & Reiser, 2007) and evidence- centered.
1 Copyright © 2011 by Saunders, an imprint of Elsevier Inc. Chapter 8 Clarifying Quantitative Research Designs.
1 THEORETICAL FRAMEWORK AND RESEARCH METHODOLOGY CHAPTER 3 PREPARED BY: MUKUNDA KUMAR.
Evaluating Impacts of MSP Grants Hilary Rhodes, PhD Ellen Bobronnikov February 22, 2010 Common Issues and Recommendations.
AN ANALYSIS OF THE EFFECTIVENESS OF THE EARLY READING INTERVENTION FOR SELF-EFFICACY (E-RISE) ON FIRST, SECOND, AND THIRD GRADE STUDENTS IN AN AT-RISK.
1 The Theoretical Framework. A theoretical framework is similar to the frame of the house. Just as the foundation supports a house, a theoretical framework.
ISECON 2006 The Work System Model as a Tool for Understanding the Problem in an Introductory IS Project Doncho Petkov Eastern Connecticut State University.
4-Day Agenda and Expectations Day 2 Building Formative Assessments linked to deconstructed content and skills Building Formative Assessments linked to.
Reading (Comprehending), Interpreting and Applying Educational Research.
Research Strategies. Why is Research Important? Answer in complete sentences in your bell work spiral. Discuss the consequences of good or poor research.
Copyright © Allyn & Bacon 2008 Intelligent Consumer Chapter 14 This multimedia product and its contents are protected under copyright law. The following.
Evaluating Impacts of MSP Grants Ellen Bobronnikov Hilary Rhodes January 11, 2010 Common Issues and Recommendations.
The Research Process.  There are 8 stages to the research process.  Each stage is important, but some hold more significance than others.
S519: Evaluation of Information Systems Analyzing data: Merit Ch8.
Qualitative and Quantitative Research Methods
What Are the Characteristics of an Effective Portfolio? By Jay Barrett.
Evaluating Impacts of MSP Grants Ellen Bobronnikov January 6, 2009 Common Issues and Potential Solutions.
Evaluation Requirements for MSP and Characteristics of Designs to Estimate Impacts with Confidence Ellen Bobronnikov February 16, 2011.
Producing Data: Experiments BPS - 5th Ed. Chapter 9 1.
A. Strategies The general approach taken into an enquiry.
Choosing Questions and Planning the Evaluation. What do we mean by choosing questions? Evaluation questions are the questions your evaluation is meant.
AUTHOR: NADIRAN TANYELI PRESENTER: SAMANTHA INSTRUCTOR: KATE CHEN DATE: MARCH 10, 2010 The Efficiency of Online English Language Instruction on Students’
CONNECT WITH CAEP | | Measures of Teacher Impact on P-12 Students Stevie Chepko, Sr. VP for Accreditation.
Research methods revision The next couple of lessons will be focused on recapping and practicing exam questions on the following parts of the specification:
Critiquing Quantitative Research.  A critical appraisal is careful evaluation of all aspects of a research study in order to assess the merits, limitations,
Investigate Plan Design Create Evaluate (Test it to objective evaluation at each stage of the design cycle) state – describe - explain the problem some.
Conducting a research project. Clarify Aims and Research Questions Conduct Literature Review Describe methodology Design Research Collect DataAnalyse.
Case Study of the TOEFL iBT Preparation Course: Teacher’s perspective Jie Chen UWO.
Rigor and Transparency in Research
Classroom Assessments Checklists, Rating Scales, and Rubrics
Evaluation Requirements for MSP and Characteristics of Designs to Estimate Impacts with Confidence Ellen Bobronnikov March 23, 2011.
Writing a sound proposal
Classroom Assessments Checklists, Rating Scales, and Rubrics
11/20/2018 Study Types.
What is Research? A research study is a study conducted to collect and analyse information in order to increase our understanding of a topic or an issue.
Eloise Forster, Ed.D. Foundation for Educational Administration (FEA)
Managerial Decision Making and Evaluating Research
Regulatory Perspective of the Use of EHRs in RCTs
EDUC 2130 Quiz #10 W. Huitt.
Presentation transcript:

Evaluating Assessments of Novice Programming Environments Paul Gross and Kris Powers Tufts University

ICER '052 How did we get into this? Wanted to know: “Do novice programming environments really help students learn? And, if so, how?” Literature/web search uncovered plethora of environments… ~40 new tools in the last 5 years …but assessments of their impact sparse, disjoint No consistent questions or methods Little recognition of other assessments

ICER '053 What can we learn from them? Focus on empirical assessments Hard, objective data Repeatable and generalizable Multiplicity of approaches complicates comparison Different observation and analysis methods, data observed, questions, etc. Need evaluation tool!

ICER '054 Evaluating Assessments Objective variable coding E.g., population, study duration, questions asked, conclusions, etc. Allows comparison of study methods and pragmatics Subjective evaluation Critique of study design and reporting Fairness managed by rubric of 8 questions Adapted from Long & Godfrey (2004)

ICER '055 Evaluation Conclusions Questions asked are often too vague Studies often only conducted by developer or those closely associated Approaches tend towards outcome-based rather than process-based Data collected is naturally occurring, rarely explicitly intended for assessment study Observation instruments used are not validated Reporting of practices incomplete

ICER '056 Our Assessment Evaluations Evaluation of 5 environment assessments Alice, BlueJ, Jeliot 2000, Lego Mindstorms with Ada, RAPTOR Represent a cross-section of environment types Variety of approaches to assessment Evaluated using Objective variable coding Rubric of 8 questions

ICER '057 Evaluative Rubric (1/8) 1. How appropriate is the question asked and is the question of reasonable scope? Example (Alice; Moskal et al., 2004) Does exposure to the Alice course improve student performance in CS1? Evaluation Appropriate as Alice course expected to prepare students for CS1 Reasonable as question addresses very specific, measurable effect

ICER '058 Evaluative Rubric (2/8) 2. What theoretical framework guides or informs the study and how is it reflected in the methodology? Example (Jeliot 2000; Levy et al., 2003) Authors cite previous results showing that animation’s impact is more noticeable in labs rather than exams Evaluation Study incorporates previous results by deliberately integrating Jeliot 2000 into lab assignments

ICER '059 Evaluative Rubric (3/8) 3. Is the reporting of the observation and analysis methods adequate? Example (BlueJ; Ragonis & Ben-Ari, 2005) Investigated teaching objects-first approach to young novices, BlueJ chosen tool Analyzed audio/video recordings and student artifacts to identify “difficulties” with program flow Evaluation Inadequate reporting of the analysis methods “Difficulties” are said to occur “frequently” with no discussion about how difficulties were recognized or what might constitute frequent occurrence

ICER '0510 Evaluative Rubric (4/8) 4. Are the observation and analysis methods valid and appropriate to the question? Example (Lego Mindstorms with Ada; Fagin & Merkle, 2002) Asked what is the impact of using robotics on student exam performance? Observed midterm and final exam scores Analyzed scores for statistically significant difference between robotics and control sections Evaluation Both valid and adequate as question is simple and straightforward to satisfy

ICER '0511 Evaluative Rubric (5/8) 5. Do the authors outline potential sources of bias? Example (RAPTOR; Carlisle et al., 2005) Treatment group performed worse than control group on exam question for one semester Evaluation No, sources of bias not adequately addressed Performance result attributed to difficult lab No discussion about other possible factors including lack of grading standardization, instructor bias, or other variables between courses and semesters

ICER '0512 Evaluative Rubric (6/8) 6. To what degree is the study generalizable and repeatable? Example (Alice; Moskal et al., 2004) Study determines “at risk” students, intervenes with Alice course, measures CS1 grades, retention, and attitudes Evaluation Easily generalizable as observation and analysis methods are not explicitly dependent on Alice Mostly repeatable as most of the methods are discussed (not “at risk” measure and focus group methods) and materials are available

ICER '0513 Evaluative Rubric (7/8) 7. Is there a coherent chain of reasoning from the analysis results to the assessment conclusions? Example (Jeliot 2000; Levy et al., 2003) Concludes animation students used a different and better vocabulary describing solutions in interview questions than control students Evaluation Not particularly strong Need to clarify interview methodology and criteria for how the solution descriptions were classified and evaluated

ICER '0514 Evaluative Rubric (8/8) 8. Do the conclusions answer the original questions? Example (Lego Mindstorms with Ada; Fagin & Merkle, 2002) Ask what is the effect of using robotics in CS1 course on exam performance? Concludes that robotics had negative effect on exam performance Evaluation Yes and to a great degree as they account for other factors that could bias exam performance data

ICER '0515 Future Work Refine questions asked in assessments Consider individual features Ask how and why impact occurs Develop validated instruments Multi-institutional studies of a single environment

ICER '0516 Thank you! Questions?