Towards a Concept Inventory for Algorithm Analysis Topics

Slides:



Advertisements
Similar presentations
Questionnaire Development
Advertisements

Part II Sigma Freud & Descriptive Statistics
MEQ Analysis. Outline Validity Validity Reliability Reliability Difficulty Index Difficulty Index Power of Discrimination Power of Discrimination.
Evaluating Online Tutorials for Data Structures and Algorithms Courses June 24, Simin Hall, PhD Mechanical Engineering Prof. Clifford A. Shaffer,
Developing a Statistics Teaching and Beliefs Survey Jiyoon Park Audbjorg Bjornsdottir Department of Educational Psychology The National Statistics Teaching.
Test Construction Processes 1- Determining the function and the form 2- Planning( Content: table of specification) 3- Preparing( Knowledge and experience)
ICE Evaluations Some Suggestions for Improvement.
Concept of Reliability and Validity. Learning Objectives  Discuss the fundamentals of measurement  Understand the relationship between Reliability and.
Item Response Theory. Shortcomings of Classical True Score Model Sample dependence Limitation to the specific test situation. Dependence on the parallel.
Formative and Summative Evaluations
Review of SUNY Oneonta Course Evaluation Form Report and Recommendations from The Committee on Instruction: Part II October 4, 2010.
FOUNDATIONS OF NURSING RESEARCH Sixth Edition CHAPTER Copyright ©2012 by Pearson Education, Inc. All rights reserved. Foundations of Nursing Research,
In the name of Allah. Development and psychometric Testing of a new Instrument to Measure Affecting Factors on Women’s Behaviors to Breast Cancer Prevention:
Instructional Design Dr. Lam TECM 5180.
Classical Test Theory By ____________________. What is CCT?
Validity and Validation: An introduction Note: I have included explanatory notes for each slide. To access these, you will probably have to save the file.
Technical Issues Two concerns Validity Reliability
Conducting a Job Analysis to Establish the Examination Content Domain Patricia M. Muenzen Associate Director of Research Programs Professional Examination.
Development of Evaluation Instruments based on ICT Skills Standards for Teachers (ISST) 7th May, 2014.
University Assessment Committee Comprehensive Standard (CS) The institution identifies expected outcomes, assesses the extent to which it achieves.
Using Authentic Discovery Projects to Improve Student Outcomes in Statistics Joint Mathematics Meetings January 16, 2010 Dianna Spence Brad Bailey Robb.
Engineering Education Conference - Spring 2009 Increasing Assessment Effectiveness in a Time of Decreasing Budgets Increasing Assessment Effectiveness.
FAEIS Project User Opinion Survey 2005 Thursday, June 23, 2005 Washington, D.C. H. Dean SutphinYasamin Miller ProfessorDirector, SRI Agriculture & Extension.
Instrumentation.
The Genetics Concept Assessment: a new concept inventory for genetics Michelle K. Smith, William B. Wood, and Jennifer K. Knight Science Education Initiative.
Teacher Evaluation System Part II: Student Learning Data May 5, 2015.
The Analysis of the quality of learning achievement of the students enrolled in Introduction to Programming with Visual Basic 2010 Present By Thitima Chuangchai.
Is the Script-Concordance Test a Valid Instrument for Assessment of Intra-operative Decision-making Skills? Brent Zabolotny 1, Robert Gagnon 2, Bernard.
Validity and Reliability THESIS. Validity u Construct Validity u Content Validity u Criterion-related Validity u Face Validity.
A Validation Study of the School Leader Dispositions Inventory © Teri Melton, Ed.D. Barbara Mallory, Ed.D. James Green, Ph.D. Georgia Southern University.
Appraisal and Its Application to Counseling COUN 550 Saint Joseph College For Class # 3 Copyright © 2005 by R. Halstead. All rights reserved.
VALIDITY AND VALIDATION: AN INTRODUCTION Note: I have included explanatory notes for each slide. To access these, you will probably have to save the file.
HCI – NYGH IP ED SEMINAR th Jan 2011 Diagnostic Teaching through Identification of Scientific Misconceptions Using Just-In-Time Teaching (JITT)
University of Georgia – Chemistry Department JExam - A Method to Measure Outcomes Assessment Charles H. Atwood, Kimberly D. Schurmeier, and Carrie G. Shepler.
By: HANIM MOHAMED (MP ) SITI FATIMAH ZAINI (MP091421)
Validity: Introduction. Reliability and Validity Reliability Low High Validity Low High.
Validity and Item Analysis Chapter 4. Validity Concerns what the instrument measures and how well it does that task Not something an instrument has or.
Validity and Item Analysis Chapter 4.  Concerns what instrument measures and how well it does so  Not something instrument “has” or “does not have”
Measurement Issues General steps –Determine concept –Decide best way to measure –What indicators are available –Select intermediate, alternate or indirect.
Building the NCSC Summative Assessment: Towards a Stage- Adaptive Design Sarah Hagge, Ph.D., and Anne Davidson, Ed.D. McGraw-Hill Education CTB CCSSO New.
Authentic Discovery Projects in Statistics GCTM Conference October 16, 2009 Dianna Spence NGCSU Math/CS Dept, Dahlonega, GA.
By: HANIM MOHAMED (MP ) SITI FATIMAH ZAINI (MP091421)
Reliability performance on language tests is also affected by factors other than communicative language ability. (1) test method facets They are systematic.
Advanced Science and Technology Letters Vol.47 (Education 2014), pp Instructor’s Evaluation on Importance.
Algorithm Visualization (AV)  AVs are used for motivating students in exploring the core concepts of data structure and algorithms.  Instructors report.
Item Analysis: Classical and Beyond SCROLLA Symposium Measurement Theory and Item Analysis Heriot Watt University 12th February 2003.
Writing A Review Sources Preliminary Primary Secondary.
11-Item Quiz Algorithmic (5) Le Chatelier’s Principle (3) Equilibrium Systems (2) Conceptual (5) Le Chatelier’s Principle (1) Equilibrium Systems (4) General.
Instructional Design Course Evaluation Phase. Agenda The Evaluation Process Expert Review Small Group Review The Pilot Feedback and Revision Evaluation.
National framework for student assessment student assessment is laid down in the School Act (Act no. 561/2004 Coll. The School Act determines basic conditions.
Assistant Instructor Nian K. Ghafoor Feb Definition of Proposal Proposal is a plan for master’s thesis or doctoral dissertation which provides the.
Methods for the Development and Validation of New Assessment Instruments in Nursing Education Francisco A. Jimenez, PhD A.J. Kleinheksel, PhD Presentation.
Pedagogical Content Knowledge in primary technology education Ellen J. Rohaan Constructing a multiple choice test to measure teachers’
Looking at the both ‘ends’ of the social aptitude dimension
ASSESSMENT OF THE QUALITY OF SERVICE OFFERED BY THE LIFESTYLE LEARNING CENTER OF CENTRAL PHILIPPINE UNIVERSITY by Jay Melvin C. Dayaday Ronald Christopher.
Lecture 5 Validity and Reliability
Evaluating a Task-based English Course: A Proposed Model
Evaluating Multi-Item Scales
Classroom Analytics.
UMDNJ-New Jersey Medical School
Evaluating the Effectiveness of Algorithm Analysis Visualizations
Evaluating the pedagogical effectiveness of educational interventions
الاختبارات محكية المرجع بناء وتحليل (دراسة مقارنة )
Analyzing Reliability and Validity in Outcomes Assessment Part 1
UTEAC Update 11/7.
By ____________________
Challenges of Piloting Test Items
Department of Educational Psychology
Analyzing Reliability and Validity in Outcomes Assessment
Chapter 8 VALIDITY AND RELIABILITY
Presentation transcript:

Towards a Concept Inventory for Algorithm Analysis Topics Mohammed Farghally Information Systems Department, Assiut University, Egypt Kyu Han Koh Department of Computer Science, CSU Stanislaus Jeremy V. Ernst School of Education, Virginia Tech Clifford A. Shaffer Department of Computer Science, Virginia Tech Friday, March10, 2017 SIGGCSE 2017 Seattle WA

Agenda The Algorithm Analysis Concept Inventory (AACI) Defining the Scope Identifying Misconceptions Pilot AACI item Design AACI Validity and Reliability Item Analysis Conclusions

The Algorithm Analysis Concept Inventory To support evaluation of AAV: a standardized instrument to measure student performance. A Concept Inventory for algorithm analysis can serve as the AA part of the final exam. CI

The Algorithm Analysis Concept Inventory (2) Defining the scope Start with an initial list of concepts By consulting experts, a final list of concepts is composed Delphi process Identifying misconceptions Initial list Experts are consulted for validation List is validated through student responses Developing Questions Pilot AACI items were created. Based on fundamental concepts and misconceptions. Validity and reliability Face Validity through Delphi experts Student Content Validity Single administration reliability measures (Cronbach’s alpha) Item Response Theory (ICC and TIF)

Defining the Scope Started with an initial list of 21 concepts. 10 experts were consulted through a Delphi process. Delphi process : a structured process for collecting information and reaching consensus among a group of experts Experts are anonymous (to each other)

Phase1 : Initial Rating 1- Initial list of concepts 2- Concept ratings (Importance, Difficulty) + New Concepts 3- Medians and IQRs Phase2 : Negotiation 4- Calculated medians and IQRs for each concept 5- Concept ratings (Importance, Difficulty) 6- Rating justification if outside IQR 7- Medians and IQRs Phase3 : Final Rating 8- Calculated medians and IQRs for each concept 9- Anonymous rating justifications 10- Concept ratings (Importance, Difficulty) 11- Medians and IQRs Researchers Experts

Region Of Interest (ROI)

Identifying Misconceptions Started with an initial list of 17 misconceptions. Experience and Literature Post-test at Fall 2014 Validated by our Delphi experts.

Pilot AACI Item Design A set of 10 items were developed. Some contains sub-items for a total of 29 items. MCQ, TF, and open-ended questions. Covering all important and difficult concepts. Probing student misconceptions.

AACI Validity and Reliability To be widely accepted, a CI should be both valid and reliable. A CI validation test answers “Are we measuring what we think we are measuring?” “Have we covered the required concepts?” A CI reliability test answers “Do test items give appropriately consistent results?”

AACI Validity and Reliability The pilot AACI was administered 4 times Virginia Tech during Fall 2015 (N = 67) and Spring 2016 (N = 155). Christopher Newport during Fall 2015 (N = 40) and Spring 2016 (N = 32).

AACI Reliability Single administration reliability measure (Internal Consistency). Cronbach’s-alpha Coefficient The pilot AACI has α = 0.82 (Good) (Nunnally, J. C. 1978).

AACI Validity Expert Content Validity (Face Validity). For all items, at least 80% of the experts rated it as good. Otherwise, some rewording was suggested.

AACI Validity (2) Student Content Validity.

Item Analysis Provides information about the usefulness of a test item and how it relates to other items in the test. Item Response Theory (IRT) Item Characteristic Curves (ICC) Test Information function (TIF) Two parameter logistic model

Item Analysis (2) Difficulty: 0.51 (Moderate) Difficulty: -1.54 (Easy) Discrimination: 1.39 (High) Discrimination: 1.30 (Moderate) Discrimination is the slope of the ICC at its inflection point. Difficulty: -0.17 (Moderate) Difficulty: -0.84 (Moderate) Discrimination: 8.62 (Very High) Discrimination: 2.74 (Very High) Baker, Frank B. The basics of item response theory. For full text: http://ericae. net/irt/baker., 2001.

Item Analysis (3) Difficulty: -1.31 (Easy) Difficulty: -0.79 (Moderate) Discrimination: 0.62 (Low) Discrimination: 0.49 (Low) Difficulty: -6.72 (Very Easy) Difficulty: 1.05 (Above Average) Discrimination: 0.20 (Very Low) Discrimination: -0.17 (Very Low)

Item Analysis (4)

Conclusions The pilot AACI did a good job in detecting the identified student misconceptions. Most of the items are good according to expert feedback and IRT analysis. It is a Valid and Reliable measure of student knowledge of Algorithm Analysis.

Future Work Alpha AACI Improve items. (Get more feedback from instructors.) More administrations. (Get more performance feedback.) Conduct student interviews. (Get more feedback from students.) Thank you! To NSF funding under grants DUE-1139861, IIS-1258471, and DUE-1432008.