Moving beyond the psychometric discourse: A model for programmatic assessment Researching Medical Education Association for the Study of Medical education.

Slides:



Advertisements
Similar presentations
Educational Supervision & Find Your Way Around in the E-portfolio Dr Jane Mamelok RCGP WPBA Clinical Lead.
Advertisements

Intelligence Step 5 - Capacity Analysis Capacity Analysis Without capacity, the most innovative and brilliant interventions will not be implemented, wont.
Introduction to Competency-Based Residency Education
Wynne Harlen. What do you mean by assessment? Is there assessment when: 1. A teacher asks pupils questions to find out what ideas they have about a topic.
The role of teachers in the assessment of learning
Evaluating E-Learning TECH665. Evaluation Evaluation defined Evaluation defined Evaluation defined Evaluation defined Formative evaluation Formative evaluation.
Chapter 3 Flashcards. obligation of an individual to other individuals based on a social or legal contract to justify his or her actions; the processes.
Logic, observation, representation, dialectic and ethical values: what counts as evidence in educational research? Dylan Wiliam Annual conference of the.
Statistics / Sadistics Juggling Basic Statistical Notions Important For Teachers While Maintaining Sanity 1.
Workplace-based Assessment. Overview Types of assessment Assessment for learning Assessment of learning Purpose of WBA Benefits of WBA Miller’s Pyramid.
Beg, Borrow, or Steal: Using Multidisciplinary Approaches in Empirical Software Engineering Research An ICSE 2000 Workshop.
Reliability, Validity, Trustworthiness If a research says it must be right, then it must be right,… right??
An overview of Assessment. Aim of the presentation Define and conceptualise assessment Consider the purposes of assessment Describe the key elements of.
An overview of Assessment. Aim of the presentation Define and conceptualise assessment Consider the purposes of assessment Describe the key elements of.
Assessment of Clinical Competence in Health Professionals Education
FORMATIVE ASSESSMENT Nisreen Ahmed Wilma Teviotdale.
Lecture 7 Evaluation. Purpose Assessment of the result Against requirements Qualitative Quantitative User trials Etc Assessment of and Reflection on process.
Reliability & Validity Qualitative Research Methods.
Accountability in adult literacy, language and numeracy in England; with reference to the other UK countries Dr Janine Eldred Associate Director NIACE.
Chapter 1. Introduction: The World of Educational Assessment 1.
LOGO Teacher evaluation Dr Kia Karavas Session 5 Evaluation and testing in language education.
Assessment Tools. Contents Overview Objectives What makes for good assessment? Assessment methods/Tools Conclusions.
Work based assessment Challenges and opportunities.
Evaluation: A Challenging Component of Teaching Darshana Shah, PhD. PIES
Qualitative Research.
Objectives understand the feedback process and its’ importance to learning understand the feedback process and its’ importance to learning know factors.
Validation of Non-formal Adult Education courses REACTION No CP LT-GRUNDTVIG-G1.
Program Evaluation. Program evaluation Methodological techniques of the social sciences social policy public welfare administration.
Verification: Quality Assurance in Assessment Verification is the main quality assurance process associated with assessment systems and practice - whether.
Classroom Assessments Checklists, Rating Scales, and Rubrics
Unit 1 – Preparation for Assessment LO 1.1&1.2&1.3.
Universiteit Maastricht Barcelona, 6 – 9 July th Ottawa conference on Medical Education.
Workshops to support the implementation of the new languages syllabuses in Years 7-10.
Assessment Tools.
Dr. Roxana Reichman Gordon Academic College Haifa, ISRAEL.
1 Instructional Practices Task Group Chicago Meeting Progress Report April 20, 2007.
Qualitative Research EDUC 7741/Paris/Terry.
Base Paper : Sub Group III Accreditation Procedure and Mechanisms of Grading of VOs 2 nd Meeting of the Task Force September 23, 2009.
Qualitative Research Intro for Educational Technologists.
Government Social Research Unit Quality and purpose: issues and approaches Teresa Williams Head, Government Social Research Unit Research Methods Festival.
Shifting Paradigms: Competency-based Medical Education and the Quality of Care Problem Reynolds Meeting 2012.
Fitz-Albert R. Russell, PhD JP Establishing Standards for Students’ Assessment.
Welcome! A-Team Session 4 Fall Overview for Today 1 Minute Message review Outcomes Feedback Methodology Review Assessment Steps  Sampling  Validity/Reliability.
Development of the Egyptian Code of Practice for Student Assessment Lamis Ragab, MD, MHPE Hala Salah, MD.
Guidelines Recommandations. Role Ideal mediator for bridging between research findings and actual clinical practice Ideal tool for professionals, managers,
OUNL’s assessment model January the 10th 2006 Colin Tattersall & Henry Hermans.
From rote to realism: The role of clinical legal education in providing best practice in assessment and feedback Carol Boothby, Director of the Student.
Elspeth Slayter, Ph.D. Assistant Professor, Salem State University Lecture notes on threats to validity, threats to trustworthiness and the optimization.
INTRODUCTION TO ASSESSMENT METHODS USED IN MEDICAL EDUCATION AND THEIR RATIONALE.
Reliability in assessment Cees van der Vleuten Maastricht University Certificate Course on Assessment 6 May 2015.
.  Evaluators are not only faced with methodological challenges but also ethical challenges on a daily basis.
PGCE Evaluation of Assessment Methods. Why do we assess? Diagnosis: establish entry behaviour, diagnose learning needs/difficulties. Diagnosis: establish.
Click to edit Master subtitle style Competence by Design (CBD) Foundations of Assessment.
Henry M. Sondheimer, MD Association of American Medical Colleges 7 August 2013 A Common Taxonomy of Competency Domains for the Health Professions and Competencies.
The International Society for Quality in Health Care (ISQua) – Guidelines.
Five Doors to Excellence in Assessment India Broyles EdD, Director, Master of Science in Medical Education Leadership University of New England College.
Presented by John Versluis What is Assessment? The Career and Technology Education and Training Framework.
EDU704 – Assessment and Evaluation
Programmatic Assessment for Learning
Classroom Assessments Checklists, Rating Scales, and Rubrics
Classroom Assessments Checklists, Rating Scales, and Rubrics
Trustworthiness in Qualitative Research
Assessment 101 Zubair Amin MD MHPE.
Competency and Performance: What Should We Measure and How
To Entrust or Not to Entrust
Assessment of Clinical Competencies
TESTING AND EVALUATION IN EDUCATION GA 3113 lecture 1
Assessing educational/training competencies of trainers of trainers
The High Stakes Assessment
Presentation transcript:

Moving beyond the psychometric discourse: A model for programmatic assessment Researching Medical Education Association for the Study of Medical education London, 23 November 2010 Cees van der Vleuten Maastricht University The Netherlands Powerpoint:

The first step is to measure whatever can be easily measured. This is ok as far as it goes. The second step is to disregard that which can't be easily measured or to give it an arbitrary quantitative value. This is artificial and misleading. The third step is to presume that what can't be measured easily really isn't important. This is blindness. The fourth step is to say that what can't be easily measured really doesn't exist. This is suicide. —Charles Handy, The Empty Raincoat, page 219.

 The Mc Namara fallacy refers to Robert McNamara, The Secretary of defence from and his belief that the body count was a good way of measuring how the war was going.  As long as more Vietcong were being killed than US forces, the war was being won

 Planned arrangement of individual assessments in a learning program  Quality compromises are made for individual methods/data points but not for the program as a whole  Programmatic assessment is fit for purpose: ◦ Assessment for learning ◦ Robust decision making over learner’s performance ◦ Improving the learned curriculum

 Proposed in 2005 (Van der Vleuten & Schuwirth, 2005)  Multitude of quality criteria and self- assessment instrument of program quality (Baartman et al, 2006, 2007)  Framework of design (Dijkstra et al, 2009) and design guidelines (Dijkstra et al, in prep)  Need for theoretical model of programmatic assessment for program in action

 Any point measurement or single data point of assessment is flawed  Standardized methods of assessment can have ´built-in´ validity ◦ Quality control around test construction and administration ◦ Assessment ‘technology’ is available  Validity in unstandardized methods lies in the users of the instruments, more than in the instruments 1 Theoretical/empirical account: Van der Vleuten, C. P., Schuwirth, L. W., Scheele, F., Driessen, E. W., & Hodges, B. (2010). The assessment of professional competence: building blocks for theory development. Best Pract Res Clin Obstet Gynaecol, 24,

 Assessment drives learning ◦ Requires richness or meaningfulness of data ◦ Qualitative, narrative information carries a lot of weight ◦ Theoretical understanding emerging (Cilliers et al., 2010, under editorial review)  S takes (formative and summative assessment) is a continuum  N of datapoints need to be proportional to stakes  Expert judgment is imperative for assessing complex competencies and when diverse information is to be combined ◦ Sampling strategies can reduce random error ◦ Procedural strategies can reduce bias 1 Van der Vleuten et al., 2010

Time Assessment Activities Training Activities Supporting Activities Artifacts of learning - Outcome artifacts: Products of learning tasks - Process artifacts: Learning or working activities vv Learning task - PBL case - Patient encounter - Operation - Project - Lecture - Self-study

Time Assessment Activities Training Activities Supporting Activities Individual data points of assessment - Fit for purpose - Multiple/all levels of Miller - Learning oriented, Information rich documentation, meaningful (quantitative, qualitative) - Low stake Certification of mastery-oriented learning tasks - Rescuscitation - Normal delivery of infant vv

Time Assessment Activities Training Activities Supporting Activities Supportive social interaction - Coaching/mentoring/supervision - Peer interaction (intervision) (P)Reflective activity by learner - Interpretation of feedback - Planning new learning objectives and tasks vv

Time Assessment Activities Training Activities Supporting Activities Intermediate evaluation - Aggregate information held against performance standard - Committee of examiners - Decision making: diagnostic, therapeutic, prognostic - Remediation oriented, not repetition oriented - Informative - Longitudinal - Intermediate stake vv Firewall dilemma - Dilemma between access to rich information and compromising relationship supporting person(s) and learner

Time Assessment Activities Training Activities Supporting Activities vvvv

Time Assessment Activities Training Activities Supporting Activities vvvvvv Final evaluation - Aggregate information held against performance standard - Committee of examiners - Pass/fail(/distinction) high stake decision - Based on many data points and rich information - Decision trustworthiness optimized though procedural measures, inspired qualitative methodology strategies - High stake

QuantitativeQualitative Criterionapproachapproach Truth valueInternal validityCredibility ApplicabilityExternal validityTransferability ConsistencyReliabilityDependability NeutralityObjectivityConfirmability

Strategy to establish trustworthinessCriteria Potential Assessment Strategy (sample) CredibilityProlonged engagementTraining of examiners TriangulationTailored volume of expert judgment based on certainty of information Peer examinationBenchmarking examiners Member checkingIncorporate learner view Structural coherenceScrutiny of committee inconsistencies TransferabilityTime samplingJudgment based on broad sample of data points Thick descriptionJustify decisions DependabilityStepwise replicationUse multiple assessors who have credibility ConfirmabilityAuditGive learners the possibility to appeal to the assessment decision

vv Time Assessment Activities Training Activities Supporting Activities vvvv

vv Time Assessment Activities Training Activities Supporting Activities vvvv vv Time Assessment Activities Training Activities Supporting Activities vvvv

 Resources/cost (do fewer things well rather than doing more but poorly)  Bureaucracy, Trivialization, Reductionism  Legal restrictions  Novelty/unknown

 Assessment for learning combined with rigorous decision making  Post-psychometric era of individual instruments  Theory driven assessment design  Infinite research opportunities ◦ Multiple formalized models of assessment (e.g. psychometrics, Bayesian approaches to information gathering, new conceptions of validity……) ◦ Judgment (bias, expertise, learning….) ◦ How and why is learning facilitated (theory of assessment driving learning) ◦ ………..