Presented By Dr / Said Said Elshama  Distinguish between validity and reliability.  Describe different evidences of validity.  Describe methods of.

Slides:



Advertisements
Similar presentations
1 COMM 301: Empirical Research in Communication Kwan M Lee Lect4_1.
Advertisements

© McGraw-Hill Higher Education. All rights reserved. Chapter 3 Reliability and Objectivity.
© 2006 The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill Validity and Reliability Chapter Eight.
Assessment Procedures for Counselors and Helping Professionals, 7e © 2010 Pearson Education, Inc. All rights reserved. Chapter 5 Reliability.
VALIDITY AND RELIABILITY
MGT-491 QUANTITATIVE ANALYSIS AND RESEARCH FOR MANAGEMENT
Assessment Procedures for Counselors and Helping Professionals, 7e © 2010 Pearson Education, Inc. All rights reserved. Chapter 6 Validity.
Assessment: Reliability, Validity, and Absence of bias
Chapter 4 Validity.
Reliability or Validity Reliability gets more attention: n n Easier to understand n n Easier to measure n n More formulas (like stats!) n n Base for validity.
Concept of Measurement
FOUNDATIONS OF NURSING RESEARCH Sixth Edition CHAPTER Copyright ©2012 by Pearson Education, Inc. All rights reserved. Foundations of Nursing Research,
Research Methods in MIS
Validity and Reliability EAF 410 July 9, Validity b Degree to which evidence supports inferences made b Appropriate b Meaningful b Useful.
Classroom Assessment A Practical Guide for Educators by Craig A
Reliability and Validity. Criteria of Measurement Quality How do we judge the relative success (or failure) in measuring various concepts? How do we judge.
Validity Lecture Overview Overview of the concept Different types of validity Threats to validity and strategies for handling them Examples of validity.
Validity and Reliability Neither Valid nor Reliable Reliable but not Valid Valid & Reliable Fairly Valid but not very Reliable Think in terms of ‘the purpose.
Technical Issues Two concerns Validity Reliability
Measurement and Data Quality
Validity and Reliability
Measurement in Exercise and Sport Psychology Research EPHE 348.
Foundations of Educational Measurement
Educational Research: Competencies for Analysis and Application, 9 th edition. Gay, Mills, & Airasian © 2009 Pearson Education, Inc. All rights reserved.
MEASUREMENT CHARACTERISTICS Error & Confidence Reliability, Validity, & Usability.
Copyright © 2012 Wolters Kluwer Health | Lippincott Williams & Wilkins Chapter 14 Measurement and Data Quality.
McMillan Educational Research: Fundamentals for the Consumer, 6e © 2012 Pearson Education, Inc. All rights reserved. Educational Research: Fundamentals.
Unanswered Questions in Typical Literature Review 1. Thoroughness – How thorough was the literature search? – Did it include a computer search and a hand.
Technical Adequacy Session One Part Three.
Induction to assessing student learning Mr. Howard Sou Session 2 August 2014 Federation for Self-financing Tertiary Education 1.
Student assessment AH Mehrparvar,MD Occupational Medicine department Yazd University of Medical Sciences.
Chapter 4: Test administration. z scores Standard score expressed in terms of standard deviation units which indicates distance raw score is from mean.
Validity Is the Test Appropriate, Useful, and Meaningful?
6. Evaluation of measuring tools: validity Psychometrics. 2012/13. Group A (English)
Chapter 8 Validity and Reliability. Validity How well can you defend the measure? –Face V –Content V –Criterion-related V –Construct V.
Selecting a Sample. Sampling Select participants for study Select participants for study Must represent a larger group Must represent a larger group Picked.
Validity and Reliability Neither Valid nor Reliable Reliable but not Valid Valid & Reliable Fairly Valid but not very Reliable Think in terms of ‘the purpose.
Assessment Specifications Gronlund, Chapter 4 Gronlund, Chapter 5.
RELIABILITY AND VALIDITY OF ASSESSMENT
Evaluating Survey Items and Scales Bonnie L. Halpern-Felsher, Ph.D. Professor University of California, San Francisco.
Chapter 4 Validity Robert J. Drummond and Karyn Dayle Jones Assessment Procedures for Counselors and Helping Professionals, 6 th edition Copyright ©2006.
Copyright © 2008 Wolters Kluwer Health | Lippincott Williams & Wilkins Chapter 17 Assessing Measurement Quality in Quantitative Studies.
Validity Validity is an overall evaluation that supports the intended interpretations, use, in consequences of the obtained scores. (McMillan 17)
Validity and Item Analysis Chapter 4. Validity Concerns what the instrument measures and how well it does that task Not something an instrument has or.
Validity and Item Analysis Chapter 4.  Concerns what instrument measures and how well it does so  Not something instrument “has” or “does not have”
Criteria for selection of a data collection instrument. 1.Practicality of the instrument: -Concerns its cost and appropriateness for the study population.
©2011 Cengage Learning. All Rights Reserved. May not be scanned, copied or duplicated, or posted to a publicly accessible website, in whole or in part.
Chapter 6 - Standardized Measurement and Assessment
Classroom Assessment Chapters 4 and 5 ELED 4050 Summer 2007.
Reliability EDUC 307. Reliability  How consistent is our measurement?  the reliability of assessments tells the consistency of observations.  Two or.
TEST SCORES INTERPRETATION - is a process of assigning meaning and usefulness to the scores obtained from classroom test. - This is necessary because.
Slides to accompany Weathington, Cunningham & Pittenger (2010), Chapter 10: Correlational Research 1.
Copyright © 2014 Wolters Kluwer Health | Lippincott Williams & Wilkins Chapter 11 Measurement and Data Quality.
Copyright © Springer Publishing Company, LLC. All Rights Reserved. DEVELOPING AND USING TESTS – Chapter 11 –
Measurement and Scaling Concepts
1 Measurement Error All systematic effects acting to bias recorded results: -- Unclear Questions -- Ambiguous Questions -- Unclear Instructions -- Socially-acceptable.
ESTABLISHING RELIABILITY AND VALIDITY OF RESEARCH TOOLS Prof. HCL Rawat Principal UCON,BFUHS Faridkot.
Reliability and Validity
Ch. 5 Measurement Concepts.
Lecture 5 Validity and Reliability
DUMMIES RELIABILTY AND VALIDITY FOR By: Jeremy Starkey Lijia Zhang
Concept of Test Validity
RELIABILITY OF QUANTITATIVE & QUALITATIVE RESEARCH TOOLS
Validity and Reliability
Reliability & Validity
Week 3 Class Discussion.
Reliability and Validity of Measurement
PSY 614 Instructor: Emily Bullock, Ph.D.
Chapter 4 Characteristics of a Good Test
Chapter 8 VALIDITY AND RELIABILITY
Presentation transcript:

Presented By Dr / Said Said Elshama

 Distinguish between validity and reliability.  Describe different evidences of validity.  Describe methods of estimating test reliability.  Mention factors that decrease exam validity and reliability and how to avoid them.

 Appropriateness, meaningful and usefulness of results interpretations.  It is unitary concept.  It is expressed by degree  It is specific use  It is related to assessment using consequences  It refers to inferences only, not the instrument  It depends on different types of evidence.

 Content representativeness  Criterion relationships  Construct evidence  Consequences of assessment using

CONTENT REPRESENTATIVENESS  Adequate sampling  Representative broad range of content  Relevance content Application Application We prepared the valid test content by :- We prepared the valid test content by :- 1. Assessed learning outcomes identification 2. Specification the sample of used tasks 3. Suitable assessment procedure for specifications

CRITERION RELATIONSHIPS  Relationship degree between test scores and predicted or estimated criterion  It is expressed by correlation coefficients or expectancy table  Positive relationship  Negative relationship

CONSTRUCT EVIDENCE  It measures all possible influences on the scores  It clarifies and verifies the inferences of results  It depends on other different types of evidence (content, criterion) Relevant and Representative Predicting scores Relationship Factors affecting the score

CONSEQUENCES OF ASSESSMENT RESULTS USING  It is using the extracted inferences from assessment resultsApplication  Motivation was improved by assessment results using?  Performance was improved by assessment results using?  Self assessment skills were improved by assessment results using?  Independent learning was encouraged by assessment results using?  Good study habits were encouraged by assessment results using?

 Face validity  Construct validity  Content validity  Criterion validity 1. Predictive 2. Concurrent  Consequences validity

 Insufficient of assessed tasks sample  Using of improper tasks types  Using of irrelevant or ambiguity tasks  Improper tasks arrangement  Using of unclear directions  Few tasks (test items)  Improper administration (inadequacy of allowed time)  Objectivity of scoring  Inadequacy of scoring guides

Application Application  Relevance  Balance  Efficiency  Specificity  Difficulty  Discrimination  Variability  Clear directions  Proper administration  Adequacy of scoring guides

 It is consistency of results  The scores free from measurement errors  It is determined by reliability coefficient or standard error of measurement  Using of different sample of the same type of task leads to the same results  Generalizable results over similar samples of tasks, time periods and raters  Using of assessment in the different time leads to the same results  Different evaluators rate the performance assessment in the same way

 Test –retest method Twice administrations of the same test to the same group with time interval in between  Equivalent –forms method Administration of two equivalent forms of the test to the same group in close succession during the same session  Test-retest with equivalent forms (combination ) Administration of two equivalent forms of the same test with time interval in between  Internal consistency method One single administration of the test and account the response consistency within the test

Factors lower test scores reliability  Few items of test scores  Limited scores range  Inadequacy of testing conditions  Subjectivity of scoring Avoid the factors which lower test scores reliability by the following:- Use longer tests Gather scores from many short tests Item difficulty adjustment for larger scores spread Arrange test administration time Interrupted factors elimination Scoring key preparation

1) Comparison between two obtaining scores of two judges who scored the performance independently 2) Assessment performance reliability is percentage of agreement between two judges scores

Factors lowers performance assessments reliability  Inadequacy of number of tasks  Poor structured procedures of assessment  Specific performance dimensions to tasks  Insufficient scoring guides  Personal bias affecting scoring judgments Avoid the factors which lower performance assessments reliability by :- Gather results from many assessments Identify nature of tasks Identify scoring criteria Increase performance generalizability Use specific rating scales according to levels and criteria of quality Train for judging and rating Check scores with independent judge scores

 Reliability refers to obtained results and instrument evaluation and not to instrument itself.  Reliability refers to type of consistency.  Reliability is a necessary but not sufficient for validity.  Reliability provides the consistency which makes validity.

Heterogeneous groupModerate difficult test items More included items less scoring reliability High reliability

 Accuracy and appropriateness of results interpretations.  High, moderate or low degree.  Does the instrument measure what it tells it does? Factors lower validity  Unclear directions  Inadequate time limits  Inappropriate difficulty level  Poor constructed test items  Test items inappropriate for measured outcomes  Too short tests  Improper arrangement of items (complex to easy)  Administration and scoring  Nature of criterion

1- Gronlund, NE. (2006). Assessment of Student Achievement. 8 th Edition. Pearson, USA. Chapters Carter, DE, &Porter, S. (2000) Validity and reliability. In Cormack D (ed.) The Research Process in Nursing. Fourth edition, Oxford, Blackwell Science Knapp, TR. (1998). Quantitative Nursing Research. Thousand Oaks, Sage. 4- Peat, J. (2002). Health Services Research: A Handbook of Quantitative Methods. London, Sage. 5- Moskal, BM, & Leydens, JA. (2000). Scoring rubric development: Validity and reliability. Practical Assessment, Research & Evaluation, 7(10). [Available online: