Presentation is loading. Please wait.

Presentation is loading. Please wait.

Unanswered Questions in Typical Literature Review 1. Thoroughness – How thorough was the literature search? – Did it include a computer search and a hand.

Similar presentations


Presentation on theme: "Unanswered Questions in Typical Literature Review 1. Thoroughness – How thorough was the literature search? – Did it include a computer search and a hand."— Presentation transcript:

1 Unanswered Questions in Typical Literature Review 1. Thoroughness – How thorough was the literature search? – Did it include a computer search and a hand search? – In a computer search, what were the descriptors used? – Which journals were searched? – Were theses and dissertations searched and included?

2 2. Inclusion/Exclusion – On what basis were studies included or excluded from the written review? – Were theses and dissertations arbitrarily excluded? – Did the author make decisions about the inclusion or exclusion based on the perceived internal validity of the research? sample size? research design? use of appropriate statistics?

3 3. Conclusions – Were conclusions based on the number of studies supporting or refuting a point (called vote counting)? – Were studies weighted differently according to sample size? meaningfulness of the results? quality of the journal? internal validity of the research?

4 Establishing Cause and Effect The selection of a good theoretical framework The application of an appropriate experimental design The use of the correct statistical model and analyses The proper selection and control of the independent variable The appropriate selection and measurement of the dependent variable The use of appropriate subjects The correct interpretation of the results

5 Reliability and Validity Reliability - the consistency or repeatability of a measure; is the measure reproducible Validity - the truthfulness of a measure; validity depends on reliability and relevance

6 Reliability The observed measure is the summation of the true score and the error score The more reliable a test, the less error is involved Reliability is defined as the proportion of observed score variance that is true score variance Reliability is determined using Pearson’s correlation coefficient or ANOVA

7 Types of Reliability Interclass reliability - Pearson’s Product Moment (Correlation of only two variables) Test -retest reliability (stability)-determines if a single test is stable over time Equivalence reliability - are two tests similar in measuring the same item or trait Split halves reliability - estimates the reliable of a test based on the scores of the odd and even test items Spearman Brown prophecy - estimates test reliability based on addition or deletion of test items

8 Intraclass Reliability Reliability within an individual’s scores of more than two measures Cronbach’s alpha estimates the reliability of tests. Uses AVOVA to determine mean differences within and between an individual’s scores

9 Indices of Reliability Index of reliability - the theoretical correlation between true and observed scores; IR=  r xx Standard Error of Measure - SEM the degree of fluctuation of an individual’s observed score from their true score SEM = s  (1-r xx )

10 Factors Affecting Reliability Fatigue Practice Ability Time between tests Testing Circumstances Test Difficulty Type of measurement Environment

11 Validity Content validity - face validity, logical validity Criterion Validity- measures are related to specific criterion Predictive validity Concurrent validity Construct validity - test validity as a measure of psychological constraints

12 The Relationship between Reliability and Validity

13 Possible Test Items 1 Be able to define and differentiate between reliability and validity; what types of error do each try to explain. Know the different classes and types of reliability; be able to differentiate between the different scores that are included in reliability. Be able to calculate the reliability of test examples and describe what type/class of reliability is being defined.

14 Possible Test Items 2 Be able to define/describe/determine Cronbach’s Alpha, Reliability Index, and the Standard Error of Measurement. Know what factors affect test reliability and how to compensate to make a more reliable test. Be able to describe, define and differentiate between the types of validity Be able to describe the different methods for developing a criterion for the evaluation of validity and give examples of the different types of criterion validity.


Download ppt "Unanswered Questions in Typical Literature Review 1. Thoroughness – How thorough was the literature search? – Did it include a computer search and a hand."

Similar presentations


Ads by Google