Download presentation

Presentation is loading. Please wait.

Published bySaige Butrum Modified over 2 years ago

1
Testing Theories: Three Reasons Why Data Might not Match the Theory

2
Theory testing Part of what differentiates science from non-science is the process of theory testing. When a theory has been articulated carefully, it should be possible to derive quantitative hypotheses from the theory—propositions about what should be observed under specific conditions.

3
Example Example: Imagine that we have a theory concerning the relationship between television habits and obesity. According to our theory, there is a correlation between the amount of television that people watch and their obesity levels. Our theory, however, does not assume that this correlation is due to a causal relationship between the two variables. Rather, our theory assumes that exercise causally influences obesity, and that people tend to get less exercise when they are watching TV.

4
Hypotheses Given this simple theory, there are a number of hypotheses that we can derive. Here are four: –there will be a positive correlation between television viewing and obesity –there will be a negative correlation between time spent exercising and obesity –there will be a negative correlation between television viewing and time spent exercising –if we hold exercise constant (statistical control), we will not observe a correlation between television viewing and obesity

5
Hypotheses Notice that each of these implications is inherently quantitative: –people who watch more television will be more obese (greater than; less than; correlation—all of these are quantitative statements) –if we were to hold exercise constant, we would not observe a relationship between television viewing and weight (precise numerical prediction: zero correlation)

6
Directional predictions The first kind of hypothesis is what we call a directional hypothesis or a directional prediction. A directional prediction concerns the direction of a difference between two groups or the sign (+ vs. -) of a correlation between two variables. Example: If we had two groups of people, those who watch TV and those who do not, we would predict that the average weight of the TV group to be higher than that of the no-TV group (M TV > M no-TV )

7
Testing directional predictions How do we test directional predictions? Virtually all approaches to testing quantitative hypotheses are based on the logic that the difference between a prediction and an empirical observation should be small. (T – O)

8
Testing directional predictions We can apply this logic easily in this circumstance. We simply find the average weight for the two groups and compare them. If our hypothesis (that M TV > M no- TV ) is correct, then M TV - M no-TV should be > 0. Thus, we find support for our hypothesis if this difference is greater than zero (i.e., positive). We disconfirm the hypothesis if the difference is less than or equal to zero.

9
The weakness of directional tests It is important to keep in mind that testing directional hypotheses is a relatively weak way to test a theory. Imagine for a moment that we had a computer randomly generate a prediction for us. Since one group will naturally have a higher score than the other due to randomness, the computer, picking blindly, has a 50% chance of getting the prediction correct. In short, testing a theory via a single directional prediction is a weak way to test a theory.

10
Point predictions Sometimes our hypothesis may be precise enough to make a specific, rather than a directional, quantitative prediction For example, in the previous example, our theory allows us to derive a point prediction: the difference between the groups will be zero (exactly zero) when exercise is held constant. How do we test point predictions?

11
Testing point predictions PersonXY a11 b33 c62 d124 r =.16 The correlation between these two variables is.16. Clearly the prediction is incorrect, but not by much. How do we quantify how much? (predicted – observed) 2 (.00 -.16) 2 =.025

12
Riskiness Point predictions are much more risky for a theory. Why? Precise predictions are simply more likely to be wrong. Thus, when the predictions turn out to be pretty close to what is observed, the theory gets “more credit.” 10 10 1/20 1/2

13
Evaluating point predictions How much of an error is too much of an error when a point prediction is being tested? There is no “standard” for making this decision, and, arguably, the amount of error that one is willing to tolerate may vary from one research context to the next. It is useful, however, to examine the comparative accuracy of alternative or competing hypotheses— hypotheses derived from different theories.

14
Competing theories and hypotheses In our example, there could be an alternative theory that predicts that there should be a correlation between television viewing and obesity equal to.50. If we were to observe a correlation of.16, the original theory clearly is more accurate than the other, despite the fact that it was slightly off. (If the alternative theory predicted a correlation of about.32, the evidence is clearly equivocal. (.16 falls between the two predictions of.00 and.32.))

15
What do errors mean? What does it mean when there is a difference between the value predicted and the value observed? –There are variables that matter that were not included in the theory –Imprecision in measurement (noise in the data) –Sampling error

16
What do errors mean? Where does error come from? –(1) Psychologically interesting variables that matter, but are not included in the theory This represents a problem with the theory. It might be incomplete (not too much of a problem), or just dead wrong (big problem). Let’s assume that, in reality, some variable of interest (e.g., happiness) was a positive function of x (e.g., coffee consumption) and the square of x: y = 2 + 2x + 2x 2

17
Psychologically interesting variables that matter, but are not included in the model real model: y = 2 + 2x + 2x 2 Thus, if we had 5 people with scores on x of –2, -1, 0, 1, and 2, their values of y would be 6, 2, 2, 6, and 14, respectively.

18
Psychologically interesting variables that matter, but are not included in the model If we were to test a model with just one predictor variable (e.g., x), we would clearly make some errors in prediction. y = 2 + 2x In this case, the errors are not huge, but they exist nonetheless. These errors are due to the fact that the theory is incomplete. (Note: This does not necessarily mean the theory is horribly flawed.)

19
What do errors mean? Where does error come from? –(2) Imprecision in measurement (noise in the data) This represents a problem with the data, and the measurement process that produced it. It is not a problem with the theory per se.

20
Imprecision in measurement (noise in the data) As we discussed previously, a measured score can be broken down into three components: O = T + E + S –O = observed score –T = “True” score –E = random error component –S = systematic error component (we’ll ignore this component for now since we could construe it as a variable that was omitted from the model) O = T + E

21
Another Example Here we have the same model, but random errors of measurement, e, have been added to each observation y = 2 + 2x + 2x 2 y e (y+e) [1] 6 -1 5 [2] 2 2 4 [3] 2 -2 0 [4] 6 0 6 [5] 14 -1 13

22
What do errors mean? Where does error come from? –(3) Sampling error This represents a problem with the data too, and the sampling process that generated it. It is not a problem with the theory per se.

23
Sampling error Error that occurs in data due to inadequacies in sampling from a population –Population: the group of interest (e.g., all people with access to televisions) –Sample: a subset of the population that is studied (i.e., people in this class) Note: For some reason, the problem of sampling error is the one that worries psychologists the most, and we’ll discuss some of the methods psychologists use to deal with it in our next lecture.

24
One of the advantages of comparative theory testing It is important to note that these last two sources of error (measurement errors and sampling errors) are specific to a data set. They are properties of the data (i.e., the measures used or the way the sample was obtained), not the theory. As such, these two problems are not too fatal if we are testing two competing theories. Why? Because the errors will count against both theories, and what is left over is the differences in the predictions of the two theories.

Similar presentations

© 2017 SlidePlayer.com Inc.

All rights reserved.

Ads by Google