Statistics II: An Overview of Statistics. Outline for Statistics II Lecture: SPSS Syntax – Some examples. Normal Distribution Curve. Sampling Distribution.

Slides:



Advertisements
Similar presentations
AP Statistics Course Review.
Advertisements

Lecture (11,12) Parameter Estimation of PDF and Fitting a Distribution Function.
Hypothesis Testing Steps in Hypothesis Testing:
Hypothesis: It is an assumption of population parameter ( mean, proportion, variance) There are two types of hypothesis : 1) Simple hypothesis :A statistical.
6-1 Introduction To Empirical Models 6-1 Introduction To Empirical Models.
Regression Analysis Once a linear relationship is defined, the independent variable can be used to forecast the dependent variable. Y ^ = bo + bX bo is.
Ch11 Curve Fitting Dr. Deshi Ye
Chapter 15 (Ch. 13 in 2nd Can.) Association Between Variables Measured at the Interval-Ratio Level: Bivariate Correlation and Regression.
Statistical Significance What is Statistical Significance? What is Statistical Significance? How Do We Know Whether a Result is Statistically Significant?
© 2010 Pearson Prentice Hall. All rights reserved Least Squares Regression Models.
LINEAR REGRESSION: Evaluating Regression Models Overview Assumptions for Linear Regression Evaluating a Regression Model.
LINEAR REGRESSION: Evaluating Regression Models. Overview Assumptions for Linear Regression Evaluating a Regression Model.
HYPOTHESIS TESTING Four Steps Statistical Significance Outcomes Sampling Distributions.
Chapter Seventeen HYPOTHESIS TESTING
PSY 307 – Statistics for the Behavioral Sciences
Chapter 10 Simple Regression.
Hypothesis Testing Steps of a Statistical Significance Test. 1. Assumptions Type of data, form of population, method of sampling, sample size.
DATA ANALYSIS I MKT525. Plan of analysis What decision must be made? What are research objectives? What do you have to know to reach those objectives?
Statistical Significance What is Statistical Significance? How Do We Know Whether a Result is Statistically Significant? How Do We Know Whether a Result.
PPA 415 – Research Methods in Public Administration Lecture 9 – Bivariate Association.
Chapter Topics Types of Regression Models
Chapter 11 Multiple Regression.
Ch 15 - Chi-square Nonparametric Methods: Chi-Square Applications
Aaker, Kumar, Day Seventh Edition Instructor’s Presentation Slides
Chapter 9 Hypothesis Testing.
PPA 501 – Analytical Methods in Administration Lecture 9 – Bivariate Association.
Correlation and Regression Analysis
Summary of Quantitative Analysis Neuman and Robson Ch. 11
Review for Exam 2 Some important themes from Chapters 6-9 Chap. 6. Significance Tests Chap. 7: Comparing Two Groups Chap. 8: Contingency Tables (Categorical.
Richard M. Jacobs, OSA, Ph.D.
Copyright ©2006 Brooks/Cole, a division of Thomson Learning, Inc. More About Regression Chapter 14.
PSY 307 – Statistics for the Behavioral Sciences
Leedy and Ormrod Ch. 11 Gray Ch. 14
AM Recitation 2/10/11.
Aaker, Kumar, Day Ninth Edition Instructor’s Presentation Slides
Chapter 13: Inference in Regression
Hypothesis Testing Charity I. Mulig. Variable A variable is any property or quantity that can take on different values. Variables may take on discrete.
OPIM 303-Lecture #8 Jose M. Cruz Assistant Professor.
Statistics 11 Correlations Definitions: A correlation is measure of association between two quantitative variables with respect to a single individual.
Production Planning and Control. A correlation is a relationship between two variables. The data can be represented by the ordered pairs (x, y) where.
1 Chapter 12 Simple Linear Regression. 2 Chapter Outline  Simple Linear Regression Model  Least Squares Method  Coefficient of Determination  Model.
Inference for Regression Simple Linear Regression IPS Chapter 10.1 © 2009 W.H. Freeman and Company.
Statistics - methodology for collecting, analyzing, interpreting and drawing conclusions from collected data Anastasia Kadina GM presentation 6/15/2015.
Section 9-1: Inference for Slope and Correlation Section 9-3: Confidence and Prediction Intervals Visit the Maths Study Centre.
© Copyright McGraw-Hill Correlation and Regression CHAPTER 10.
CHI SQUARE TESTS.
Copyright ©2011 Brooks/Cole, Cengage Learning Inference about Simple Regression Chapter 14 1.
Inferential Statistics. The Logic of Inferential Statistics Makes inferences about a population from a sample Makes inferences about a population from.
© Copyright McGraw-Hill 2004
Copyright © 2013, 2009, and 2007, Pearson Education, Inc. Chapter 11 Analyzing the Association Between Categorical Variables Section 11.2 Testing Categorical.
T Test for Two Independent Samples. t test for two independent samples Basic Assumptions Independent samples are not paired with other observations Null.
Chapter 15 The Chi-Square Statistic: Tests for Goodness of Fit and Independence PowerPoint Lecture Slides Essentials of Statistics for the Behavioral.
Statistical Inference Drawing conclusions (“to infer”) about a population based upon data from a sample. Drawing conclusions (“to infer”) about a population.
Statistical Inference Statistical inference is concerned with the use of sample data to make inferences about unknown population parameters. For example,
Chapter 13- Inference For Tables: Chi-square Procedures Section Test for goodness of fit Section Inference for Two-Way tables Presented By:
1 1 Slide The Simple Linear Regression Model n Simple Linear Regression Model y =  0 +  1 x +  n Simple Linear Regression Equation E( y ) =  0 + 
Chapter 14 – 1 Chi-Square Chi-Square as a Statistical Test Statistical Independence Hypothesis Testing with Chi-Square The Assumptions Stating the Research.
Chapter 13 Understanding research results: statistical inference.
SUMMARY EQT 271 MADAM SITI AISYAH ZAKARIA SEMESTER /2015.
Educational Research Inferential Statistics Chapter th Chapter 12- 8th Gay and Airasian.
Chapter 11: Categorical Data n Chi-square goodness of fit test allows us to examine a single distribution of a categorical variable in a population. n.
26134 Business Statistics Week 4 Tutorial Simple Linear Regression Key concepts in this tutorial are listed below 1. Detecting.
Lecture Nine - Twelve Tests of Significance.
Math 4030 – 10b Inferences Concerning Variances: Hypothesis Testing
Chapter 14 in 1e Ch. 12 in 2/3 Can. Ed.
BIVARIATE REGRESSION AND CORRELATION
Statistics II: An Overview of Statistics
RES 500 Academic Writing and Research Skills
Correlation and Simple Linear Regression
Correlation and Simple Linear Regression
Presentation transcript:

Statistics II: An Overview of Statistics

Outline for Statistics II Lecture: SPSS Syntax – Some examples. Normal Distribution Curve. Sampling Distribution Hypothesis Testing Type I and Type II Errors Linking Z to, alpha, and hypothesis testing. Bivariate Measures of Association Bivariate Regression/Correlation

The Normal Distribution

The standard normal distribution: a bell-shaped symmetrical distribution having a mean of 0 and a standard deviation of 1.

Z scores. A z score (or standard score): a transformed score expressed as a deviation from an expected value that has a standard deviation as its unit of measurement. A standard score belonging to the standard normal distribution. Y-μ z = — σ

Sampling Distribution

That is, the spread of the sampling distribution depends on the sample size n, and the spread of the population distribution. As the sample size n increases the standard error decreases. The reason for this is that the denominator of the ratio increases as n increases, whereas the numerator is the population standard deviation, which is a constant and is not dependent on the value of n.

Central Limit Theorem For random sampling, as the sample size n grows, the sampling distribution of approaches a normal distribution. The approximate normality of the sampling distribution applies no matter what the shape of the population distribution.

Hypothesis Testing

Steps of a Statistical Significance Test. 1. Assumptions Type of data, form of population, method of sampling, sample size 2. Hypotheses Null hypothesis, Ho (parameter value for “no effect”) Alternative hypothesis, Ha (alternative parameter values) 3. Test statistic Compares point estimate to null hypothesized parameter value 4. P-value Weight of evidence about Ho; smaller P is more contradictory

5. Conclusion Report P-value Formal decision

Alpha or significance levels: The α - level is a number such that one reject if Ho if the P-value is less than or equal to it. The α - level is also called the significance level of the test. The most common α - levels are.05 and.01.

Type I and Type II Errors: A Type I error occurs when Ho is rejected, even though it is true. A Type II error occurs when Ho is not rejected, even though it is false.

Bivariate Statistics

PROPORTIONAL REDUCTION IN ERROR (PRE) “all good measures of association use a proportionate reduction in error (PRE) approach. The PRE family of statistics is based on comparing the errors made in predicting the dependent variable with knowledge of the independent variable, to the errors made without information about the independent variable. In other words, PRE measures indicate how knowing the values of the independent variable (first variable) increase our ability to accurately predict the dependent variable (second variable).

Error without Error with decision rule - decision rule PRE statistic = _____________________________ Error without decision rule

Another way of stating this is: E1 - E2 PRE value = _____ E1 Where E1 = number of errors made by the first prediction method. E2 = number of errors made by the second prediction method.

PRE measures are more versatile and more informative than are the chi-square-based measures. All PRE measures are normed; they use a standardized scale where the value 0 means there is no association and 1 means there is perfect association. Any value between these extremes indicates the relative degree of association in a ratio comparison sense. E.g., a PRE measure with a value of.50 represents an association that is twice as strong as one that has a PRE value of.25. The number of cases, the table size, and the variables being measured do not interfere with the interpretation that can be given to them.

Chi Square The Chi-square test examines whether two nominal variables are associated. It is NOT a PRE measure. The chi-square test is based on a comparison between the frequencies that are observed in the cells of a cross-classification table and those that we would expect to observe if the null hypothesis were true. The hypotheses for the chi-square are: Ho: the variables are statistically independent. Ha: the variables are statistically dependent.

Goodman and Kruska’s Gamma (G) A measure of association for data grouped in ordered categories. G is a PRE measure. G compares two measures of a prediction: 1st it randomly predicts all untied scores to be either in agreement or disagreement. 2nd it predicts all untied pairs to be of the same type. Agreement or disagreement is determined by the direction of the bivariate distribution. For a positive pattern we expect untied pairs to be in agreement For a negative pattern we expect untied pairs to be in disagreement.

Pa: we find the number of agreement pairs by multiplying the frequency for each cell by the sum of the frequencies from all cells that are both to the right and below it. Pd: is found by multiplying the frequency for each cell in the table by the sum of the frequencies from all cells that are both to the left and below it.

Bivariate Regression and Correlation

BIVARIATE REGRESSION AND CORRELATION WHY AND WHEN TO USE REGRESSION/CORRELATION? WHAT DOES REGRESSION/CORRELATION MEAN?

You should be able to interpret: The least squares equation. R 2 and Adjusted R 2 F and significance. The unstandardized regression coefficient. The standardized regression coefficient. t and significance. The 95% confidence interval. A graph of the regression line.

ASSUMPTIONS UNDERLYING REGRESSION/CORRELATION NORMALITY OF VARIANCE IN Y FOR EACH VALUE OF X For any fixed value of the independent variable X, the distribution of the dependent variable Y is normal. NORMALITY OF VARIANCE FOR THE ERROR TERM The error term is normally distributed. (Many authors argue that this is more important than normality in the distribution of Y). THE INDEPENDENT VARIABLE IS UNCORRELATED WITH THE ERROR TERM

ASSUMPTIONS UNDERLYING REGRESSION/CORRELATION (Continued) HOMOSCEDASTICITY It is assumed that there is equal variances for Y, for each fixed value of X. LINEARITY The relationship between X and Y is linear. INDEPENDENCE The Y’s are statistically independent of each other.