Models of Choice. Agenda Administrivia –Readings –Programming –Auditing –Late HW –Saturated –HW 1 Models of Choice –Thurstonian scaling –Luce choice theory.

Slides:



Advertisements
Similar presentations
Vector Spaces A set V is called a vector space over a set K denoted V(K) if is an Abelian group, is a field, and For every element vV and K there exists.
Advertisements

Modeling of Data. Basic Bayes theorem Bayes theorem relates the conditional probabilities of two events A, and B: A might be a hypothesis and B might.
Introduction Simple Random Sampling Stratified Random Sampling
Let X 1, X 2,..., X n be a set of independent random variables having a common distribution, and let E[ X i ] = . then, with probability 1 Strong law.
General Linear Model With correlated error terms  =  2 V ≠  2 I.
Mean, Proportion, CLT Bootstrap
PROBABILITY. Uncertainty  Let action A t = leave for airport t minutes before flight from Logan Airport  Will A t get me there on time ? Problems :
FTP Biostatistics II Model parameter estimations: Confronting models with measurements.
© by Kenneth H. Rosen, Discrete Mathematics & its Applications, Sixth Edition, Mc Graw-Hill, 2007 Chapter 1: (Part 2): The Foundations: Logic and Proofs.
6-1 Introduction To Empirical Models 6-1 Introduction To Empirical Models.
Probability Dr. Deshi Ye Outline  Introduction  Sample space and events  Probability  Elementary Theorem.
The Simple Linear Regression Model: Specification and Estimation
1-1 Regression Models  Population Deterministic Regression Model Y i =  0 +  1 X i u Y i only depends on the value of X i and no other factor can affect.
Chapter 10 Simple Regression.
Chapter 4 Multiple Regression.
SYSTEMS Identification
Programme in Statistics (Courses and Contents). Elementary Probability and Statistics (I) 3(2+1)Stat. 101 College of Science, Computer Science, Education.
Probability theory 2010 Outline  The need for transforms  Probability-generating function  Moment-generating function  Characteristic function  Applications.
Elementary hypothesis testing Purpose of hypothesis testing Type of hypotheses Type of errors Critical regions Significant levels Hypothesis vs intervals.
Lecture 24: Thurs. Dec. 4 Extra sum of squares F-tests (10.3) R-squared statistic (10.4.1) Residual plots (11.2) Influential observations (11.3,
Probability theory 2008 Conditional probability mass function  Discrete case  Continuous case.
Chapter 11 Multiple Regression.
Probability theory 2010 Conditional distributions  Conditional probability:  Conditional probability mass function: Discrete case  Conditional probability.
The moment generating function of random variable X is given by Moment generating function.
Continuous Random Variables and Probability Distributions
Approximations to Probability Distributions: Limit Theorems.
Standard error of estimate & Confidence interval.
The Neymann-Pearson Lemma Suppose that the data x 1, …, x n has joint density function f(x 1, …, x n ;  ) where  is either  1 or  2. Let g(x 1, …,
Regression Analysis Regression analysis is a statistical technique that is very useful for exploring the relationships between two or more variables (one.
1 9/8/2015 MATH 224 – Discrete Mathematics Basic finite probability is given by the formula, where |E| is the number of events and |S| is the total number.
Chapter 1 Probability and Distributions Math 6203 Fall 2009 Instructor: Ayona Chatterjee.
1 9/23/2015 MATH 224 – Discrete Mathematics Basic finite probability is given by the formula, where |E| is the number of events and |S| is the total number.
Theory of Probability Statistics for Business and Economics.
Use of moment generating functions 1.Using the moment generating functions of X, Y, Z, …determine the moment generating function of W = h(X, Y, Z, …).
ECE 8443 – Pattern Recognition LECTURE 07: MAXIMUM LIKELIHOOD AND BAYESIAN ESTIMATION Objectives: Class-Conditional Density The Multivariate Case General.
Probability & Statistics I IE 254 Exam I - Reminder  Reminder: Test 1 - June 21 (see syllabus) Chapters 1, 2, Appendix BI  HW Chapter 1 due Monday at.
Multiple Regression and Model Building Chapter 15 Copyright © 2014 by The McGraw-Hill Companies, Inc. All rights reserved.McGraw-Hill/Irwin.
11 Chapter 12 Quantitative Data Analysis: Hypothesis Testing © 2009 John Wiley & Sons Ltd.
Chapter 7 Sampling and Sampling Distributions ©. Simple Random Sample simple random sample Suppose that we want to select a sample of n objects from a.
EMIS 7300 SYSTEMS ANALYSIS METHODS FALL 2005 Dr. John Lipp Copyright © Dr. John Lipp.
Academic Research Academic Research Dr Kishor Bhanushali M
Lecture V Probability theory. Lecture questions Classical definition of probability Frequency probability Discrete variable and probability distribution.
1 What Practitioners Need to know... By Mark Kritzman Holding Period Return Holding Period Return  HPR = (Ending Price – Beginning Price + Income) / Beginning.
Statistical Decision Theory Bayes’ theorem: For discrete events For probability density functions.
The Simple Linear Regression Model: Specification and Estimation ECON 4550 Econometrics Memorial University of Newfoundland Adapted from Vera Tabakova’s.
"Classical" Inference. Two simple inference scenarios Question 1: Are we in world A or world B?
Axiomatic Theory of Probabilistic Decision Making under Risk Pavlo R. Blavatskyy University of Zurich April 21st, 2007.
SYSTEMS Identification Ali Karimpour Assistant Professor Ferdowsi University of Mashhad Reference: “System Identification Theory For The User” Lennart.
Point Estimation of Parameters and Sampling Distributions Outlines:  Sampling Distributions and the central limit theorem  Point estimation  Methods.
Chapter 5 Joint Probability Distributions and Random Samples  Jointly Distributed Random Variables.2 - Expected Values, Covariance, and Correlation.3.
Continuous Random Variables and Probability Distributions
Central Limit Theorem Let X 1, X 2, …, X n be n independent, identically distributed random variables with mean  and standard deviation . For large n:
Statistical NLP: Lecture 4 Mathematical Foundations I: Probability Theory (Ch2)
Sampling and Sampling Distributions
Introduction to Quantitative Research
Bayesian Semi-Parametric Multiple Shrinkage
CORRELATION.
Appendix A: Probability Theory
Sample Mean Distributions
Chapter 7: Sampling Distributions
CHAPTER 29: Multiple Regression*
Laboratory and Field Experiments
Review of Hypothesis Testing
Statistical NLP: Lecture 4
COMP60611 Fundamentals of Parallel and Distributed Systems
Welcome to the wonderful world of Probability
COMP60621 Designing for Parallelism
Review of Hypothesis Testing
Experiments, Outcomes, Events and Random Variables: A Revisit
Introductory Statistics
Presentation transcript:

Models of Choice

Agenda Administrivia –Readings –Programming –Auditing –Late HW –Saturated –HW 1 Models of Choice –Thurstonian scaling –Luce choice theory –Restle choice theory Quantitative vs. qualitative tests of models. Rumelhart & Greeno (1971) Conditioning… Next assignment

Choice The same choice is not always made in the “same” situation. Main assumption: Choice alternatives have choice probabilities.

Overview of 3 Models Thurstone & Luce –Responses have an associated ‘strength’. –Choice probability results from the strengths of the choice alternatives. Restle –The factors in the probability of a choice cannot be combined into a simple strength, but must be assessed individually.

Thurstone Scaling Assumptions –The strongest of a set of alternatives will be selected. –All alternatives gives rise to a probabilistic distribution (discriminal dispersions) of strengths.

Thurstone Scaling Let x j denote the discriminal process produced by stimulus j. The probability that Object k is preferred to Stimulus j is given by –P(x k > x j ) = P(x k - x j > 0)

Thurstone Scaling Assume x j & x k are normally distributed with means  j &  k, variances  j &  k, and correlation r jk. Then the distribution of x k - x j is normal with –mean  k -  j –variance  j 2 +  k r jk  j  k =  jk 2

Thurstone Scaling

Special cases: –Case III: r = 0 If n stimuli, n means, n variances, 2n parameters. –Case V: r = 0,  j 2 =  k 2 If n stimuli, n means, n parameters.

Luce’s Choice Theory Classical strength theory explains variability in choices by assuming that response strengths oscillate. Luce assumed that response strengths are constant, but that there is variability in the process of choosing. –The probability of each response is proportional to the strength of that response.

A Problem with Thurstone Scaling Works well for 2 alternatives, not more.

Luce’s Choice Theory For Thurstone with 3 or more alternatives, it can be difficult to predict how often B will be selected over A. The probabilities of choice may depend on what other alternatives are available. Luce is based on the assumption that the relative frequency of choices of B over C should not change with the mere availability of other choices.

Luce’s Choice Axiom Mathematical probability theory cannot extend from one set of alternatives to another. For example, it might be possible for: –T1 = {ice cream, sausages} P(ice cream) > P(sausage) –T2 = {ice cream, sausages, sauerkraut} P(sausage) > P(ice cream) Need a psychological theory.

Luce’s Choice Axiom Assumption: The relative probabilities of any two alternatives would remain unchanged as other alternatives are introduced. –Menu: 20% choose beef, 30% choose chicken. –New menu with only beef & chicken: 40% choose beef, 60% choose chicken.

Luce’s Choice Axiom P T (S) is the probability of choosing any element of S given a choice from T. –P {chicken, beef, pork, veggies} (chicken, pork)

Luce’s Choice Axiom Let T be a finite subset of U such that, for every S  T, Ps is defined, Then: –(i) If P(x, y)  0, 1 for all x, y  T, then for R  S  T, P T (R) = P S (R) P T (S) –(ii) If P(x, y) = 0 for some x, y in T, then for every S  T, P T (S) = P T-{x} (S-{x})

Luce’s Choice Axiom S R T (i) If P(x, y)  0, 1 for all x, y  T, then for R  S  T, P T (R) = P S (R) P T (S)

Luce’s Choice Axiom (ii) If P(x, y) = 0 for some x, y in T, then for every S  T, P T (S) = P T-{x} (S-{x}) Why? If x is dominated by any element in T, it is dominated by all elements. Causes division problems. S T X

Luce’s Choice Theorem Theorem: There exists a positive real- valued function v on T, which is unique up to multiplication by a positive constant, such that for every S  T,

Luce’s Choice Theorem Proof: Define v(x) = kP T (x), for k > 0. Then, by the choice axiom (proof of uniqueness left to reader),

Thurstone & Luce Thurstone's Case V model becomes equivalent to the Choice Axiom if its discriminal processes are assumed to be independent double exponential random variables –This is true for 2 and 3 choice situations. –For 2 choice situations, other discriminal processes will work.

Restle A choice between 2 complex and overlapping choices depends not on their common elements, but on their differential elements. –$10 + an apple –$10 XXX X XXX P($10+A, $10) = (4 - 3)/( ) = 1

Quantitative vs. Qualitative Tests Dimensions StimulusLegsEyeHeadBody A11110 A21010 A31011 A41101 A50111 B11100 B20110 B30001 B40000

Quantitative vs. Qualitative Tests Dimensions StimulusLegsEyeHeadBody A11110 A21010 A31011 A41101 A50111 B11100 B20110 B30001 B40000 Prototype vs. Exemplar Theories

Quantitative Test P(Correct) StimulusDataPrototypeExemplar A A A A A B B B B GOF Made-up #s

Qualitative Test Dimensions StimulusLegsEyeHeadBody A11110 A21010 A31011 A41101 A50111 B11100 B20110 B30001 B40000 <- More ‘protypical’ <- Less ‘prototypcial’

Qualitative Test Dimensions StimulusLegsEyeHeadBody A11110 A21010 A31011 A41101 A50111 B11100 B20110 B30001 B40000 <- Similar to A1, A3 <- Similar to A2, B6, B7 Prototype: A1>A2 Exemplar: A2>A1

Quantitative Test P(Correct) StimulusDataPrototypeExemplar A A A A A B B B B GOF Made-up #s

Quantitative vs. Qualitative Tests You ALWAYS have to figure out how to split up your data. –Batchelder & Riefer, 1980 used E1, E2, etc instead of raw outputs. –Rumelhart & Greeno, 1971 looked at particular triples.

Caveat Qualitative tests are much more compelling and, if used properly, telling, but –qualitative tests can be viewed as specialized quantitative tests, i.e., on a subset of the data. –“qualitative” tests often rely on quantitative comparisons.