Using Informative Priors to Enhance Wisdom in Small Crowds.

Slides:



Advertisements
Similar presentations
Online Multi-camera Tracking with a Switiching State-Space Model Wojciech Zajdel, A. Taylan Cemgil, and Ben KrÄose ICPR 2004.
Advertisements

Bayesian inference of normal distribution
Next Semester CSCI 5622 – Machine learning (Matt Wilder)  great text by Hastie, Tibshirani, & Friedman great text ECEN 5018 – Game Theory ECEN 5322 –
Bayesian Estimation in MARK
Sampling Distributions (§ )
THE CENTRAL LIMIT THEOREM The “World is Normal” Theorem.
Parameter Estimation using likelihood functions Tutorial #1
1 Vertically Integrated Seismic Analysis Stuart Russell Computer Science Division, UC Berkeley Nimar Arora, Erik Sudderth, Nick Hay.
Distribution of Tau Distances Assessing Episodic and Semantic Contributions in Serial Recall Pernille Hemmer, Brent Miller & Mark Steyvers University of.
3.3 Toward Statistical Inference. What is statistical inference? Statistical inference is using a fact about a sample to estimate the truth about the.
Wisdom of Crowds in Human Memory: Reconstructing Events by Aggregating Memories across Individuals Mark Steyvers Department of Cognitive Sciences University.
Wisdom of Crowds and Rank Aggregation Wisdom of crowds phenomenon: aggregating over individuals in a group often leads to an estimate that is better than.
Ch 6 Introduction to Formal Statistical Inference.
Matching Experiment Class Results. Experiment Analyzed 114 subjects after removal of subjects who completed fewer than 4 problems 8 problems 2.
CENTER FOR BIOLOGICAL SEQUENCE ANALYSIS Bayesian Inference Anders Gorm Pedersen Molecular Evolution Group Center for Biological Sequence Analysis Technical.
CS 8751 ML & KDDEvaluating Hypotheses1 Sample error, true error Confidence intervals for observed hypothesis error Estimators Binomial distribution, Normal.
Chapter 6 Introduction to Sampling Distributions
Fall 2006 – Fundamentals of Business Statistics 1 Chapter 6 Introduction to Sampling Distributions.
Statistics, data, and deterministic models NRCSE.
Econ 140 Lecture 31 Univariate Populations Lecture 3.
Chapter 7: Variation in repeated samples – Sampling distributions
1 Introduction to Biostatistics (PUBHLTH 540) Sampling.
Computer vision: models, learning and inference
Inferential Statistics
STA291 Statistical Methods Lecture 16. Lecture 15 Review Assume that a school district has 10,000 6th graders. In this district, the average weight of.
Bayesian inference review Objective –estimate unknown parameter  based on observations y. Result is given by probability distribution. Bayesian inference.
Ch4 Describing Relationships Between Variables. Pressure.
Applied Quantitative Analysis and Practices LECTURE#11 By Dr. Osman Sadiq Paracha.
Learning Geographical Preferences for Point-of-Interest Recommendation Author(s): Bin Liu Yanjie Fu, Zijun Yao, Hui Xiong [KDD-2013]
Estimating  0 Estimating the proportion of true null hypotheses with the method of moments By Jose M Muino.
- 1 - Bayesian inference of binomial problem Estimating a probability from binomial data –Objective is to estimate unknown proportion (or probability of.
June 11, 2008Stat Lecture 10 - Review1 Midterm review Chapters 1-5 Statistics Lecture 10.
Processing Sequential Sensor Data The “John Krumm perspective” Thomas Plötz November 29 th, 2011.
Statistical Inference for the Mean Objectives: (Chapter 9, DeCoursey) -To understand the terms: Null Hypothesis, Rejection Region, and Type I and II errors.
Three Frameworks for Statistical Analysis. Sample Design Forest, N=6 Field, N=4 Count ant nests per quadrat.
Not in FPP Bayesian Statistics. The Frequentist paradigm Defines probability as a long-run frequency independent, identical trials Looks at parameters.
Probability Course web page: vision.cis.udel.edu/cv March 19, 2003  Lecture 15.
Statistics: Unlocking the Power of Data Lock 5 Exam 2 Review STAT 101 Dr. Kari Lock Morgan 11/13/12 Review of Chapters 5-9.
Detectability Lab. Outline I.Brief Discussion of Modeling, Sampling, and Inference II.Review and Discussion of Detection Probability and Point Count Methods.
The famous “sprinkler” example (J. Pearl, Probabilistic Reasoning in Intelligent Systems, 1988)
Bias and Variance of the Estimator PRML 3.2 Ethem Chp. 4.
Simple examples of the Bayesian approach For proportions and means.
The generalization of Bayes for continuous densities is that we have some density f(y|  ) where y and  are vectors of data and parameters with  being.
Latent regression models. Where does the probability come from? Why isn’t the model deterministic. Each item tests something unique – We are interested.
Tracking with dynamics
1 Chapter 8: Model Inference and Averaging Presented by Hui Fang.
Bayesian Approach Jake Blanchard Fall Introduction This is a methodology for combining observed data with expert judgment Treats all parameters.
Statistical NLP: Lecture 4 Mathematical Foundations I: Probability Theory (Ch2)
BAYCLONE: BAYESIAN NONPARAMETRIC INFERENCE OF TUMOR SUBCLONES USING NGS DATA SUBHAJIT SENGUPTA, JIN WANG, JUHEE LEE, PETER MULLER, KAMALAKAR GULUKOTA,
Model Comparison. Assessing alternative models We don’t ask “Is the model right or wrong?” We ask “Do the data support a model more than a competing model?”
Prediction and Missing Data. Summarising Distributions ● Models are often large and complex ● Often only interested in some parameters – e.g. not so interested.
Markov Chain Monte Carlo in R
Sampling Distributions
Understanding Sampling Distributions: Statistics as Random Variables
A Bayesian approach to recommender systems
Bayes Net Learning: Bayesian Approaches
THE CENTRAL LIMIT THEOREM
Week 10 Chapter 16. Confidence Intervals for Proportions
CONCEPTS OF HYPOTHESIS TESTING
Using Informative Priors to Enhance Wisdom in Small Crowds
Effect of Prior Knowledge and Wisdom of Crowds
Confidence Interval Estimation
Statistical NLP: Lecture 4
Probability Theory.
A Switching Observer for Human Perceptual Estimation
A Switching Observer for Human Perceptual Estimation
Bayesian vision Nisheeth 14th February 2019.
Matching Experiment Class Results.
Yalchin Efendiev Texas A&M University
Presentation transcript:

Using Informative Priors to Enhance Wisdom in Small Crowds

Eyewitness testimony Common problems Reliance on error-prone eyewitnesses Reliance on very small number of eyewitnesses (often one) This research Integrating recalled memories across multiple individuals Focus on situations involving small number of individuals with potentially poor memory Approach Incorporate informative priors into a wisdom of crowd aggregation model Prior knowledge is extracted from a different group of individuals 2

Illustrative example: Galton’s Ox Galton’s original experiment estimate weight of ox (true answer is 1198 pounds) large number of individuals (800) median answer came within 9 pounds of true weight Thought experiment Suppose we only have small numbers of individuals Suppose we know something about ox weight a priori 3

Galton’s Ox with Prior Information 4 j=1,..,N Goal: infer the true mean from the observed estimates Assumptions:

Bayesian Inference Standard solution: Measure mean absolute difference 5

Simulation Vary the number of individuals (N) and strength of the assumed prior by the researcher (σ 0 ) 6 uninformative prior (based on 50,000 repetitions of Galton’s experiment in each cell)

Application to Episodic Memory We apply the same ideas to an episodic memory task: serial recall Serial recall experiment (N=28) Study event sequence in video Order the test images from memory Norming experiment (N=16) No study phase Order test images in as natural order as possible Allows us to build a model for prior probability of each sequence 7

Materials Type I 3 videos with stereotyped event sequences (e.g. wedding) associated with “strong prior” Type II 3 videos with less predictable event sequence associated with “weak prior” Extracted 10 images for testing 8

Example Type I Sequence: Bus video 9 priorserial recall Note: first row is subject id, second row is Kendall tau to true ordering

Example Type II Sequence: Clay Video 10 priorserial recall

Empirical Results: distribution of tau distances 11 Note: bin width is two so first bar is the relative proportion to tau = 0 and tau = 1 Type I Type II

Mallows Model 12 Kendall tau distance Scaling parameter Normalizing constant Latent truth Recalled order

ωyjyj θjθj j=1,..,M Model 1: Mallows Model with Uninformative Prior

ωyjyj θjθj j=1,..,M θ*θ* ωoωo yojyoj θojθoj j=1,..,N Model 2: Mallows Model with Informative Prior (prior knowledge data) (memory data)

Example calibration result 15 (clay video; all subs) Model 1 Model 2

Wisdom of Crowds Effect (Model 1) 16

Pick worst K individuals (from episodic task) 17 (means of taus across mcmc samples; averaged across 3 videos for each type)

Modeling Results (pick worst K individuals; means of taus across mcmc samples) 18

Pick random K individuals 19 (means of taus across mcmc samples; averaged across 3 videos for each type)

Pick random K individuals 20 means of taus across mcmc samples

Left over 21

Pick worst K individuals (single tau) 22

Modeling Results (pick worst K individuals; single tau – use Borda count to aggregate samples) 23

This research Wisdom of crowds Recover original sequence of events based on recollected order from a group of individuals Prior knowledge We measure prior knowledge about event sequences Incorporate this prior knowledge into our wisdom of crowd aggregation model  informative priors 24