HW1 Beta(p;4,10),Beta(p;9,15), Beta(6,6) likelihood Beta(p;1,1)

Slides:



Advertisements
Similar presentations
Legacy of Ed Jaynes -- approaches to uncertainty management. Stefan Arnborg, KTH.
Advertisements

Lecture (11,12) Parameter Estimation of PDF and Fitting a Distribution Function.
Week 11 Review: Statistical Model A statistical model for some data is a set of distributions, one of which corresponds to the true unknown distribution.
Statistical Modeling and Data Analysis Given a data set, first question a statistician ask is, “What is the statistical model to this data?” We then characterize.
Bayesian inference “Very much lies in the posterior distribution” Bayesian definition of sufficiency: A statistic T (x 1, …, x n ) is sufficient for 
AP Statistics – Chapter 9 Test Review
Bayesian inference Gil McVean, Department of Statistics Monday 17 th November 2008.
Maximum Likelihood. Likelihood The likelihood is the probability of the data given the model.
Statistical Significance What is Statistical Significance? What is Statistical Significance? How Do We Know Whether a Result is Statistically Significant?
© 2010 Pearson Prentice Hall. All rights reserved Least Squares Regression Models.
EPIDEMIOLOGY AND BIOSTATISTICS DEPT Esimating Population Value with Hypothesis Testing.
Deciding, Estimating, Computing, Checking How are Bayesian posteriors used, computed and validated?
Business Statistics - QBM117
Statistical Significance What is Statistical Significance? How Do We Know Whether a Result is Statistically Significant? How Do We Know Whether a Result.
PSY 1950 Confidence and Power December, Requisite Quote “The picturing of data allows us to be sensitive not only to the multiple hypotheses that.
Fall 2006 – Fundamentals of Business Statistics 1 Chapter 8 Introduction to Hypothesis Testing.
Chapter 11 Multiple Regression.
Chapter 9 Hypothesis Testing.
G. Cowan Lectures on Statistical Data Analysis 1 Statistical Data Analysis: Lecture 7 1Probability, Bayes’ theorem, random variables, pdfs 2Functions of.
5-3 Inference on the Means of Two Populations, Variances Unknown
Inference about Population Parameters: Hypothesis Testing
Three Common Misinterpretations of Significance Tests and p-values 1. The p-value indicates the probability that the results are due to sampling error.
Multiple testing correction
Hypothesis Testing – Introduction
1 © Lecture note 3 Hypothesis Testing MAKE HYPOTHESIS ©
Review of Statistical Inference Prepared by Vera Tabakova, East Carolina University.
Section 9.1 Introduction to Statistical Tests 9.1 / 1 Hypothesis testing is used to make decisions concerning the value of a parameter.
Review of Statistical Inference Prepared by Vera Tabakova, East Carolina University ECON 4550 Econometrics Memorial University of Newfoundland.
Hypothesis Testing.
Statistical Decision Theory
Chapter 8 Introduction to Hypothesis Testing
1 Institute of Engineering Mechanics Leopold-Franzens University Innsbruck, Austria, EU H.J. Pradlwarter and G.I. Schuëller Confidence.
CHAPTER 17: Tests of Significance: The Basics
Multiple Regression Petter Mostad Review: Simple linear regression We define a model where are independent (normally distributed) with equal.
1 Chapter 10: Introduction to Inference. 2 Inference Inference is the statistical process by which we use information collected from a sample to infer.
Week 71 Hypothesis Testing Suppose that we want to assess the evidence in the observed data, concerning the hypothesis. There are two approaches to assessing.
Maximum Likelihood - "Frequentist" inference x 1,x 2,....,x n ~ iid N( ,  2 ) Joint pdf for the whole random sample Maximum likelihood estimates.
Bayesian vs. frequentist inference frequentist: 1) Deductive hypothesis testing of Popper--ruling out alternative explanations Falsification: can prove.
Statistical Hypotheses & Hypothesis Testing. Statistical Hypotheses There are two types of statistical hypotheses. Null Hypothesis The null hypothesis,
Confidence intervals and hypothesis testing Petter Mostad
Lecture 16 Section 8.1 Objectives: Testing Statistical Hypotheses − Stating hypotheses statements − Type I and II errors − Conducting a hypothesis test.
1 Chapter 9 Hypothesis Testing. 2 Chapter Outline  Developing Null and Alternative Hypothesis  Type I and Type II Errors  Population Mean: Known 
Correlation Assume you have two measurements, x and y, on a set of objects, and would like to know if x and y are related. If they are directly related,
Statistical Inference for the Mean Objectives: (Chapter 9, DeCoursey) -To understand the terms: Null Hypothesis, Rejection Region, and Type I and II errors.
Uncertainty Management in Rule-based Expert Systems
Section A Confidence Interval for the Difference of Two Proportions Objectives: 1.To find the mean and standard error of the sampling distribution.
Statistical Decision Theory Bayes’ theorem: For discrete events For probability density functions.
Simple examples of the Bayesian approach For proportions and means.
Suppose we have T genes which we measured under two experimental conditions (Ctl and Nic) in n replicated experiments t i * and p i are the t-statistic.
© Copyright McGraw-Hill 2004
Multi-target Detection in Sensor Networks Xiaoling Wang ECE691, Fall 2003.
G. Cowan Lectures on Statistical Data Analysis Lecture 8 page 1 Statistical Data Analysis: Lecture 8 1Probability, Bayes’ theorem 2Random variables and.
Introduction to Hypothesis Testing
Statistical Inference Statistical inference is concerned with the use of sample data to make inferences about unknown population parameters. For example,
Education 793 Class Notes Inference and Hypothesis Testing Using the Normal Distribution 8 October 2003.
Testing a Single Mean Module 16. Tests of Significance Confidence intervals are used to estimate a population parameter. Tests of Significance or Hypothesis.
- 1 - Outline Introduction to the Bayesian theory –Bayesian Probability –Bayes’ Rule –Bayesian Inference –Historical Note Coin trials example Bayes rule.
Jump to first page Inferring Sample Findings to the Population and Testing for Differences.
Hypothesis Testing. Statistical Inference – dealing with parameter and model uncertainty  Confidence Intervals (credible intervals)  Hypothesis Tests.
BIOL 582 Lecture Set 2 Inferential Statistics, Hypotheses, and Resampling.
Fundamentals of Data Analysis Lecture 4 Testing of statistical hypotheses pt.1.
Statistical Inference for the Mean Objectives: (Chapter 8&9, DeCoursey) -To understand the terms variance and standard error of a sample mean, Null Hypothesis,
Copyright © 2013, 2009, and 2007, Pearson Education, Inc. 1 FINAL EXAMINATION STUDY MATERIAL III A ADDITIONAL READING MATERIAL – INTRO STATS 3 RD EDITION.
© 2010 Pearson Prentice Hall. All rights reserved Chapter Hypothesis Tests Regarding a Parameter 10.
General Linear Model & Classical Inference
Hypothesis Testing and Confidence Intervals (Part 1): Using the Standard Normal Lecture 8 Justin Kern October 10 and 12, 2017.
Hypothesis Testing (Cont.)
Contrasts & Statistical Inference
Contrasts & Statistical Inference
Contrasts & Statistical Inference
Presentation transcript:

HW1 Beta(p;4,10),Beta(p;9,15), Beta(6,6) likelihood Beta(p;1,1)

FDR, Evidence Theory, Robustness

function PV=testdep(D,N) % General dependecy test TESTDEP(D,N) % D: n by d Data Matrix % N : Monte Carlo comparison sample F=[]; [n,d]=size(D); mv=mean(D); D=D-repmat(mv,n,1); %remove mean st=std(D); D=D./repmat(st,n,1); %standardize variance for i=1:d for j=1:i-1 q=mean(D(:,i).*D(:,j)); F=[F q]; end General Dependency test

EE=[]; for iN=1:N-1 E=[]; for i=1:d q=[]; if i>1 D(:,i)=D(randperm(n),i); for j=1:i-1 q=mean(D(:,i).*D(:,j)); E=[E q]; end EE=[EE;E]; end Empirical no-dependency distribution

%Sorting twice gives value ranks of EE - test statistics EE=[F ; EE]; [EEs,iix]=sort(EE); [EEs,iix]=sort(iix); % p-value is proportional to value rank PV=iix(1,:)/N; % reshuffle to matrix PVM(ix)=PV Computing P-value

Correlation coefficient >> D=[1:100]'; >> D=[D -D D.^2 D+200*rand(size(D)) randn(size(D))]; >> [c pv]=corrcoef(D) c =

Correlation coefficient >> [c pv]=corrcoef(D) pv = >> [,pv]=testdep(D,N) pv = >>

Multiple testing The probability of rejecting a true null hypothesis at 99% is 1%. Thus, if you repeat test 100 times, each time with new data, you will reject sometime with probability 0.63 Bonferroni correction, FWE control: in order to reach significance level 1% in an experiment involving 1000 tests, each test should be checked with significance 1/1000 %

Multiple testing Several approaches try to verify an excess of small p-values Sort set of p-values and test if there is an excess of small values - this is an indication of false null hypotheses

Approaches to multiple testing

Definition of FDR, positive correlation

No significance Lower envelope FDR corrected

Some significance One of 15 first tests not null-- On 5% significanc

More significance FDR: 95% of first 3 tests not null hypothesis

Even more significance 95% of first 14 tests not null - worth effort to investigate all

FDR Example - independence Fdrex(pv,0.05,0) 10 signals suggested. Smallest p-value not significant with Bonferroni correction (0.019 vs 0.013)

FDR Example - dependency Fdrex(pv,0.05,1) 10 signals suggested assuming independence all disappear with correction term

Ed Jaynes devoted a large part of his career to promote Bayesian inference. He also championed the use of Maximum Entropy in physics Outside physics, he received resistance from people who had already invented other methods. Why should statistical mechanics say anything about our daily human world??

Generalisation of Bayes/Kalman: What if: You have no prior? Likelihood infeasible to compute (imprecision)? Parameter space vague, i.e., not the same for all likelihoods? (Fuzziness, vagueness)? Parameter space has complex structure (a simple structure is e.g., a Cartesian product of reals, R, and some finite sets)?

Philippe Smets ( ) Developed Dempster’s and Shafer’s method in uncertainty management into the Transferable Belief Method, that combines imprecise ‘evidence’ (likelihood or prior) using Dempster’ rule, and uses pignistic transformation to get a sharp decision criterion

Some approaches... Robust Bayes: replace distributions by convex sets of distributions (Berger m fl) Dempster/Shafer/TBM: Describe imprecision with random sets DSm: Transform parameter space to capture vagueness. (Dezert/Smarandache, controversial) FISST: FInite Set STatistics: Generalises observation- and parameter space to product of spaces described as random sets. (Goodman, Mahler, Ngyuen)

Combining Evidence

Robust Bayes Priors and likelihoods are convex sets of probability distributions (Berger, de Finetti, Walley,...): imprecise probability: Every member of posterior is a ’parallell combination’ of one member of likelihood and one member of prior. For decision making: Jaynes recommends to use that member of posterior with maximum entropy (Maxent estimate).

Ellsberg’s Paradox: Ambiguity Avoidance ? ? ? ? Urna A innehåller 4 vita och 4 svarta kulor, och 4 av okänd färg (svart eller vit) Urna B innehåller 6 vita och 6 svarta kulor Du får en krona om du drar en svart kula. Ur vilken urna vill du dra den? En precis Bayesian bör först anta hur ?-kulorna är färgade och sedan svara. Men en majoritet föredrar urna B även om svart byts mot vit

Prospect Theory: Kahneman, Tversky Safety belts eliminate car collision injuries at low speed completely (I BUY IT!!!) Safety belts eliminate 90% of injuries in car accidents. In 10% the speed is to high (So belts are not that good!???)

Hur används imprecisa sannolikheter? Förväntad nytta för beslutsalternativ blir intervall i stället för punkter: maximax, maximin, maximedel? u a pessimist optimist Bayesian

Dempster/Shafer/Smets Evidence is random set over over . I.e., probability distribution over. Probability of singleton: ‘Belief’ allocated to alternative, i.e., probability. Probability of non-singelton: ‘Belief’ allocated to set of alternatives, but not to any part of it. Evidences combined by random intersection conditioned to be non-empty (Dempster’s rule).

Logic of Dempster’s rule Each observer has a private state space and assesses the posterior over it. Each private state can correspond to one or more global or common states, multivalued mapping Observers state spaces are assumed independent.

Correspondence DS-structure-- set of probability distributions For a pdf (bba) m over 2^ , consider all ways of reallocating the probability mass of non-singletons to their member atoms: This gives a convex set of probability distributions over . Example:  ={A,B,C} A: 0.1 B: 0.3 C: 0.1 AB: 0.5 A: *x B: *(1-x) C: 0.1 Can we regard any set of pdf:s as a bba? Answer is NO!! There are more convex sets of pdf:s than DS-structures for all x  [0,1] bba set of pdfs

Representing probability set as bba: 3-element universe Rounding up: use lower envelope. Rounding down: Linear programming Rounding is not unique!! Black: convex set Blue: rounded up Red: rounded down

Another appealing conjecture Precise pdf can be regarded as (singleton) random set. Bayesian combination of precise pdf:s corresponds to random set intersection (conditioned on non-emptiness) DS-structure corresponds to Choquet capacity (set of pdf:s) Is it reasonable to combine Choquet capacities by (nonempty) random set intersection (Dempster’s rule)?? Answer is NO!! Counterexample: Dempster’s combination cannot be obtained by combining members of prior and likelihood: Arnborg: JAIF vol 1, No 1, 2006

Consistency of fusion operators DS rule MDS rule Rounded robust Operands (evidence) Robust Fusion Dempster’s rule Modified Dempster’s rule Axes are probabilities of A and B in a 3-element universe P(A) P(B) P(C )=1-P(A)-P(B)

Deciding target type Attack aircraft: small, dynamic Bomber aircraft: large, dynamic Civilian: Large, slow dynamics Prior: (0.5,0.4,0.1); Observer 1: probably small, likelihood (0.8,0.1,0.1); Observer 2: probably fast, likelihood (0.4,0.4,0.2);

Estimators Pignistic MaxEnt Center encl sphere 3 states: P( C ) = 1-P(A)-P(B)

What about Smets’ TBM?? TBM combines the original Dempster’s rule with the pignistic transformation. This is not compatible with precise Bayesian analysis. However, there is nothing against claiming TBM to be some kind of Robust Bayesian scheme. Main problem: Dempster’s rule and its motivation using multi-valued mappings is against the dominant argumentation used in introductions and tutorials: TBM is incompatible with the Capacity interpretation of DS structures