Introduction to Probability Theory

Slides:



Advertisements
Similar presentations
Bayes rule, priors and maximum a posteriori
Advertisements

Lecture Discrete Probability. 5.3 Bayes’ Theorem We have seen that the following holds: We can write one conditional probability in terms of the.
Lecture 4A: Probability Theory Review Advanced Artificial Intelligence.
Review of Probability. Definitions (1) Quiz 1.Let’s say I have a random variable X for a coin, with event space {H, T}. If the probability P(X=H) is.
Acknowledgement: Thanks to Professor Pagano
ฟังก์ชั่นการแจกแจงความน่าจะเป็น แบบไม่ต่อเนื่อง Discrete Probability Distributions.
Lec 18 Nov 12 Probability – definitions and simulation.
Chapter 4 Discrete Random Variables and Probability Distributions
Chapter 1 Probability Theory (i) : One Random Variable
Business Statistics: A Decision-Making Approach, 6e © 2005 Prentice-Hall, Inc. Chap 4-1 Introduction to Statistics Chapter 5 Random Variables.
Introduction to Probability Theory Rong Jin. Outline  Basic concepts in probability theory  Bayes’ rule  Random variable and distributions.
Machine Learning CMPT 726 Simon Fraser University CHAPTER 1: INTRODUCTION.
Probability Distributions
Basic Probability and Statistics CIS 8590 – Fall 2008 NLP 1.
Statistics for Managers Using Microsoft Excel, 4e © 2004 Prentice-Hall, Inc. Chap 5-1 Chapter 5 Some Important Discrete Probability Distributions Statistics.
C82MCP Diploma Statistics School of Psychology University of Nottingham 1 Overview Parameters and Statistics Probabilities The Binomial Probability Test.
Tues. March 9, 1999 n The Logic of Probability Theory – Foundations, Notation and Definitions – Axioms and Theorems – Conditional Probability, Independence.
Discrete Probability Distributions
Relevance Feedback Users learning how to modify queries Response list must have least some relevant documents Relevance feedback `correcting' the ranks.
Chapter 9 Introducing Probability - A bridge from Descriptive Statistics to Inferential Statistics.
PPT ON BUSINESS STATISTICS
Chapter 5 Discrete Probability Distribution I. Basic Definitions II. Summary Measures for Discrete Random Variable Expected Value (Mean) Variance and Standard.
Prof. SankarReview of Random Process1 Probability Sample Space (S) –Collection of all possible outcomes of a random experiment Sample Point –Each outcome.
Problem A newly married couple plans to have four children and would like to have three girls and a boy. What are the chances (probability) their desire.
Welcome To B ASIC C ONCEPT O F P ROBABILITY D EPARTMENT OF S TATISTICS D R. M D. M ESBAHUL A LAM, A SSOCIATE P ROFESSOR Tel:
Sets, Combinatorics, Probability, and Number Theory Mathematical Structures for Computer Science Chapter 3 Copyright © 2006 W.H. Freeman & Co.MSCS SlidesProbability.
Sets, Combinatorics, Probability, and Number Theory Mathematical Structures for Computer Science Chapter 3 Copyright © 2006 W.H. Freeman & Co.MSCS SlidesProbability.
1 Intro to Probability Zhi Wei 2 Outline  Basic concepts in probability theory  Random variable and probability distribution  Bayes ’ rule.
PBG 650 Advanced Plant Breeding
CSE 531: Performance Analysis of Systems Lecture 2: Probs & Stats review Anshul Gandhi 1307, CS building
Statistics & Econometrics Statistics & Econometrics Statistics & Econometrics Statistics & Econometrics Statistics & Econometrics Statistics & Econometrics.
Lecture 5a: Bayes’ Rule Class web site: DEA in Bioinformatics: Statistics Module Box 1Box 2Box 3.
1 Basic probability theory Professor Jørn Vatn. 2 Event Probability relates to events Let as an example A be the event that there is an operator error.
Bernoulli Trials Two Possible Outcomes –Success, with probability p –Failure, with probability q = 1  p Trials are independent.
BINOMIALDISTRIBUTION AND ITS APPLICATION. Binomial Distribution  The binomial probability density function –f(x) = n C x p x q n-x for x=0,1,2,3…,n for.
COMP 170 L2 L17: Random Variables and Expectation Page 1.
Chapter 5: Probability Analysis of Randomized Algorithms Size is rarely the only property of input that affects run time Worst-case analysis most common.
Week 21 Conditional Probability Idea – have performed a chance experiment but don’t know the outcome (ω), but have some partial information (event A) about.
Math b (Discrete) Random Variables, Binomial Distribution.
Sixth lecture Concepts of Probabilities. Random Experiment Can be repeated (theoretically) an infinite number of times Has a well-defined set of possible.
IE 300, Fall 2012 Richard Sowers IESE. 8/30/2012 Goals: Rules of Probability Counting Equally likely Some examples.
Probability (outcome k) = Relative Frequency of k
2003/02/19 Chapter 2 1頁1頁 Chapter 2 : Basic Probability Theory Set Theory Axioms of Probability Conditional Probability Sequential Random Experiments Outlines.
Probability Theory, Bayes’ Rule & Random Variable Lecture 6.
Welcome to MM305 Unit 3 Seminar Prof Greg Probability Concepts and Applications.
1 Part Three: Chapters 7-9 Performance Modeling and Estimation.
Statistical NLP: Lecture 4 Mathematical Foundations I: Probability Theory (Ch2)
Business Statistics, A First Course (4e) © 2006 Prentice-Hall, Inc. Chap 5-1 Chapter 5 Some Important Discrete Probability Distributions Business Statistics,
Random Variables Lecture Lecturer : FATEN AL-HUSSAIN.
Conditional Probability 423/what-is-your-favorite-data-analysis-cartoon 1.
4. Overview of Probability Network Performance and Quality of Service.
Welcome to MM305 Unit 3 Seminar Dr
What is Probability? Quantification of uncertainty.
Introduction to Probability Theory
Review of Probabilities and Basic Statistics
Basic Probability aft A RAJASEKHAR YADAV.
Econ 113 Lecture Module 2.
Discrete Probability Distributions
Daniela Stan Raicu School of CTI, DePaul University
Probability & Statistics Probability Theory Mathematical Probability Models Event Relationships Distributions of Random Variables Continuous Random.
Advanced Artificial Intelligence
Statistical NLP: Lecture 4
CS590I: Information Retrieval
Elementary Statistics
Discrete Random Variables: Basics
Mathematical Foundations of BME Reza Shadmehr
Discrete Random Variables: Basics
Daniela Stan Raicu School of CTI, DePaul University
A random experiment gives rise to possible outcomes, but any particular outcome is uncertain – “random”. For example, tossing a coin… we know H or T will.
Sets, Combinatorics, Probability, and Number Theory
Presentation transcript:

Introduction to Probability Theory Rong Jin

Outline Basic concepts in probability theory Bayes’ rule Random variable and distributions

Definition of Probability Experiment: toss a coin twice Sample space: possible outcomes of an experiment S = {HH, HT, TH, TT} Event: a subset of possible outcomes A={HH}, B={HT, TH} Probability of an event : an number assigned to an event Pr(A) Axiom 1: Pr(A)  0 Axiom 2: Pr(S) = 1 Axiom 3: For every sequence of disjoint events Example: Pr(A) = n(A)/N: frequentist statistics

Joint Probability For events A and B, joint probability Pr(AB) stands for the probability that both events happen. Example: A={HH}, B={HT, TH}, what is the joint probability Pr(AB)?

Independence Two events A and B are independent in case Pr(AB) = Pr(A)Pr(B) A set of events {Ai} is independent in case

Independence Two events A and B are independent in case Pr(AB) = Pr(A)Pr(B) A set of events {Ai} is independent in case Example: Drug test A = {A patient is a Women} B = {Drug fails} Will event A be independent from event B ? Women Men Success 200 1800 Failure

Independence Consider the experiment of tossing a coin twice Example I: A = {HT, HH}, B = {HT} Will event A independent from event B? Example II: A = {HT}, B = {TH} Disjoint  Independence If A is independent from B, B is independent from C, will A be independent from C?

Conditioning If A and B are events with Pr(A) > 0, the conditional probability of B given A is

Conditioning If A and B are events with Pr(A) > 0, the conditional probability of B given A is Example: Drug test A = {Patient is a Women} B = {Drug fails} Pr(B|A) = ? Pr(A|B) = ? Women Men Success 200 1800 Failure

Conditioning If A and B are events with Pr(A) > 0, the conditional probability of B given A is Example: Drug test Given A is independent from B, what is the relationship between Pr(A|B) and Pr(A)? A = {Patient is a Women} B = {Drug fails} Pr(B|A) = ? Pr(A|B) = ? Women Men Success 200 1800 Failure

Which Drug is Better ?

Simpson’s Paradox: View I Drug II is better than Drug I A = {Using Drug I} B = {Using Drug II} C = {Drug succeeds} Pr(C|A) ~ 10% Pr(C|B) ~ 50% Drug I Drug II Success 219 1010 Failure 1801 1190

Simpson’s Paradox: View II Female Patient A = {Using Drug I} B = {Using Drug II} C = {Drug succeeds} Pr(C|A) ~ 20% Pr(C|B) ~ 5%

Simpson’s Paradox: View II Female Patient A = {Using Drug I} B = {Using Drug II} C = {Drug succeeds} Pr(C|A) ~ 20% Pr(C|B) ~ 5% Male Patient A = {Using Drug I} B = {Using Drug II} C = {Drug succeeds} Pr(C|A) ~ 100% Pr(C|B) ~ 50%

Simpson’s Paradox: View II Drug I is better than Drug II Female Patient A = {Using Drug I} B = {Using Drug II} C = {Drug succeeds} Pr(C|A) ~ 20% Pr(C|B) ~ 5% Male Patient A = {Using Drug I} B = {Using Drug II} C = {Drug succeeds} Pr(C|A) ~ 100% Pr(C|B) ~ 50%

Conditional Independence Event A and B are conditionally independent given C in case Pr(AB|C)=Pr(A|C)Pr(B|C) A set of events {Ai} is conditionally independent given C in case

Conditional Independence (cont’d) Example: There are three events: A, B, C Pr(A) = Pr(B) = Pr(C) = 1/5 Pr(A,C) = Pr(B,C) = 1/25, Pr(A,B) = 1/10 Pr(A,B,C) = 1/125 Whether A, B are independent? Whether A, B are conditionally independent given C? A and B are independent  A and B are conditionally independent

Outline Important concepts in probability theory Bayes’ rule Random variables and distributions

Bayes’ Rule Given two events A and B and suppose that Pr(A) > 0. Then Example: Pr(R) = 0.8 R: It is a rainy day W: The grass is wet Pr(R|W) = ? Pr(W|R) R R W 0.7 0.4 W 0.3 0.6

Bayes’ Rule R: It rains W: The grass is wet R R W 0.7 0.4 W 0.3 0.6 Information Pr(W|R) R W Inference Pr(R|W)

Bayes’ Rule Posterior Likelihood Prior R R W 0.7 0.4 W 0.3 0.6 R: It rains W: The grass is wet Information: Pr(E|H) Hypothesis H Evidence E Posterior Likelihood Prior Inference: Pr(H|E)

Bayes’ Rule: More Complicated Suppose that B1, B2, … Bk form a partition of S: Suppose that Pr(Bi) > 0 and Pr(A) > 0. Then

Bayes’ Rule: More Complicated Suppose that B1, B2, … Bk form a partition of S: Suppose that Pr(Bi) > 0 and Pr(A) > 0. Then

Bayes’ Rule: More Complicated Suppose that B1, B2, … Bk form a partition of S: Suppose that Pr(Bi) > 0 and Pr(A) > 0. Then

A More Complicated Example It rains W The grass is wet U People bring umbrella Pr(UW|R)=Pr(U|R)Pr(W|R) Pr(UW| R)=Pr(U| R)Pr(W| R) R W U Pr(R) = 0.8 Pr(W|R) R R W 0.7 0.4 W 0.3 0.6 Pr(U|R) R R U 0.9 0.2 U 0.1 0.8 Pr(U|W) = ?

A More Complicated Example It rains W The grass is wet U People bring umbrella Pr(UW|R)=Pr(U|R)Pr(W|R) Pr(UW| R)=Pr(U| R)Pr(W| R) R W U Pr(R) = 0.8 Pr(W|R) R R W 0.7 0.4 W 0.3 0.6 Pr(U|R) R R U 0.9 0.2 U 0.1 0.8 Pr(U|W) = ?

A More Complicated Example It rains W The grass is wet U People bring umbrella Pr(UW|R)=Pr(U|R)Pr(W|R) Pr(UW| R)=Pr(U| R)Pr(W| R) R W U Pr(R) = 0.8 Pr(W|R) R R W 0.7 0.4 W 0.3 0.6 Pr(U|R) R R U 0.9 0.2 U 0.1 0.8 Pr(U|W) = ?

Outline Important concepts in probability theory Bayes’ rule Random variable and probability distribution

Random Variable and Distribution A random variable X is a numerical outcome of a random experiment The distribution of a random variable is the collection of possible outcomes along with their probabilities: Discrete case: Continuous case:

Random Variable: Example Let S be the set of all sequences of three rolls of a die. Let X be the sum of the number of dots on the three rolls. What are the possible values for X? Pr(X = 5) = ?, Pr(X = 10) = ?

Expectation A random variable X~Pr(X=x). Then, its expectation is In an empirical sample, x1, x2,…, xN, Continuous case: Expectation of sum of random variables

Expectation: Example Let S be the set of all sequence of three rolls of a die. Let X be the sum of the number of dots on the three rolls. What is E(X)? Let S be the set of all sequence of three rolls of a die. Let X be the product of the number of dots on the three rolls.

Variance The variance of a random variable X is the expectation of (X-E[x])2 :

Bernoulli Distribution The outcome of an experiment can either be success (i.e., 1) and failure (i.e., 0). Pr(X=1) = p, Pr(X=0) = 1-p, or E[X] = p, Var(X) = p(1-p)

Binomial Distribution n draws of a Bernoulli distribution Xi~Bernoulli(p), X=i=1n Xi, X~Bin(p, n) Random variable X stands for the number of times that experiments are successful. E[X] = np, Var(X) = np(1-p)

Plots of Binomial Distribution

Poisson Distribution Coming from Binomial distribution Fix the expectation =np Let the number of trials n A Binomial distribution will become a Poisson distribution E[X] = , Var(X) = 

Plots of Poisson Distribution

Normal (Gaussian) Distribution X~N(,) E[X]= , Var(X)= 2 If X1~N(1,1) and X2~N(2,2), X= X1+ X2 ?