Primer on Probability Sushmita Roy BMI/CS 576 Sushmita Roy Oct 2nd, 2012 BMI/CS 576.

Slides:



Advertisements
Similar presentations
Bayes rule, priors and maximum a posteriori
Advertisements

Section 10.2: Tests of Significance
CS CS1512 Foundations of Computing Science 2 Lecture 23 Probability and statistics (4) © J.
Class 6: Hypothesis testing and confidence intervals
Markov models and applications
Sequence comparison: Significance of similarity scores Genome 559: Introduction to Statistical and Computational Genomics Prof. James H. Thomas.
Pairwise Sequence Alignment Sushmita Roy BMI/CS 576 Sushmita Roy Sep 10 th, 2013 BMI/CS 576.
Commonly Used Distributions
Pick Me. Basic Probability Population – a set of entities concerning which statistical inferences are to be drawn Random Sample – all member of population.
MOMENT GENERATING FUNCTION AND STATISTICAL DISTRIBUTIONS
Probabilistic models Haixu Tang School of Informatics.
Random Variables ECE460 Spring, 2012.
Probability Distributions CSLU 2850.Lo1 Spring 2008 Cameron McInally Fordham University May contain work from the Creative Commons.
Scores and substitution matrices in sequence alignment Sushmita Roy BMI/CS 576 Sushmita Roy Sep 11 th,
Simulation with ArenaAppendix C – A Refresher on Probability and StatisticsSlide 1 of 33 A Refresher on Probability and Statistics Appendix C.
Business Statistics for Managerial Decision
.. . Parameter Estimation using likelihood functions Tutorial #1 This class has been cut and slightly edited from Nir Friedman’s full course of 12 lectures.
Primer on Probability for Discrete Variables BMI/CS 576 Colin Dewey Fall 2010.
OUTLINE Scoring Matrices Probability of matching runs Quality of a database match.
Probability - 1 Probability statements are about likelihood, NOT determinism Example: You can’t say there is a 100% chance of rain (no possibility of.
Parameter Estimation using likelihood functions Tutorial #1
Chapter 4 Probability.
This presentation has been cut and slightly edited from Nir Friedman’s full course of 12 lectures which is available at Changes.
C82MCP Diploma Statistics School of Psychology University of Nottingham 1 Overview Parameters and Statistics Probabilities The Binomial Probability Test.
Class 3: Estimating Scoring Rules for Sequence Alignment.
Continuous Random Variables and Probability Distributions
Alignment III PAM Matrices. 2 PAM250 scoring matrix.
Probability (cont.). Assigning Probabilities A probability is a value between 0 and 1 and is written either as a fraction or as a proportion. For the.
Chapter 6 Probability.
Chapter Nine: Evaluating Results from Samples Review of concepts of testing a null hypothesis. Test statistic and its null distribution Type I and Type.
Alignment Statistics and Substitution Matrices BMI/CS 576 Colin Dewey Fall 2010.
Review of Probability Concepts ECON 4550 Econometrics Memorial University of Newfoundland Adapted from Vera Tabakova’s notes.
Chapter 1 Basics of Probability.
Short Resume of Statistical Terms Fall 2013 By Yaohang Li, Ph.D.
1 Probability and Statistics  What is probability?  What is statistics?
Introduction to Probability Theory March 24, 2015 Credits for slides: Allan, Arms, Mihalcea, Schutze.
Mid-Term Review Final Review Statistical for Business (1)(2)
Theory of Probability Statistics for Business and Economics.
Introduction to Bayesian statistics Yves Moreau. Overview The Cox-Jaynes axioms Bayes’ rule Probabilistic models Maximum likelihood Maximum a posteriori.
Applied Business Forecasting and Regression Analysis Review lecture 2 Randomness and Probability.
1 Dr. Jerrell T. Stracener, SAE Fellow Leadership in Engineering EMIS 7370/5370 STAT 5340 : PROBABILITY AND STATISTICS FOR SCIENTISTS AND ENGINEERS Systems.
1 Chapters 6-8. UNIT 2 VOCABULARY – Chap 6 2 ( 2) THE NOTATION “P” REPRESENTS THE TRUE PROBABILITY OF AN EVENT HAPPENING, ACCORDING TO AN IDEAL DISTRIBUTION.
From Randomness to Probability
Review of Probability Concepts ECON 6002 Econometrics Memorial University of Newfoundland Adapted from Vera Tabakova’s notes.
Comp. Genomics Recitation 3 The statistics of database searching.
Chapter 4 Probability ©. Sample Space sample space.S The possible outcomes of a random experiment are called the basic outcomes, and the set of all basic.
Intro to Probability for Discrete Variables BMI/CS 576 Colin Dewey Fall 2015.
Probability Course web page: vision.cis.udel.edu/cv March 19, 2003  Lecture 15.
1 CS 552/652 Speech Recognition with Hidden Markov Models Winter 2011 Oregon Health & Science University Center for Spoken Language Understanding John-Paul.
Essential Statistics Chapter 91 Introducing Probability.
Lesson 6 - R Discrete Probability Distributions Review.
CHAPTER 10 Introducing Probability BPS - 5TH ED.CHAPTER 10 1.
POSC 202A: Lecture 4 Probability. We begin with the basics of probability and then move on to expected value. Understanding probability is important because.
BPS - 3rd Ed. Chapter 91 Introducing Probability.
Random Variables Learn how to characterize the pattern of the distribution of values that a random variable may have, and how to use the pattern to find.
V7 Foundations of Probability Theory „Probability“ : degree of confidence that an event of an uncertain nature will occur. „Events“ : we will assume that.
Discrete Random Variables. Introduction In previous lectures we established a foundation of the probability theory; we applied the probability theory.
Random Variables. Numerical Outcomes Consider associating a numerical value with each sample point in a sample space. (1,1) (1,2) (1,3) (1,4) (1,5) (1,6)
The statistics of pairwise alignment BMI/CS 576 Colin Dewey Fall 2015.
1 Chapter 10 Probability. Chapter 102 Idea of Probability u Probability is the science of chance behavior u Chance behavior is unpredictable in the short.
1 Review of Probability and Random Processes. 2 Importance of Random Processes Random variables and processes talk about quantities and signals which.
1 Dr. Jerrell T. Stracener, SAE Fellow Leadership in Engineering EMIS 7370/5370 STAT 5340 : PROBABILITY AND STATISTICS FOR SCIENTISTS AND ENGINEERS Systems.
AP Statistics From Randomness to Probability Chapter 14.
Substitution Matrices and Alignment Statistics BMI/CS 776 Mark Craven February 2002.
From last time: on-policy vs off-policy Take an action Observe a reward Choose the next action Learn (using chosen action) Take the next action Off-policy.
Review of Probability and Estimators Arun Das, Jason Rebello
Review of Probability Concepts
Honors Statistics From Randomness to Probability
M248: Analyzing data Block A UNIT A3 Modeling Variation.
Essential Statistics Introducing Probability
Presentation transcript:

Primer on Probability Sushmita Roy BMI/CS Sushmita Roy Oct 2nd, 2012 BMI/CS 576

Definition of probability frequentist interpretation: the probability of an event from a random experiment is the proportion of the time events of same kind will occur in the long run, when the experiment is repeated examples – the probability my flight to Chicago will be on time – the probability this ticket will win the lottery – the probability it will rain tomorrow always a number in the interval [0,1] 0 means never occurs 1 means always occurs

Sample spaces sample space: a set of possible outcomes for some event event: a subset of sample space examples – flight to Chicago: {on time, late} – lottery: {ticket 1 wins, ticket 2 wins,…,ticket n wins} – weather tomorrow: {rain, not rain} or {sun, rain, snow} or {sun, clouds, rain, snow, sleet} or…

Random variables random variable: a function associating a value with an attribute of the outcome of an experiment example – X represents the outcome of my flight to Chicago – we write the probability of my flight being on time as P(X = on-time) – or when its clear which variable were referring to, we may use the shorthand P(on-time)

Notation uppercase letters and capitalized words denote random variables lowercase letters and uncapitalized words denote values well denote a particular value for a variable as follows well also use the shorthand form for Boolean random variables, well use the shorthand

Probability distributions if X is a random variable, the function given by P(X = x) for each x is the probability distribution of X requirements: sun clouds rain snow sleet

Joint distributions joint probability distribution: the function given by P(X = x, Y = y) read X equals x and Y equals y example x, y P(X = x, Y = y) sun, on-time 0.20 rain, on-time 0.20 snow, on-time 0.05 sun, late 0.10 rain, late 0.30 snow, late 0.15 probability that its sunny and my flight is on time

Marginal distributions the marginal distribution of X is defined by the distribution of X ignoring other variables this definition generalizes to more than two variables, e.g.

Marginal distribution example x, y P(X = x, Y = y) sun, on-time 0.20 rain, on-time 0.20 snow, on-time 0.05 sun, late 0.10 rain, late 0.30 snow, late 0.15 xP(X = x) sun 0.3 rain 0.5 snow 0.2 joint distributionmarginal distribution for X

Conditional distributions the conditional distribution of X given Y is defined as: the distribution of X given that we know the value of Y

Conditional distribution example x, y P(X = x, Y = y) sun, on-time 0.20 rain, on-time 0.20 snow, on-time 0.05 sun, late 0.10 rain, late 0.30 snow, late 0.15 xP(X = x|Y=on-time) sun 0.20/0.45 = rain 0.20/0.45 = snow 0.05/0.45 = joint distribution conditional distribution for X given Y=on-time

Independence two random variables, X and Y, are independent if

Independence example #1 x, y P(X = x, Y = y) sun, on-time 0.20 rain, on-time 0.20 snow, on-time 0.05 sun, late 0.10 rain, late 0.30 snow, late 0.15 xP(X = x) sun 0.3 rain 0.5 snow 0.2 joint distributionmarginal distributions yP(Y = y) on-time 0.45 late 0.55 Are X and Y independent here?NO.

Independence example #2 x, y P(X = x, Y = y) sun, fly-United 0.27 rain, fly-United 0.45 snow, fly-United 0.18 sun, fly-Northwest 0.03 rain, fly-Northwest 0.05 snow, fly-Northwest 0.02 xP(X = x) sun 0.3 rain 0.5 snow 0.2 joint distributionmarginal distributions yP(Y = y) fly-United 0.9 fly-Northwest 0.1 Are X and Y independent here?YES.

two random variables X and Y are conditionally independent given Z if once you know the value of Z, knowing Y doesnt tell you anything about X alternatively Conditional independence

Conditional independence example FluFeverHeadacheP true 0.04 true false0.04 truefalsetrue0.01 truefalse 0.01 falsetrue falsetruefalse0.081 false true0.081 false Are Fever and Headache independent? NO.

Conditional independence example FluFeverHeadacheP true 0.04 true false0.04 truefalsetrue0.01 truefalse 0.01 falsetrue falsetruefalse0.081 false true0.081 false Are Fever and Headache conditionally independent given Flu : YES.

Chain rule of probability for two variables for three variables etc. to see that this is true, note that

Bayes theorem this theorem is extremely useful there are many cases when it is hard to estimate P(x | y) directly, but its not too hard to estimate P(y | x) and P(x)

Bayes theorem example MDs usually arent good at estimating P(Disorder| Symptom) theyre usually better at estimating P(Symptom | Disorder) if we can estimate P(Fever | Flu) and P(Flu) we can use Bayes Theorem to do diagnosis

Expected values the expected value of a random variable that takes on numerical values is defined as: this is the same thing as the mean we can also talk about the expected value of a function of a random variable

Expected value examples Suppose each lottery ticket costs $1 and the winning ticket pays out $100. The probability that a particular ticket is the winning ticket is

distribution over the number of successes in a fixed number n of independent trials (with same probability of success p in each) e.g. the probability of x heads in n coin flips The binomial distribution P(X=x) p=0.5p=0.1 xx

k possible outcomes on each trial probability p i for outcome x i in each trial distribution over the number of occurrences x i for each outcome in a fixed number n of independent trials e.g. with k=6 (a six-sided die) and n=30 The multinomial distribution vector of outcome occurrences

Continuous distributions Needed for real-valued attributes – Weight of an object – Height of a tree Pobability distribution needs a probability density function which must be integrated over an interval to derive probability values

Cumulative distribution function (CDF) To define a probability distribution for a continuous variable, we need to integrate f(x)

Examples of continuous distributions Uniform distribution over [a,b] Gaussian distribution

Statistics of alignment scores Q: How do we assess whether an alignment provides good evidence for homology? A: determine how likely it is that such an alignment score would result from chance. What is chance? – real but non-homologous sequences – real sequences shuffled to preserve compositional properties – sequences generated randomly based upon a DNA/protein sequence model

Model for unrelated sequences well assume that each position in the alignment is sampled randomly from some distribution of amino acids let be the probability of amino acid a the probability of an n -character alignment of x and y is given by

Model for related sequences well assume that each pair of aligned amino acids evolved from a common ancestor let be the probability that evolution gave rise to amino acid a in one sequence and b in another sequence the probability of an alignment of x and y is given by

Probabilistic model of alignments How can we decide which possibility (U or R) is more likely? one principled way is to consider the relative likelihood of the two possibilities taking the log, we get

Probabilistic model of alignments the score for an alignment is thus given by: the substitution matrix score for the pair a, b should thus be given by:

Scores from random alignments suppose we assume sequence lengths m and n a particular substitution matrix and amino-acid frequencies and we consider generating random sequences of lengths m and n and finding the best alignment of these sequences this will give us a distribution over alignment scores for random pairs of sequences

The extreme value distribution were picking the best alignments, so we want to know what the distribution of max scores for alignments against a random set of sequences looks like this is given by an extreme value distribution The one that is of most interest to us is called the Gumbel distribution PDF CDF

x is a given score threshold K and are constants that can be calculated from the substitution matrix the frequencies of the individual amino acids Assessing significance of scores the probability of observing a max score at least x from N random sequence alignments is : Effective lengths of sequences

Statistics of alignment scores E value in BLAST is the expected number of sequences one would expect to see of max score x – If p is the probability of observing a score > x, E=p*m*n – n is the sum of all sequences in a database. theory for gapped alignments not as well developed computational experiments suggest this analysis holds for gapped alignments (but K and must be estimated from data)