OUTLINE Probability Theory Linear Algebra Probability makes extensive use of set operations, A set is a collection of objects, which are the elements.

Slides:



Advertisements
Similar presentations
CS433: Modeling and Simulation
Advertisements

Random Variables ECE460 Spring, 2012.
2 – In previous chapters: – We could design an optimal classifier if we knew the prior probabilities P(wi) and the class- conditional probabilities P(x|wi)
Week 21 Basic Set Theory A set is a collection of elements. Use capital letters, A, B, C to denotes sets and small letters a 1, a 2, … to denote the elements.
Chapter 4 Probability and Probability Distributions
Chapter 4 Discrete Random Variables and Probability Distributions
Probability Review 1 CS479/679 Pattern Recognition Dr. George Bebis.
Probability Theory Part 2: Random Variables. Random Variables  The Notion of a Random Variable The outcome is not always a number Assign a numerical.
Chapter 2: Probability.
Probability Distributions
Stats Probability Theory.
Data Basics. Data Matrix Many datasets can be represented as a data matrix. Rows corresponding to entities Columns represents attributes. N: size of the.
1 Engineering Computation Part 5. 2 Some Concepts Previous to Probability RANDOM EXPERIMENT A random experiment or trial can be thought of as any activity.
Visual Recognition Tutorial1 Random variables, distributions, and probability density functions Discrete Random Variables Continuous Random Variables.
Probability and Statistics Review Thursday Sep 11.
Random Variable and Probability Distribution
Lecture II-2: Probability Review
Joint Probability distribution
NIPRL Chapter 2. Random Variables 2.1 Discrete Random Variables 2.2 Continuous Random Variables 2.3 The Expectation of a Random Variable 2.4 The Variance.
Separate multivariate observations
Prof. SankarReview of Random Process1 Probability Sample Space (S) –Collection of all possible outcomes of a random experiment Sample Point –Each outcome.
: Appendix A: Mathematical Foundations 1 Montri Karnjanadecha ac.th/~montri Principles of.
Sets, Combinatorics, Probability, and Number Theory Mathematical Structures for Computer Science Chapter 3 Copyright © 2006 W.H. Freeman & Co.MSCS SlidesProbability.
Sets, Combinatorics, Probability, and Number Theory Mathematical Structures for Computer Science Chapter 3 Copyright © 2006 W.H. Freeman & Co.MSCS SlidesProbability.
Census A survey to collect data on the entire population.   Data The facts and figures collected, analyzed, and summarized for presentation and.
Chapter 1 Probability and Distributions Math 6203 Fall 2009 Instructor: Ayona Chatterjee.
Theory of Probability Statistics for Business and Economics.
0 K. Salah 2. Review of Probability and Statistics Refs: Law & Kelton, Chapter 4.
BINOMIALDISTRIBUTION AND ITS APPLICATION. Binomial Distribution  The binomial probability density function –f(x) = n C x p x q n-x for x=0,1,2,3…,n for.
Probability theory Petter Mostad Sample space The set of possible outcomes you consider for the problem you look at You subdivide into different.
Discrete Random Variables. Numerical Outcomes Consider associating a numerical value with each sample point in a sample space. (1,1) (1,2) (1,3) (1,4)
STA347 - week 51 More on Distribution Function The distribution of a random variable X can be determined directly from its cumulative distribution function.
Computer Vision Group Prof. Daniel Cremers Autonomous Navigation for Flying Robots Lecture 5.2: Recap on Probability Theory Jürgen Sturm Technische Universität.
Appendix : Probability Theory Review Each outcome is a sample point. The collection of sample points is the sample space, S. Sample points can be aggregated.
CS433 Modeling and Simulation Lecture 03 – Part 01 Probability Review 1 Dr. Anis Koubâa Al-Imam Mohammad Ibn Saud University
The two way frequency table The  2 statistic Techniques for examining dependence amongst two categorical variables.
Stats Probability Theory Summary. The sample Space, S The sample space, S, for a random phenomena is the set of all possible outcomes.
Probability Review-1 Probability Review. Probability Review-2 Probability Theory Mathematical description of relationships or occurrences that cannot.
PROBABILITY, PROBABILITY RULES, AND CONDITIONAL PROBABILITY
Conditional Probability Mass Function. Introduction P[A|B] is the probability of an event A, giving that we know that some other event B has occurred.
Review of Probability. Important Topics 1 Random Variables and Probability Distributions 2 Expected Values, Mean, and Variance 3 Two Random Variables.
1 Probability and Statistical Inference (9th Edition) Chapter 4 Bivariate Distributions November 4, 2015.
V7 Foundations of Probability Theory „Probability“ : degree of confidence that an event of an uncertain nature will occur. „Events“ : we will assume that.
Probability Theory Modelling random phenomena. Permutations the number of ways that you can order n objects is: n! = n(n-1)(n-2)(n-3)…(3)(2)(1) Definition:
Random Variables. Numerical Outcomes Consider associating a numerical value with each sample point in a sample space. (1,1) (1,2) (1,3) (1,4) (1,5) (1,6)
Lecture 6 Dustin Lueker.  Standardized measure of variation ◦ Idea  A standard deviation of 10 may indicate great variability or small variability,
+ Chapter 5 Overview 5.1 Introducing Probability 5.2 Combining Events 5.3 Conditional Probability 5.4 Counting Methods 1.
Chapter 2: Probability. Section 2.1: Basic Ideas Definition: An experiment is a process that results in an outcome that cannot be predicted in advance.
Basic Probability. Introduction Our formal study of probability will base on Set theory Axiomatic approach (base for all our further studies of probability)
Dr. Michael Nasief 1. MathematicsMeasure TheoryProbability TheoryRandom Process theory 2.
1 Objective To provide background material in support of topics in Digital Image Processing that are based on matrices and/or vectors. Review Matrices.
Pattern Recognition Probability Review
Virtual University of Pakistan
PROBABILITY AND PROBABILITY RULES
What is Probability? Quantification of uncertainty.
Probability for Machine Learning
Appendix A: Probability Theory
Review of Probability and Estimators Arun Das, Jason Rebello
Advanced Artificial Intelligence
Chapter 2. Random Variables
Discrete Random Variables: Basics
Sets and Probabilistic Models
Discrete Random Variables: Joint PMFs, Conditioning and Independence
Experiments, Outcomes, Events and Random Variables: A Revisit
Discrete Random Variables: Expectation, Mean and Variance
Quick Review of Probability
Sets and Probabilistic Models
Discrete Random Variables: Basics
Sets and Probabilistic Models
Sets, Combinatorics, Probability, and Number Theory
Presentation transcript:

OUTLINE Probability Theory Linear Algebra

Probability makes extensive use of set operations, A set is a collection of objects, which are the elements of the set, If S is a set and x is an element of S, we write x ∈ S. If x is not an element of S, we write x / ∈ S. A set can have no elements, in which case it is called the empty set, denoted by Ø. Probability Theory: Sets

Sets can be specified as For example, – the set of possible outcomes of a die roll is {1, 2, 3, 4, 5, 6}, – The set of possible outcomes of a coin toss is {H, T}, where H stands for “heads” and T stands for “tails.” Probability Theory: Sets

Alternatively, we can consider the set of all x that have a certain property P, and denote it by (The symbol “|” is to be read as “such that.”) For example, – the set of even integers can be written as {k | k/2 is integer}. Probability Theory: Sets

Complement: – The complement of a set S, with respect to the universe Ω, is the set {x ∈ Ω|x / ∈ S} of all elements of Ω that do not belong to S, and is denoted by S c. Union: Intersection: Probability Theory: Sets Operations

Disjoint: – several sets are said to be disjoint if no two of them have a common element Partition: – A collection of sets is said to be a partition of a set S if the sets in the collection are disjoint and their union is S. Probability Theory: Sets Operations

The set of all possible outcomes of an experiment is the sample space, denoted Ω. An event A is a (set of) possible outcomes of the experiment, and corresponds to a subset of Ω. Probability Theory

A probability law / measure is a function P(A) – with the argument A, – that assigns a value to A based on the expected proportion of number of times that event A is actually likely to happen. Probability Theory

Probability Theory: Axioms of Probability

The probability function P(A) must satisfy the following: Probability Theory: Axioms of Probability

Probability Theory: Axioms of Probability (Example)

Probability Theory

In many probabilistic models, the outcomes are of a numerical nature, e.g., if they correspond to instrument readings or stock prices. In other experiments, the outcomes are not numerical, but they may be associated with some numerical values of interest. For example, if the experiment is the selection of students from a given population, we may wish to consider their grade point average. When dealing with such numerical values, it is often useful to assign probabilities to them. This is done through the notion of a random variable. Probability Theory: Random Variables

Briefly: – A random variable X is a function that maps every possible eventin the space Ωof a random experiment to a real number. Probability Theory: Random Variables

Random variables can discrete, e.g., the number of heads in three consecutive coin tosses, or continuous, the weight of a class member. Probability Theory: Random Variables

Random variables can – discrete, e.g., the number of heads in three consecutive coin tosses, or – continuous, the weight of a class member. Probability Theory: Random Variables

A probability mass (distribution) function is a function that tells us the probability of x, an observation of X, assuming a specific value. – P(X=x) The cumulative mass (distribution) function indicates the probability of X assuming a value less then or equal to x. – F(x) = P(X<=x) Probability Theory: Probability/Cumulative Mass Function

For continuos variables: – Probability Mass Function  Probability Density function, – Cumulative Mass Function  Cumulative Density Function Probability Theory: Probability/Cumulative Density Function

The PMF of a random variable X provides us with several numbers, the probabilities of all the possible values of X. It would be desirable to summarize this information in a single representative number. This is accomplished by the expectation of X, which is a weighted (in proportion to probabilities) average of the possible values of X. Expectation

Example: – suppose you spin a wheel of fortune many times, – at each spin, one of the numbers m1,m2,..., mn comes up with corresponding probability p1, p2,..., pn, and – this is your reward from that spin. Expectation

Example: – What is the amount of money that you “expect” to get “per spin”? Expectation

Example: – Suppose that you spin the wheel k times, and that ki is the number of times that the outcome is mi. Then, the total amount received is m1k1 +m2k2 +· · ·+ mnkn. The amount received per spin is Expectation

Example: – If the number of spins k is very large, and if we are willing to interpret probabilities as relative frequencies, it is reasonable to anticipate that mi comes up a fraction of times that is roughly equal to pi: Expectation

Example: – Thus, the amount of money per spin that you “expect” to receive is Expectation

The expected value, or average, of a random variable X, whose possible values are {x1,…,xm} with respective probabilities p1,…,pm, is given as: Expectation

Moments of a random variable: Expectation

the variance, the average dispersion of the data from the mean Expectation

How ?

Expectation

Mean, second moment, variance ?

Expectation

The variance provides a measure of dispersion of X around its mean. An-other measure of dispersion is the standard deviation of X, which is defined as the square root of the variance Expectation

The standard deviation is often easier to interpret, because it has the same units as X. For example, if X measures length in meters, the units of variance are square meters, while the units of the standard deviation are meters. Expectation

Consider two discrete random variables X and Y associated with the same experiment. The joint PMF of X and Y is defined by Pairs of Random Variables

Joint probability also need to satisfy the axioms of the probability theory Everything that relate to X, or Y – individually or together – can be obtained from the P(x,y). In particular, the individual pmfs, called the marginal distribution functions can be obtained as Pairs of Random Variables

Random variables X and Y are said to be statistically independent, if and only if That is, if the outcome of one event does not effect the outcome of the other, they are statistically independent. For example, the outcome of two individual dice are independent, as one does not affect the other. Statistical Independence

Example

Example (cont.)

Expected values, moments and variances of joint distributions can be computed similar to single variable cases: Pairs of Random Variables

A cross-moment can also be defined as the covariance Covariance defines how the variables vary together as a pair – are they both increasing together, does one increase when the other decrease, etc Co-Variance

the covariance matrix, denoted by ∑ Co-Variance

If ρ=1, then the variables are identical, they move together, If ρ=-1, then the variables are negatively correlated, one decreases as the other increases at the same rate If ρ=0 the variables are uncorrelated. The variation of one, has no effect on the other. Correlation Coefficient

The conditional PMF of a random variable X, conditioned on a particular event A with P(A) > 0, is defined by Conditional Probability

A is a legitimate PMF. As an example, let X be the roll of a die and let A be the event that the roll is an even number. Then, by applying the preceding formula, we obtain Conditional Probability

The conditional probability of X=x given the Y=y has been observed is given as Conditional Probability

Example

Example (cont.)

Law of Total Probability

Bayes Rule

Example

Many Dimensions

Random Vectors

Gaussian Distribution

Multivariate Gaussian Distribution

Linear Algebra x, u, v, w (bold face, lower case) d-dimensional column vector xi ith element of the vector X (bold face upper case) dxk dimensional matrix

Linear Algebra A linear combination of vectors is another vector: A collection of vectors are linearly dependent, if any one of them can be written as a linear combination of others with at least one non-zero scalar.

Linear Algebra A basis of V is a collection of linearly independent vectors such that any vector v in V can be written as a linear combination of these basis vectors. That if B={u1,u2,…,un} is a basis for V, than any v in V can be written as

Linear Algebra

An inner product in a vector space, is a way to multiply vectors together, with the result of this multiplication being a scalar.vector spacevectorsscalar More precisely, for a real vector space, an inner product satisfies the following four properties:real vector space

Linear Algebra

Orthogonality: Two vectors, x and y, in an inner product space, V, are orthogonal if their inner product, is zero.vectorsinner product spaceinner product

Linear Algebra Gradient: Jacobian:

Linear Algebra Hessian:

Linear Algebra Taylor expansions for vector valued functions:

References R.O. Duda, P.E. Hart, and D.G. Stork, Pattern Classification, New York: John Wiley, Dimitri P. Bertsekas and John N. Tsitsiklis, Introduction to Probability, 2nd Edition, Dimitri P. BertsekasJohn N. Tsitsiklis Robi Polikar, Rowan University, Glassboro, Pattern Recognition Course Materials, 2001.