Presentation is loading. Please wait.

Presentation is loading. Please wait.

Exam 2: Rules Section 2.1 Bring a cheat sheet. One page 2 sides. Bring a calculator. Bring your book to use the tables in the back.

Similar presentations


Presentation on theme: "Exam 2: Rules Section 2.1 Bring a cheat sheet. One page 2 sides. Bring a calculator. Bring your book to use the tables in the back."— Presentation transcript:

1 Exam 2: Rules Section 2.1 Bring a cheat sheet. One page 2 sides. Bring a calculator. Bring your book to use the tables in the back.

2 Exam 2: Rules Section 2.1 Five question: One fill in the blanks One multiple choice Three to solve One of those three is from your homework Show work; if you do you might get partial credit

3 In studying uncertainty: 1)Identify the experiment of interest and understand it well (including the associated population) 2)Identify the sample space (all possible outcomes) 3)Identify an appropriate random variable that reflects what you are studying (and simple events based on this random variable) 4)Construct the probability distribution associated with the simple events based on the random variable

4 Ch3: 3.1Random Variables 3.2Probability distributions for discrete random variables 3.3Expected values 3.4The Binomial Distribution 3.5Negative Binomial and Hypergeometric 3.6The Poisson

5 Random Variables Section 3.1 A Random Variable: is a function on the outcomes of an experiment; i.e. a function on outcomes in S. A discrete random variable is one with a sample space that is finite or countably infinite. (Countably infinite => infinite yet can be matched to the integer line) A continuous random variable is one with a continuous sample space.

6 For discrete random variables, we call P(X = x) = P(x) the probability mass function (pmf). From the axioms of probability, we can show that: 1. 2. A CDF, F(x) is defined to be, Probability distributions for discrete rvs Section 3.2

7 We can also find the pmf using the CDF if we note that: Probability distributions for discrete rvs Section 3.2 So, for any two numbers a, b where a < b,

8 Expected values Section 3.3 The variance of a discrete random variable is the weighted average of the squared distance from the mean, The standard deviation, Let h(X) be a function, a and b be constants then, The expected value E(X) of a discrete random variable is the weighted average or the mean of that random variable,

9 Discrete Probability & Expected values Section 3.2-6 Talked about the following distributions: Bernoulli Binomial Hypergeometric Geometric Negative Binomial Poisson

10 Discrete Probability & Expected values Section 3.2-6 Bernoulli Two possible outcomes S and F, probability of success = p. S = {S, F}

11 Discrete Probability & Expected values Section 3.2-6 Binomial The experiment consists of a group of n independent Bernoulli sub-experiments, where n is fixed in advance of the experiment and the probability of a success is p. What we are interested in studying is the number of successes that we may observe in any run of such an experiment.

12 Discrete Probability & Expected values Section 3.2-6 Binomial The binomial random variable X = the number of successes (S’s) among n Bernoulli trials or sub- experiments. We say X is distributed Binomial with parameters n and p, The pmf can become (depending on the book),

13 The CDF can become (also depending on the book), Discrete Probability & Expected values Section 3.2-6 Binomial Tabulated in Table A.1, page 664-666

14 The Binomial Distribution Section 3.4 When to use the binomial distribution? 1.When we have n independent Bernoulli trials 2.When each Bernoulli trial is formed from a sample n of individuals (parts, animals, …) from a population with replacement. 3.When each Bernoulli trial is formed from a sample of n individuals (parts, animals, …) from a population of size N without replacement if n/N < 5%.

15 Hypergeometric Discrete Probability & Expected values Section 3.2-6 The experiment consists of a group of n dependent Bernoulli sub-experiments, where n is fixed in advance of the experiment and the probability of a success is p. What we are interested in studying is the number of successes that we may observe in any run of such an experiment.

16 The hypergeometric random variable X = the number of successes (S’s) among n trials or sub-experiments. We say X is distributed Hypergeometric with parameters N, M and n Discrete Probability & Expected values Section 3.2-6 Hypergeometric

17 The pmf can become (depending on the book), The CDF Discrete Probability & Expected values Section 3.2-6 Hypergeometric

18 Discrete Probability & Expected values Section 3.2-6 Hypergeometric

19 Discrete Probability & Expected values Section 3.2-6 Negative Binomial The experiment consists of a group of independent Bernoulli sub-experiments, where r (not n), the number of successes we are looking to observe, is fixed in advance of the experiment and the probability of a success is p. What we are interested in studying is the number of failures that precede the r th success. Called negative binomial because instead of fixing the number of trials n we fix the number of successes r.

20 The negative binomial random variable X = the number of failures (F’s) until the r th success. We say X is distributed negative Binomial with parameters r and p, Discrete Probability & Expected values Section 3.2-6 Negative Binomial pmf is:

21 CDF is Discrete Probability & Expected values Section 3.2-6 Negative Binomial

22 Discrete Probability & Expected values Section 3.2-6 Geometric A special case of the negative binomial is when r = 1, then we call the distribution geometric. The geometric random variable X = the number of failures (F’s) until the 1 st success. We say X is distributed geometric with parameter p, pmf is:

23 Discrete Probability & Expected values Section 3.2-6 Geometric CDF is

24 Discrete Probability & Expected values Section 3.2-6 Poisson We can get to the Poisson model in two ways: 1.As an approximation of the Binomial distribution 2.As a model describing the Poisson process

25 1.Approximating the Binomial distribution Rules for approximation: The math ones are: If,, and then In practice: If n is large (>50) and p is small such as np < 5, then we can approximate with, where Discrete Probability & Expected values Section 3.2-6 Poisson

26 pmf: Poisson random variable X = the number of successes (S). We say X is distributed Poisson with parameter, 1.Approximating the Binomial distribution Discrete Probability & Expected values Section 3.2-6 Poisson

27 CDF: Tabulated in Table A.2, page 667 1.Approximating the Binomial distribution Discrete Probability & Expected values Section 3.2-6 Poisson

28 2.As a model describing the Poisson process This is a process of counting events, usually, over time Assumptions: a.There exists a parameter  such that, b.There is a very small chance that 2 or more events will occur in, b.The number of events observed in is independent from that occurring in any other period. Discrete Probability & Expected values Section 3.2-6 Poisson

29 pmf: Poisson random variable X = the number of successes (S) within time period t. We say X is distributed Poisson with parameter  t, 2.As a model describing the Poisson process Discrete Probability & Expected values Section 3.2-6 Poisson

30 CDF: Tabulated in Table A.2, page 667 Discrete Probability & Expected values Section 3.2-6 Poisson

31 Ch4: 4.1Probability Density Functions 4.2CDFs and Expected Values 4.3The Normal Distribution 4.4The Exponential Distribution

32 Continuous pdfs, CDFs and Expectation Section 4.1-2 For continuous random variables, we call f(x) the probability density function (pdf). From the axioms of probability, we can show that: 1. 2. CDF

33 We can also find the pdf using the CDF if we note that: Probability distributions for discrete rvs Section 3.2 So, for any two numbers a, b where a < b,

34 Expected values Section 3.3

35 Continuous random variables Section4.2-6 Talked about the following distributions: Uniform Normal Exponential

36 Section4.2-6 Uniform Continuous random variables

37 Normal Section4.2-6 Continuous random variables The most important distribution of classical and applied statistics. CDF Expectation

38 The standard Normal Z is said to have a standard normal distribution with mean = μ = 0 and standard deviation = σ = 1, pdf, A CDF,, as provided by Table A.3 pages 668-669 Normal Section4.2-6 Continuous random variables

39 Percentiles Normal Section4.2-6 Continuous random variables z α = x(1-α) = equal to the (1-α) th percentile. If, with, then we can use the normal distribution to approximate this distribution as follows,

40 Exponential Section4.2-6 Continuous random variables Commonly used to model component life time (if that component can be assumed not to change over time) and times between occurrence of multiple events in a Poisson process. A good approximation to the geometric distribution

41 We say that a random variable is exponentially distributed,, governed by parameter λ if the pdf of its distribution is, CDF, Expectation, Exponential Section4.2-6 Continuous random variables

42 Chebyshev’s rule: Says that no matter what probability distribution you are looking at the chance that an observed simple event of an experiment (from now on we will hand waive it and call it an outcome) will be between k standard deviations from the mean of the distribution is going to be at least 1 – 1/k 2 In simple math: For any type of random variables


Download ppt "Exam 2: Rules Section 2.1 Bring a cheat sheet. One page 2 sides. Bring a calculator. Bring your book to use the tables in the back."

Similar presentations


Ads by Google