Download presentation

Presentation is loading. Please wait.

Published byTrevor Doyle Modified about 1 year ago

1
Chapter 4 Probability and Probability Distributions Now that you have learned to describe a data set, how can you use sample data to draw conclusions about the sampled populations? The technique involves a statistical tool called probability. To use this tool correctly, you must first understand how it works. The first part of this chapter will teach you the new language of probability, presenting the basic concepts with simple examples. © 1998 Brooks/Cole Publishing/ITP

2
The variables that we measured in Chapters 1 and 2 can now be redefined as random variables, whose values depend on the chance selection of the elements in the sample. Using probability as a tool, you can develop probability distributions that serve as models for discrete random variables, and you can describe these random variables using a mean and standard deviation similar to those in Chapter 2. © 1998 Brooks/Cole Publishing/ITP

3
Specific Topics 1. Experiments and Events 2. Relative frequency definition of probability 3. Counting Rules (Combinations) 4. Intersections, unions, and complements 5. Conditional probability and independence 6.Additive and Multiplicative Rules of Probability 7. 8. Random variables 9. Probability distributions for discrete random variables 10.The mean and standard deviation for a discrete random variable © 1998 Brooks/Cole Publishing/ITP

4
4.1 and 4.2 The Role of Probability in Statistics and Events and the Sample Space n When a population is known, probability is used to describe the likelihood of observing a particular sample outcome, e.g., a 50% or.5 chance of getting a head or a tail in a fair toss of a coin. n When the population is unknown and only a sample from that population is available, probability is used in making statements about the makeup of the population, that is, in making statistical inferences. n We use the term experiment to describe either uncontrolled events in nature or controlled situations in a laboratory. © 1998 Brooks/Cole Publishing/ITP

5
Definition: An experiment is the process by which an observation (or measurement) is obtained. Definition: An event is an outcome of an experiment. n See Example 4.1 for a listing of some events. Definition: Two events are mutually exclusive if, when one event occurs, the other cannot, and vice versa, e.g., a head or a tail in a toss of a coin. Definition: An event that cannot be decomposed is called a simple event, e.g., a head or a tail in the toss of a coin. Definition: A set of all simple events is called the sample space, e.g., a head and a tail in the toss of a coin. Definition: An event is a collection of one or more simple events, e.g., the toss of two heads in a row. © 1998 Brooks/Cole Publishing/ITP

6
Example 4.1 Experiment: Toss a die and observe the number that appears on the upper face. Solution Event A : Observe an odd number Event B : Observe a number less than 4 Event E 1 : Observe a 1 Event E 2 : Observe a 2 Event E 3 : Observe a 3 Event E 4 : Observe a 4 Event E 5 : Observe a 5 Event E 6 : Observe a 6

7
© 1998 Brooks/Cole Publishing/ITP Figure 4.1 Venn diagram for die tossing

8
n Venn diagram: The outer box represents the sample space, which contains all of the simple events; the inner circles represent events and contain simple events. n See Figure 4.1 for a Venn diagram for die tossing. n Also see Examples 4.2 and 4.3 for examples of experiments. n Tree diagram: For displaying the sample space of an experiment, each successive level of branching in the tree corresponds to a step required to generate the final outcome. n See Example 4.4 for an example of a tree diagram. © 1998 Brooks/Cole Publishing/ITP

9
Figure 4.2 Tree diagram for Example 4.4

10
4.3 Calculating Probabilities Using Simple Events Relative frequency Frequency / n n n Requirements for Simple-Event Probabilities, - Each probability must lie between 0 and 1. - The sum of the probabilities for all simple events in S equals 1. Definition: The probability of an event A is equal to the sum of the probabilities of the simple events contained in A. © 1998 Brooks/Cole Publishing/ITP

11
Example 4.5 Toss two fair coins and record the outcome. Find the probability of observing exactly one head in the two tosses. Solution © 1998 Brooks/Cole Publishing/ITP

12
Calculating the probability of an event: 1. List all simple events in the sample space. 2. Assign an appropriate probability to each simple event. 3. Determine which simple events results in the event of interest. 4. Sum the probabilities of the simple events that result in the event of interest/ n Be careful to satisfy two conditions in your calculation: - Include all simple events in the sample space. - Assign realistic probabilities to the simple events. © 1998 Brooks/Cole Publishing/ITP

13
Counting Rule for Combinations: n See Examples 4.14 and 4.15 for examples of the counting rules, including the use of counting rules to solve a probability problem. © 1998 Brooks/Cole Publishing/ITP

14
4.5 Event Composition and Event Relations n Compound events can be formed by unions or intersections of other events. Definition: The intersection of events A and B, denoted by A B, is the event that A or B occur. Definition: The union of events A and B, denoted by A B, is the event that A or B or both occur. n See Figures 4.7 and 4.8 for Venn diagrams illustrating union and intersection. © 1998 Brooks/Cole Publishing/ITP

15
Figure 4.8 Venn diagram of A B Figure 4.7 Venn diagram of A B

16
n Example 4.16 illustrates the use of a Venn diagram to determine probabilities. Definition: When two events A and B are mutually exclusive, it means that when A occurs, B cannot, and vice versa. Mutually exclusive events are also referred to as disjoint events. Figure 4.9 Two disjoint events © 1998 Brooks/Cole Publishing/ITP

17
When events A and B are mutually exclusive: P(A B) 0 and P(A B) P(A) P(B) If P(A) and P(B) are known, we do not need to break (A B) down into simple events—we can simply sum them. n See Example 4.17. Definition: The complement of an event A, denoted A C, consists of all the simple events in the sample space S that are not in A. © 1998 Brooks/Cole Publishing/ITP Figure 4.10 The complement of an event

18
4.6 Conditional Probability and Independence The conditional probability of A, given that B has occurred, is denoted as P(A | B), where the vertical bar is read “given” and the events appearing to the right of the bar are those that you know have occurred. Definition: The conditional probability of B, given that A has occurred, is n The conditional probability of A, given that B has occurred, is © 1998 Brooks/Cole Publishing/ITP

19
Definition: Two events A and B are said to be independent if and only if either P(A| B) P(A) or P(B | A) P(B) otherwise, the events are said to be dependent. n Two events are independent if the occurrence or nonoccurrence of one of the events does not change the probability of the occurrence of the other event. © 1998 Brooks/Cole Publishing/ITP

20
Additive Rule of Probability: Given two events, A and B, the probability of their union, A B, is equal to n See Figure 4.12 for a representation of the Additive Rule. © 1998 Brooks/Cole Publishing/ITP Figure 4.12

21
Multiplicative Rule of Probability: The probability that both of the two events, A and B, occur is If A and B are independent, Similarly, if A, B, and C are mutually independent events, then the probability that A, B, and C occur is n See Example 4.21 for an example of the Multiplicative Rule of probability. © 1998 Brooks/Cole Publishing/ITP

22
Example 4.24 Consider the experiment in which three coins are tossed. Let A be the event that the toss results in at least one head. Find P (A ). Solution A C is the collection of simple events implying the event “three tails,” and because A C is the complement of A, P (A ) 1 P (A C ) Then, applying the Multiplicative Rule, we have and

23
4.8 Discrete Random Variables and Their Probability Distributions Definition: A variable x is a random variable if the value that it assumes, corresponding to the outcome of an experiment, is a chance or random event. Definition: The probability distribution for a discrete random variable is a formula, table, or graph that provides p(x), the probability associated with each of the values of x. Requirements for a Discrete Probability Distribution: – n See Example 4.25 for an example involving probability distributions. © 1998 Brooks/Cole Publishing/ITP

24
Definition: Let x be a discrete random variable with probability distribution p(x). The mean or expected value of x is given as E(x) xp(x) where the elements are summed over all values of the random variable x. Definition: Let x be a discrete random variable with probability distribution p(x) and mean . The variance of x is E [(x ) 2 ] (x ) 2 p(x) where the summation is over all values of the random variable x. Definition: The standard deviation of a random variable x is equal to the square root of its variance. n See Examples 4.26, 4.27, and 4.28 for examples of the calculation of the mean variance and standard deviation. © 1998 Brooks/Cole Publishing/ITP

Similar presentations

© 2017 SlidePlayer.com Inc.

All rights reserved.

Ads by Google