570 likes | 784 Views
Combining Probabilities and Conditional Probability. Stat 700 Lecture 4 9/11/2001-9/13/2001. Overview of Lecture. Algebra of Events and Probabilities Conditional Probability and its Importance Probability Updating: Bayes Rule Independent Events Discrete Random Variables
E N D
Combining Probabilities and Conditional Probability Stat 700 Lecture 4 9/11/2001-9/13/2001
Overview of Lecture • Algebra of Events and Probabilities • Conditional Probability and its Importance • Probability Updating: Bayes Rule • Independent Events • Discrete Random Variables • Introduction to Probability Distributions • Parameters of Distributions Lecture 4: More Probabilities and Conditioning
Algebra of Events and Probabilities • Given an event A, its complement, denoted by Ac, is the event whose elements are those that are not in A. Thus, the event Ac is the opposite of event A. • Note that (Ac)c = A. (Rule of double complementation) • Rule 1 (Complementation Rule): For any event A, P(A) = 1 - P(Ac). • From this rule it follows immediately that P() = 0, since P(S) = 1. Lecture 4: More Probabilities and Conditioning
A Simple Application • Example: Consider the experiment of rolling two fair dice simultaneously, and let A be the event that the sum of the outcomes is at most 11. We seek P(A). • Solution: Since there are (6)(6) = 36 simple events, and we can assume that these events are equally likely, then each has probability of 1/36. Out of these 36 events, only one of them, the simple event {(6,6)}, has a sum that exceeds 11, so Ac = {(6,6)}. Therefore, by the complementation rule, Lecture 4: More Probabilities and Conditioning
Algebra … continued • Given two events, A and B, the union of A and B, denoted by (A B), is the event whose elements are those that belong to either A or B or both. It represents the occurrence of at least one of the events A or B. • We also write (A or B) for (A B). • We could generalize this to the union of several events, e.g., (A B C D) which would then represent the occurrence of at least one of these 4 events. Lecture 4: More Probabilities and Conditioning
Algebra … continued • Given two events, A and B, the intersection of A and B, denoted by (A B), is the event whose elements are those that belong to both A and B. It represents the simultaneous occurrence of A and B. • We also write (A and B) and (AB) for (A B). • We could generalize this to the intersection of several events, e.g., (A B C D) which would then represent the simultaneous occurrence of all 4 events. • If A B = = {empty event}, we say that A and B are disjoint or mutually exclusive. • Generalization: notion of pairwise disjoint events. Lecture 4: More Probabilities and Conditioning
Finite Additivity Properties • Rule 2: Given disjoint events A and B, then P(A B) = P(A) + P(B). • Extended Rule 2: Given pairwise disjoint events A1, A2, …, Ak, then Lecture 4: More Probabilities and Conditioning
Addition and Generalized Addition Rules • Rule 3 (Addition Rule): Given events A and B (which are not necessarily disjoint), then • P(A B) = P(A) + P(B) - P(A B). • Extended Rule 3 (Inclusion-Exclusion Principle): Given three events A, B, and C (which are not necessarily pairwise disjoint), then Lecture 4: More Probabilities and Conditioning
Some Concrete Applications of the Probability Rules • Example 1: A study is to be performed to examine the association between the occurrence of lung cancer and smoking. Suppose that one person is to be randomly chosen and classified into either a smoker or a nonsmoker, and whether he/she has lung cancer or not. For this experiment, the sample space is: • S ={(Nonsmoker, No lung cancer), (Nonsmoker, With lung cancer), (Smoker, No lung cancer), (Smoker, With lung cancer)}. • Assume that the proportion in the population who are smokers is 0.15, while the proportion in the population who have lung cancer is 0.05. Furthermore, assume that the proportion in the population who are smokers and with lung cancer is 0.009. • We seek the probability that the person chosen will either be a smoker or has lung cancer. Lecture 4: More Probabilities and Conditioning
Information for Example • The table below shows the information provided for in the description of the problem. • Such a table could also have been used to compute the desired probabilities … illustrated in class. Lecture 4: More Probabilities and Conditioning
Applications … continued • Solution of Example 1: Let A be the event that the person is a smoker, and B be the event that the person has lung cancer. From the given information we have that P(A) = 0.15, P(B) = 0.05, and P(AB) = 0.009. Therefore, by the addition rule, the desired probability is Lecture 4: More Probabilities and Conditioning
DeMorgan’s Rules • DeMorgan’s rules state that: • the complement of the union of events equals the intersection of their complements; • the complement of the intersections of events equals the union of their complements. • Formally, for two events A and B, we have: Lecture 4: More Probabilities and Conditioning
Example … continued • Suppose that in the preceding example we were instead interested in the probability that the person is neither a smoker nor has lung cancer. Then, in formal notation, we want: • P(Ac Bc) = P{(not a smoker) and (is free of lung cancer)}. • By virtue of deMorgan’s rule, we have that Ac Bc = (A B)c. • Applying the complementation rule, we therefore obtain Lecture 4: More Probabilities and Conditioning
Another Approach: By Completing the Table of Probabilities • The answers obtained in the preceding slides could also be derived by simply completing the table of probabilities: Lecture 4: More Probabilities and Conditioning
A Matching Problem • Example 2: Four people, Peter, Paul, Mary, and Magdalene, write their names on identically-sized chips of papers. These chips are then placed on a box and thoroughly shuffled. Each of them (in the order given above) then randomly draws a chip from the box, with the drawing being without replacement. We seek the probabilities of the following events: • a) probability that Paul draws his name; • b) probability that either Paul or Magdalene (or both) draw their respective names; and • c) probability that at least one of them draws his/her name. Lecture 4: More Probabilities and Conditioning
Matching … continued • Solution: We shall let A = {Peter draws his name}; B = {Paul draws his name}; C = {Mary draws her name}; and D = {Magdalene draws her name}. Observe that the number of outcomes {e.g., (Paul, Mary, Magdalene, Peter)} in the sample space S is (4)(3)(2)(1) = 4! = 24, and these outcomes are equally likely since the draws are done at random. • Now, for problem (a), we want: P(B) = N(B)/N(S). But N(B) = (3)(1)(2)(1) = 6 since we need “Paul” to be the second chip drawn. Therefore, P(B) = 6/24 = 1/4 = 0.25. Lecture 4: More Probabilities and Conditioning
Continued ... • For problem (b) we want P(B or D). By the addition rule, we have P(B or D) = P(B) + P(D) - P(B andD). Analogously to problem (a), we have P(D) = (3)(2)(1)(1)/24 = 1/4. • On the other hand, P(B andD) = (2)(1)(1)(1)/24 = 1/12 since both Paul and Magdalene must get their names, respectively. • Therefore, P(B or D) = 1/4 + 1/4 - 1/12 = 5/12 = 0.4167. Lecture 4: More Probabilities and Conditioning
Solution … continued • For problem (c) where we are seeking the probability of having at least one of them draw their own name, we want: • P{A B C D} = P{at least one of A, B, C, or D occurs}. • But by the generalized addition rule (inclusion-exclusion principle) for 4 events, this is: • P{A B C D} = {P(A)+P(B)+P(C)+P(D)} - {P(AB) + P(AC) + P(AD) + P(BC) + P(BD) + P(CD)} + {P(ABC) + P(ABD) + P(ACD) + P(BCD)} - P(ABCD}. Lecture 4: More Probabilities and Conditioning
Continued ... • Similar calculations as in the previous two problems now yield that (coloreds means the number of summands): • P{A B C D} = (4)(1/4) - (6)(1/12) + (4)(1/24) - (1)(1/24) = 1 - 1/2 + 1/6 - 1/24 = 1 - 1/2! + 1/3! - 1/4! = 0.625. • Food for Thought!! What do you think will happen to this probability if there were 1000 people, instead of 4 people?? Lecture 4: More Probabilities and Conditioning
Another Solution: The “Brute Force” Approach Lecture 4: More Probabilities and Conditioning
Event B: “Paul” is Matched • Therefore, P(B) = 6/24 = 1/4 = 0.25. Lecture 4: More Probabilities and Conditioning
Event (B or D): Either “Paul” or “Magdalene” or Both Are Matched • Therefore, P(B or D) = 10/24 = 5/12 = 0.4167. Lecture 4: More Probabilities and Conditioning
Event (A or B or C or D): At Least One of Them Gets a Match • Thus, P(A or B or C or D) = 15/24 = 5/8 = 0.625. Lecture 4: More Probabilities and Conditioning
Conditional Probability: Motivation • In many situations occurring in the sciences, both natural and social, one is interested in the probability of an event given other information. • For example, one may not be interested in the probability of a person getting lung cancer, but rather might be interested in the probability that the person will get lung cancer given the information that this person is a cigarette smoker. • Or, one maybe interested in knowing the probability that someone is HIV-infected given a positive result from an ELISA test (a test for HIV-infection). Lecture 4: More Probabilities and Conditioning
Definition of Conditional Probability • In such situations, we are interested in the conditional probability of an event. • Given events A and B (for some experiment), the conditional probability of B given A is: Lecture 4: More Probabilities and Conditioning
Justification of the Definition • If we are given the information that event A has occurred, then we know that the outcome of the experiment is in event A. • Therefore, given this information, event B could have occurred only if the outcome is in the intersection of A and B. Dividing P(A and B) by P(A) serves to standardize P(A and B) since, given that A has occurred, the effective sample space becomes A. • If P(A) = 0, then conditioning on an event that never occurs is clearly not of any interest! Lecture 4: More Probabilities and Conditioning
A Simple Example • Consider the experiment of tossing three fair coins simultaneously. • Thus, S = {HHH, HHT, HTH, THH, HTT, THT, TTH, TTT}, and each outcome has probability of 1/8. • Define A = identical outcomes = {HHH, TTT}, and B = at least two heads = {HHT, HTH, THH, HHH}. Thus, P(A) = 2/8 and P(B) = 4/8. Note that (A B) = {HHH} so P(A B) = 1/8. Lecture 4: More Probabilities and Conditioning
Continued ... • P(B|A) = P(A B)/P(A) = (1/8)/(2/8) = 1/2. • This conditional probability is clearly intuitive since if we know that A has occurred [so the outcome was either (HHH) or (TTT)], then the only way that B could have occurred is if the outcome was (HHH) Lecture 4: More Probabilities and Conditioning
Example … continued • On the other hand, if we are given the information that B has occurred, then • P(A|B) = P(A B)/P(B) = (1/8)/(4/8) = 1/4. • Again, this is intuitive because the information that B occurred tells us that the outcome is either (HHT), (HTH), (THH), or (HHH). In order for A to have occurred, then (HHH) must be the outcome, hence the conditional probability of A given B is 1/4. • From these two examples, note that: • P(B|A) and P(A|B) need not be identical. Lecture 4: More Probabilities and Conditioning
Another Example • Consider the hypothetical population proportions regarding smoking and the presence of lung cancer which were used in previous examples. The information is reproduced in the table below. Recall that the experiment is to randomly choose one person from this population. Lecture 4: More Probabilities and Conditioning
Example … continued • Let us denote by A the event that the person chosen is a smoker, and by B the event that the person chosen has lung cancer. • Suppose that we are given the information that the person chosen is a smoker. Then • P(has lung cancer | smoker) = P(B|A) = P(AB)/P(A) = 0.009/0.15 = 0.06. • On the other hand, if we know that the person is a nonsmoker, then • P(has lung cancer | nonsmoker) = P(B|Ac) = P(AcB)/P(Ac) = 0.041/0.85 = 0.048. Lecture 4: More Probabilities and Conditioning
Example … continued • By comparing the conditional probabilities P(B|A) = 0.060 and P(B|Ac) = 0.048 [and we could do such a comparison since the process of conditioning has standardized the probabilities for the two groups], one can conclude (in this hypothetical situation) that the prevalence of lung cancer among smokers is slightly higher than the prevalence of lung cancer among nonsmokers. • One may also look at the “inverse” probabilities: • P(A|B) = P(A and B)/P(B) = 0.009/0.05 = 0.18; and • P(A|Bc) = P(A and Bc)/P(Bc) = 0.141/0.95 = 0.148. Lecture 4: More Probabilities and Conditioning
Multiplication Rule • Given events A and B, if we are given P(A) > 0 and the conditional probability P(B|A), then the probability P(A B) could be obtained by inverting the conditional probability formula to get the multiplication rule: Lecture 4: More Probabilities and Conditioning
Utility of Multiplication Rule • The multiplication rule is what enables us to multiply the (conditional) probabilities in the branches of a tree diagram to get the (joint) probabilities of the outcomes of the experiment. • On the other hand, the multiplication rule is only usable if we are able to determine the conditional probability of B given A without recourse to the conditional probability rule [since this latter rule requires P(AB)]. • Obtaining P(B|A) this way is usually done by examining the situation at hand. Lecture 4: More Probabilities and Conditioning
An Illustration • Example: Suppose you have two boxes with Box I containing 4 red and 7 blue balls, and Box II containing 6 red and 2 blue balls. • The two-step experiment is to pick a ball at random from Box I which is then transferred to Box II. A ball is then drawn from Box II. • Let A = event that a red ball is transferred from I to II; and • let B = event that a red ball is drawn from II. • We seek: P(A and B) = P({both balls are red}) Lecture 4: More Probabilities and Conditioning
Example … continued • Clearly, P(A) = 4/11, while without using the conditional probability formula, we obtain: • P(B|A) = 7/9 since, after the transfer of a red ball, there will be 7 reds and 2 blues in Box II. • Consequently, by the multiplication rule: • P(A and B) = P(A)P(B|A) = (4/11)(7/9) = 28/99 = 0.2828. • In class we illustrate this computation via a tree diagram. Lecture 4: More Probabilities and Conditioning
Example … continued • Similarly, we obtain the joint probability of getting a blue ball from Box I and a red ball from Box II, which is symbolically represented by (Ac B) as follows: • P(Ac) = P(blue from Box I) = 7/11; • P(B|Ac) = P(red from Box II | blue ball transferred) = 6/9, so • P(Ac B) = P(Ac)P(B|Ac) = (7/11)(6/9) = 42/99 = 0.4242. Lecture 4: More Probabilities and Conditioning
Example … continued • Let us now consider the problem of determining the (marginal) probability of getting a red ball from Box II, that is, we want P(B). • Notice that we are not anymore directly interested in what happens in step 1 (getting a ball from Box I), but at the same time we see that the occurrence of event B depends on what will happen in step 1. • We are therefore faced with the problem of combining the probabilities arising from whether we transfer a red ball to Box II or we transfer a blue ball to Box II. • The question is: how do we combine? Lecture 4: More Probabilities and Conditioning
Analyzing the Situation • First we note that the event B could arise in two ways: • (draw a red from I, then draw a red from II); or • (draw a blue from I, then draw a red from II). • Symbolically, this is could be represented via: • B = (A B) (Ac B) • Furthermore, note that since A and Ac are disjoint (could not occur simultaneously), then (A B) and (Ac B) are also disjoint events. • We could therefore apply the addition rule. Lecture 4: More Probabilities and Conditioning
The Combined Probability • By the addition rule (or the finite additivity property), we obtain: • P(B) = P{(A B) (Ac B)} = P(A B) + P(Ac B) = P(A)P(B|A) + P(Ac)P(B|Ac). • From our earlier calculations where we used the multiplication rule, we obtained: • P(A B) = 28/99 and P(Ac B) = 42/99. • Therefore, P(B) = 28/99 + 42/99 = 70/99 = .7070. This is the probability of getting a red ball from Box II (taking into proper account what could happen with the draw from Box I). Lecture 4: More Probabilities and Conditioning
Theorem of Total Probabilities • The calculation of P(B), where B is an event pertaining to the result of the second step of the experiment, is a special case of what is referred to as the “Theorem of Total Probabilities,”a method for combining probabilities from the different possibilities arising from the first step of the experiment. • In its simplest form, when there are only two possibilities, A and Ac, from the first step of the experiment, the theorem states that: Lecture 4: More Probabilities and Conditioning
Updating of Probabilities • Let us go back to the example we were considering. Evidently, it is immediate that • P(A) = P(red from Box I) = 4/11 = .3636; and • P(Ac) = P(blue from Box I) = 7/11 = .6363. • These are our prior probabilities for A and Ac. • Suppose that when we performed the experiment, we did not look at the color of the ball that we transferred from Box I to Box II. Furthermore, suppose that when we looked at the ball drawn from Box II it is red. • Given this information, how do we update our knowledge of what we transferred from Box I to II? Lecture 4: More Probabilities and Conditioning
Updating … continued • Since we are interested in determining the probability of event A (and also of Ac), given event B, then the desired probabilities are just the conditional probabilities. Therefore, applying the conditional probability rule, we have: Lecture 4: More Probabilities and Conditioning
The Updating … continued • From our earlier calculations, we have obtained: • P(A B) = 28/99 and P(Ac B) = 42/99 • P(B) = 70/99 • Substituting these values in the formulas of the preceding slide, we obtain our updated probabilities, also called the posterior probabilities, to be: Lecture 4: More Probabilities and Conditioning
Comparison: Priors and Posteriors • We may compare the posterior probabilities of P(A|B) = 0.40 and P(Ac|B) = 0.60 with the prior probabilities of P(A) = 0.3636 and P(Ac) = 0.6363. • These values indicate that the information that we got a red ball from Box II had increased the probability that we transferred a red ball from Box I to Box II from 0.3636 to 0.40, and decreased the probability that we transferred a blue ball from 0.6363 to 0.60. • The directions of change in the values are clearly intuitive, but the exact magnitudes of the changes cannot be obtained without recourse to the formulas and reasoning we had employed. Lecture 4: More Probabilities and Conditioning
(Reverend) Bayes Theorem • The procedure we have just employed to update our prior probabilities of A and Ac, given the occurrence of event B, is a special case of Bayes Theorem. • For the situation where in the first step of the experiment we only have two possibilities: A and Ac, and B is an event pertaining to the outcome of the second step of the experiment, Bayes Theorem states that: Lecture 4: More Probabilities and Conditioning
Another Example • Situation: A medical test for HIV-infection has the following characteristics: • if the person is HIV-infected, the (conditional) probability that the test will be positive is 0.98, so the (conditional) probability that it will turn up negative is 0.02; while • if the person is not HIV-infected, the (conditional) probability that the test will be negative is 0.99, so the conditional probability that it will turn up positive is 0.01. • Assume that the prevalence of HIV-infection in the population is 0.005. Lecture 4: More Probabilities and Conditioning
Characteristics of the Medical Test for Infection • Below we present in tabular form the characteristics of the test for HIV-infection. Note that the probabilities are conditional probabilities. Lecture 4: More Probabilities and Conditioning
Example … continued • The Experiment: Suppose that a person is randomly chosen from the population, and this person is subjected to the test for HIV-infection. We are interested in: • a) The probability that the person will be HIV-infected (prior to the test). • b) The probability that the test for HIV-infection will show a positive result. • C) Given that the test showed a positive result, the (updated) probability that the person is HIV-infected. [QUESTION: Without computing the probability, what is your best estimate??] Lecture 4: More Probabilities and Conditioning
Example … continued • Solution: We let A be the event that the person chosen is HIV-infected, and by B the event that the test for HIV-infection will show a positive result. From the given information, we have: • P(A) = 0.005 so P(Ac) = 1 - P(A) = 1-0.005 = 0.995. • P(B|A) = 0.98 and P(B|Ac) = 0.01. • By Theorem of Total Probabilities: • P(B) = P(A)P(B|A) + P(Ac) P(B|Ac) = (.005)(.98) + (.995)(.01) = .0049 + .00995 = .01485 • By Bayes Rule: • P(A|B) = P(A)P(B|A)/P(B) = (.005)(.98)/.01485 = 0.33. Lecture 4: More Probabilities and Conditioning