Library homepage

  • school Campus Bookshelves
  • menu_book Bookshelves
  • perm_media Learning Objects
  • login Login
  • how_to_reg Request Instructor Account
  • hub Instructor Commons

Margin Size

  • Download Page (PDF)
  • Download Full Book (PDF)
  • Periodic Table
  • Physics Constants
  • Scientific Calculator
  • Reference & Cite
  • Tools expand_more
  • Readability

selected template will load here

This action is not available.

Statistics LibreTexts

2.4: Conditional Probability

  • Last updated
  • Save as PDF
  • Page ID 10132

  • Kyle Siegrist
  • University of Alabama in Huntsville via Random Services

\( \newcommand{\vecs}[1]{\overset { \scriptstyle \rightharpoonup} {\mathbf{#1}} } \)

\( \newcommand{\vecd}[1]{\overset{-\!-\!\rightharpoonup}{\vphantom{a}\smash {#1}}} \)

\( \newcommand{\id}{\mathrm{id}}\) \( \newcommand{\Span}{\mathrm{span}}\)

( \newcommand{\kernel}{\mathrm{null}\,}\) \( \newcommand{\range}{\mathrm{range}\,}\)

\( \newcommand{\RealPart}{\mathrm{Re}}\) \( \newcommand{\ImaginaryPart}{\mathrm{Im}}\)

\( \newcommand{\Argument}{\mathrm{Arg}}\) \( \newcommand{\norm}[1]{\| #1 \|}\)

\( \newcommand{\inner}[2]{\langle #1, #2 \rangle}\)

\( \newcommand{\Span}{\mathrm{span}}\)

\( \newcommand{\id}{\mathrm{id}}\)

\( \newcommand{\kernel}{\mathrm{null}\,}\)

\( \newcommand{\range}{\mathrm{range}\,}\)

\( \newcommand{\RealPart}{\mathrm{Re}}\)

\( \newcommand{\ImaginaryPart}{\mathrm{Im}}\)

\( \newcommand{\Argument}{\mathrm{Arg}}\)

\( \newcommand{\norm}[1]{\| #1 \|}\)

\( \newcommand{\Span}{\mathrm{span}}\) \( \newcommand{\AA}{\unicode[.8,0]{x212B}}\)

\( \newcommand{\vectorA}[1]{\vec{#1}}      % arrow\)

\( \newcommand{\vectorAt}[1]{\vec{\text{#1}}}      % arrow\)

\( \newcommand{\vectorB}[1]{\overset { \scriptstyle \rightharpoonup} {\mathbf{#1}} } \)

\( \newcommand{\vectorC}[1]{\textbf{#1}} \)

\( \newcommand{\vectorD}[1]{\overrightarrow{#1}} \)

\( \newcommand{\vectorDt}[1]{\overrightarrow{\text{#1}}} \)

\( \newcommand{\vectE}[1]{\overset{-\!-\!\rightharpoonup}{\vphantom{a}\smash{\mathbf {#1}}}} \)

The purpose of this section is to study how probabilities are updated in light of new information, clearly an absolutely essential topic. If you are a new student of probability, you may want to skip the technical details.

Definitions and Interpretations

The basic definition.

As usual, we start with a random experiment modeled by a probability space \((S, \mathscr S, \P)\). Thus, \( S \) is the set of outcomes, \( \mathscr S \) the collection of events, and \( \P \) the probability measure on the sample space \( (S, \mathscr S) \). Suppose now that we know that an event \(B\) has occurred. In general, this information should clearly alter the probabilities that we assign to other events. In particular, if \(A\) is another event then \(A\) occurs if and only if \(A\) and \(B\) occur; effectively, the sample space has been reduced to \(B\). Thus, the probability of \(A\), given that we know \(B\) has occurred, should be proportional to \(\P(A \cap B)\).

Events A and B

However, conditional probability, given that \(B\) has occurred, should still be a probability measure, that is, it must satisfy the axioms of probability. This forces the proportionality constant to be \(1 \big/ \P(B)\). Thus, we are led inexorably to the following definition:

Let \(A\) and \(B\) be events with \(\P(B) \gt 0\). The conditional probability of \(A\) given \(B\) is defined to be \[\P(A \mid B) = \frac{\P(A \cap B)}{\P(B)}\]

The Law of Large Numbers

The definition above was based on the axiomatic definition of probability. Let's explore the idea of conditional probability from the less formal and more intuitive notion of relative frequency (the law of large numbers). Thus, suppose that we run the experiment repeatedly. For \( n \in \N_+ \) and an event \(E\), let \(N_n(E)\) denote the number of times \(E\) occurs (the frequency of \( E \)) in the first \(n\) runs. Note that \(N_n(E)\) is a random variable in the compound experiment that consists of replicating the original experiment. In particular, its value is unknown until we actually run the experiment \( n \) times.

If \(N_n(B)\) is large, the conditional probability that \(A\) has occurred, given that \(B\) has occurred, should be close to the conditional relative frequency of \(A\) given \(B\), namely the relative frequency of \(A\) for the runs on which \(B\) occurred: \(N_n(A \cap B) / N_n(B)\). But note that \[\frac{N_n(A \cap B)}{N_n(B)} = \frac{N_n(A \cap B) / n}{N_n(B) / n}\] The numerator and denominator of the main fraction on the right are the relative frequencies of \( A \cap B \) and \( B \), respectively. So by the law of large numbers again, \( N_n(A \cap B) / n \to \P(A \cap B) \) as \( n \to \infty \) and \( N_n(B) \to \P(B) \) as \( n \to \infty \). Hence \[\frac{N_n(A \cap B)}{N_n(B)} \to \frac{\P(A \cap B)}{\P(B)} \text{ as } n \to \infty\] and we are led again to the definition above.

In some cases, conditional probabilities can be computed directly , by effectively reducing the sample space to the given event. In other cases, the formula in the mathematical definition is better. In some cases, conditional probabilities are known from modeling assumptions, and then are used to compute other probabilities. We will see examples of all of these situations in the computational exercises below.

It's very important that you not confuse \(\P(A \mid B)\), the probability of \(A\) given \(B\), with \(\P(B \mid A)\), the probability of \(B\) given \(A\). Making that mistake is known as the fallacy of the transposed conditional . (How embarrassing!)

Conditional Distributions

Suppose that \(X\) is a random variable for the experiment with values in \(T\). Mathematically, \( X \) is a function from \( S \) into \( T \), and \( \{X \in A\} \) denotes the event \( \{s \in S: X(s) \in A\} \) for \(A \subseteq T \). Intuitively, \( X \) is a variable of interest in the experiment, and every meaningful statement about \( X \) defines an event. Recall that the probability distribution of \(X\) is the probability measure on \(T\) given by \[A \mapsto \P(X \in A), \quad A \subseteq T\] This has a natural extension to a conditional distribution, given an event.

If \(B\) is an event with \( \P(B) \gt 0 \), then the conditional distribution of \(X\) given \(B\) is the probability measure on \(T\) given by \[A \mapsto \P(X \in A \mid B), \quad A \subseteq T\]

Recall that \( T \) will come with a \( \sigma \)-algebra of admissible subsets so that \( (T, \mathscr T) \) is a measurable space, just like the sample space \( (S, \mathscr S) \). Random variable \( X \) is required to be measurable as a function from \( S \) into \( T \). This ensures that \( \{X \in A\} \) is a valid event for each \( A \in \mathscr T \), so that the definition makes sense.

Basic Theory

Preliminary results.

Our first result is of fundamental importance, and indeed was a crucial part of the argument for the definition of conditional probability.

Suppose again that \( B \) is an event with \( \P(B) \gt 0 \). Then \(A \mapsto \P(A \mid B)\) is a probability measure on \( S \).

Clearly \( \P(A \mid B) \ge 0 \) for every event \( A \), and \( \P(S \mid B) = 1 \). Thus, suppose that \( \{A_i: i \in I\} \) is a countable collection of pairwise disjoint events. Then \[ \P\left(\bigcup_{i \in I} A_i \biggm| B\right) = \frac{1}{\P(B)} \P\left[\left(\bigcup_{i \in I} A_i\right) \cap B\right] = \frac{1}{\P(B)} \P\left(\bigcup_{i \in I} (A_i \cap B)\right) \] But the collection of events \( \{A_i \cap B: i \in I\} \) is also pairwise disjoint, so \[ \P\left(\bigcup_{i \in I} A_i \biggm| B\right) = \frac{1}{\P(B)} \sum_{i \in I} \P(A_i \cap B) = \sum_{i \in I} \frac{\P(A_i \cap B)}{\P(B)} = \sum_{i \in I} \P(A_i \mid B) \]

It's hard to overstate the importance of the last result because this theorem means that any result that holds for probability measures in general holds for conditional probability, as long as the conditioning event remains fixed. In particular the basic probability rules in the section on Probability Measure have analogs for conditional probability. To give two examples, \begin{align} \P\left(A^c \mid B\right) & = 1 - \P(A \mid B) \\ \P\left(A_1 \cup A_2 \mid B\right) & = \P\left(A_1 \mid B\right) + \P\left(A_2 \mid B\right) - \P\left(A_1 \cap A_2 \mid B\right) \end{align} By the same token, it follows that the conditional distribution of a random variable with values in \( T \), given in above , really does define a probability distribution on \( T \). No further proof is necessary. Our next results are very simple.

Suppose that \(A\) and \(B\) are events with \( \P(B) \gt 0 \).

  • If \(B \subseteq A\) then \(\P(A \mid B) = 1\).
  • If \(A \subseteq B\) then \(\P(A \mid B) = \P(A) / \P(B)\).
  • If \(A\) and \(B\) are disjoint then \(\P(A \mid B) = 0\).

These results follow directly from the definition of conditional probability. In part (a), note that \( A \cap B = B \). In part (b) note that \( A \cap B = A \). In part (c) note that \( A \cap B = \emptyset \).

Parts (a) and (c) certainly make sense. Suppose that we know that event \( B \) has occurred. If \( B \subseteq A \) then \( A \) becomes a certain event. If \( A \cap B = \emptyset \) then \( A \) becomes an impossible event. A conditional probability can be computed relative to a probability measure that is itself a conditional probability measure. The following result is a consistency condition.

Suppose that \(A\), \(B\), and \(C\) are events with \( \P(B \cap C) \gt 0 \). The probability of \(A\) given \(B\), relative to \(\P(\cdot \mid C)\), is the same as the probability of \(A\) given \(B\) and \(C\) (relative to \(\P\)). That is, \[\frac{\P(A \cap B \mid C)}{\P(B \mid C)} = \P(A \mid B \cap C)\]

From the definition, \[\frac{\P(A \cap B \mid C)}{\P(B \mid C)} = \frac{\P(A \cap B \cap C) \big/ \P(C)}{\P(B \cap C) \big/ \P(C)} = \frac{\P(A \cap B \cap C)}{\P(B \cap C)} = \P(A \mid B \cap C)\]

Correlation

Our next discussion concerns an important concept that deals with how two events are related, in a probabilistic sense.

Suppose that \(A\) and \(B\) are events with \( \P(A) \gt 0 \) and \( \P(B) \gt 0 \).

  • \(\P(A \mid B) \gt \P(A)\) if and only if \(\P(B \mid A) \gt \P(B)\) if and only if \(\P(A \cap B) \gt \P(A) \P(B)\). In this case, \( A \) and \( B \) are positively correlated .
  • \(\P(A \mid B) \lt \P(A)\) if and only if \(\P(B \mid A) \lt \P(B)\) if and only if \(\P(A \cap B) \lt \P(A) \P(B)\). In this case, \( A \) and \( B \) are negatively correlated .
  • \(\P(A \mid B) = \P(A)\) if and only if \(\P(B \mid A) = \P(B)\) if and only if \(\P(A \cap B) = \P(A) \P(B)\). In this case, \( A \) and \( B \) are uncorrelated or independent .

These properties following directly from the definition of conditional probability and simple algebra. Recall that multiplying or dividing an inequality by a positive number preserves the inequality.

Intuitively, if \( A \) and \( B \) are positively correlated, then the occurrence of either event means that the other event is more likely. If \(A\) and \(B\) are negatively correlated, then the occurrence of either event means that the other event is less likely. If \(A\) and \(B\) are uncorrelated, then the occurrence of either event does not change the probability of the other event. Independence is a fundamental concept that can be extended to more than two events and to random variables; these generalizations are studied in the next section on Independence. A much more general version of correlation, for random variables, is explored in the section on Covariance and Correlation in the chapter on Expected Value.

Suppose that \( A \) and \( B \) are events. Note from (4) that if \( A \subseteq B \) or \( B \subseteq A \) then \( A \) and \( B \) are positively correlated. If \( A \) and \( B \) are disjoint then \( A \) and \( B \) are negatively correlated.

Suppose that \(A\) and \(B\) are events in a random experiment.

  • \(A\) and \(B\) have the same correlation (positive, negative, or zero) as \(A^c\) and \(B^c\).
  • \(A\) and \(B\) have the opposite correlation as \(A\) and \(B^c\) (that is, positive-negative, negative-positive, or 0-0).
  • Using DeMorgan's law and the complement law. \[ \P(A^c \cap B^c) - \P(A^c) \P(B^c) = \P\left[(A \cup B)^c\right] - \P(A^c) \P(B^c) = \left[1 - \P(A \cup B)\right] - \left[1 - \P(A)\right]\left[1 - \P(B)\right] \] Using the inclusion-exclusion law and algebra, \[ \P(A^c \cap B^c) - \P(A^c) \P(B^c) = \P(A \cap B) - \P(A) \P(B) \]
  • Using the difference rule and the complement law: \[ \P(A \cap B^c) - \P(A) \P(B^c) = \P(A) - \P(A \cap B) - \P(A) \left[1 - \P(B)\right] = -\left[\P(A \cap B) - \P(A) \P(B)\right]\]

The Multiplication Rule

Sometimes conditional probabilities are known and can be used to find the probabilities of other events. Note first that if \( A \) and \( B \) are events with positive probability, then by the very definition of conditional probability, \[ \P(A \cap B) = \P(A) \P(B \mid A) = \P(B) \P(A \mid B) \] The following generalization is known as the multiplication rule of probability. As usual, we assume that any event conditioned on has positive probability.

Suppose that \((A_1, A_2, \ldots, A_n)\) is a sequence of events. Then \[\P\left(A_1 \cap A_2 \cap \cdots \cap A_n\right) = \P\left(A_1\right) \P\left(A_2 \mid A_1\right) P\left(A_3 \mid A_1 \cap A_2\right) \cdots \P\left(A_n \mid A_1 \cap A_2 \cap \cdots \cap A_{n-1}\right)\]

The product on the right a collapsing product in which only the probability of the intersection of all \(n\) events survives. The product of the first two factors is \( \P\left(A_1 \cap A_2\right) \), and hence the product of the first three factors is \( \P\left(A_1 \cap A_2 \cap A_3\right) \), and so forth. The proof can be made more rigorous by induction on \( n \).

The multiplication rule is particularly useful for experiments that consist of dependent stages, where \(A_i\) is an event in stage \(i\). Compare the multiplication rule of probability with the multiplication rule of combinatorics.

As with any other result, the multiplication rule can be applied to a conditional probability measure. In the context above, if \(E\) is another event, then \[\P\left(A_1 \cap A_2 \cap \cdots \cap A_n \mid E\right) = \P\left(A_1 \mid E\right) \P\left(A_2 \mid A_1 \cap E\right) P\left(A_3 \mid A_1 \cap A_2 \cap E\right) \cdots \P\left(A_n \mid A_1 \cap A_2 \cap \cdots \cap A_{n-1} \cap E\right)\]

Conditioning and Bayes' Theorem

Suppose that \(\mathscr{A} = \{A_i: i \in I\}\) is a countable collection of events that partition the sample space \(S\), and that \( \P(A_i) \gt 0 \) for each \( i \in I \).

Image: Total.png

The following theorem is known as the law of total probability .

If \( B \) is an event then \[\P(B) = \sum_{i \in I} \P(A_i) \P(B \mid A_i)\]

Recall that \(\{A_i \cap B: i \in I\}\) is a partition of \(B\). Hence \[ \P(B) = \sum_{i \in I} \P(A_i \cap B) = \sum_{i \in I} \P(A_i) \P(B \mid A_i) \]

The following theorem is known as Bayes' Theorem , named after Thomas Bayes:

If \( B \) is an event then \[\P(A_j \mid B) = \frac{\P(A_j) \P(B \mid A_j)}{\sum_{i \in I}\P(A_i) \P(B \mid A_i)}, \quad j \in I\]

Again the numerator is \(\P(A_j \cap B)\) while the denominator is \(\P(B)\) by the law of total probability .

These two theorems are most useful, of course, when we know \(\P(A_i)\) and \(\P(B \mid A_i)\) for each \(i \in I\). When we compute the probability of \(\P(B)\) by the law of total probability , we say that we are conditioning on the partition \(\mathscr{A}\). Note that we can think of the sum as a weighted average of the conditional probabilities \(\P(B \mid A_i)\) over \(i \in I\), where \(\P(A_i)\), \(i \in I\) are the weight factors. In the context of Bayes theorem , \(\P(A_j)\) is the prior probability of \(A_j\) and \(\P(A_j \mid B)\) is the posterior probability of \(A_j\) for \( j \in I \). We will study more general versions of conditioning and Bayes theorem in the section on Discrete Distributions in the chapter on Distributions, and again in the section on Conditional Expected Value in the chapter on Expected Value.

Once again, the law of total probability and Bayes' theorem can be applied to a conditional probability measure. So, if \(E\) is another event with \( \P(A_i \cap E) \gt 0 \) for \( i \in I \) then \begin{align} \P(B \mid E) & = \sum_{i \in I} \P(A_i \mid E) \P(B \mid A_i \cap E) \\ \P(A_j \mid B \cap E) & = \frac{\P(A_j \mid E) \P(B \mid A_j \cap E)}{\sum_{i \in I}\P(A_i \cap E) \P(B \mid A_i \cap E)}, \quad j \in I \end{align}

Examples and Applications

Basic rules.

Suppose that \(A\) and \(B\) are events in an experiment with \(\P(A) = \frac{1}{3}\), \(\P(B) = \frac{1}{4}\), \(\P(A \cap B) = \frac{1}{10}\). Find each of the following:

  • \(\P(A \mid B)\)
  • \(\P(B \mid A)\)
  • \(\P(A^c \mid B)\)
  • \(\P(B^c \mid A)\)
  • \(\P(A^c \mid B^c)\)
  • \(\frac{2}{5}\)
  • \(\frac{3}{10}\)
  • \(\frac{3}{5}\)
  • \(\frac{7}{10}\)
  • \(\frac{31}{45}\)

Suppose that \(A\), \(B\), and \(C\) are events in a random experiment with \(\P(A \mid C) = \frac{1}{2}\), \(\P(B \mid C) = \frac{1}{3}\), and \(\P(A \cap B \mid C) = \frac{1}{4}\). Find each of the following:

  • \(\P(B \setminus A \mid C)\)
  • \(\P(A \cup B \mid C)\)
  • \(\P(A^c \cap B^c \mid C)\)
  • \(\P(A^c \cup B^c \mid C)\)
  • \(\P(A^c \cup B \ \mid C)\)
  • \(\P(A \mid B \cap C)\)
  • \(\frac{1}{12}\)
  • \(\frac{7}{12}\)
  • \(\frac{5}{12}\)
  • \(\frac{3}{4}\)

Suppose that \(A\) and \(B\) are events in a random experiment with \(\P(A) = \frac{1}{2}\), \(\P(B) = \frac{1}{3}\), and \(\P(A \mid B) =\frac{3}{4}\).

  • Find \(\P(A \cap B)\)
  • Find \(\P(A \cup B)\)
  • Find \(\P(B \cup A^c)\)
  • Find \(\P(B \mid A)\)
  • Are \(A\) and \(B\) positively correlated, negatively correlated, or independent?
  • \(\frac{1}{4}\)
  • \(\frac{1}{2}\)
  • positively correlated.

Open the conditional probability experiment.

  • Given \( \P(A) \), \( \P(B) \), and \( \P(A \cap B) \), in the table, verify all of the other probabilities in the table.
  • Run the experiment 1000 times and compare the probabilities with the relative frequencies.

Simple Populations

In a certain population, 30% of the persons smoke cigarettes and 8% have COPD ( Chronic Obstructive Pulmonary Disease ). Moreover, 12% of the persons who smoke have COPD.

  • What percentage of the population smoke and have COPD?
  • What percentage of the population with COPD also smoke?
  • Are smoking and COPD positively correlated, negatively correlated, or independent?

A company has 200 employees: 120 are women and 80 are men. Of the 120 female employees, 30 are classified as managers, while 20 of the 80 male employees are managers. Suppose that an employee is chosen at random.

  • Find the probability that the employee is female.
  • Find the probability that the employee is a manager.
  • Find the conditional probability that the employee is a manager given that the employee is female.
  • Find the conditional probability that the employee is female given that the employee is a manager.
  • Are the events female and manager positively correlated, negatively correlated, or indpendent?
  • \(\frac{120}{200}\)
  • \(\frac{50}{200}\)
  • \(\frac{30}{120}\)
  • \(\frac{30}{50}\)
  • independent

Dice and Coins

Consider the experiment that consists of rolling 2 standard, fair dice and recording the sequence of scores \(\bs{X} = (X_1, X_2)\). Let \(Y\) denote the sum of the scores. For each of the following pairs of events, find the probability of each event and the conditional probability of each event given the other. Determine whether the events are positively correlated, negatively correlated, or independent.

  • \(\{X_1 = 3\}\), \(\{Y = 5\}\)
  • \(\{X_1 = 3\}\), \(\{Y = 7\}\)
  • \(\{X_1 = 2\}\), \(\{Y = 5\}\)
  • \(\{X_1 = 3\}\), \(\{X_1 = 2\}\)

In each case below, the answers are for \( \P(A) \), \( \P(B) \), \( \P(A \mid B) \), and \( \P(B \mid A) \)

  • \(\frac{1}{6}\), \(\frac{1}{9}\), \(\frac{1}{4}\), \(\frac{1}{6}\). Positively correlated.
  • \(\frac{1}{6}\), \(\frac{1}{6}\), \(\frac{1}{6}\), \(\frac{1}{6}\). Independent.
  • \(\frac{1}{6}\), \(\frac{1}{6}\), \(0\), \(0\). Negatively correlated.

Note that positive correlation is not a transitive relation. From the previous exercise, for example, note that \(\{X_1 = 3\}\) and \(\{Y = 5\}\) are positively correlated, \(\{Y = 5\}\) and \(\{X_1 = 2\}\) are positively correlated, but \(\{X_1 = 3\}\) and \(\{X_1 = 2\}\) are negatively correlated (in fact, disjoint).

In dice experiment, set \(n = 2\). Run the experiment 1000 times. Compute the empirical conditional probabilities corresponding to the conditional probabilities in the last exercise.

Consider again the experiment that consists of rolling 2 standard, fair dice and recording the sequence of scores \(\bs{X} = (X_1, X_2)\). Let \(Y\) denote the sum of the scores, \(U\) the minimum score, and \(V\) the maximum score.

  • Find \(\P(U = u \mid V = 4)\) for the appropriate values of \(u\).
  • Find \(\P(Y = y \mid V = 4)\) for the appropriate values of \(y\).
  • Find \(\P(V = v \mid Y = 8)\) for appropriate values of \(v\).
  • Find \(\P(U = u \mid Y = 8)\) for the appropriate values of \(u\).
  • Find \(\P[(X_ 1, X_2) = (x_1, x_2) \mid Y = 8]\) for the appropriate values of \((x_1, x_2)\).
  • \(\frac{2}{7}\) for \(u \in \{1, 2, 3\}\), \(\frac{1}{7}\) for \( u = 4 \)
  • \(\frac{2}{7}\) for \(y \in \{5, 6, 7\}\), \(\frac{1}{7}\) for \( y = 8 \)
  • \(\frac{1}{5}\) for \( v = 4 \), \(\frac{2}{5}\) for \(v \in \{5, 6\}\)
  • \(\frac{2}{5}\) for \(u \in \{2, 3\}\), \(\frac{1}{5}\) for \( u = 4 \)
  • \(\frac{1}{5}\) for \((x_1, x_2) \in \{(2,6), (6,2), (3,5), (5,3), (4,4)\}\)

In the die-coin experiment, a standard, fair die is rolled and then a fair coin is tossed the number of times showing on the die. Let \(N\) denote the die score and \(H\) the event that all coin tosses result in heads.

  • Find \(\P(H)\).
  • Find \(\P(N = n \mid H)\) for \(n \in \{1, 2, 3, 4, 5, 6\}\).
  • Compare the results in (b) with \(\P(N = n)\) for \(n \in \{1, 2, 3, 4, 5, 6\}\). In each case, note whether the events \(H\) and \(\{N = n\}\) are positively correlated, negatively correlated, or independent.
  • \(\frac{21}{128}\)
  • \(\frac{64}{63} \frac{1}{2^n}\) for \(n \in \{1, 2, 3, 4, 5, 6\}\)
  • positively correlated for \(n \in \{1, 2\}\) and negatively correlated for \(n \in \{3, 4, 5, 6\}\)

Run the die-coin experiment 1000 times. Let \(H\) and \(N\) be as defined in the previous exercise.

  • Compute the empirical probability of \(H\). Compare with the true probability in the previous exercise.
  • Compute the empirical probability of \(\{N = n\}\) given \(H\), for \(n \in \{1, 2, 3, 4, 5, 6\}\). Compare with the true probabilities in the previous exercise.

Suppose that a bag contains 12 coins: 5 are fair, 4 are biased with probability of heads \(\frac{1}{3}\); and 3 are two-headed. A coin is chosen at random from the bag and tossed.

  • Find the probability that the coin is heads.
  • Given that the coin is heads, find the conditional probability of each coin type.
  • \(\frac{41}{72}\)
  • \( \frac{15}{41} \) that the coin is fair, \( \frac{8}{41} \) that the coin is biased, \( \frac{18}{41} \) that the coin is two-headed

Compare die-coin experiment and bag of coins experiment . In the die-coin experiment, we toss a coin with a fixed probability of heads a random number of times. In the bag of coins experiment, we effectively toss a coin with a random probability of heads a fixed number of times. The random experiment of tossing a coin with a fixed probability of heads \(p\) a fixed number of times \(n\) is known as the binomial experiment with parameters \(n\) and \(p\). This is a very basic and important experiment that is studied in more detail in the section on the binomial distribution in the chapter on Bernoulli Trials. Thus, the die-coin and bag of coins experiments can be thought of as modifications of the binomial experiment in which a parameter has been randomized . In general, interesting new random experiments can often be constructed by randomizing one or more parameters in another random experiment.

In the coin-die experiment, a fair coin is tossed. If the coin lands tails, a fair die is rolled. If the coin lands heads, an ace-six flat die is tossed (faces 1 and 6 have probability \(\frac{1}{4}\) each, while faces 2, 3, 4, and 5 have probability \(\frac{1}{8}\) each). Let \(H\) denote the event that the coin lands heads, and let \(Y\) denote the score when the chosen die is tossed.

  • Find \(\P(Y = y)\) for \(y \in \{1, 2, 3, 4, 5, 6\}\).
  • Find \(\P(H \mid Y = y)\) for \(y \in \{1, 2, 3, 4, 5, 6,\}\).
  • Compare each probability in part (b) with \(\P(H)\). In each case, note whether the events \(H\) and \(\{Y = y\}\) are positively correlated, negatively correlated, or independent.
  • \(\frac{5}{24}\) for \(y \in \{1, 6\}\), \(\frac{7}{48}\) for \(y \in \{2, 3, 4, 5\}\)
  • \(\frac{3}{5}\) for \(y \in \{1, 6\}\), \(\frac{3}{7}\) for \(y \in \{2, 3, 4, 5\}\)
  • Positively correlated for \( y \in \{1, 6\} \), negatively correlated for \( y \in \{2, 3, 4, 5\} \)

Run the coin-die experiment 1000 times. Let \(H\) and \(Y\) be as defined in the previous exercise.

  • Compute the empirical probability of \(\{Y = y\}\), for each \(y\), and compare with the true probability in the previous exercise
  • Compute the empirical probability of \(H\) given \(\{Y = y\}\) for each \(y\), and compare with the true probability in the previous exercise.

Consider the card experiment that consists of dealing 2 cards from a standard deck and recording the sequence of cards dealt. For \(i \in \{1, 2\}\), let \(Q_i\) be the event that card \(i\) is a queen and \(H_i\) the event that card \(i\) is a heart. For each of the following pairs of events, compute the probability of each event, and the conditional probability of each event given the other. Determine whether the events are positively correlated, negatively correlated, or independent.

  • \(Q_1\), \(H_1\)
  • \(Q_1\), \(Q_2\)
  • \(Q_2\), \(H_2\)
  • \(Q_1\), \(H_2\)

The answers below are for \( \P(A) \), \( \P(B) \), \( \P(A \mid B) \), and \( \P(B \mid A) \) where \( A \) and \( B \) are the given events

  • \(\frac{1}{13}\), \(\frac{1}{4}\), \(\frac{1}{13}\), \(\frac{1}{4}\), independent.
  • \(\frac{1}{13}\), \(\frac{1}{13}\), \(\frac{3}{51}\), \(\frac{3}{51}\), negatively correlated.

In the card experiment, set \(n = 2\). Run the experiment 500 times. Compute the conditional relative frequencies corresponding to the conditional probabilities in the last exercise.

Consider the card experiment that consists of dealing 3 cards from a standard deck and recording the sequence of cards dealt. Find the probability of the following events:

  • All three cards are all hearts.
  • The first two cards are hearts and the third is a spade.
  • The first and third cards are hearts and the second is a spade.
  • \(\frac{11}{850}\)
  • \(\frac{13}{850}\)

In the card experiment, set \(n = 3\) and run the simulation 1000 times. Compute the empirical probability of each event in the previous exercise and compare with the true probability.

Bivariate Uniform Distributions

Recall that Buffon's coin experiment consists of tossing a coin with radius \(r \le \frac{1}{2}\) randomly on a floor covered with square tiles of side length 1. The coordinates \((X, Y)\) of the center of the coin are recorded relative to axes through the center of the square, parallel to the sides. Since the needle is dropped randomly, the basic modeling assumption is that \( (X, Y) \) is uniformly distributed on the square \( [-1/2, 1/2]^2 \).

Buffon's coin experiment

In Buffon's coin experiment,

  • Find \(\P(Y \gt 0 \mid X \lt Y)\)
  • Find the conditional distribution of \((X, Y)\) given that the coin does not touch the sides of the square.
  • Given \((X, Y) \in [r - \frac{1}{2}, \frac{1}{2} - r]^2\), \((X, Y)\) is uniformly distributed on this set.

Run Buffon's coin experiment 500 times. Compute the empirical probability that \(Y \gt 0\) given that \(X \lt Y\) and compare with the probability in the last exercise.

In the conditional probability experiment, the random points are uniformly distributed on the rectangle \( S \). Move and resize events \( A \) and \( B \) and note how the probabilities change. For each of the following configurations, run the experiment 1000 times and compare the relative frequencies with the true probabilities.

  • \( A \) and \( B \) in general position
  • \( A \) and \( B \) disjoint
  • \( A \subseteq B \)
  • \( B \subseteq A \)

Reliability

A plant has 3 assembly lines that produces memory chips. Line 1 produces 50% of the chips and has a defective rate of 4%; line 2 has produces 30% of the chips and has a defective rate of 5%; line 3 produces 20% of the chips and has a defective rate of 1%. A chip is chosen at random from the plant.

  • Find the probability that the chip is defective.
  • Given that the chip is defective, find the conditional probability for each line.
  • 0.541 for line 1, 0.405 for line 2, 0.054 for line 3

Suppose that a bit (0 or 1) is sent through a noisy communications channel. Because of the noise, the bit sent may be received incorrectly as the complementary bit. Specifically, suppose that if 0 is sent, then the probability that 0 is received is 0.9 and the probability that 1 is received is 0.1. If 1 is sent, then the probability that 1 is received is 0.8 and the probability that 0 is received is 0.2. Finally, suppose that 1 is sent with probability 0.6 and 0 is sent with probability 0.4. Find the probability that

  • 1 was sent given that 1 was received
  • 0 was sent given that 0 was received

Suppose that \(T\) denotes the lifetime of a light bulb (in 1000 hour units), and that \(T\) has the following exponential distribution, defined for measurable \(A \subseteq [0, \infty)\):

  • Find \(\P(T \gt 3)\)
  • Find \(\P(T \gt 5 \mid T \gt 2)\)

Suppose again that \(T\) denotes the lifetime of a light bulb (in 1000 hour units), but that \(T\) is uniformly distributed on the interal \([0, 10]\).

  • \(\frac{5}{8}\)

Please refer to the discussion of genetics in the section on random experiments if you need to review some of the definitions in this section.

Recall first that the ABO blood type in humans is determined by three alleles: \(a\), \(b\), and \(o\). Furthermore, \(a\) and \(b\) are co-dominant and \(o\) is recessive. Suppose that the probability distribution for the set of blood genotypes in a certain population is given in the following table:

Suppose that a person is chosen at random from the population. Let \(A\), \(B\), \(AB\), and \(O\) be the events that the person is type \(A\), type \(B\), type \(AB\), and type \(O\) respectively. Let \(H\) be the event that the person is homozygous, and let \(D\) denote the event that the person has an \(o\) allele. Find each of the following:

  • \(\P(A)\), \(\P(B)\), \(\P(AB)\), \(\P(O)\), \(\P(H)\), \(\P(D)\)
  • \(P(A \cap H)\), \(P(A \mid H)\), \(P(H \mid A)\). Are the events \(A\) and \(H\) positively correlated, negatively correlated, or independent?
  • \(P(B \cap H)\), \(P(B \mid H)\), \(P(H \mid B)\). Are the events \(B\) and \(H\) positively correlated, negatively correlated, or independent?
  • \(P(A \cap D)\), \(P(A \mid D)\), \(P(D \mid A)\). Are the events \(A\) and \(D\) positively correlated, negatively correlated, or independent?
  • \(P(B \cap D)\), \(P(B \mid D)\), \(P(D \mid B)\). Are the events \(B\) and \(D\) positively correlated, negatively correlated, or independent?
  • \(P(H \cap D)\), \(P(H \mid D)\), \(P(D \mid H)\). Are the events \(H\) and \(D\) positively correlated, negatively correlated, or independent?
  • 0.360, 0.123, 0.038, 0.479, 0.536, 0.905
  • 0.050, 0.093, 0.139. \(A\) and \(H\) are negatively correlated.
  • 0.007, 0.013, 0.057. \(B\) and \(H\) are negatively correlated.
  • 0.310, 0.343, 0.861. \(A\) and \(D\) are negatively correlated.
  • 0.116, 0.128, 0.943. \(B\) and \(D\) are positivley correlated.
  • 0.479, 0.529, 0.894. \(H\) and \(D\) are negatively correlated.

Suppose next that pod color in certain type of pea plant is determined by a gene with two alleles: \(g\) for green and \(y\) for yellow, and that \(g\) is dominant and \(y\) recessive.

Suppose that a green-pod plant and a yellow-pod plant are bred together. Suppose further that the green-pod plant has a \(\frac{1}{4}\) chance of carrying the recessive yellow-pod allele.

  • Find the probability that a child plant will have green pods.
  • Given that a child plant has green pods, find the updated probability that the green-pod parent has the recessive allele.
  • \(\frac{7}{8}\)
  • \(\frac{1}{7}\)

Suppose that two green-pod plants are bred together. Suppose further that with probability \(\frac{1}{3}\) neither plant has the recessive allele, with probability \(\frac{1}{2}\) one plant has the recessive allele, and with probability \(\frac{1}{6}\) both plants have the recessive allele.

  • Find the probability that a child plant has green pods.
  • Given that a child plant has green pods, find the updated probability that both parents have the recessive gene.
  • \(\frac{23}{24}\)
  • \(\frac{3}{23}\)

Next consider a sex-linked hereditary disorder in humans (such as colorblindness or hemophilia). Let \(h\) denote the healthy allele and \(d\) the defective allele for the gene linked to the disorder. Recall that \(h\) is dominant and \(d\) recessive for women.

Suppose that in a certain population, 50% are male and 50% are female. Moreover, suppose that 10% of males are color blind but only 1% of females are color blind.

  • Find the percentage of color blind persons in the population.
  • Find the percentage of color blind persons that are male.

Since color blindness is a sex-linked hereditary disorder, note that it's reasonable in the previous exercise that the probability that a female is color blind is the square of the probability that a male is color blind. If \(p\) is the probability of the defective allele on the \(X\) chromosome, then \(p\) is also the probability that a male will be color blind. But since the defective allele is recessive, a woman would need two copies of the defective allele to be color blind, and assuming independence, the probability of this event is \(p^2\).

A man and a woman do not have a certain sex-linked hereditary disorder, but the woman has a \(\frac{1}{3}\) chance of being a carrier.

  • Find the probability that a son born to the couple will be normal.
  • Find the probability that a daughter born to the couple will be a carrier.
  • Given that a son born to the couple is normal, find the updated probability that the mother is a carrier.
  • \(\frac{5}{6}\)
  • \(\frac{1}{6}\)
  • \(\frac{1}{5}\)

Urn 1 contains 4 red and 6 green balls while urn 2 contains 7 red and 3 green balls. An urn is chosen at random and then a ball is chosen at random from the selected urn.

  • Find the probability that the ball is green.
  • Given that the ball is green, find the conditional probability that urn 1 was selected.
  • \(\frac{9}{20}\)
  • \(\frac{2}{3}\)

Urn 1 contains 4 red and 6 green balls while urn 2 contains 6 red and 3 green balls. A ball is selected at random from urn 1 and transferred to urn 2. Then a ball is selected at random from urn 2.

  • Find the probability that the ball from urn 2 is green.
  • Given that the ball from urn 2 is green, find the conditional probability that the ball from urn 1 was green.
  • \(\frac{9}{25}\)

An urn initially contains 6 red and 4 green balls. A ball is chosen at random from the urn and its color is recorded. It is then replaced in the urn and 2 new balls of the same color are added to the urn. The process is repeated. Find the probability of each of the following events:

  • Balls 1 and 2 are red and ball 3 is green.
  • Balls 1 and 3 are red and ball 2 is green.
  • Ball 1 is green and balls 2 and 3 are red.
  • Ball 2 is red.
  • Ball 1 is red given that ball 2 is red.
  • \(\frac{4}{35}\)

Think about the results in the previous exercise. Note in particular that the answers to parts (a), (b), and (c) are the same, and that the probability that the second ball is red in part (d) is the same as the probability that the first ball is red. More generally, the probabilities of events do not depend on the order of the draws. For example, the probability of an event involving the first, second, and third draws is the same as the probability of the corresponding event involving the seventh, tenth and fifth draws. Technically, the sequence of events \((R_1, R_2, \ldots)\) is exchangeable. The random process described in this exercise is a special case of Pólya's urn scheme , named after George Pólya. We sill study Pólya's urn in more detail in the chapter on Finite Sampling Models

An urn initially contains 6 red and 4 green balls. A ball is chosen at random from the urn and its color is recorded. It is then replaced in the urn and two new balls of the other color are added to the urn. The process is repeated. Find the probability of each of the following events:

  • \(\frac{6}{35}\)
  • \(\frac{16}{105}\)
  • \(\frac{17}{30}\)
  • \(\frac{9}{17}\)

Think about the results in the previous exercise, and compare with Pólya's urn . Note that the answers to parts (a), (b), and (c) are not all the same, and that the probability that the second ball is red in part (d) is not the same as the probability that the first ball is red. In short, the sequence of events \((R_1, R_2, \ldots)\) is not exchangeable.

Diagnostic Testing

Suppose that we have a random experiment with an event \(A\) of interest. When we run the experiment, of course, event \(A\) will either occur or not occur. However, suppose that we are not able to observe the occurrence or non-occurrence of \(A\) directly. Instead we have a diagnostic test designed to indicate the occurrence of event \(A\); thus the test that can be either positive for \(A\) or negative for \(A\). The test also has an element of randomness, and in particular can be in error. Here are some typical examples of the type of situation we have in mind:

  • The event is that a person has a certain disease and the test is a blood test for the disease.
  • The event is that a woman is pregnant and the test is a home pregnancy test.
  • The event is that a person is lying and the test is a lie-detector test.
  • The event is that a device is defective and the test consists of a sensor reading.
  • The event is that a missile is in a certain region of airspace and the test consists of radar signals.
  • The event is that a person has committed a crime, and the test is a jury trial with evidence presented for and against the event.

Let \(T\) be the event that the test is positive for the occurrence of \(A\). The conditional probability \(\P(T \mid A)\) is called the sensitivity of the test. The complementary probability \[\P(T^c \mid A) = 1 - \P(T \mid A)\] is the false negative probability. The conditional probability \(\P(T^c \mid A^c)\) is called the specificity of the test. The complementary probability \[\P(T \mid A^c) = 1 - \P(T^c \mid A^c)\] is the false positive probability. In many cases, the sensitivity and specificity of the test are known, as a result of the development of the test. However, the user of the test is interested in the opposite conditional probabilities, namely \(\P(A \mid T)\), the probability of the event of interest, given a positive test, and \(\P(A^c \mid T^c)\), the probability of the complementary event, given a negative test. Of course, if we know \( \P(A \mid T) \) then we also have \( \P(A^c \mid T) = 1 - \P(A \mid T) \), the probability of the complementary event given a positive test. Similarly, if we know \( \P(A^c \mid T^c) \) then we also have \( \P(A \mid T^c) \), the probability of the event given a negative test. Computing the probabilities of interest is simply a special case of Bayes' theorem.

The probability that the event occurs, given a positive test is \[\P(A \mid T) = \frac{\P(A) \P(T \mid A)}{\P(A) \P(T \mid A) + \P(A^c) \P(T \mid A^c)}\] The probability that the event does not occur, given a negative test is \[\P(A^c \mid T^c) = \frac{\P(A^c) \P(T^c \mid A^c)}{\P(A) \P(T^c \mid A) + \P(A^c) \P(T^c \mid A^c)}\]

There is often a trade-off between sensitivity and specificity. An attempt to make a test more sensitive may result in the test being less specific, and an attempt to make a test more specific may result in the test being less sensitive. As an extreme example, consider the worthless test that always returns positive, no matter what the evidence. Then \( T = S \) so the test has sensitivity 1, but specificity 0. At the opposite extreme is the worthless test that always returns negative, no matter what the evidence. Then \( T = \emptyset \) so the test has specificity 1 but sensitivity 0. In between these extremes are helpful tests that are actually based on evidence of some sort.

Suppose that the sensitivity \( a = \P(T \mid A) \in (0, 1)\) and the specificity \( b = \P(T^c \mid A^c) \in (0, 1) \) are fixed. Let \( p = \P(A) \) denote the prior probability of the event \( A \) and \( P = \P(A \mid T) \) the posterior probability of \( A \) given a positive test.

\( P \) as a function of \( p \) is given by \[ P = \frac{a p}{(a + b - 1) p + (1 - b)}, \quad p \in [0, 1] \]

  • \( P \) increases continuously from 0 to 1 as \( p \) increases from 0 to 1.
  • \( P \) is concave downward if \( a + b \gt 1 \). In this case \( A \) and \( T \) are positively correlated.
  • \( P \) is concave upward if \( a + b \lt 1 \). In this case \( A \) and \( T \) are negatively correlated.
  • \( P = p \) if \( a + b = 1 \). In this case, \( A \) and \( T \) are uncorrelated (independent).

The formula for \( P \) in terms of \( p \) follows from (42) and algebra. For part (a), note that \[ \frac{dP}{dp} = \frac{a (1 - b)}{[(a + b - 1) p + (1 - b)]^2} \gt 0\] For parts (b)-(d), note that \[ \frac{d^2 P}{dp^2} = \frac{-2 a (1 - b)(a + b - 1)}{[(1 + b - 1)p + (1 - b)]^3} \] If \( a + b \gt 1 \), \( d^2P/dp^2 \lt 0 \) so \( P \) is concave downward on \( [0, 1] \) and hence \( P \gt p \) for \( 0 \lt p \lt 1 \). If \( a + b \lt 1 \), \( d^2P/dp^2 \gt 0 \) so \( P \) is concave upward on \( [0, 1] \) and hence \( P \lt p \) for \( 0 \lt p \lt 1 \). Trivially if \( a + b = 1 \), \( P = p \) for \( 0 \le p \le 1 \).

Of course, part (b) is the typical case, where the test is useful. In fact, we would hope that the sensitivity and specificity are close to 1. In case (c), the test is worse than useless since it gives the wrong information about \( A \). But this case could be turned into a useful test by simply reversing the roles of positive and negative . In case (d), the test is worthless and gives no information about \( A \). It's interesting that the broad classification above depends only on the sum of the sensitivity and specificity.

Diagnostic.png

Suppose that a diagnostic test has sensitivity 0.99 and specificity 0.95. Find \( \P(A \mid T) \) for each of the following values of \( \P(A) \):

With sensitivity 0.99 and specificity 0.95, the test in the last exercise superficially looks good. However the small value of \(\P(A \mid T)\) for small values of \(\P(A)\) is striking (but inevitable given the properties above ). The moral, of course, is that \(\P(A \mid T)\) depends critically on \(\P(A)\) not just on the sensitivity and specificity of the test. Moreover, the correct comparison is \(\P(A \mid T)\) with \(\P(A)\), as in the exercise, not \(\P(A \mid T)\) with \(\P(T \mid A)\)—Beware of the fallacy of the transposed conditional! In terms of the correct comparison, the test does indeed work well; \(\P(A \mid T)\) is significantly larger than \(\P(A)\) in all cases.

A woman initially believes that there is an even chance that she is or is not pregnant. She takes a home pregnancy test with sensitivity 0.95 and specificity 0.90 (which are reasonable values for a home pregnancy test). Find the updated probability that the woman is pregnant in each of the following cases.

  • The test is positive.
  • The test is negative.

Suppose that 70% of defendants brought to trial for a certain type of crime are guilty. Moreover, historical data show that juries convict guilty persons 80% of the time and convict innocent persons 10% of the time. Suppose that a person is tried for a crime of this type. Find the updated probability that the person is guilty in each of the following cases:

  • The person is convicted.
  • The person is acquitted.

The Check Engine light on your car has turned on. Without the information from the light, you believe that there is a 10% chance that your car has a serious engine problem. You learn that if the car has such a problem, the light will come on with probability 0.99, but if the car does not have a serious problem, the light will still come on, under circumstances similar to yours, with probability 0.3. Find the updated probability that you have an engine problem.

The standard test for HIV is the ELISA (Enzyme-Linked Immunosorbent Assay) test. It has sensitivity and specificity of 0.999. Suppose that a person is selected at random from a population in which 1% are infected with HIV, and given the ELISA test. Find the probability that the person has HIV in each of the following cases:

The ELISA test for HIV is a very good one. Let's look another test, this one for prostate cancer, that's rather bad.

The PSA test for prostate cancer is based on a blood marker known as the Prostate Specific Antigen . An elevated level of PSA is evidence for prostate cancer. To have a diagnostic test, in the sense that we are discussing here, we must decide on a definite level of PSA, above which we declare the test to be positive. A positive test would typically lead to other more invasive tests (such as biopsy) which, of course, carry risks and cost. The PSA test with cutoff 2.6 ng/ml has sensitivity 0.40 and specificity 0.81. The overall incidence of prostate cancer among males is 156 per 100000. Suppose that a man, with no particular risk factors, has the PSA test. Find the probability that the man has prostate cancer in each of the following cases:

Diagnostic testing is closely related to a general statistical procedure known as hypothesis testing . A separate chapter on hypothesis testing explores this procedure in detail.

Data Analysis Exercises

For the M&M data set, find the empirical probability that a bag has at least 10 reds, given that the weight of the bag is at least 48 grams.

\(\frac{10}{23}\).

Consider the Cicada data.

  • Find the empirical probability that a cicada weighs at least 0.25 grams given that the cicada is male.
  • Find the empirical probability that a cicada weighs at least 0.25 grams given that the cicada is the tredecula species.
  • \(\frac{2}{45}\)
  • \(\frac{7}{44}\)

Conditional Probability

How to handle Dependent Events

Life is full of random events! You need to get a "feel" for them to be a smart and successful person.

Independent Events

Events can be " Independent ", meaning each event is not affected by any other events.

Example: Tossing a coin.

Each toss of a coin is a perfect isolated thing.

What it did in the past will not affect the current toss.

The chance is simply 1-in-2, or 50%, just like ANY toss of the coin.

So each toss is an Independent Event .

Dependent Events

But events can also be "dependent" ... which means they can be affected by previous events ...

Example: Marbles in a Bag

2 blue and 3 red marbles are in a bag.

What are the chances of getting a blue marble?

The chance is 2 in 5

But after taking one out the chances change!

So the next time:

This is because we are removing marbles from the bag.

So the next event depends on what happened in the previous event, and is called dependent .

Replacement

Note: if we replace the marbles in the bag each time, then the chances do not change and the events are independent :

  • With Replacement: the events are Independent (the chances don't change)
  • Without Replacement: the events are Dependent (the chances change)

Dependent events are what we look at here.

Tree Diagram

A Tree Diagram is a wonderful way to picture what is going on, so let's build one for our marbles example.

There is a 2/5 chance of pulling out a Blue marble, and a 3/5 chance for Red:

We can go one step further and see what happens when we pick a second marble:

If a blue marble was selected first there is now a 1/4 chance of getting a blue marble and a 3/4 chance of getting a red marble.

If a red marble was selected first there is now a 2/4 chance of getting a blue marble and a 2/4 chance of getting a red marble.

Now we can answer questions like "What are the chances of drawing 2 blue marbles?"

Answer: it is a 2/5 chance followed by a 1/4 chance :

Did you see how we multiplied the chances? And got 1/10 as a result.

The chances of drawing 2 blue marbles is 1/10

We love notation in mathematics! It means we can then use the power of algebra to play around with the ideas. So here is the notation for probability:

P(A) means "Probability Of Event A"

In our marbles example Event A is "get a Blue Marble first" with a probability of 2/5:

And Event B is "get a Blue Marble second" ... but for that we have 2 choices:

  • If we got a Blue Marble first the chance is now 1/4
  • If we got a Red Marble first the chance is now 2/4

So we have to say which one we want , and use the symbol "|" to mean "given":

P(B|A) means "Event B given Event A"

In other words, event A has already happened, now what is the chance of event B?

P(B|A) is also called the "Conditional Probability" of B given A.

And in our case:

P(B|A) = 1/4

So the probability of getting 2 blue marbles is:

And we write it as

"Probability of event A and event B equals the probability of event A times the probability of event B given event A "

Let's do the next example using only notation:

Example: Drawing 2 Kings from a Deck

Event A is drawing a King first, and Event B is drawing a King second.

For the first card the chance of drawing a King is 4 out of 52 (there are 4 Kings in a deck of 52 cards):

P(A) = 4/52

But after removing a King from the deck the probability of the 2nd card drawn is less likely to be a King (only 3 of the 51 cards left are Kings):

P(B|A) = 3/51

P(A and B) = P(A) x P(B|A) =(4/52)x (3/51) = 12/2652 = 1/221

So the chance of getting 2 Kings is 1 in 221, or about 0.5%

Finding Hidden Data

Using Algebra we can also "change the subject" of the formula, like this:

And we have another useful formula:

"The probability of event B given event A equals the probability of event A and event B divided by the probability of event A "

Example: Ice Cream

70% of your friends like Chocolate, and 35% like Chocolate AND like Strawberry.

What percent of those who like Chocolate also like Strawberry?

P(Strawberry|Chocolate) = P(Chocolate and Strawberry) / P(Chocolate)

50% of your friends who like Chocolate also like Strawberry

soccer teams

Big Example: Soccer Game

You are off to soccer, and want to be the Goalkeeper, but that depends who is the Coach today:

  • with Coach Sam the probability of being Goalkeeper is 0.5
  • with Coach Alex the probability of being Goalkeeper is 0.3

Sam is Coach more often ... about 6 out of every 10 games (a probability of 0.6 ).

So, what is the probability you will be a Goalkeeper today?

Let's build a tree diagram . First we show the two possible coaches: Sam or Alex:

The probability of getting Sam is 0.6, so the probability of Alex must be 0.4 (together the probability is 1)

Now, if you get Sam, there is 0.5 probability of being Goalie (and 0.5 of not being Goalie):

If you get Alex, there is 0.3 probability of being Goalie (and 0.7 not):

The tree diagram is complete, now let's calculate the overall probabilities. Remember that:

P(A and B) = P(A) x P(B|A)

Here is how to do it for the "Sam, Yes" branch:

(When we take the 0.6 chance of Sam being coach times the 0.5 chance that Sam will let you be Goalkeeper we end up with an 0.3 chance.)

But we are not done yet! We haven't included Alex as Coach:

With 0.4 chance of Alex as Coach, followed by the 0.3 chance gives 0.12

And the two "Yes" branches of the tree together make:

0.3 + 0.12 = 0.42 probability of being a Goalkeeper today

(That is a 42% chance)

One final step: complete the calculations and make sure they add to 1:

0.3 + 0.3 + 0.12 + 0.28 = 1

Yes, they add to 1 , so that looks right.

Friends and Random Numbers

Here is another quite different example of Conditional Probability.

4 friends (Alex, Blake, Chris and Dusty) each choose a random number between 1 and 5. What is the chance that any of them chose the same number?

Let's add our friends one at a time ...

First, what is the chance that Alex and Blake have the same number?

Blake compares his number to Alex's number. There is a 1 in 5 chance of a match.

As a tree diagram :

Note: "Yes" and "No" together  makes 1 (1/5 + 4/5 = 5/5 = 1)

Now, let's include Chris ...

But there are now two cases to consider:

  • If Alex and Blake did match, then Chris has only one number to compare to.
  • But if Alex and Blake did not match then Chris has two numbers to compare to.

And we get this:

For the top line (Alex and Blake did match) we already have a match (a chance of 1/5).

But for the "Alex and Blake did not match" there is now a 2/5 chance of Chris matching (because Chris gets to match his number against both Alex and Blake).

And we can work out the combined chance by multiplying the chances it took to get there:

Following the "No, Yes" path ... there is a 4/5 chance of No, followed by a 2/5 chance of Yes:

Following the "No, No" path ... there is a 4/5 chance of No, followed by a 3/5 chance of No:

Also notice that when we add all chances together we still get 1 (a good check that we haven't made a mistake):

(5/25) + (8/25) + (12/25) = 25/25 = 1

Now what happens when we include Dusty?

It is the same idea, just more of it:

OK, that is all 4 friends, and the "Yes" chances together make 101/125:

Answer: 101/125

But here is something interesting ... if we follow the "No" path we can skip all the other calculations and make our life easier:

The chances of not matching are:

(4/5) × (3/5) × (2/5) = 24/125

So the chances of matching are:

1 - (24/125) = 101/125

(And we didn't really need a tree diagram for that!)

And that is a popular trick in probability:

It is often easier to work out the "No" case (and subtract from 1 for the "Yes" case)

(This idea is shown in more detail at Shared Birthdays .)

Library homepage

  • school Campus Bookshelves
  • menu_book Bookshelves
  • perm_media Learning Objects
  • login Login
  • how_to_reg Request Instructor Account
  • hub Instructor Commons

Margin Size

  • Download Page (PDF)
  • Download Full Book (PDF)
  • Periodic Table
  • Physics Constants
  • Scientific Calculator
  • Reference & Cite
  • Tools expand_more
  • Readability

selected template will load here

This action is not available.

Mathematics LibreTexts

8.4.1: Conditional Probability (Exercises)

  • Last updated
  • Save as PDF
  • Page ID 37915

  • Rupinder Sekhon and Roberta Bloom
  • De Anza College

\( \newcommand{\vecs}[1]{\overset { \scriptstyle \rightharpoonup} {\mathbf{#1}} } \)

\( \newcommand{\vecd}[1]{\overset{-\!-\!\rightharpoonup}{\vphantom{a}\smash {#1}}} \)

\( \newcommand{\id}{\mathrm{id}}\) \( \newcommand{\Span}{\mathrm{span}}\)

( \newcommand{\kernel}{\mathrm{null}\,}\) \( \newcommand{\range}{\mathrm{range}\,}\)

\( \newcommand{\RealPart}{\mathrm{Re}}\) \( \newcommand{\ImaginaryPart}{\mathrm{Im}}\)

\( \newcommand{\Argument}{\mathrm{Arg}}\) \( \newcommand{\norm}[1]{\| #1 \|}\)

\( \newcommand{\inner}[2]{\langle #1, #2 \rangle}\)

\( \newcommand{\Span}{\mathrm{span}}\)

\( \newcommand{\id}{\mathrm{id}}\)

\( \newcommand{\kernel}{\mathrm{null}\,}\)

\( \newcommand{\range}{\mathrm{range}\,}\)

\( \newcommand{\RealPart}{\mathrm{Re}}\)

\( \newcommand{\ImaginaryPart}{\mathrm{Im}}\)

\( \newcommand{\Argument}{\mathrm{Arg}}\)

\( \newcommand{\norm}[1]{\| #1 \|}\)

\( \newcommand{\Span}{\mathrm{span}}\) \( \newcommand{\AA}{\unicode[.8,0]{x212B}}\)

\( \newcommand{\vectorA}[1]{\vec{#1}}      % arrow\)

\( \newcommand{\vectorAt}[1]{\vec{\text{#1}}}      % arrow\)

\( \newcommand{\vectorB}[1]{\overset { \scriptstyle \rightharpoonup} {\mathbf{#1}} } \)

\( \newcommand{\vectorC}[1]{\textbf{#1}} \)

\( \newcommand{\vectorD}[1]{\overrightarrow{#1}} \)

\( \newcommand{\vectorDt}[1]{\overrightarrow{\text{#1}}} \)

\( \newcommand{\vectE}[1]{\overset{-\!-\!\rightharpoonup}{\vphantom{a}\smash{\mathbf {#1}}}} \)

SECTION 8.4 PROBLEM SET: CONDITIONAL PROBABILITY

Questions 1 - 4: Do these problems using the conditional probability formula: \(P(A | B)=\frac{P(A \cap B)}{P(B)}\).

Questions 5 - 8 refer to the following: The table shows the distribution of Democratic and Republican U.S. Senators by gender in the 114 th Congress as of January 2015.

Use this table to determine the following probabilities:

Do the following conditional probability problems.

At a college, 72% of courses have final exams and 46% of courses require research papers. 32% of courses have both a research paper and a final exam. Let \(F\) be the event that a course has a final exam and \(R\) be the event that a course requires a research paper.

Consider a family of three children. Find the following probabilities.

Questions 21 - 26 refer to the following: The table shows highest attained educational status for a sample of US residents age 25 or over:

empower potential ™

Solutions that inspire boundless possibilities: every day, every student.

two high school students working together

Core Curriculum

  • Encourage student-centered learning
  • Build essential skills
  • Nurture lifelong learners

teacher and student working together

  • Accelerate personalized learning
  • Facilitate credit recovery
  • Support students’ academic goals

happy children in classroom

Supplemental and Intervention

  • Deliver personalized learning
  • Encourage skill practice and mastery
  • Tailor learning for grade-level success

student doing at-home learning

Instructional School Services

  • Fill your teacher vacancies with Certified Instructors
  • Expand your course offerings with our standards-aligned course catalog
  • Build your virtual programs with support

In a year, Imagine Learning serves:

U.S. schools

of U.S. districts

Learning Breakthrough Stories

Our latest research.

Cover photo of Imagine Math research paper

Imagine Math Users Achieve Greater Renaissance Star Math Test Growth Than Peers

Cover photo of Imagine Learning Twig Science research paper

California Schools That Use Twig Science Have Significantly More Students Who Meet California Science Test Standards

Cover photo of Imagine Language & Literacy research paper

Grade 4–5 Imagine Language & Literacy Users Showed Greater Literacy Growth Than Peers

Cover photo of Imagine Edgenuity research paper

Imagine Edgenuity Students Achieved Higher Course Grades and More Credit Than Peers

Ignite learning breakthroughs.

Get started with Imagine Learning

  • News & Press

conditional probability assignment edgenuity

The Summer of YES: Custom Summer School Solutions

conditional probability assignment edgenuity

Celebrate 20 Years of Imagine Language & Literacy

conditional probability assignment edgenuity

Imagine Simply Teaching Virtual Symposium

conditional probability assignment edgenuity

Imagine Learning Launches Artificial Intelligence Focused Venture Fund

Community & careers.

teachers having a conversation

The Teachers’ Lounge

Network, collaborate, grow, and innovate with peers in a community space designed for educators.

Person interviewing

Imagine Learning Careers

Seeking lifelong learners who want to empower educators, engage students, and connect families to learning.

If you're seeing this message, it means we're having trouble loading external resources on our website.

If you're behind a web filter, please make sure that the domains *.kastatic.org and *.kasandbox.org are unblocked.

To log in and use all the features of Khan Academy, please enable JavaScript in your browser.

AP®︎/College Statistics

Course: ap®︎/college statistics   >   unit 7, conditional probability and independence.

  • Conditional probability with Bayes' Theorem
  • Conditional probability using two-way tables
  • Calculate conditional probability
  • Conditional probability tree diagram example
  • Tree diagrams and conditional probability

conditional probability assignment edgenuity

Example 1: Income and universities

  • Your answer should be
  • an integer, like 6 ‍  
  • a simplified proper fraction, like 3 / 5 ‍  
  • a simplified improper fraction, like 7 / 4 ‍  
  • a mixed number, like 1   3 / 4 ‍  
  • an exact decimal, like 0.75 ‍  
  • a multiple of pi, like 12   pi ‍   or 2 / 3   pi ‍  
  • a proper fraction, like 1 / 2 ‍   or 6 / 10 ‍  
  • (Choice A)   Yes A Yes
  • (Choice B)   No B No

Example 2: Income and universities (continued)

What if the probabilities are close, want to join the conversation.

  • Upvote Button navigates to signup page
  • Downvote Button navigates to signup page
  • Flag Button navigates to signup page

Incredible Answer

Conditional Probability Calculator

Conditional probability definition - what is conditional probability, how do i calculate conditional probability, conditional probability examples and solutions.

The conditional probability calculator helps you to determine the probability of an event occurring, provided it is conditional on another event .

Read further and we explain:

  • What conditional probability is;
  • How to calculate conditional probability ; and
  • In addition, we show you a real-life conditional probability example where you can also learn how to find it in practice.

If you would like to discover the connection between conditional probability and Bayes' theorem, you may check our Bayes' theorem calculator .

In probability theory, conditional probability quantifies the probability (or likelihood) of an event occurring given that another event has already occurred . This requires that probability of the second event occurring is affected by the first event happening. It is computed by multiplying the probability of the first event by the updated probability of the second, or conditional, event.

Which situations involve conditional probability? Keep reading the article to find it out!

You need the take the following steps to compute the conditional probability of P(A|B):

Determine the total probability of a given final event, B:

P(B) = P(A∩B) + P(Ā∩B) = P(A) * P(B|A) + P(Ā) * P(B|Ā)

Compute the probability of that event :

P(A∩B) = P(A) * P(B|A)

Divide the two numbers :

P(A|B) = P(A∩B) / P(B)

Let's consider a real-life example to demonstrate how to find conditional probability and show the relevance of conditional probability .

Assume that in a population, 5 percent are affected by a disease (denoted by D D D ). A test is available to check if a given person is infected.

The test's sensitivity (also known as the true positive rate ) is 91% , meaning that if the person is infected , the probability for a negative test result (denoted by □ \Box □ ) is 0.09 . On the other hand, the test's specificity (or true negative rate ) is 95% , implying that the probability of having a positive test result (assigned by ⊕ \oplus ⊕ ) when the person doesn't have the disease is 0.05 .

Now, let's consider a random person from the population taking the test and receiving a positive result. How do you find the conditional probability that the person really does have the disease ? We formulate it as P ( D ∩ ⊕ ) P(D \cap \oplus) P ( D ∩ ⊕ ) , that you read as the conditional probability of being infected given that the person has a positive test result .

We can summarize the above problem with the following conditional probability tree diagram :

Conditional probability tree diagram showing the example.

Translating into mathematical formulas, we can obtain the following equations explaining how to find conditional probability:

So, how to solve conditional probability in this situation?

You need to take the following steps :

  • Compute the total probability of having a positive result:
  • Compute the probability of an event when the random person is infected and the test result is positive. You can use the conditional probability definition:
  • Divide the two numbers , taking into account the conditional probability rules:

Thus, the conditional probability that a random person is infected that has a positive test result is 0.4892 , which is almost fifty percent - nearly analogical to tossing a fair coin. You can try to verify the result with our conditional probability calculator if you want.

Note, that the above-described situation may represent the early stages of the COVID pandemic. Even though only a tiny portion of the population has been infected (thus having antibodies), some countries strongly supported wide-range testing based on antibodies with nearly the same parameters. Therefore, it is particularly important to consider the entire clinical state of a patient with signs and symptom before taking the test, as they may present a positive result without being infected. Taking these steps increases the accuracy of the test result.

Which situations involve conditional probability?

There are many real-life situations where conditional probability matters. For example, the probability of winning the second round of a game given that you won the first round . Or the probability that it will rain given that it is cloudy.

What is the conditional probability rule?

Conditional probability measures the chances that an event occurs, given that another event has also occurred . Typically, it is stated as P(B|A) (read as the probability of B given A), where the probability of B depends on the probability of A's occurrence.

Can the conditional probability be zero?

Yes. If the probability of a given final event is zero, the conditional probability of a previous event given that final event is zero.

  • Sum of Squares Calculator
  • Midrange Calculator
  • Coefficient of Variation Calculator

conditional probability assignment edgenuity

Circle skirt

Humans vs vampires, pie chart angle.

  • Biology (101)
  • Chemistry (100)
  • Construction (145)
  • Conversion (295)
  • Ecology (30)
  • Everyday life (262)
  • Finance (571)
  • Health (440)
  • Physics (511)
  • Sports (106)
  • Statistics (184)
  • Other (183)
  • Discover Omni (40)

IMAGES

  1. PPT

    conditional probability assignment edgenuity

  2. Edgenuity

    conditional probability assignment edgenuity

  3. Conditional Probability

    conditional probability assignment edgenuity

  4. SOLUTION: Conditional probability formula notes

    conditional probability assignment edgenuity

  5. Lesson Video: Conditional Probability

    conditional probability assignment edgenuity

  6. Conditional Probability Formula For Independent Events

    conditional probability assignment edgenuity

VIDEO

  1. Conditional probability

  2. Normal probability curve assignment of Assessment and Learning #anuradhabsr #assignment #subscribe

  3. Probability and Statistic Assignment 2

  4. Finite Math (7.4.a) Conditional Probability

  5. STA301 Statistics & Probability Assignment 1 Spring 2024 Virtual University of Pakistan

  6. Introduction to probability and Statistics Week 2 Quiz Assignment Solution

COMMENTS

  1. Conditional Probability: assignment Flashcards

    A: The conditional probability formula is. P (X │ Y) = p (x n y)/p (y) C: The notation P (R │ S) indicates the probability of event R, given that event S has already occurred. E: Conditional probabilities can be calculated using a Venn diagram. Click the card to flip 👆. 1 / 9.

  2. PDF Warm-Up Conditional Probability

    events in which the outcome of one does not affect the outcome of the other. the likelihood that a given event will occur. the probability that a specific event will occur, given that another specific event has already occurred. the occurrence of a possible outcome. . Probability Notation. Consider two events, A and B.

  3. Conditional Probability Flashcards

    Study with Quizlet and memorize flashcards containing terms like Seventy-five percent of the flowers in the arrangement are roses and the rest are tulips. Of the tulips, 50 percent are pink. To the nearest whole percent, what is the probability that a randomly chosen flower from the arrangement is a pink tulip?, Nadia's bookshelf contains 10 fiction books, two reference books, and five ...

  4. PDF Conditional Probabilityriable Warm-Up

    Conditional Probability Applying the Conditional Probability Formula. Example: A pet store has fifteen puppies. Nine of the puppies are male, eight of the puppies are terriers, and six of the terriers are male. A puppy is chosen randomly. Given that the puppy chosen is male, what is the probability that it is a terrier? Event M: choosing a

  5. PDF Conditional Probabilities Warm-Up

    Calculation and Interpretation of Conditional Probability A survey of all residents in a large apartment complex reveals that 52% have a cat, 30% have a dog, and 27% have both. Suppose we select a resident at random. Given that the resident has a cat, what is the probability that the person has a dog?

  6. Conditional Probabilit quiz(I go 100%) Flashcards

    Study with Quizlet and memorize flashcards containing terms like Seventy-five percent of the flowers in the arrangement are roses and the rest are tulips. Of the tulips, 50 percent are pink. To the nearest whole percent, what is the probability that a randomly chosen flower from the arrangement is a pink tulip? 13% 38% 50% 67%, Lola pulls two marbles from a bag containing four red marbles ...

  7. PDF Concepts in Probability and Statistics

    Describe the probability of an event as a number between 0 and 1, which represents the likelihood of the event. Identify an event with a given probability as impossible, unlikely, likely, or certain. Use the fact that the sum of the probabilities of all possible outcomes is 1 to find the probabilities of complementary events.

  8. PDF AP Statistics

    Identify the benefits of using the principle of random assignment within an experimental design. Identify the placebo effect, as well as the benefits of blindness, within an experimental design. Reading Lesson 4.2, Part 2 How to Experiment Well Describe the randomization step within an experimental design using a random number generator.

  9. 2.4: Conditional Probability

    The conditional probability P(Tc ∣ Ac) is called the specificity of the test. The complementary probability P(T ∣ Ac) = 1 − P(Tc ∣ Ac) is the false positive probability. In many cases, the sensitivity and specificity of the test are known, as a result of the development of the test.

  10. 5.5: Conditional Probabilities

    We want to draw a face card and an ace so use multiplication. P(1st face card and 2nd ace) = 12 52 ⋅ 4 51 = 48 2652 ≈ 0.018 P ( 1st face card and 2nd ace) = 12 52 ⋅ 4 51 = 48 2652 ≈ 0.018. The probability that the first card is a face card and the second card an ace is approximately 0.018 or 1.8%. Solution c. There are two ways for this ...

  11. Conditional Probability

    P (B|A) is also called the "Conditional Probability" of B given A. And in our case: P (B|A) = 1/4. So the probability of getting 2 blue marbles is: And we write it as. "Probability of event A and event B equals. the probability of event A times the probability of event B given event A".

  12. 7.10: Conditional Probability and the Multiplication Rule

    Step 7: We can compute the probability of landing on any final node by multiplying the probabilities along the path we would take to get there. For example, the probability of drawing a suspect first and a weapon second (i.e., ending up on the node labeled "SW") is Unexpected text node: 'Figure 7.43.'. Figure 7.43.

  13. Probability and Two-Way Tables- assignment Flashcards

    Consider the two-way table below. Find P (B|C). B. -P (B|C) = = 0.38. Out of 106 total sophomores, there were forty-eight boys who received either an A, B, or C on their first math test. Out of the twenty-eight total A's, sixteen girls received A's. Out of the total fifty-four B's, thirty girls received B's. Twelve boys received a C out of the ...

  14. 8.4.1: Conditional Probability (Exercises)

    This page titled 8.4.1: Conditional Probability (Exercises) is shared under a CC BY 4.0 license and was authored, remixed, and/or curated by Rupinder Sekhon and Roberta Bloom via source content that was edited to the style and standards of the LibreTexts platform; a detailed edit history is available upon request.

  15. PDF Calculate a probability

    The probability that one event happens given that another event is known to have happened is called a conditional probability. The conditional probability that event A happens given that event B has happened is denoted by P(A|B). The event "A and B" is called the intersection of events A and B. It consists of all outcomes that are common to

  16. PDF 3/23/20

    Complete the second ½ the assignment Theoretical Probability in Edgenuity. Grades for Quarter 3 close at 5pm TODAY so that I can finalize and send to administration by the 25 th deadline.

  17. PDF Concepts in Probability and Statistics

    When you log into Edgenuity, you can view the entire course map—an interactive scope and sequence of all topics you will study. The units of study are summarized below: Unit 1: Understanding Probability Unit 2: Probability Unit 3: Unit 9: Sampling and Comparing Populations Unit 4: Data Distributions Unit 5: Data Analysis

  18. PDF Warm-Up Conditional Statements and Equivalence

    of a conditional statement. This is a statement. 6 Logical Statements with Negations • Hidden conditional statement: We will have a picnic if it does not rain. • Conditional statement: If it does not rain, then we will have a . • Converse: If we have a picnic, then it will rain.

  19. Conditional probability and independence (article)

    What is conditional probability and how does it relate to independence? Learn how to use formulas and tables to calculate conditional probabilities and check if two events are independent. Khan Academy is a free online learning platform that covers various topics in math, science, and more.

  20. Algebra 2B: Probability of Independent Events Flashcards

    Edgenuity instruction Learn with flashcards, games, and more — for free. ... Conditional Probability: assignment. 9 terms. TeenTicciToby. Preview. Independent and Mutually Exclusive Events: Assignment. ... possible outcomes The probability the first drawn marble will be blue is ---The probability the second marble drawn will be yellow is ...

  21. Conditional Probability Calculator

    Thus, the conditional probability that a random person is infected that has a positive test result is 0.4892, which is almost fifty percent - nearly analogical to tossing a fair coin.You can try to verify the result with our conditional probability calculator if you want. Note, that the above-described situation may represent the early stages of the COVID pandemic.

  22. PDF Complete Calculate

    • What is the probability that a randomly selected student took the training course and passed the driving test? Circle the cell in the table that represents a randomly selected student who took the training course and passed the driving test. The frequency is 0.41. So, the probability that a randomly selected student took

  23. PDF Warm-Up Theoretical and Experimental Probability

    Favorable outcomes: A = {1, 2, 3, } S. Find the probability of rolling less than 5. 4. (< 5) = =. 6. The probability of rolling a number less than 5 is 2 out of 3, or about 67%. Because this number is greater than half, this event is actually more likely to occur than not to occur. Geometric Probability.