Introduction to Probability

Site: Saylor Academy
Course: MA121: Introduction to Statistics
Book: Introduction to Probability
Printed by: Guest user
Date: Saturday, July 27, 2024, 12:25 AM

Description

First, we will discuss experiments where outcomes are equally likely to occur and the frequency approach to assigning probabilities. Then, we will focus on the concept of events and touch on the issue of conditional probability.


Learning Objectives

  1. Define symmetrical outcomes
  2. Distinguish between frequentist and subjective approaches
  3. Determine whether the frequentist or subjective approach is better suited for a given situation

Inferential statistics is built on the foundation of probability theory, and has been remarkably successful in guiding opinion about the conclusions to be drawn from data. Yet (paradoxically) the very idea of probability has been plagued by controversy from the beginning of the subject to the present day. In this section we provide a glimpse of the debate about the interpretation of the probability concept.

One conception of probability is drawn from the idea of symmetrical outcomes. For example, the two possible outcomes of tossing a fair coin seem not to be distinguishable in any way that affects which side will land up or down. Therefore the probability of heads is taken to be 1/2, as is the probability of tails. In general, if there are N symmetrical outcomes, the probability of any given one of them occurring is taken to be 1/N. Thus, if a six-sided die is rolled, the probability of any one of the six sides coming up is 1/6.

Probabilities can also be thought of in terms of relative frequencies. If we tossed a coin millions of times, we would expect the proportion of tosses that came up heads to be pretty close to 1/2. As the number of tosses increases, the proportion of heads approaches 1/2. Therefore, we can say that the probability of a head is 1/2.

If it has rained in Seattle on 62% of the last 100,000 days, then the probability of it raining tomorrow might be taken to be 0.62. This is a natural idea but nonetheless unreasonable if we have further information relevant to whether it will rain tomorrow. For example, if tomorrow is August 1, a day of the year on which it seldom rains in Seattle, we should only consider the percentage of the time it rained on August 1. But even this is not enough since the probability of rain on the next August 1 depends on the humidity. (The chances are higher in the presence of high humidity.) So, we should consult only the prior occurrences of August 1 that had the same humidity as the next occurrence of August 1. Of course, wind direction also affects probability ... You can see that our sample of prior cases will soon be reduced to the empty set. Anyway, past meteorological history is misleading if the climate is changing.

For some purposes, probability is best thought of as subjective. Questions such as "What is the probability that Ms. Garcia will defeat Mr. Smith in an upcoming congressional election?" do not conveniently fit into either the symmetry or frequency approaches to probability. Rather, assigning probability 0.7 (say) to this event seems to reflect the speaker's personal opinion – perhaps his willingness to bet according to certain odds. Such an approach to probability, however, seems to lose the objective content of the idea of chance; probability becomes mere opinion.

Two people might attach different probabilities to the election outcome, yet there would be no criterion for calling one "right" and the other "wrong". We cannot call one of the two people right simply because she assigned higher probability to the outcome that actually transpires. After all, you would be right to attribute probability 1/6 to throwing a six with a fair die, and your friend who attributes 2/3 to this event would be wrong. And you are still right (and your friend is still wrong) even if the die ends up showing a six! The lack of objective criteria for adjudicating claims about probabilities in the subjective perspective is an unattractive feature of it for many scholars.

Like most work in the field, the present text adopts the frequentist approach to probability in most cases. Moreover, almost all the probabilities we shall encounter will be nondogmatic, that is, neither zero nor one. An event with probability 0 has no chance of occurring; an event of probability 1 is certain to occur. It is hard to think of any examples of interest to statistics in which the probability is either 0 or 1. (Even the probability that the Sun will come up tomorrow is less than 1.)

The following example illustrates our attitude about probabilities. Suppose you wish to know what the weather will be like next Saturday because you are planning a picnic. You turn on your radio, and the weather person says, "There is a 10% chance of rain". You decide to have the picnic outdoors and, lo and behold, it rains. You are furious with the weather person. But was she wrong? No, she did not say it would not rain, only that rain was unlikely. She would have been flatly wrong only if she said that the probability is 0 and it subsequently rained. However, if you kept track of her weather predictions over a long period of time and found that it rained on 50% of the days that the weather person said the probability was 0.10, you could say her probability assessments are wrong.

So when is it accurate to say that the probability of rain is 0.10? According to our frequency interpretation, it means that it will rain 10% of the days on which rain is forecast with this probability.



Source: David M. Lane, https://onlinestatbook.com/2/probability/probability_intro.html
Public Domain Mark This work is in the Public Domain.

 

 

Question 1 out of 3.
Select all that apply. Probability can be thought of as:

 symmetrical outcomes

 relative frequencies

 subjective

Question 2 out of 3.
The paper says there is an 80% chance of rain today, so you plan indoor activities. Then it doesn't rain. Was the forecast wrong?

 yes

 no

Question 3 out of 3.
Most probabilities we will deal with in psychology are zero or one.

 true

 false


  1. There is a debate about how to interpret probability. All three of these can be ways to think of probability.

  2. The forecast would have only been wrong if it had predicted a 100% chance of rain. However, the probability assessments are wrong if, over time, it usually doesn't rain on days where an 80% chance of rain is predicted.

  3. We will deal with very few probabilities that are zero (definitely will not happen) or one (definitely will happen).

Learning Objectives

  1. Compute probability in a situation where there are equally-likely outcomes
  2. Apply concepts to cards and dice
  3. Compute the probability of two independent events both occurring
  4. Compute the probability of either of two independent events occurring
  5. Do problems that involve conditional probabilities
  6. Compute the probability that in a room of N people, at least two share a birthday
  7. Describe the gambler's fallacy
  8. Probability of a Single Event

If you roll a six-sided die, there are six possible outcomes, and each of these outcomes is equally likely. A six is as likely to come up as a three, and likewise for the other four sides of the die. What, then, is the probability that a one will come up? Since there are six possible outcomes, the probability is 1/6. What is the probability that either a one or a six will come up? The two outcomes about which we are concerned (a one or a six coming up) are called favorable outcomes. Given that all outcomes are equally likely, we can compute the probability of a one or a six using the formula:

probability =\frac{\text { Number of favorable outcomes }}{\text { Number of possible equally-likely outcomes }}

In this case there are two favorable outcomes and six possible outcomes. So the probability of throwing either a one or six is 1/3. Don't be misled by our use of the term "favorable," by the way. You should understand it in the sense of "favorable to the event in question happening". That event might not be favorable to your well-being. You might be betting on a three, for example.

The above formula applies to many games of chance. For example, what is the probability that a card drawn at random from a deck of playing cards will be an ace? Since the deck has four aces, there are four favorable outcomes; since the deck has 52 cards, there are 52 possible outcomes. The probability is therefore 4/52 = 1/13. What about the probability that the card will be a club? Since there are 13 clubs, the probability is 13/52 = 1/4.

Let's say you have a bag with 20 cherries: 14 sweet and 6 sour. If you pick a cherry at random, what is the probability that it will be sweet? There are 20 possible cherries that could be picked, so the number of possible outcomes is 20. Of these 20 possible outcomes, 14 are favorable (sweet), so the probability that the cherry will be sweet is 14/20 = 7/10. There is one potential complication to this example, however. It must be assumed that the probability of picking any of the cherries is the same as the probability of picking any other. This wouldn't be true if (let us imagine) the sweet cherries are smaller than the sour ones. (The sour cherries would come to hand more readily when you sampled from the bag.) Let us keep in mind, therefore, that when we assess probabilities in terms of the ratio of favorable to all potential cases, we rely heavily on the assumption of equal probability for all outcomes.

Here is a more complex example. You throw 2 dice. What is the probability that the sum of the two dice will be 6? To solve this problem, list all the possible outcomes. There are 36 of them since each die can come up one of six ways. The 36 possibilities are shown below.

Die 1 Die 2 Total   Die 1 Die 2 Total   Die 1 Die 2 Total
1 1 2   3 1 4   5 1 6
1 2 3   3 2 5   5 2 7
1 3 4   3 3 6   5 3 8
1 4 5   3 4 7   5 4 9
1 5 6   3 5 8   5 5 10
1 6 7   3 6 9   5 6 11
2 1 3   4 1 5   6 1 7
2 2 4   4 2 6   6 2 8
2 3 5   4 3 7   6 3 9
2 4 6   4 4 8   6 4 10
2 5 7   4 5 9   6 5 11
2 6 8   4 6 10   6 6 12

You can see that 5 of the 36 possibilities total 6. Therefore, the probability is 5/36.

If you know the probability of an event occurring, it is easy to compute the probability that the event does not occur. If P(A) is the probability of Event A, then 1 - P(A) is the probability that the event does not occur. For the last example, the probability that the total is 6 is 5/36. Therefore, the probability that the total is not 6 is 1 - 5/36 = 31/36.

Events A and B are independent events if the probability of Event B occurring is the same whether or not Event A occurs. Let's take a simple example. A fair coin is tossed two times. The probability that a head comes up on the second toss is 1/2 regardless of whether or not a head came up on the first toss. The two events are (1) first toss is a head and (2) second toss is a head. So these events are independent. Consider the two events (1) "It will rain tomorrow in Houston" and (2) "It will rain tomorrow in Galveston" (a city near Houston). These events are not independent because it is more likely that it will rain in Galveston on days it rains in Houston than on days it does not.

Probability of A and B

When two events are independent, the probability of both occurring is the product of the probabilities of the individual events. More formally, if events A and B are independent, then the probability of both A and B occurring is:

P(A\;and\;B) = P(A) x P(B)

where P(A\;and\;B) is the probability of events A and B both occurring, P(A) is the probability of event A occurring, and P(B) is the probability of event B occurring.

If you flip a coin twice, what is the probability that it will come up heads both times? Event A is that the coin comes up heads on the first flip and Event B is that the coin comes up heads on the second flip. Since both P(A) and P(B) equal 1/2, the probability that both events occur is

1/2\times1/2 = 1/4.

Let's take another example. If you flip a coin and roll a six-sided die, what is the probability that the coin comes up heads and the die comes up 1? Since the two events are independent, the probability is simply the probability of a head (which is 1/2) times the probability of the die coming up 1 (which is 1/6). Therefore, the probability of both events occurring is 1/2 x 1/6 = 1/12.

One final example: You draw a card from a deck of cards, put it back, and then draw another card. What is the probability that the first card is a heart and the second card is black? Since there are 52 cards in a deck and 13 of them are hearts, the probability that the first card is a heart is 13/52 = 1/4. Since there are 26 black cards in the deck, the probability that the second card is black is 26/52 = 1/2. The probability of both events occurring is therefore 1/4 x 1/2 = 1/8.

See the section on conditional probabilities on this page to see how to compute P(A\;and\;B) when A and B are not independent.

Probability of A or B

If Events A and B are independent, the probability that either Event A or Event B occurs is:

P(A\;or\;B) = P(A) + P(B) - P(A\;and\;B)

In this discussion, when we say "A or B occurs" we include three possibilities:

  1. A occurs and B does not occur
  2. B occurs and A does not occur
  3. Both  A and B occur

This use of the word "or" is technically called inclusive or because it includes the case in which both A and B occur. If we included only the first two cases, then we would be using an exclusive or.

(Optional) We can derive the law for P(A-or-B) from our law about P(A-and-B). The event "A-or-B" can happen in any of the following ways::

  1. A-and-B happens
  2. A-and-not-B happens
  3. not-A-and-B happens.

The simple event A can happen if either A-and-B happens or A-and-not-B happens. Similarly, the simple event B happens if either A-and-B happens or not-A-and-B happens. P(A)+P(B) is therefore P(A-and-B )+P(A-and-not-B )+ P(A-and-B)+P( not-A-and-B ), whereas P(A-or-B ) is P(A-and-B )+P(A-and-notB) +P( not-A-and-B ). We can make these two sums equal by subtracting one occurrence of P(A-and-B) from the first. Hence, P(A-or-B)=P(A)+P(B)-P(A-and-B).

Now for some examples. If you flip a coin two times, what is the probability that you will get a head on the first flip or a head on the second flip (or both)? Letting Event A be a head on the first flip and Event B be a head on the second flip, then P(A) = 1/2, P(B) = 1/2, and P(A\;and\;B) = 1/4. Therefore,

P(A\;or\;B) = 1/2 + 1/2 - 1/4 = 3/4.

If you throw a six-sided die and then flip a coin, what is the probability that you will get either a 6 on the die or a head on the coin flip (or both)? Using the formula,


    P(6\;or\;head) = P(6) + P(head) - P(6\;and\;head)
                         = (1/6) + (1/2) - (1/6)(1/2)
                         = 7/12

An alternate approach to computing this value is to start by computing the probability of not getting either a 6 or a head. Then subtract this value from 1 to compute the probability of getting a 6 or a head. Although this is a complicated method, it has the advantage of being applicable to problems with more than two events. Here is the calculation in the present case. The probability of not getting either a 6 or a head can be recast as the probability of

(not getting a 6) AND (not getting a head).

This follows because if you did not get a 6 and you did not get a head, then you did not get a 6 or a head. The probability of not getting a six is 1 - 1/6 = 5/6. The probability of not getting a head is 1 - 1/2 = 1/2. The probability of not getting a six and not getting a head is 5/6 x 1/2 = 5/12. This is therefore the probability of not getting a 6 or a head. The probability of getting a six or a head is therefore (once again) 1 - 5/12 = 7/12.

If you throw a die three times, what is the probability that one or more of your throws will come up with a 1? That is, what is the probability of getting a 1 on the first throw OR a 1 on the second throw OR a 1 on the third throw? The easiest way to approach this problem is to compute the probability of

NOT getting a 1 on the first throw

AND not getting a 1 on the second throw

AND not getting a 1 on the third throw.

The answer will be 1 minus this probability. The probability of not getting a 1 on any of the three throws is 5/6 x 5/6 x 5/6 = 125/216. Therefore, the probability of getting a 1 on at least one of the throws is 1 - 125/216 = 91/216.

Often it is required to compute the probability of an event given that another event has occurred. For example, what is the probability that two cards drawn at random from a deck of playing cards will both be aces? It might seem that you could use the formula for the probability of two independent events and simply multiply 4/52 x 4/52 = 1/169. This would be incorrect, however, because the two events are not independent. If the first card drawn is an ace, then the probability that the second card is also an ace would be lower because there would only be three aces left in the deck.

Once the first card chosen is an ace, the probability that the second card chosen is also an ace is called the conditional probability of drawing an ace. In this case, the "condition" is that the first card is an ace. Symbolically, we write this as:

P(ace on second draw | an ace on the first draw)

The vertical bar "|" is read as "given," so the above expression is short for: "The probability that an ace is drawn on the second draw given that an ace was drawn on the first draw". What is this probability? Since after an ace is drawn on the first draw, there are 3 aces out of 51 total cards left. This means that the probability that one of these aces will be drawn is 3/51 = 1/17.

If Events A and B are not independent, then P(A\;and\;B) = P(A) \times P(B|A).

Applying this to the problem of two aces, the probability of drawing two aces from a deck is 4/52 x 3/51 = 1/221.

One more example: If you draw two cards from a deck, what is the probability that you will get the Ace of Diamonds and a black card? There are two ways you can satisfy this condition: (1) You can get the Ace of Diamonds first and then a black card or (2) you can get a black card first and then the Ace of Diamonds. Let's calculate Case A. The probability that the first card is the Ace of Diamonds is 1/52. The probability that the second card is black given that the first card is the Ace of Diamonds is 26/51 because 26 of the remaining 51 cards are black. The probability is therefore 1/52 x 26/51 = 1/102. Now for Case 2: the probability that the first card is black is 26/52 = 1/2. The probability that the second card is the Ace of Diamonds given that the first card is black is 1/51. The probability of Case 2 is therefore 1/2 x 1/51 = 1/102, the same as the probability of Case 1. Recall that the probability of A or B is P(A) + P(B) - P(A\;and\;B). In this problem, P(A\;and\;B) = 0 since a card cannot be the Ace of Diamonds and be a black card. Therefore, the probability of Case 1 or Case 2 is 1/102 + 1/102 = 2/102 = 1/51. So, 1/51 is the probability that you will get the Ace of Diamonds and a black card when drawing two cards from a deck.

If there are 25 people in a room, what is the probability that at least two of them share the same birthday. If your first thought is that it is 25/365 = 0.068, you will be surprised to learn it is much higher than that. This problem requires the application of the sections on P(A\;and\;B) and conditional probability.

This problem is best approached by asking what is the probability that no two people have the same birthday. Once we know this probability, we can simply subtract it from 1 to find the probability that two people share a birthday.

If we choose two people at random, what is the probability that they do not share a birthday? Of the 365 days on which the second person could have a birthday, 364 of them are different from the first person's birthday. Therefore the probability is 364/365. Let's define P2 as the probability that the second person drawn does not share a birthday with the person drawn previously. P2 is therefore 364/365. Now define P3 as the probability that the third person drawn does not share a birthday with anyone drawn previously given that there are no previous birthday matches. P3 is therefore a conditional probability. If there are no previous birthday matches, then two of the 365 days have been "used up," leaving 363 non-matching days. Therefore P3 = 363/365. In like manner, P4 = 362/365, P5 = 361/365, and so on up to P25 = 341/365.

In order for there to be no matches, the second person must not match any previous person and the third person must not match any previous person, and the fourth person must not match any previous person, etc. Since P(A\;and\;B) = P(A)P(B), all we have to do is multiply P2, P3, P4 ... P25 together. The result is 0.431. Therefore the probability of at least one match is 0.569.

A fair coin is flipped five times and comes up heads each time. What is the probability that it will come up heads on the sixth flip? The correct answer is, of course, 1/2. But many people believe that a tail is more likely to occur after throwing five heads. Their faulty reasoning may go something like this: "In the long run, the number of heads and tails will be the same, so the tails have some catching up to do". The flaws in this logic are exposed in the simulation in this chapter.

Question 1 out of 12.
You have a bag of marbles. There are 3 red marbles, 2 green marbles, 7 yellow marbles, and 3 blue marbles. What is the probability of drawing a blue marble?

Question 2 out of 12.
You have a bag of marbles. There are 3 red marbles, 2 green marbles, 7 yellow marbles, and 3 blue marbles. What is the probability of drawing a yellow or red marble?

Question 3 out of 12.
You have a bag of marbles. There are 3 red marbles, 2 green marbles, 7 yellow marbles, and 3 blue marbles. What is the probability of drawing something other than a red marble?

Question 4 out of 12.
You throw 2 dice. What is the probability that the sum of the two dice will be 5?

Question 5 out of 12.
Select all that apply. Which of the following pairs are independent events?

 two coin flips

 a student's midterm and final grades in a class

 draw an ace, leave it out, then draw an ace again

 draw an ace, put it back, then draw an ace again

Question 6 out of 12.
You win a game if you roll a die and get an odd number and flip a coin and get tails. What is the probability that you win?

Question 7 out of 12.
You win $1 every time you flip a coin and get heads. You win $3 every time you roll a die and get a 5. You roll the die and flip the coin one time each. What is the probability that you win money from at least one game?

Question 8 out of 12.
A survey showed that 60% of all adults in your city take public transportation to work. If 3 people are chosen at random, what is the probability that they will all take public transportation to work?

Question 9 out of 12.
You win a game if you roll a 3 on at least one of your two dice. What is the probability that you win?

Question 10 out of 12.
If you throw a die four times, what is the probability that one or more of your throws will come up with a 4?

Question 11 out of 12.
What is the probability that you draw two cards from a deck and both of them are spades?

Question 12 out of 12.
You flip a coin three times. Is it more likely to get heads all three times or heads, then tails, then heads?

 HHH

 HTH

 They are both equally likely

 

 


  1. 3 out of the 15 possible outcomes are favorable, so 3/15 = 0.2

  2. There are 7+3 = 10 favorable outcomes out of a possible 15. 10/15 = 0.667

  3. The probability of drawing a red marble is 3/15 = 0.2, so the probability of NOT drawing a red marble is: 1 - 0.2 = 0.8.

  4. To solve this problem, list all the possible outcomes. There are 36 of them since each die can come up one of six ways. There are four ways to get a sum of five: (1) Get a 1 then a 4 (2) Get a 4 then a 1 (3) Get a 2 then a 3 (4) Get a 3 then a 2; 4/36 = 0.11

  5. Two events are independent if the occurrence of one has no effect on the probability of the occurrence of the other. A student's test grades are not independent because a student who does well on the midterm understands the material and is therefore more likely to do well on the final. Two draws of an ace without replacement are not independent because what you get on your first draw affects the probability of getting an ace on the second draw. This is a conditional probability.

  6. P(Odd # AND Tails) = 3/6 x 1/2 = .5 x .5 = 0.25

  7. The probability that you win $1 OR $3 (or both games) is: (1/2) + (1/6) - (1/2)(1/6) = 7/12 = 0.58

  8. The probability that person 1 AND person 2 AND person 3 will take public transportation is: .6 x .6 x.6 = 0.216

  9. The probability of getting a 3 on the first roll OR a 3 on the second roll is: (1/6) + (1/6) - (1/6)(1/6) = 11/36 = 0.31

  10. The probability of NOT getting a 4 on any roll is: (5/6)(5/6)(5/6)(5/6) = .48, so the probability of rolling at least one 4 is: 1 - .48 = 0.52

  11. P(spade on the first draw) = 13/52 = .25, P(spade on the second draw | spade on the first draw) = 12/51 = .235, .25*.235 = 0.0588

  12. They are both equally likely. There are 8 possible outcomes for flipping a coin 3 times, so each has a 1/8 probability.