Marquette University

Department of Mathematics, Statistics and Computer Science

Wim Ruitenburg's Fall 2010 MATH 1300-101

Last updated: November 2010
Comments and suggestions: Email   wimr@mscs.mu.edu

Book, chapter 15 on probability, plus extra notes


Probability from chapter 15

From the chapter we can learn: Some rules of computing the number of elements of a sample space are: These formulas are particularly useful when the probability space is equiprobable (page 569). In class we did two examples:

Taking small risks often

Here is a special limit case of the product rule. Suppose we take a same small risk of say 1/n to our lives, and take this risk n independent times. What is our chance to survive?

The birthday paradox

During the class we had about 60 people. What are the hidden assumptions about the distribution of birthdays of people?

The other child

Suppose a parent has two children. Again there are hidden assumptions. In this case, the `reasonable' hidden assumption is that each next child has an equal and independent chance of being a boy or a girl. Assuming this, the two chances above are 1/2 and 2/3.

Three Doors

This problem is also referred to as the Monty Hall problem. The game host shows us three doors, a red door, a white door, and a blue door. Behind one of these doors there is a big prize. We are asked to pick one of the three doors. Once our choice is final, and the prize is behind the door, we have won. Obviously, our chance of winning is one-third, or 1/3. What happens with our chances when the game show host adapts the rules a bit? Suppose we have picked a door, say the red door. Then the game host opens another door, say the white door, and reveals that there is no prize behind the white door. Now we are offered the options of sticking with the red door, or switch to the blue door. Should we switch, or should we stay? What are our chances for the red door, what for the blue door? Let us change the problem in a seemingly irrelevant way as follows. As before, we are asked to pick one of the three doors. Once our choice is final, and the prize is behind the door, we have won. Obviously, our chance of winning is 1/3. Suppose we have picked the red door. Then a storm blows through the hall, and one of the doors is randomly blown open. Suppose that the white door blew open. Behind it there is no prize. Now we are offered the options of sticking with the red door, or switch to the blue door. Should we switch, or should we stay?

Spinning Wheels

Page 580 illustrates some spinning wheels. We are interested in finding probabilities and expected values of such spinning wheels. When is the game fair? When is the game unfair in our favor? When is the game unfair in favor of the wheel owner?

The Best of Three Dice

Suppose we have three fair but unusual dice, as follows. Instead of having six sides with values 1, 2, 3, 4, 5, and 6, their sides are allowed to have other values. We have a red, a white, and a blue one. The red one has numbers 2, 2, 2, 2, 6, and 6 on its six sides. The white one has numbers 1, 1, 5, 5, 5, and 5 on its six sides. The blue one has numbers 3, 3, 3, 3, 3, and 3 on its six sides. When we list the six sides in tables, we have a situation like this:
2
2
2
2
6
6
1
1
5
5
5
5
3
3
3
3
3
3
The following game is played between two players. The first player picks one on the dice. Then the second player picks another. Next, both players throw their dice simultaneously. Whoever throws the higher number, wins. With the three dice above, draws are not possible. Question: Which one of these three dice is best? In the game above, it is good to be the second player. No matter which color the first player picks, the second player can always pick another color which is better than the first color.

Winning a Losing Game

Suppose you play a color at roulette. You can pick either black or red. Your chance for doubling your money is 18/38, your chance for losing your money is 20/38.

Check Swapping

Suppose we have a stack of checks with denominations of 1 dollar, 2 dollars, 4 dollars, 8 dollars, 16 dollars, 32 dollars, 64 dollars, and 128 dollars. (We rather arbitrarily stop at 128; you may change the problem by extending the pile of possible checks with many more doublings.) We also have two blank envelops. In each of the envelops I put one check such that, with equal probability, I either pick 1 and 2, or 2 and 4, or 4 and 8, or 8 and 16, or 16 and 32, or 32 and 64, or 64 and 128. You can see the two envelops. You do not know which two consecutive checks I picked. You know that one envelop contains twice the amount of the other, but you don't know which one.