(29 intermediate revisions by 3 users not shown) | |||
Line 1: | Line 1: | ||
[[Category:math]] | [[Category:math]] | ||
[[Category:tutorial]] | [[Category:tutorial]] | ||
+ | [[Category:bayes rule]] | ||
+ | [[Category:conditional probability]] | ||
+ | [[Category:math squad]] | ||
+ | |||
== Bayes' Theorem == | == Bayes' Theorem == | ||
− | by Maliha Hossain | + | by [[user:Mhossain | Maliha Hossain]], proud Member of [[Math_squad | the Math Squad]]. |
− | <pre> keyword: probability, Bayes' Theorem, Bayes' Rule </pre> | + | ---- |
+ | <pre>keyword: probability, Bayes' Theorem, Bayes' Rule </pre> | ||
'''INTRODUCTION''' | '''INTRODUCTION''' | ||
− | Bayes' Theorem (or Bayes' Rule) allows us to calculate P(A|B) from P(B|A) given that P(A) and P(B) are also known | + | Bayes' Theorem (or Bayes' Rule) allows us to calculate P(A|B) from P(B|A) given that P(A) and P(B) are also known. In this tutorial, we will derive Bayes' Theorem and illustrate it with a few examples. After going over the examples, if you have any questions or if you find any mistakes please leave me a comment at the end of the relevant section. |
− | Note that this tutorial assumes familiarity with conditional probability and the axioms of probability. | + | Note that this tutorial assumes familiarity with conditional probability and the axioms of probability. If you interested in the derivation of the conditional distributions for continuous and discrete random variables, you may wish to go over Professor Mary Comer's [[ECE600_F13_rv_conditional_distribution_mhossain|notes]] on the subject. |
<pre> Contents | <pre> Contents | ||
- Bayes' Theorem | - Bayes' Theorem | ||
- Proof | - Proof | ||
− | - Example | + | - Example Problems |
− | + | ||
− | + | ||
- References | - References | ||
</pre> | </pre> | ||
Line 31: | Line 34: | ||
== Proof == | == Proof == | ||
− | We will now derive Bayes' | + | We will now derive Bayes' Theorem as it is expressed in the second form, which simply takes the expression one step further than the first. |
Let <math>A</math> and <math>B_j</math> be as defined above. By definition of the conditional probability, we have that | Let <math>A</math> and <math>B_j</math> be as defined above. By definition of the conditional probability, we have that | ||
Line 37: | Line 40: | ||
<math>P[A|B_j] = \frac{P[A\cap B_j]}{P[B_j]}</math> | <math>P[A|B_j] = \frac{P[A\cap B_j]}{P[B_j]}</math> | ||
− | + | Multiplying both sides with <math>B_j</math>, we get | |
− | <math>P[A\cap B_j] = P[A|B_j]P[B_j]</math> | + | <math>P[A\cap B_j] = P[A|B_j]P[B_j] \ </math> |
− | + | Using the same argument as above, we have that | |
− | <math>P[B_j|A] = \frac{P[B_j\cap A]}{P[A]} | + | <math> |
+ | \begin{align} | ||
+ | P[B_j|A] & = \frac{P[B_j\cap A]}{P[A]} \\ | ||
− | + | \Rightarrow P[B_j\cap A] &= P[B_j|A]P[A] | |
+ | \end{align} | ||
+ | </math> | ||
Because of the commutativity property of intersection, we can say that | Because of the commutativity property of intersection, we can say that | ||
− | <math>P[B_j|A]P[A] = P[A|B_j]P[B_j]</math> | + | <math> P[B_j|A]P[A] = P[A|B_j]P[B_j] \ </math> |
+ | Dividing both sides by <math>P[A]</math>, we get | ||
+ | |||
+ | <math> P[B_j|A] = \frac{P[A|B_j]P[B_j]}{P[A]}</math> | ||
+ | |||
+ | Finally, the denominator can be broken down further using the theorem of total probability so that we have the following expression | ||
+ | |||
+ | <math>P[B_j|A] = \frac{P[A|B_j]P[B_j]}{\sum_{k=1}^n P[A|B_k]P[B_k]}</math> | ||
---- | ---- | ||
+ | |||
+ | == Example Problems == | ||
+ | |||
+ | [[bayes_theorem_eg1_S13|Example 1: Quality Control]] | ||
+ | |||
+ | [[bayes_theorem_eg2_S13|Example 2: False Positive Paradox]] | ||
+ | |||
+ | [[bayes_theorem_eg3_S13|Example 3: Monty Hall Problem]] | ||
+ | ---- | ||
+ | |||
+ | == References == | ||
+ | |||
+ | * Alberto Leon-Garcia, ''Probability, Statistics, and Random Processes for Electrical Engineering,'' Third Edition | ||
+ | ---- | ||
+ | |||
+ | ==Questions and comments== | ||
+ | |||
+ | If you have any questions, comments, etc. please post them below: | ||
+ | |||
+ | * Comment / question 1 | ||
+ | |||
+ | ---- | ||
+ | |||
+ | [[Math_squad|Back to Math Squad page]] | ||
+ | |||
+ | |||
+ | <div style="font-family: Verdana, sans-serif; font-size: 14px; text-align: justify; width: 70%; margin: auto; border: 1px solid #aaa; padding: 2em;"> | ||
+ | The Spring 2013 Math Squad 2013 was supported by an anonymous [https://www.projectrhea.org/learning/donate.php gift] to [https://www.projectrhea.org/learning/about_Rhea.php Project Rhea]. If you enjoyed reading these tutorials, please help Rhea "help students learn" with a [https://www.projectrhea.org/learning/donate.php donation] to this project. Your [https://www.projectrhea.org/learning/donate.php contribution] is greatly appreciated. | ||
+ | </div> |
Latest revision as of 12:08, 25 November 2013
Contents
Bayes' Theorem
by Maliha Hossain, proud Member of the Math Squad.
keyword: probability, Bayes' Theorem, Bayes' Rule
INTRODUCTION
Bayes' Theorem (or Bayes' Rule) allows us to calculate P(A|B) from P(B|A) given that P(A) and P(B) are also known. In this tutorial, we will derive Bayes' Theorem and illustrate it with a few examples. After going over the examples, if you have any questions or if you find any mistakes please leave me a comment at the end of the relevant section.
Note that this tutorial assumes familiarity with conditional probability and the axioms of probability. If you interested in the derivation of the conditional distributions for continuous and discrete random variables, you may wish to go over Professor Mary Comer's notes on the subject.
Contents - Bayes' Theorem - Proof - Example Problems - References
Bayes' Theorem
Let $ B_1, B_2, ..., B_n $ be a partition of the sample space $ S $, i.e. $ B_1, B_2, ..., B_n $ are mutually exclusive events whose union equals the sample space S. Suppose that the event $ A $ occurs. Then, by Bayes' Theorem, we have that
$ P[B_j|A] = \frac{P[A|B_j]P[B_j]}{P[A]}, j = 1, 2, . . . , n $
Bayes' Theorem is also often expressed in the following form:
$ P[B_j|A] = \frac{P[A|B_j]P[B_j]}{\sum_{k=1}^n P[A|B_k]P[B_k]} $
Proof
We will now derive Bayes' Theorem as it is expressed in the second form, which simply takes the expression one step further than the first.
Let $ A $ and $ B_j $ be as defined above. By definition of the conditional probability, we have that
$ P[A|B_j] = \frac{P[A\cap B_j]}{P[B_j]} $
Multiplying both sides with $ B_j $, we get
$ P[A\cap B_j] = P[A|B_j]P[B_j] \ $
Using the same argument as above, we have that
$ \begin{align} P[B_j|A] & = \frac{P[B_j\cap A]}{P[A]} \\ \Rightarrow P[B_j\cap A] &= P[B_j|A]P[A] \end{align} $
Because of the commutativity property of intersection, we can say that
$ P[B_j|A]P[A] = P[A|B_j]P[B_j] \ $
Dividing both sides by $ P[A] $, we get
$ P[B_j|A] = \frac{P[A|B_j]P[B_j]}{P[A]} $
Finally, the denominator can be broken down further using the theorem of total probability so that we have the following expression
$ P[B_j|A] = \frac{P[A|B_j]P[B_j]}{\sum_{k=1}^n P[A|B_k]P[B_k]} $
Example Problems
Example 2: False Positive Paradox
References
- Alberto Leon-Garcia, Probability, Statistics, and Random Processes for Electrical Engineering, Third Edition
Questions and comments
If you have any questions, comments, etc. please post them below:
- Comment / question 1
The Spring 2013 Math Squad 2013 was supported by an anonymous gift to Project Rhea. If you enjoyed reading these tutorials, please help Rhea "help students learn" with a donation to this project. Your contribution is greatly appreciated.