(4 intermediate revisions by the same user not shown) | |||
Line 1: | Line 1: | ||
[[Category:2010 Spring ECE 662 mboutin]] | [[Category:2010 Spring ECE 662 mboutin]] | ||
− | = | + | =Details of Lecture 5, [[ECE662]] Spring 2010= |
+ | In Lecture 5, we observed that [[Bayes_Decision_Theory|Bayes rule]] minimizes the probability of error. One student pointed out an example that raises doubts about the validity of this statement. This gave rise to an interesting [[ECE662_topic2_discussions|online discussion]]. In particular, a student [[EE662Sp10OptimalPrediction|proved that the example previously proposed performs worse]] than following Bayes rule. | ||
− | + | We extended the discussion of [[Bayes_Decision_Theory|Bayes rule]] to the case of continuous-valued feature vectors, including a discussion of the expected loss (called "risk") when following [[Bayes_Decision_Theory|Bayes rule]]. | |
+ | Previous: [[Lecture4ECE662S10|Lecture 4]] | ||
+ | Next: [[Lecture6ECE662S10|Lecture 6]] | ||
+ | ---- | ||
+ | [[OutlineECE662S10|Back to course outline]] | ||
+ | [[ 2010 Spring ECE 662 mboutin|Back to 2010 Spring ECE 662 mboutin]] | ||
− | + | [[ECE662|Back to ECE662]] | |
− | [[ | + |
Latest revision as of 10:42, 13 April 2010
Details of Lecture 5, ECE662 Spring 2010
In Lecture 5, we observed that Bayes rule minimizes the probability of error. One student pointed out an example that raises doubts about the validity of this statement. This gave rise to an interesting online discussion. In particular, a student proved that the example previously proposed performs worse than following Bayes rule.
We extended the discussion of Bayes rule to the case of continuous-valued feature vectors, including a discussion of the expected loss (called "risk") when following Bayes rule.
Previous: Lecture 4 Next: Lecture 6