(3 intermediate revisions by the same user not shown) | |||
Line 4: | Line 4: | ||
− | In | + | In Lecture 5, we observed that [[Bayes_Decision_Theory|Bayes rule]] minimizes the probability of error. One student pointed out an example that raises doubts about the validity of this statement. This gave rise to an interesting [[ECE662_topic2_discussions|online discussion]]. In particular, a student [[EE662Sp10OptimalPrediction|proved that the example previously proposed performs worse]] than following Bayes rule. |
+ | We extended the discussion of [[Bayes_Decision_Theory|Bayes rule]] to the case of continuous-valued feature vectors, including a discussion of the expected loss (called "risk") when following [[Bayes_Decision_Theory|Bayes rule]]. | ||
+ | Previous: [[Lecture4ECE662S10|Lecture 4]] | ||
Next: [[Lecture6ECE662S10|Lecture 6]] | Next: [[Lecture6ECE662S10|Lecture 6]] | ||
---- | ---- |
Latest revision as of 10:42, 13 April 2010
Details of Lecture 5, ECE662 Spring 2010
In Lecture 5, we observed that Bayes rule minimizes the probability of error. One student pointed out an example that raises doubts about the validity of this statement. This gave rise to an interesting online discussion. In particular, a student proved that the example previously proposed performs worse than following Bayes rule.
We extended the discussion of Bayes rule to the case of continuous-valued feature vectors, including a discussion of the expected loss (called "risk") when following Bayes rule.
Previous: Lecture 4 Next: Lecture 6