Line 4: Line 4:
  
  
In the second lecture, we
+
In Lecture 5, we observed that Bayes rule minimizes the probability of error. One student pointed out an example that raises doubts about the validity of this statement. This gave rise to an interesting [[ECE662_topic2_discussions|online discussion]]. In particular, a student [[EE662Sp10OptimalPrediction|proved that the example previously proposed performs worse]] than following Bayes rule.
  
 +
We extended the discussion of Bayes rule to the case of continuous-valued feature vectors, including a discussion of the expected loss (called "risk") when following Bayes rule.
  
 
Previous: [[Lecture4ECE662S10|Lecture 4]]
 
Previous: [[Lecture4ECE662S10|Lecture 4]]
 
 
Next: [[Lecture6ECE662S10|Lecture 6]]
 
Next: [[Lecture6ECE662S10|Lecture 6]]
 
----
 
----

Revision as of 07:17, 12 April 2010


Details of Lecture 5, ECE662 Spring 2010

In Lecture 5, we observed that Bayes rule minimizes the probability of error. One student pointed out an example that raises doubts about the validity of this statement. This gave rise to an interesting online discussion. In particular, a student proved that the example previously proposed performs worse than following Bayes rule.

We extended the discussion of Bayes rule to the case of continuous-valued feature vectors, including a discussion of the expected loss (called "risk") when following Bayes rule.

Previous: Lecture 4 Next: Lecture 6


Back to course outline

Back to 2010 Spring ECE 662 mboutin

Back to ECE662

Alumni Liaison

Sees the importance of signal filtering in medical imaging

Dhruv Lamba, BSEE2010