(5 intermediate revisions by 3 users not shown)
Line 1: Line 1:
[[Category:Classification using Bayes Rule in 1-dimensional and N-dimensional feature spaces Review]][[Category:Classification using Bayes Rule in 1-dimensional and N-dimensional feature spaces Review]][[Category:Classification using Bayes Rule in 1-dimensional and N-dimensional feature spaces Review]][[Category:Classification using Bayes Rule in 1-dimensional and N-dimensional feature spaces Review]][[Category:Classification using Bayes Rule in 1-dimensional and N-dimensional feature spaces Review]]
+
<br>
 +
<center><font size="4"></font>
 +
<font size="4">Review on KNN to [[Slecture_from_KNN_to_nearest_neighbor|Nearest Neighbor Slecture by Jonathan Manring]] </font>
  
=KNN_to_nearest_neighbor_slecture_review_comment=
+
A [https://www.projectrhea.org/learning/slectures.php slecture] by graduate student Jonathan Manring
 +
</center>
 +
----
 +
----
  
 +
----
  
 +
== Comments/Feedback ==
  
Put your content here . . .
+
This slecture reviewed by Ben Foster
  
 +
This slecture provides a brief overview of the KNN classification method and transitions from KNN into a brief description of the nearest neighbor classification method. A few comments/suggestions:
  
 +
*It is alright that there is not an abundance of theoretical background for KNN presented in this slecture, but it does seem like there could be a bit more. This is especially true since the idea of the slecture is to set the stage for the nearest neighbor method with KNN as a support. With this in mind, the author could use to make a more structured transition from KNN to the nearest neighbor method.
  
 +
*There could also be some more discussion of the unbiased nature of KNN as an estimator for the local density <math>\rho(\vec{X_0})</math>. Why is this a point of interest? Are other local estimators of <math>\rho(\vec{X_0})</math> biased? Those that have taken the class are likely familiar with the answers to these questions, but this may be a point of confusion for a reader that is just becoming familiar with the material.
  
[[ Classification using Bayes Rule in 1-dimensional and N-dimensional feature spaces Review|Back to Classification using Bayes Rule in 1-dimensional and N-dimensional feature spaces Review]]
+
*The discussion of distance metrics is similarly sparse. Perhaps the author could add links/references to sites where more information on these topics can be found.
 +
 
 +
----
 +
 
 +
[[Slecture_from_KNN_to_nearest_neighbor|Back to KNN to Nearest Neighbor Slecture]]
 +
 
 +
[[Category:Slecture]] [[Category:ECE662Spring2014Boutin]] [[Category:ECE]] [[Category:ECE662]] [[Category:Pattern_recognition]]

Latest revision as of 10:20, 7 May 2014


Review on KNN to Nearest Neighbor Slecture by Jonathan Manring

A slecture by graduate student Jonathan Manring




Comments/Feedback

This slecture reviewed by Ben Foster

This slecture provides a brief overview of the KNN classification method and transitions from KNN into a brief description of the nearest neighbor classification method. A few comments/suggestions:

  • It is alright that there is not an abundance of theoretical background for KNN presented in this slecture, but it does seem like there could be a bit more. This is especially true since the idea of the slecture is to set the stage for the nearest neighbor method with KNN as a support. With this in mind, the author could use to make a more structured transition from KNN to the nearest neighbor method.
  • There could also be some more discussion of the unbiased nature of KNN as an estimator for the local density $ \rho(\vec{X_0}) $. Why is this a point of interest? Are other local estimators of $ \rho(\vec{X_0}) $ biased? Those that have taken the class are likely familiar with the answers to these questions, but this may be a point of confusion for a reader that is just becoming familiar with the material.
  • The discussion of distance metrics is similarly sparse. Perhaps the author could add links/references to sites where more information on these topics can be found.

Back to KNN to Nearest Neighbor Slecture

Alumni Liaison

has a message for current ECE438 students.

Sean Hu, ECE PhD 2009