(New page: Hw assignment 2 Due Tuesday April 1, 2008 Guidelines: Write a short report to present your results. Be sure to include all the relevant graphs as well as a copy of your code. Teamwork ...) |
|||
Line 1: | Line 1: | ||
− | Hw assignment 2 | + | == Hw assignment 2 == |
+ | === Assignment Description === | ||
Due Tuesday April 1, 2008 | Due Tuesday April 1, 2008 | ||
Line 38: | Line 39: | ||
d) Compare the three approaches. | d) Compare the three approaches. | ||
− | + | === Data === | |
− | + | Publicly available data sources are listed [[Data_Sources_OldKiwi|here]]. | |
− | [[ | + |
Revision as of 10:58, 5 March 2008
Hw assignment 2
Assignment Description
Due Tuesday April 1, 2008
Guidelines:
Write a short report to present your results. Be sure to include all the relevant graphs as well as a copy of your code. Teamwork is encouraged, but the write up of your report must be your own. Please write the names of ALL your collaborators on the cover page of your report.
Question 1: In the Parametric Method section of the course, we learned how to draw a separation hyperplane between two classes by obtaining w0, the argmax of the cost function J(w)=wTSBw / wTSww. The solution was found to be w0= Sw-1(m1-m2), where m1 and m2 are the sample means of each class, respectively.
Some students raised the question: can one simply use J(w)= wTSBw instead (i.e. setting Sw as the identity matrix in the solution wo? Investigate this question by numerical experimentation.
Question 2: Obtain a set of training data. Divide the training data into two sets. Use the first set as training data and the second set as test data.
a) Experiment with designing a classifier using the neural network approach.
b) Experiment with designing a classifier using the support vector machine approach.
c) Compare the two approaches.
Note: you may use code downloaded from the web, but if you do so, please be sure to explain what the code does in your report and give the reference.
Question 3: Using the same data as for question 2 (perhaps projected to one or two dimensions for better visualization),
a) Design a classifier using the Parzen window technique.
b) Design a classifier using the K-nearest neighbor technique
c) Design a classifier using the nearest neighbor technique.
d) Compare the three approaches.
Data
Publicly available data sources are listed here.