Assignment-1 Solution

$35.00 $29.00

Please use Google Classroom to upload your submission by the deadline mentioned above. Your submission should comprise of a single le (PDF/ZIP), named <Your Roll No> Assign1, with all your solutions. For late submissions, 10% is deducted for each day (including weekend) late after an assign-ment is due. Note that each student begins the course…

5/5 – (2 votes)

You’ll get a: zip file solution

 

Description

5/5 – (2 votes)

Please use Google Classroom to upload your submission by the deadline mentioned above. Your submission should comprise of a single le (PDF/ZIP), named <Your Roll No> Assign1, with all your solutions.

For late submissions, 10% is deducted for each day (including weekend) late after an assign-ment is due. Note that each student begins the course with 7 grace days for late submission of assignments. Late submissions will automatically use your grace days balance, if you have any left. You can see your balance on the CS6510 Marks and Grace Days document under the course Google drive (soon to be shared).

You have to use Python for the programming questions.

Please read the department plagiarism policy. Do not engage in any form of cheating – strict penalties will be imposed for both givers and takers. Please talk to instructor or TA if you have concerns.

      • Questions

  1. k-NN: (7 marks) In k-nearest neighbors (k-NN), the classi cation is achieved by majority vote in the vicinity of data. Given n points, imagine two classes of data each of n=2 points, which are overlapped to some extent in a 2-dimensional space.

    1. (1 mark) Describe what happens to the training error (using all available data) when the neighbor size k varies from n to 1.

    1. (2 marks) Predict and explain with a sketch how the generalization error (e.g. holding out some data for testing) would change when k varies? Explain your reasoning.

1

    1. (2 marks) Give two reasons (with sound justi cation) why k-NN may be undesirable when the input dimension is high.

    1. (2 marks) Is it possible to build a univariate decision tree (with decisions at each node of the form is x > a, is x < b, is y > c, or is y < d for any real constants a; b; c; d) which classi es exactly similar to a 1-NN using the Euclidean distance measure ? If so, explain how. If not, explain why not.

  1. Bayes Classi er: (6 marks)

    1. (3 marks) A training set consists of one dimensional examples from two classes. The training examples from class 1 are f0:5; 0:1; 0:2; 0:4; 0:3; 0:2; 0:2; 0:1; 0:35; 0:25g and from class 2 are f0:9; 0:8; 0:75; 1:0g. Fit a (one dimensional) Gaussian using Maximum Likelihood to each of these two classes. You can assume that the variance for class 1 is 0.0149, and the variance for class 2 is 0.0092. Also estimate the class probabilities p1 and p2 using Maximum Likelihood. What is the probability that the test point x = 0:6 belongs to class 1?

    1. (3 marks) You are now going to make a text classi er. To begin with, you attempt to classify documents as either sport or politics. You decide to represent each document as a (row) vector of attributes describing the presence or absence of the following words.

x = (goal,football,golf,defence,o ence,wicket,o ce,strategy)

Training data from sport documents and from politics documents is represented below in a matrix in which each row represents the 8 attributes.

20

0

0

1

0

0

1

13

1

0

1

1

1

0

1

1

xpolitics =

60

1

0

0

1

1

0

17

6

7

1

0

0

1

1

0

1

0

60

0

0

1

1

0

1

17

6

7

60

0

0

1

1

0

0

17

6

7

4

5

20

0

1

0

0

0

0

03

1

1

0

0

0

0

0

0

xsport =

61

1

0

1

0

0

0

17

6

7

1

1

0

1

0

0

0

0

61

1

0

1

1

0

0

07

6

7

60

0

0

1

0

1

0

07

6

7

4

5

Using a maximum likelihood naive Bayes classi er, what is the probability that the document x = (1; 0; 0; 1; 1; 1; 1; 0) is about politics?

  1. Decision Trees: (15 marks) In this question, you will use the Wine dataset1, a popular dataset to evaluate classi cation algorithms. The classi cation task is to determine, based on various parameters, whether a wine quality is over 7. The dataset has already been preprocessed to convert this into a binary classi cation problem (scores less than 7 belong to the \zero” class, and scores greater than or equal to 7 belong to the \one” class). Each line

  • https://archive.ics.uci.edu/ml/datasets/Wine+Quality

2

describes a wine, using 12 columns: the rst 11 describe the wines characteristics (details), and the last column is a ground truth label for the quality of the wine (0/1). You must not use the last column as an input feature when you classify the data.

  1. (5 marks) Decision Tree Implementation: Implement your own version of the decision tree using binary univariate split, entropy and information gain. (If you are using Python, you can use the skeleton code provided.)

  1. (5 marks) Cross-Validation: Evaluate your decision tree using 10-fold cross valida-tion. Please see the lecture slides for details. In a nutshell, you will rst make a split of the provided data into 10 parts. Then hold out 1 part as the test set and use the remaining 9 parts for training. Train your decision tree using the training set and use the trained decision tree to classify entries in the test set. Repeat this process for all 10 parts, so that each entry will be used as the test set exactly once. To get the nal ac-curacy value, take the average of the 10 folds accuracies. With correct implementation of both parts (decision tree and cross validation), your classi cation accuracy should be around 0.78 or higher.

  1. (5 marks) Improvement Strategies: Now, try and improve your decision tree algorithm. Some things you could do include (not exhaustive):

Use Gini index instead of entropy

Use multi-way split (instead of binary split) Use multivariate split (instead of univariate)

Prune the tree after splitting for better generalization

Report your performance as an outcome of the improved strategies.

Deliverables:

Code

Brief report (PDF) with: (i) Accuracy of your initial implementation; (ii) Accuracy of your improved implementation, along with your explanation of why the accuracy improved with this change.

  1. (12 marks) Kaggle – What’s cooking: The next task of this assignment is to work

on a (completed) Kaggle challenge: \what’s cooking?” As part of this task, please visit https://www.kaggle.com/c/whats-cooking to know more about this problem, and download the data. (You may have to create a Kaggle account to download the data, if you don’t have one already.)

In this assignment, you are allowed to use any existing machine learning library of your choice: scikitlearn, pandas, Weka (we recommend scikitlearn) – but you should use only the decision tree, k-NN or the naive Bayes classi er (to align with what we have covered in class so far, random forests not allowed too). Use train.json to train your classi er. Predict the cuisine on the data in test.json, and report your best 2 scores in your report. (Note that Kaggle will not publish the scores of a completed contest on its leaderboard, but will reveal the scores to you – please report them. We will also upload your codes randomly to con rm the scores.)

Deliverables:

Code

3

Brief report (PDF) with top-2 scores of your methods, and a brief description of the methods that resulted in the top 2 scores.

Your report should also include your analysis of why your best 2 methods performed better than others you tried.

4

Assignment-1 Solution
$35.00 $29.00