SVM Assignment ABA Course To Be Returned With Your Answers
SVM Assignment ABA Course To Be Returned With Your Answers
______________________
Excel Exercise (25 marks): Construct an Excel worksheet to perform Hard Margin
SVM calculations and optimization of the margins. Your answer should fill out all columns that
you will copy from Excel on the table shown below. Indicate the rough positions of the decision
boundaries on the graph shown below. Hint: Use Solver® to optimize w1, w2 and gamma.
Do not submit the Excel Worksheet. Use the table below to fill your answers and return.
Input Data
Max Upper Lower Class
Margin Bounding Bounding found by
x1 x2 d W1 W2 gamma Separator Plane Plane SVM
1 1 -1
2 1 -1
1 2 -1
2 2 -1
4.5 4.5 -1
4 4 1
4 5 1
5 4 1
5 5 1
1.5 1.5 1
1
Multiple Choice Questions (25 marks)
Question Context: 1 – 2
Suppose you are using a Linear SVM classifier with 2 class classification problem. Now
you have been given the following data in which some points are circled red that are
representing support vectors.
1) If you remove the following any one red points from the data. Does the decision
boundary will change?
A) Yes
B) No
Answer:
2) [True or False] If you remove the non-red circled points from the data, the decision
boundary will change?
A) True
B) False
Answer:
2
3) What do you mean by generalization error in terms of the SVM?
Answer:
4) When the C parameter is set to infinite, which of the following holds true?
A) The optimal hyperplane if exists, will be the one that completely separates the da ta
B) The soft-margin classifier will separate the data
C) None of the above
Answer:
Answer:
6) The minimum time complexity for training an SVM is O(n2). According to this fact, what
sizes of datasets are not best suited for SVM’s?
A) Large datasets
B) Small datasets
C) Medium sized datasets
D) Size does not matter
Answer:
3
7) The effectiveness of an SVM depends upon:
A) Selection of Kernel
B) Kernel Parameters
C) Soft Margin Parameter C
D) All of the above
Answer:
8) Support vectors are the data points that lie closest to the decision surface.
A) TRUE
B) FALSE
Answer:
Answer:
10) Suppose you are using RBF kernel in SVM with high Gamma value. What does this
signify?
A) The model would consider even far away points from hyperplane for modeling
B) The model would consider only the points close to the hyperplane for modeling
C) The model would not be affected by distance of points from hyperplane for modeling
D) None of the above
Answer:
4
11) The cost parameter C in the SVM means:
Answer:
12) Suppose you are building a SVM model on data X. The data X can be error prone
which means that you should not trust any specific data point too much. Now think that
you want to build a SVM model which has quadratic kernel function of polynomial degree
2 that uses Slack variable C as one of its hyper parameter. Based upon that give the
answer for following question.
What would happen when you use very large value of C (C->infinity) ?
Note: For small C was also classifying all data points correctly
A) We can still classify data correctly for given setting of hyper parameter C
B) We can not classify data correctly for given setting of hyper parameter C
C) Can’t Say
D) None of these
Answer:
13) What would happen when you use very small C (C~0)?
Answer:
5
14) If I am using all features of my dataset and I achieve 100% accuracy on my training set,
but ~70% on validation set, what should I look out for?
A) Underfitting
B) Nothing, the model is perfect
C) Overfitting
Answer:
15) Which of the following are real world applications of the SVM?
Answer:
Question Context: 16 – 18
Suppose you have trained an SVM with linear decision boundary after training SVM, you
correctly infer that your SVM model is under fitting.
16) Which of the following option would you more likely to consider iterating SVM next
time?
Answer:
6
17) Suppose you gave the correct answer in previous question. What do you think that is
actually happening?
A) 1 and 2
B) 2 and 3
C) 1 and 4
D) 2 and 4
Answer:
18) In above question suppose you want to change one of it’s(SVM) hyperparameter so that
effect would be same as previous questions i.e. model will not under fit?
Answer:
7
19) We usually use feature normalization before using the Gaussian kernel in SVM. What is
true about feature normalization?
A) 1
B) 1 and 2
C) 1 and 3
D) 2 and 3
Answer:
Suppose you are dealing with 4 class classification problem and you want to train a SVM
model on the data for that you are using One-vs-all method. Now answer the below
questions?
20) How many times we need to train our SVM model in such case?
A) 1
B) 2
C) 3
D) 4
Answer:
8
21) Suppose you have same distribution of classes in the data. Now, say for training 1 time
in one vs all setting the SVM is taking 10 second. How many seconds would it require to
train one-vs-all method end to end?
A) 20
B) 40
C) 60
D) 80
Answer:
22) Suppose your problem has changed now. Now, data has only 2 classes. What would you
think how many times we need to train SVM in such case?
A) 1
B) 2
C) 3
D) 4
Answer:
Question context: 23 – 24
Suppose you are using SVM with linear kernel of polynomial degree 2, Now think that you
have applied this on data and found that it perfectly fit the data that means, Training and
testing accuracy is 100%.
23) Now, think that you increase the complexity(or degree of polynomial of this kernel).
What would you think will happen?
9
C) Nothing will happen since your model was already 100% accurate
D) None of these
Answer:
24) In the previous question after increasing the complexity you found that training
accuracy was still 100%. According to you what is the reason behind that?
1. Since data is fixed and we are fitting more polynomial term or parameters so the
algorithm starts memorizing everything in the data
2. Since data is fixed and SVM doesn’t need to search in big hypothesis space
A) 1
B) 2
C) 1 and 2
D) None of these
Answer:
A) 1
B) 2
C) 1 and 2
D) None of these
Answer:
10