0% found this document useful (0 votes)
38 views2 pages

CS60050 Machine Learning EA 2015

Uploaded by

KingOfKings 007
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF or read online on Scribd
0% found this document useful (0 votes)
38 views2 pages

CS60050 Machine Learning EA 2015

Uploaded by

KingOfKings 007
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF or read online on Scribd
You are on page 1/ 2
Department of Computer Science & Engineering Indian Institute of Technology, Kharagpur End-semester Examination, Autumn 2015 $6000: Machine Learning Full Marks: 100 ime: 3 Hrs Answer ALL questions. You can use calculators. QI. Which of the following is true or false. Explain with reasons (max 2-3 sentences): a. Rule post pruning in C4.5 can sometimes produce rules which do not correspond to a decision tree [S marks] b. SMO algorithm can be used to solve any quadratic programming problem. [5 marks} ©. In SGD, the objective functicn value decreases monotonically with the iterations. [5 marks] 4d. All conditional independences in encoded in the Bayesian network are preserved in the MRF obtained after moralization, [5 marks] e. The perceptron learning algorithm is a special case SGD. [5 marks} f. The Nesterov and Vial’s proof for convergence of SGD applies to perceptron algorithm as well. [5 marks} 2. Answer the following: a. Write a stochastic gradient descent algorithm, for optimizing a function L(@) = May Eker Un Yn, 9). [5 marks] 'b. Define the expected loss function, clearly mentioning the random variables. Write the formula for convergence in expectation, clearly defining each variable. {5 marks] ©. Derive a condition on the step sizes which guarantees convergence in expectation. [5 marks] 4. Consider a situation where a biased coin with probability of head (p) is tossed N times and the outcomes (head = 1,tail = 0) are recorded in random variables, xp, = 1,...,N. Derive the log likelihood function for estimating parameter, p. Derive an SGD algorithm for solving the Max-likelihood problem derived above. [10 marks} Q3. Consider the following dataset for predicting outcome of a tennis match between Federer and Nadale: Time Mateh type | Court Best Effort | Outcome surface Moming Master Grass. 1 F ‘Afternoon Grand slam | Cla 1 F Night Friendly Hard 0 F ‘Afternoon Friendly Mixed 0 N ‘Afternoon Master Clay a N ‘Afternoon Grand slam | Grass i F ‘Afternoon ‘Grand slam | Hard i F ‘Afternoon Grand slam | Hard 1 F Morning. Master Grass 7 F Afternoon Grand slam | Clay 7 N Night Friendly Hard 0 F Night Master Mixed 1 N ‘Afternoon Master Clay 1 N ‘Afternoon Master Grass 1 F ‘Afternoon Grand slam | Hard 1 F Afternoon Grand slam [ Cla i F Write the formula for information gain for an attribute. Compute information gain for all attributes. Which is selected as a root node? [10 marks} Compute the structure of the decision tree computed by ID3 algorithm. [5 marks} Q4. Show 5 updates of perceptron algorithm for the following data, starting at w = (0,0) Qs. (6-4), HED), B,3) ELS OMS.29, -2)} [10 marks] Define d-separation in Bayesian Networks stating all the conditions. How does d-separation help in determining conditional independence? [5 marks} Consider the following Bayesian network connection various factors related to chest diseases: [10 marks} Broncrs? roves? Postve ore | om i, Are Bronchitis and Tuberculosis independent when nothing is observed? ii, Are they independent after observing x-ray? iii, Are they independent after observing both x-ray and smoking? Q6. Consider the following Markov Random Field with the corresponding potentials: OQ OOO (Xi, Xess) = 1 — XiXi4s for odd é, ard WeigaXi,Xisa) = 1 + XiXiva for even é; where X; € {+1,—1). Find P(X,) and x3 = maxP(X3]X2 = 1) using an appropriate message-passing algorithm, [20 marks}

You might also like