Open navigation menu
Close suggestions
Search
Search
en
Change Language
Upload
Sign in
Sign in
Download free for days
0 ratings
0% found this document useful (0 votes)
75 views
10 pages
Midterm 2002
Machine Learning Notes
Uploaded by
Muhammad Murtaza
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content,
claim it here
.
Available Formats
Download as PDF or read online on Scribd
Download
Save
Save Midterm 2002 For Later
Share
0%
0% found this document useful, undefined
0%
, undefined
Print
Embed
Report
0 ratings
0% found this document useful (0 votes)
75 views
10 pages
Midterm 2002
Machine Learning Notes
Uploaded by
Muhammad Murtaza
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content,
claim it here
.
Available Formats
Download as PDF or read online on Scribd
Carousel Previous
Carousel Next
Download
Save
Save Midterm 2002 For Later
Share
0%
0% found this document useful, undefined
0%
, undefined
Print
Embed
Report
Download
Save Midterm 2002 For Later
You are on page 1
/ 10
Search
Fullscreen
15-781 Midterm, Fall 2002 YOUR ANDREW USERID IN CAPITAL LETTERS: © There are 5 questions. © Questions 1-5 are worth 20 points each. ¢ The maximum possible total score is 100, © Unless otherwise stated there is no need to show your working.1 Decision Trees (20 points) Master Yoda is concerned about the number of Jedi apprentices that have turned to the Dark Side, so he’s decided to train a decision tree on some historical data to help identify problem cases in the future. ‘I'he following table summarizes whether or not each of 12 initiates turned to the Dark Side based on their age when their Jedi training began, whether or not they completed their training, their general disposition, and their species. Dark Side | Age Started Training | Completed Training | Disposition | _Sp 0 3 1 Happy Human 0 9 I Happy Gungan 0 6 0 Happy Wookie 0 6 T Sad | Mon Calamari 0 7 0 Sad Human 0 8 T ‘Angry Tuman 0 5 I Angry Ewok I 9 0 Happy Ewok 1 & a Sad Hinman 1 8 0 Sad Human I 6 0 Angry Wookie 1 7 0 Angry Mon Calamari (3 points) What is the initial entropy of Dark Sid (8 points) Which attribute would the decisi for the root of the tree? on-tree building algorithm choose to use (8 points) What is the information gain of the attribute you chose to split on in the previous question? (8 points) Draw the full decision tree that would be learned for this data (with no pruning). (Hint: The tree will have no more than three splits. The correct split at each point should be clear from just the groups it splits the data into, without having to actually compute the information gain for each possible split.)‘e) f) (2 points) Consider the possibility that the input data above is noisy and not completely accurate, so that the decision tree you learned may not accurately reflect the function you want to learn. If you were to evaluate the three initiates represented by the data points below, on which one would you be most confident of your prediction, and why’? Name_| Age Started Training | Completed Training | Disposition | __ Species Ardath 5 0 Angry Human Barbar 8 0 Angry Gungan Caldar 8 0 Happy | Mon Calamari (8 points) Assume we train a decision tree to predict Z from A, B, and C using the following data (with no pruning): Z[A[BIC o[oo{o ofofofi ofofoti ofoli{o ofolifa Tpo}i}t oft foto 1j1jo}. 1jijiijo 1jijijo o;ifiti dyajiie What would be the training set error for this dataset? Express your answer as the number of records out of 12 that would be misclassified. (3 points) Consider a dec mn tree built from an arbitrary set of data. If the output is diserete-valued and can take on k different possible values, what is the maximum training set error (expressed as a fraction) that any data set could possibly have?2 Probability and Bayes Classifiers (20 points) Lhis figure illustrates a simple class of probability density functions over pairs of real-valued variables. We call it the Rectangle PDF. re | pean. | (sy) © Rect, tes iss) ee | means Par P(x,y) i Ifo Se Saye and Yo Sy Syn} | seer: = 0 otherwise Y fase) oe (2 points) Assuming (x,y) ~ Rect(0,0,0.5,2) (as shown in | (a) the diagram to the right), compute the value of the density _y=1.0 | r= i= 4) | 0.5 yo xel/2 xel xslt (b) (8 points) Under the same assumptions, compute the density p(y = (c) (8 points) Under the same assumptions, compute the density p(x = +) (a) (3 points) Under the same assumptions, compute the density pl = "y=‘e) ) (s) Maximum Likelihood Estimation of Rectangles drawn Assume we have R datapoints (2,41), (2, yo)-(tp. ye) where each datapoint i independently from Rect( to, thos.This Yui) Suppose we want to find the MLE parameters (jo, to, tnisYai) that maximize the likelihood of the datapoints. It turns out (no proof given or required) that these MLE values define the bounding box of the datapoints: ming mingys = marry margyy Now, suppose that we use the rectangle distribution as the density estimator for each class of a Bayes Classifier that we're about to learn. ‘Ihe data is 6 5 c =p [cas] A 1j4] A 4 e 3}1| A 2/2, B 3 2 EN] B 3(2[ Cc fa . of c 405l © A 5]2[ c yu . [5] C 0 0 12 3 4 5 6 x—> Assuming we use the Rectangle Bayes Classifier learned from the data, what value will the classifier give for: (3 points) P(Class = Ax = 1.5, (8 points) P(Class = Alr = 2.5, y (8 points) P(Class = Aly = 5;3 Cross Validation (20 points) Suppose we are learning a classifier with binary output values real-valued input X. Here is our data: and Y=1. ‘There is one [v 0 2/0 1 oe ew 3 [0 4 6a | 711 yo| s+ eee . St PO rt eran Ko} o 12 3 45 6 7 8 9 10 9 {1 x—> 10 [1 Assume we will learn a decision tree on this data. Assume that when the decision tree splits on the real valued attribute x, it puts the split threshold halfway between the attributes that surround the split. For example, using information gain as the splitting criterion, the decision tree would initially choose to split at x = 5, which is halfway between the x = 4 and x = 6 datapoints Let Algorithm D12 be the method of learning a decision tree with only two leaf nodes (ie. only one split). Let Algorithm DT* be the method of learning a decision tree fully with no pruning. (a) (5 points) What will be the training set error of DT2 on onr data? In this part, and all future parts, you can express your answer as the number of misclassifications out of 10. (b) (5 points) What will be the leave-one-out-cro alidation error of DT2 on our data? {c) (5 points) What will be the training set error of DT* on our data? (a) (5 points) What will be the leave-one-ont-cross-validation error of DT* on our data?4 Computational learning theory (20 points) ‘True or false: For a-d, if false, give a counter example. If true, give a 1 sentence justification. (a) (3 points) Within the setting of the PAC model it is impossible to assure with proba- bility 1 that the concept will be learned perfectly (ie., with true error=0), regardless of how many training examples are provided. (3 points) If the Halving Algorithm has made exactly |logo|#|| mis tains the target concept, then it must have learned a hypothesis with true error: regardless of what training sequence we presented and what hypothesis space H it considered (3 points) It is impossible for the Halving Algorithm to lean any concept without it making at least VC(H) mistakes, regardless of what training sequence we present, and what hypothesis space H it considers. (3 points) The PAC bounds make a worst case assumption about the probability di tribution over the instances X, but it is possible to learn from fewer examples for some distributions over } a‘onsider the class of concepts H2p defined by conjunctions of wwo arbitrary perceptrons. ore preciseiy, each nypoti esis h(x) : X + {0,1} in H2p is of the form h(x) = pi(z) AND po(2), where pi(2x) and p2(x) are any two-input pe coptrons. The fignre illnstrates one sich possi two dimensions Je clasefier in {e) (4 points) Draw a set of three points in the plane that cannot be shattered by H2p. ‘f) (4 points) What is the VC dimension of H2p? (Partial credit will be given if you can bound it, so show your reasoning!)Regression and neural networks (20 points) (8 points) Derive a gradient descent training algorithm that minimizes the sum of squared errors for a variant of a perceptron where the output o of the unit depends on its inputs 2; as follows: = Wy + WT + Wat + Woe. + WT +. + Wyo, + WyTh Give your answer in the form w; + w; +... for 1
You might also like
SMAI Question Papers
PDF
No ratings yet
SMAI Question Papers
13 pages
University of Pennsylvania CIS 520: Machine Learning Midterm, 2016
PDF
No ratings yet
University of Pennsylvania CIS 520: Machine Learning Midterm, 2016
18 pages
Midterm 2006
PDF
No ratings yet
Midterm 2006
11 pages
ML End Sem Nov2024 Paper
PDF
No ratings yet
ML End Sem Nov2024 Paper
4 pages
Final2019 Solutions
PDF
No ratings yet
Final2019 Solutions
23 pages
07au Midterm
PDF
No ratings yet
07au Midterm
17 pages
Midterm2022 Solutions
PDF
No ratings yet
Midterm2022 Solutions
12 pages
MachineLearning MidTerm UMT Spring 2021
PDF
100% (1)
MachineLearning MidTerm UMT Spring 2021
12 pages
cs675 SS2022 Midterm Solution PDF
PDF
No ratings yet
cs675 SS2022 Midterm Solution PDF
10 pages
15-381 Spring 2007 Assignment 6: Learning
PDF
No ratings yet
15-381 Spring 2007 Assignment 6: Learning
14 pages
t4 Sol
PDF
No ratings yet
t4 Sol
8 pages
1000099853
PDF
No ratings yet
1000099853
2 pages
Midterm Solutions Machine
PDF
100% (1)
Midterm Solutions Machine
17 pages
Midterm Solutions PDF
PDF
No ratings yet
Midterm Solutions PDF
17 pages
MidA F21
PDF
No ratings yet
MidA F21
8 pages
Solutions: 10-601 Machine Learning, Midterm Exam: Spring 2008 Solutions
PDF
No ratings yet
Solutions: 10-601 Machine Learning, Midterm Exam: Spring 2008 Solutions
8 pages
Midterm Solutions
PDF
No ratings yet
Midterm Solutions
8 pages
Final 2019
PDF
No ratings yet
Final 2019
15 pages
MLvsMAP Merged
PDF
No ratings yet
MLvsMAP Merged
208 pages
Midterm
PDF
No ratings yet
Midterm
12 pages
CS725 2020 Midsem
PDF
No ratings yet
CS725 2020 Midsem
3 pages
EE2211 Past Paper
PDF
No ratings yet
EE2211 Past Paper
14 pages
Statistical Methods For ML
PDF
No ratings yet
Statistical Methods For ML
24 pages
ML Midsem 2018 Solutions
PDF
No ratings yet
ML Midsem 2018 Solutions
7 pages
Midterm Practice Questions
PDF
No ratings yet
Midterm Practice Questions
14 pages
SMAI End 2015 S
PDF
No ratings yet
SMAI End 2015 S
4 pages
Quiz2 Mock Solutions
PDF
No ratings yet
Quiz2 Mock Solutions
19 pages
601 sp09 Midterm Solutions
PDF
No ratings yet
601 sp09 Midterm Solutions
14 pages
Introduction To Machine Learning IIT KGP Week 2
PDF
100% (1)
Introduction To Machine Learning IIT KGP Week 2
14 pages
10-701 Midterm Exam Solutions, Spring 2007
PDF
No ratings yet
10-701 Midterm Exam Solutions, Spring 2007
20 pages
Quiz2 B
PDF
No ratings yet
Quiz2 B
6 pages
Machine Learning Full Question Bank
PDF
No ratings yet
Machine Learning Full Question Bank
14 pages
CS725 2021 Midsem
PDF
No ratings yet
CS725 2021 Midsem
6 pages
Exam Spring 10
PDF
No ratings yet
Exam Spring 10
10 pages
EE2211 Past Paper Ans
PDF
No ratings yet
EE2211 Past Paper Ans
19 pages
Solution of Final Exam: 10-701/15-781 Machine Learning: Fall 2004 Dec. 12th 2004
PDF
No ratings yet
Solution of Final Exam: 10-701/15-781 Machine Learning: Fall 2004 Dec. 12th 2004
27 pages
Exam 21
PDF
No ratings yet
Exam 21
17 pages
10-701/15-781 Machine Learning Mid-Term Exam Solution: Your Name
PDF
No ratings yet
10-701/15-781 Machine Learning Mid-Term Exam Solution: Your Name
12 pages
Midterm 2008s Solution
PDF
No ratings yet
Midterm 2008s Solution
12 pages
Midterm Sol
PDF
No ratings yet
Midterm Sol
16 pages
ML 20230316 1
PDF
No ratings yet
ML 20230316 1
9 pages
2023-24 AIML ML Mid-Semester Make-Up Answer-Keys
PDF
No ratings yet
2023-24 AIML ML Mid-Semester Make-Up Answer-Keys
6 pages
1160 BITS F464 20241220013738 Mid Semester Question Paper
PDF
No ratings yet
1160 BITS F464 20241220013738 Mid Semester Question Paper
4 pages
Midterm2008f Sol
PDF
No ratings yet
Midterm2008f Sol
12 pages
2nd Sessional
PDF
No ratings yet
2nd Sessional
6 pages
Midterm Solution
PDF
No ratings yet
Midterm Solution
6 pages
10-701 Midterm Exam, Fall 2007
PDF
No ratings yet
10-701 Midterm Exam, Fall 2007
25 pages
ML 20240315
PDF
No ratings yet
ML 20240315
8 pages
MedTerm Machine Learning
PDF
No ratings yet
MedTerm Machine Learning
14 pages
1 Analytical Part (3 Percent Grade) : + + + 1 N I: y +1 I 1 N I: y 1 I
PDF
No ratings yet
1 Analytical Part (3 Percent Grade) : + + + 1 N I: y +1 I 1 N I: y 1 I
5 pages
ML Assignment 2 2019 Nptel
PDF
No ratings yet
ML Assignment 2 2019 Nptel
34 pages
Machine Learning Foundations and Applications Assignment 1 Due Date: 10 October, 2021
PDF
No ratings yet
Machine Learning Foundations and Applications Assignment 1 Due Date: 10 October, 2021
3 pages
Midterm Spring13
PDF
No ratings yet
Midterm Spring13
10 pages
2CS501 IR May 2022
PDF
No ratings yet
2CS501 IR May 2022
3 pages
Time: 03 Hours
PDF
No ratings yet
Time: 03 Hours
4 pages
MLFA Spring 2024
PDF
No ratings yet
MLFA Spring 2024
11 pages
MCQ Question
PDF
No ratings yet
MCQ Question
5 pages
Sample Final AI
PDF
No ratings yet
Sample Final AI
9 pages
Midterm 2010 F
PDF
No ratings yet
Midterm 2010 F
15 pages
Midterm Sol
PDF
No ratings yet
Midterm Sol
23 pages
Final 2003
PDF
No ratings yet
Final 2003
18 pages
Homework 4
PDF
No ratings yet
Homework 4
3 pages
Homework 3
PDF
No ratings yet
Homework 3
4 pages
Final 2006
PDF
No ratings yet
Final 2006
15 pages
Homework 2
PDF
No ratings yet
Homework 2
4 pages
03 MLE MAP NBayes-1-21-2015
PDF
No ratings yet
03 MLE MAP NBayes-1-21-2015
40 pages