Machine Learning Techniques Assignment-7: Name:Ishaan Kapoor Rollno:1/15/Fet/Bcs/1/055
Machine Learning Techniques Assignment-7: Name:Ishaan Kapoor Rollno:1/15/Fet/Bcs/1/055
Ques 1:
(a) Write the properties of Expectation?
Ans : In probability theory, the expected value of a random variable, intuitively, is the
long-run average value of repetitions of the same experiment it represents. ... More
practically, the expected value of a discrete random variable is the probability-weighted
average of all possible values.
(b) Differentiate between Supervised and Unsupervised
learning? Ans :
SUPERVISED
UNSUPERVISED LEARNING
LEARNING
Reason 2: The law of averages states that the larger the set of predictors, the higher the
probability of having missing values in the data. If we chose to delete cases which have
missing values for some predictors, we may end up with a shortage of samples.
Example: A practical rule of thumb used by data miners is to have atleast 5(p+2) samples
where p is the number of predictors. If your data set is sufficiently large and this rule is
easily satisfied, then you may not be risking much by deleting cases. But if your data is
from an expensive market survey for example, a systematic procedure to actually reduce
the data set, may result in a situation where you dont have to address this problem of losing
samples. It is better to lose variables which dont impact your prediction than to lose
somewhat more expensive samples.
There are several other more technical reasons for reducing data dimensionality which
will be explored in subsequent articles. In a next article, we will discuss some common
techniques for actually implementing this process.
Ans : Factor analysis is a statistical method used to describe variability among observed,
correlated variables in terms of a potentially lower number of unobserved variables
called factors. For example, it is possible that variations in six observed variables mainly
reflect the variations in two unobserved variables.
Ensemble learning methods are included in different types of statistical software packages.
Some experts describe ensemble learning as “crowdsourcing” of data aggregation.
Ques 4: Describe how principle component analysis is carried out to
reduce dimensionality of data sets .
Ans : The main idea of principal component analysis (PCA) is to reduce the dimensionality of
a data set consisting of many variables correlated with each other, either heavily or lightly,
while retaining the variation present in the dataset, up to the maximum extent. The same is done
by transforming the variables to a new set of variables, which are known as the principal
components (or simply, the PCs) and are orthogonal, ordered such that the retention of variation
present in the original variables decreases as we move down in the order. So, in this way, the
1st principal component retains maximum variation that was present in the original
components. The principal components are the eigenvectors of a covariance matrix, and hence
they are orthogonal.
Importantly, the dataset on which PCA technique is to be used must be scaled. The results are
also sensitive to the relative scaling. As a layman, it is a method of summarizing data. Imagine
some wine bottles on a dining table. Each wine is described by its attributes like colour,
strength, age, etc. But redundancy will arise because many of them will measure related
properties. So what PCA will do in this case is summarize each wine in the stock with less
characteristics.
Intuitively, Principal Component Analysis can supply the user with a lower-dimensional
picture, a projection or "shadow" of this object when viewed from its most informative
viewpoint.