Risk Minimization
Risk Minimization
• Error - difference between the actual value and the predicted value.
• In supervised learning, we use the dataset that is representative of all the classes
for a given problem, to solve the problem.
• Since we cannot have a data of all the people in the world, we sample out data of
some people to be fed into our model.
• We use a loss function L(h, x) to find the difference between the actual diagnosis
and the predicted diagnosis. This is essentially a measurement of error.
https://www.cs.cornell.edu/courses/cs4780/2018fa/lectures/
lecturenote10.html
True Risk
• True risk is the average loss/error over all
possibilities (here, the population of the whole
world). Its formula is as follows:
Here, the gray area is the condition where the model's prediction was 0 and the red circle
represents the actual region where the sample did not yield results.
Here, the difference between the area of the gray rectangle and the area of the quadrant
gives the true risk.
True risk = 0.5 x 1 - 0.25 x 3.14 x (0.5)2 = 0.30
Empirical Risk Minimization
• While building our machine learning model,
we choose a function that reduces the
differences between the actual and the
predicted output i.e. empirical risk.