0% found this document useful (0 votes)
20 views2 pages

Random Forest

Random Forest is a supervised machine learning algorithm used for classification and regression, leveraging ensemble learning by combining multiple decision trees to enhance predictive accuracy. It builds unique trees from random data samples and feature subsets, with final predictions determined by majority vote for classification or averaging for regression. While it offers high accuracy and handles missing data well, it can be computationally expensive and result in large model sizes.

Uploaded by

userstudent7758
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as DOCX, PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
20 views2 pages

Random Forest

Random Forest is a supervised machine learning algorithm used for classification and regression, leveraging ensemble learning by combining multiple decision trees to enhance predictive accuracy. It builds unique trees from random data samples and feature subsets, with final predictions determined by majority vote for classification or averaging for regression. While it offers high accuracy and handles missing data well, it can be computationally expensive and result in large model sizes.

Uploaded by

userstudent7758
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as DOCX, PDF, TXT or read online on Scribd
You are on page 1/ 2

Random Forest:

 Random Forest is a popular machine learning algorithm that belongs to the supervised
learning technique. It can be used for both Classification and Regression problems in
ML.
 It is based on the concept of ensemble learning, which is a process of combining
multiple classifiers to solve a complex problem and to improve the performance of the
model.
 "Random Forest is a classifier that contains a number of decision trees on various
subsets of the given dataset and takes the average to improve the predictive accuracy
of that dataset."
 Instead of relying on one decision tree, the random forest takes the prediction from
each tree and based on the majority votes of predictions, and it predicts the final
output.
 The greater number of trees in the forest leads to higher accuracy and prevents
the problem of overfitting.

How Random Forest Algorithm Works?


The random Forest algorithm works in several steps:
 Random Forest builds multiple decision trees using random samples of the data. Each
tree is trained on a different subset of the data which makes each tree unique.
 When creating each tree the algorithm randomly selects a subset of features or
variables to split the data rather than using all available features at a time. This adds
diversity to the trees.
 Each decision tree in the forest makes a prediction based on the data it was trained on.
When making final prediction random forest combines the results from all the trees.
o For classification tasks the final prediction is decided by a majority vote. This
means that the category predicted by most trees is the final prediction.
o For regression tasks the final prediction is the average of the predictions from
all the trees.
 The randomness in data samples and feature selection helps to prevent the model from
overfitting making the predictions more accurate and reliable.

Advantages of Random Forest


 Random Forest provides very accurate predictions even with large datasets.
 Random Forest can handle missing data well without compromising with accuracy.
 It doesn’t require normalization or standardization on dataset.
 When we combine multiple decision trees it reduces the risk of overfitting of the
model.
Limitations of Random Forest
 It can be computationally expensive especially with a large number of trees.
 Sizes of the models created by random forest may be very large.It may take hundreds
of megabytes of memory and may be slow to evaluate

You might also like