0% found this document useful (0 votes)
30 views4 pages

Day43 KNN Intro

K-nearest neighbors (KNN) is a supervised machine learning algorithm that can be used for classification or regression. It makes predictions based on the K closest training examples in the feature space. To classify a new data point, KNN finds the K closest neighbors and assigns the data point to the most common class among those neighbors. Choosing an optimal value for K can affect the performance of the KNN algorithm.

Uploaded by

Igor Fernandes
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
30 views4 pages

Day43 KNN Intro

K-nearest neighbors (KNN) is a supervised machine learning algorithm that can be used for classification or regression. It makes predictions based on the K closest training examples in the feature space. To classify a new data point, KNN finds the K closest neighbors and assigns the data point to the most common class among those neighbors. Choosing an optimal value for K can affect the performance of the KNN algorithm.

Uploaded by

Igor Fernandes
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 4

K-Nearest Neighbor(KNN) By: Loga Aswin

➢ K-nearest neighbors (kNN) is a supervised machine


learning technique that may be used to handle both classification
and regression tasks.
➢ It is an algorithm that is used to classify a data point based on how
its neighbors are classified.
➢ The “K” value refers to the number of nearest neighbor data points
➢ It’s used in many different areas, such as handwriting detection,
image recognition, and video recognition.
➢ K-NN algorithm stores all the available data and classifies a new
data point based on the similarity.

➢ K-NN is a non-parametric algorithm, which means it does not


make any assumption on underlying data.
➢ It is also called a lazy learner algorithm because it does not learn
from the training set immediately instead it stores the dataset and at
the time of classification, it performs an action on the dataset.

How does KNN Work?


Step 1 - Assign a value to K.
Step 2 - Calculate the distance between the new data entry and all other
existing data entries (you'll learn how to do this shortly). Arrange them
in ascending order.
Step 3 - Find the K nearest neighbors to the new entry based on the
calculated distances.
Step 4 - Assign the new data entry to the majority class in the nearest
neighbors.

i. data set consisting of two classes — red and blue.

ii. A new data has been introduced (Green point)


iii. Let's assume the value of K is 3.

iv. Out of the 3 nearest neighbors, the majority class is red so the
new entry will be assigned to that class by using Euclidean
distance
Advantages of K-NN Algorithm
• It is simple to implement.
• No training is required before classification.

Disadvantages of K-NN Algorithm


• Can be cost-intensive when working with a large data set.
• A lot of memory is required for processing large data sets.
• Choosing the right value of K can be tricky.

K-Nearest Neighbor (Applications)

➢ RECOMMENDATION ENGINE
➢ CONCEPT SEARCH
➢ PATTERN RECOGNITION
➢ MISSING DATA IMPUTATION

You might also like