Classifiers
Classifiers
categories of information.
A simple practical example are spam filters
that scan incoming “raw” emails and classify them as either “spam” or “not-
spam.”
Classifiers are a concrete implementation of pattern recognition in many
forms of machine learning.
Naive Bayes classifiers use probability to predict whether an input will fit into a
certain category. The Naive Bayes algorithm family includes a range of different
classifiers based on a theorem of probability. These classifiers can determine the
probability of an input fitting into one or more categories.
In multiple category scenarios, the algorithm reviews the probability that a data point
fits into each classification. After comparing the probability of a match in each
category, it outputs the category that is most likely to match the given text.
Many companies use this type of algorithm to assign tags to text segments like email
subject lines, customer comments and articles.
2. Decision tree
Since the primary goal of a decision tree is to make increasingly specific distinctions,
it has to continuously learn new classification rules. It learns these rules by applying
if-then logic to training data. The algorithm continues the classification process until it
reaches a designated stopping condition.
Their mechanism of action mimics how human brains work, and includes a collection
of artificial neurons that transmit signals.
This makes artificial neural networks capable of solving extremely complex problems
that involve multiple layers. Because of their complexity, it can be challenging to train
and adjust ANNs, and it often requires large amounts of training data. However, a
fully trained ANN can perform tasks that would be impossible for single algorithms.
They work by finding hyperplanes within a data distribution, which you can
visualize as a line separating two different classes of data.
There are often many hyperplanes capable of separating the data, and the
algorithm will select the optimum line of separation. In the SVM model, the
optimum hyperplane is the dividing line that offers the greatest margin
between the different classes.
5. K-nearest neighbor
While the training period for these algorithms is often shorter than for
"eager learners," they're often slower to make predictions.
After storing its training data, a KNN algorithm compares it with test data
and measures the degree of similarity between them. It then stores all
instances that correspond with the training data. Next, the algorithm
attempts to predict the likelihood that future data will correspond to the
dataset it compiled. While this algorithm is common in classification, many
professionals also use it to complete regression tasks.