Credit Card Score Prediction Using Machine Learning
Credit Card Score Prediction Using Machine Learning
ISSN No:-2456-2165
DAT RUL
COMPUT
A ES
OUTP ER
The life of a machine learning programme, summed up
UT as follows
Machine Define the problem,
How does Machine learningLearning
work? Data gathering,
The brain in which all learning takes place is machine Data display,
learning. Like humans, machines learn. People learn from Algorithm training,
their own experience. The more we know, the easier it is to Algorithm testing,
predict. For example, when we face an unknown situation. Feedback collection,
The success rate is lower than the known situation. The way Algorithm optimization.
of machine learning is the same. In order to make an
accurate prediction, the machine will consider an example.
Cycle 4-7 to a satisfactory outcome.
If we give the machine a similar example, it can calculate Use the predictive model.
the result. However, just like humans, it is difficult for
machines to make predictions when you provide examples Once the algorithm is successful, the learned
that have never been seen before. Learning and reasoning is knowledge is applied to a new data set.
the main objective of machine learning. The machine first
learns by recognising patterns. Data are attributed to this Machine learning Algorithms and where they are used?
discovery. As a data scientist, selecting carefully the data
you wish to supply to the computer is the most important
thing. List of problem-solving attributes. This is known as a
vector of function. Feature vectors can be considered a
subset of data for solving the problem. In order to simplify
reality and translate it into a model, the machine uses
complex algorithms. The training phase is therefore used to
describe and consolidate the data into the model.
Classification
Assume you want to anticipate a customer's gender for
a salesperson. They'll take your client database's height,
weight, employment, salary, shopping cart, and other
information. Each customer's gender is known to you. It has
to be either male or female. The classifier's goal is to assess
whether a person is more likely to be male or female (i.e.
tags) based on the information provided (i.e the features you
collect).
Machine learning may be separated into two types of There might be two or more classes in a label. In the
learning objectives: supervised and unsupervised. There are case above, however, there are just two classes. When the
other different algorithms available. classifier wants to forecast an item, though, it has hundreds
of options (such as glass, table, shoes, etc.). A class is an
Supervised Learning. object).
The algorithm finds the association between a certain
input and a specified output using feedback and training Regression
data. Professionals, for example, can deduct marketing The issue is regression when the output is a continuous
costs. Forecasting sales is done with the use of weather value. Financial analysts, for example, may be required to
predictions. You can utilise supervised learning when the forecast the value of stocks depending on a variety of factors
outcome is known. New data can be predicted by the (such as stocks, past stock returns, and macroeconomic
programme. indexes). Errors are possible.
There are two types of supervised learning:
Ø Classification problem
Linear Enables each feature to be correlated to the result to predict future values. Regression
regression
Logistic Linear regression extension used for classification tasks. The output variable 3 is binary Classification
regression instead of continuous (e.g. just black and white) (e.g., an infinite list of potential colors)
Decision tree Classification or regression model that separates data-feature values into branches at Regression
decision nodes (e.g., if a feature is a colour, each potential colour becomes a new branch) Classification
until a final decision output is determined.
Naive Bayes Classification or regression model that separates data-feature values into branches at Regression
decision nodes (e.g., if a feature is a colour, each potential colour becomes a new branch) Classification
until a final decision output is determined.
Support For classification, the Support Vector Machine, or SVM, is commonly employed. The Regression (not
vector SVM method finds a hyperplane that divides the classes in the most efficient way. It very common)
Random The technique is based on a decision tree, which greatly improves accuracy. Random Regression
forest forest builds numerous basic decision trees and choose one label to return using Classification
the'majority vote' approach. The final forecast for the classification job will be the one
with the most votes, while the final prediction for the regression job will be the average
forecast of all the trees.
AdaBoost A classification or regression strategy that assesses a variety of models depending on their Regression
accuracy in predicting the outcome before making a choice. Classification
Gradient- Gradient-boosting trees are a cutting-edge classification and regression method. It focuses Regression
boosting on the preceding trees' mistakes and attempts to fix them. Classification
trees
Unsupervised learning
The algorithm checks input in an uncontrolled learning without explicit variables of output (for example, checking customer
demographics to identify patterns). You can use this to classify the data or to identify patterns for your algorithm and to classify
the information.
K-means Puts data into groups (k), each of which contains data with a comparable set of Clustering
clustering attributes (as determined by the model, not in advance by humans)
Gaussian mixture A k-means clustering extension that allows for more flexibility in the size and structure Clustering
model of group clusters.
Hierarchical Forms a categorization system by splitting groups along a hierarchical tree. Clustering
clustering Can be utilised for customers that have a Cluster loyalty card.
PCA/T-SNE Usually used to reduce the data's dimensionality. The techniques decrease the amount Dimension
of features to 3 or 4 greatest variance vectors. Reduction
Data On a small/medium dataset, excellent results were Excellent performance with large amount of data.
Dependencies achieved.
Feature The characteristics that represent the data must be There's no need to figure out which feature best
engineering understood. describes the data.
Execution time From few minutes to hours Several weeks are possible. A large number of
weights must be computed using a neural
network.
Compared with deep learning, machine learning requires less data to train algorithms. Deep learning requires a large number
of diverse data sets to define the infrastructure. In addition, machine learning also provides faster model training. Exercise from a
few days to a week. The advantage of deep learning over machine learning is that it is very accurate. It is not necessary to know
which characteristics the data best represents. Neural networks have learned to highlight key characteristics. With machine
learning, you need to choose the features to include in the model.
REFERENCES