Final Report Dinesh
Final Report Dinesh
Final Report Dinesh
BACHELOR OF TECHNOLOGY
IN
COMPUTER SCIENCE AND ENGINEERING
Submitted by
AKULA DINESH SAI
21121A3104
IV B. Tech I Semester
Under the esteemed supervision of
Mr. N. Balakrishna
Assistant Professor
(AUTONOMOUS)
(Affiliated to JNTUA, Anantapuramu and approved by AICTE, New Delhi) Accredited by NAAC with A Grade
Sree Sainath Nagar, Tirupati, Chittoor Dist. -517 102, A.P, INDIA
2024 - 2025.
(AUTONOMOUS)
Sree Sainath Nagar, A Rangampeta
Certificate
This is to certify that the internship report entitled “AI-ML Development Internship” is
the bonafide work done by AKULA DINESH SAI (Roll No:21121A3104) in the Department of
University Anantapur, Anantapuramu in partial fulfillment of the requirements for the award
of the degree of Bachelor of Technology in Computer Science during the academic year
2024-2025.
Head:
This report summarizes my three-month internship in AI-ML Development, with a primary focus on
enhancing ML models for real-world applications in natural language processing (NLP). Although the
internship included some foundational concepts in artificial intelligence (AI), my work was primarily
concentrated on ML techniques and processes. Key responsibilities involved preparing and preprocessing
domain-specific datasets, training, and fine-tuning models, and optimizing hyperparameters to achieve
high levels of accuracy and efficiency. I conducted detailed experiments to build and refine NLP models
for applications such as automated customer support, text classification, and content generation, leveraging
prompt engineering to increase response relevance and coherence.
My experience also included managing the end-to-end ML pipeline, from data ingestion and model
selection to deployment and continuous monitoring in production environments. In addition, I developed
skills in error analysis, performance evaluation, and troubleshooting, ensuring model reliability at scale.
Through this internship, I gained a strong understanding of how machine learning algorithms and
techniques can drive automation, streamline workflows, and provide actionable insights, reinforcing ML's
pivotal role in modern problem-solving and operational improvement.
Keywords: Machine Learning (ML), Natural Language Processing (NLP), Model Training,
Hyperparameter Optimization, Model Deployment, Error Analysis, Automation, AI
i
ACKNOWLEDGEMENT
We are extremely thankful to our beloved Chairman and founder Dr. M. Mohan Babu
who took keen interest to provide us the opportunity for carrying out the project work.
We are very much obliged to Dr. B. Narendra Kumar Rao, Professor & Head,
Department of CSE, for providing us the guidance and encouragement in completion of
this work.
21121A3104
ii
TABLE OF CONTENTS
TITLE PAGE NO
Abstract i
Acknowledgement ii
Table of contents iii
List of Tables iv
List of figures v
Chapter 1: Introduction
1. Quickintroduction to machine learning 01 – 02
2. History of machine learning 02
3. Important terms 02 - 03
Chapter 2: Machine learning classification
1. Basic machine learning classification 04
2. Difference between supervised and unsupervised 04 - 06
Chapter 3: Algorithms in machine learning
1. Regression 07 - 09
2. classification 09 - 13
3. Association 13 - 15
Chapter 4: Advantages and disadvantages
1. Advantages of machine learning 16 - 18
Conclusions 24
References 25
iii
LIST OF TABLES
TITLE PAGE NO
Difference between supervised and unsupervised 05 - 06
iv
LIST OF FIGURES
TITLE PAGE NO
Fig 1.1 introduction to machine learning 1
Fig 1.2 linear regression visualization 8
Fig 1.3 support vector machine model 10
Fig 1.4 random forest classifier 12
Fig 1.5 f-p growth algorithm in data mining 14
Fig 1.6 apriori algorithm flowchart 15
v
CHAPTER 1
INTRODUCTI0N
1
machine learning is that it can do tasks that are too complex for a person to implement
directly. As a human, we have some limitations as we cannot access the hugeamount of
data manually, so for this, we need some computer systems and here comes the machine
learning to make things easy for us.
The name machine learning was coined in 1959 by Arthur Samuel. Tom M. Mitchell
provided a widely quoted, more formal definition of the algorithms studied in the machine
learning field: "A computer program is said to learn from experience E with respect to
some class of tasks T and performance measure P if its performance at tasks in T, as
measured by P, improves with experience E." This follows Alan Turing's proposal in his
paper "Computing Machinery and Intelligence", in which the question "Can machines
think?" is replaced with the question "Can machines do what we (as thinking entities) can
do?". In Turing’s proposal the characteristics that could be possessed by a thinking machine
and the various implications in constructing one are exposed.
2
of machine learning algorithms that often use Artificial Neural Networks to
generate models. Deep Learning techniques, for example, have been very
successful in solving Image Recognition problems due to their ability to pick the
best features, as well as to express layers of representation.
4. ANN Inspired by biological neural networks; artificial neural networks are a
network of interconnected nodes that make up a model. They can be defined as
statistical learning models that are used to estimate or approximate functions that
depend on many inputs. Neural networks are usually used when the volumeof
inputs is far too large for standard machine learning approaches previously
discussed.
5. Regression Another sub-category of supervised learning used when the value being
predicted differs to a “yes or no” label as it falls somewhere on a continuous
spectrum.
3
CHAPTER-2
ML CLASSIFICATION
Machine learning is a subset of AI, which enables the machine to automatically learn from
data, improve performance from past experiences, and make predictions. Machine learning
contains a set of algorithms that work on a huge amount of data. Data is fed to these
algorithms to train them, and on the basis of training, they build the model & perform a
specific task.
These ML algorithms help to solve different business problems like Regression,
Classification, Forecasting, Clustering, and Associations, etc.
Based on the methods and way of learning, machine learning is divided into mainly four
types, which are:
1. Supervised Machine Learning
2. Unsupervised Machine Learning
3. Semi-Supervised Machine Learning
4. Reinforcement Learning
Supervised and Unsupervised learning are the two techniques of machine learning. But
both the techniques are used in different scenarios and with different datasets. Below the
explanation of both learning methods along with their difference table is given.
Supervised Machine Learning
Supervised learning is a machine learning method in which models are trained using
labeled data. In supervised learning, models need to find the mapping function to map the
input variable (X) with the output variable (Y).
Supervised learning needs supervision to train the model, which is similar to as a student
learns things in the presence of a teacher. Supervised learning can be used for two types of
problems: Classification and Regression. Example Suppose we have an image of different
types of fruits. The task of our supervised learning model is to identify the fruits and
classify them accordingly. So, to identify the image in supervised learning, we will give
4
the input data as well as output for that, which means we will train the model by the shape,
size, colour, and taste of each fruit. Once the training is completed, we will test the model
by giving the new set of fruit. The model will identify the fruit and predict the output using
a suitable algorithm.
Unsupervised learning is another machine learning method in which patterns inferred from the
unlabeled input data. The goal of unsupervised learning is to find the structure and patterns
from the input data. Unsupervised learning does not need any supervision. Instead,it finds
patterns from the data by its own.
Example To understand the unsupervised learning, we will use the example given above.
So, unlike supervised learning, here we will not provide any supervision to the model. We
will just provide the input dataset to the model and allow the model to find the patterns
from the data. With the help of a suitable algorithm, the model will train itself and divide
the fruits into different groups according to the most similar features between them.
Supervised learning model takes direct Unsupervised learning model does not
feedback to check if it is predicting correct take any feedback.
output or not.
Supervised learning model predicts the Unsupervised learning model finds the
output. hidden patterns in data.
The goal of supervised learning is to train the The goal of unsupervised learning is to find
model so that it can predict the output when the hidden patterns and useful insights
it is given new data. from the unknown dataset.
5
Supervised learning needs supervision to Unsupervised learning does not need any
train the model. supervision to train the model.
Supervised learning can be used for those Unsupervised learning can be used for
cases where we know the input as well as those cases where we have only input data
corresponding outputs. and no corresponding output data.
Supervised learning is not close to true Unsupervised learning is more close tothe
Artificial intelligence as in this, we first train true Artificial Intelligence as it learns
the model for each data, and then only it can similarly as a child learns daily routine
predict the correct output. things by his experiences.
6
CHAPTER-3
ALGORITHMS IN MACHINE LEARNING
3.1 Regression
Machine Learning (ML) has a wide range of industrial applications that are likely to
increase in the coming areas. The global Machine Learning market is expected to reach
USD 117 billion by 2027 with an impressive CAGR (Compound Annual Growth Rate) of
39%. Freshers and tech enthusiasts should know about Machine Learning concepts to
upskill and build a successful career in the ML industry. Regression algorithms in
MachineLearning are an important concept with a lot of use cases.
The future values are predicted with the help of regression algorithms in Machine Learning.
The input data/historical data is used to predict a wide range of future values using
regression. Label in ML is defined as the target variable (to be predicted) and regression
helps in defining the relationship between label and data points. Regression is a type of
supervised learning in ML that helps in mapping a predictive relationship between labels
and data points. The top types of regression algorithms in ML are linear, polynomial,
logistic, stepwise, etc. Read on to know more about the most popular regression algorithms.
1. Linear Regression
7
Figure 1.2 Linear Regression Visualization
Loss function: (Predicted output – actual output)
2. logistic Regression
Logistic regression is used to solve classification problems, and the most common use case
is binary logistic regression, where the outcome is binary (yes or no). In the real world, you
can see logistic regression applied across multiple areas and fields.
1. In health care, logistic regression can be used to predict if a tumor is likely to be
benign ormalignant.
2. In the financial industry, logistic regression can be used to predict if a transaction
isfraudulent or not.
3. In marketing, logistic regression can be used to predict if a targeted audience will
respondor not.
Are there other use cases for logistic regression aside from binary logistic regression? Yes.
There are two other types of logistic regression that depend on the number of predicted
outcomes.
3. Ridge regression
You all must be aware of the power of neural networks in making predictions/assumptions.
Each node in a neural network has a respective activation function that defines the output
of the node based on a set of inputs. The last activation function can be manipulated to
change a neural network into a regression model. One can use ‘Keras’ that is the
appropriate python library for building neural networks in ML.
The output of a neuron is mapped to a variety of values in neural network regression, thus
ensuring non-linearity. You can choose a single parameter or a range of parameters for
predicting output using neural network regression.
3.2 Classification
The Classification algorithm is a Supervised Learning technique that is used to identify the
category of new observations based on training data. In Classification, a program learns
from the given dataset or observations and then classifies new observation into several
classes or groups. Such as, Yes or No, 0 or 1, Spam or Not Spam, cat, or dog, etc. Classes
can be called as targets/labels or categories.
Unlike regression, the output variable of Classification is a category, not a value, such as
9
"Green or Blue", "fruit or animal", etc. Since the Classification algorithm is a Supervised
learning technique, hence it takes labeled input data, which means it contains input with
the corresponding output.
In classification algorithm, a discrete output function(y) is mapped to input variable(x).
The best example of an ML classification algorithm is Email Spam Detector.
The main goal of the Classification algorithm is to identify the category of a given dataset,
and these algorithms are mainly used to predict the output for the categorical data.
Classification algorithms can be better understood using the below diagram. In the below
diagram, there are two classes, class A and Class B. These classes have features that are
like each other and dissimilar to other classes.
Types of classification algorithms:
Support Vector Machine or SVM is one of the most popular Supervised Learning
algorithms, which is used for Classification as well as Regression problems. However,
primarily, it is used for Classification problems in Machine Learning.
The goal of the SVM algorithm is to create the best line or decision boundary that can
segregate n-dimensional space into classes so that we can easily put the new data point in
the correct category in the future. This best decision boundary is called a hyperplane.
SVM chooses the extreme points/vectors that help in creating the hyperplane. These
extreme cases are called as support vectors, and hence algorithm is termed as Support
Vector Machine. Consider the below diagram in which there are two different categories
that are classified using a decision boundary or hyperplane.
10
example that we have used in the KNN classifier. Suppose we see a strange cat that also
has some features of dogs, so if we want amodel that can accurately identify whether it is
a cat or dog, so such a model can be created by using the SVM algorithm. We will first
train our model with lots of images of cats and dogs so that it can learn about different
features of cats and dogs, and then we test it with this strange creature. So as support vector
creates a decision boundary between these two data (cat and dog) and choose extreme cases
(support vectors), it will see the extreme case of cat and dog. Based on the support vectors,it
will classify it as a cat.
1. DECISION TREES
Decision Tree is a Supervised learning technique that can be used for both classification
and Regression problems, but mostly it is preferred for solving Classification problems. It
is a tree-structured classifier, where internal nodes represent the features of a dataset,
branches represent the decision rules and each leaf node represents the outcome.
In a Decision tree, there are two nodes, which are the Decision Node and Leaf Node.
Decision nodes are used to make any decision and have multiple branches, whereas Leaf
nodes are the output of those decisions and do not contain any further branches. The
decisions or the test are performed based on features of the given dataset. It is a graphical
representation for getting all the possible solutions to a problem/decision based on given
conditions.
It is called a decision tree because, like a tree, it starts with the root node, which expands
on further branches and constructs a tree-like structure. To build a tree, we use the CART
algorithm, which stands for Classification and Regression Tree algorithm. A decision tree
simply asks a question, and based on the answer (Yes/No), it further split the tree into
subtrees.
2. RANDOM FOREST
Random Forest is a popular machine learning algorithm that belongs to the supervised
learning technique. It can be used for both Classification and Regression problems in ML.
It is based on the concept of ensemble learning, which is a process of combining multiple
classifiers to solve a complex problem and to improve the performance of the model.
11
As the name suggests, "Random Forest is a classifier that contains a number of decision
trees on various subsets of the given dataset and takes the average to improve the predictive
accuracy of that dataset." Instead of relying on one decision tree, the random forest takes
the prediction from each tree and based on the majority votes of predictions, and it predicts the
final output. The greater number of trees in the forest leads to higher accuracy and prevents
the problem of overfitting.
Example Suppose there is a dataset that contains multiple fruit images. So, this dataset is
given to the Random Forest classifier. The dataset is divided into subsets and given to each
decision tree. During the training phase, each decision tree produces a prediction result,
and when a new data point occurs, then based on most results, the Random Forest classifier
predicts the final decision.
3. NAVIE BIAES
1. Naïve Bayes algorithm is a supervised learning algorithm, which is based on Bayes
theorem and used for solving classification problems.
12
2. It is mainly used in text classification that includes a high-dimensional training
dataset.
3. Naïve Bayes Classifier is one of the simple and most effective Classification
algorithms which helps in building the fast machine learning models that can make
quick predictions.
4. It is a probabilistic classifier, which means it predicts based on the probability of
an object.
5. Some popular examples of Naïve Bayes Algorithm are spam filtration, Sentimental
analysis, and classifying articles.
3.3 Association
Association rule learning is a type of unsupervised learning technique that checks for the
dependency of one data item on another data item and maps accordingly so that it can be
more profitable. It tries to find some interesting relations or associations among the
variables of dataset. It is based on different rules to discover the interesting relations
between variables in the database.
The association rule learning is one of the very important concepts of machine learning,
and it is employed in Market Basket analysis, Web usage mining, continuous production,
etc. Here market basket analysis is a technique used by the various big retailer to discover
the associations between items. We can understand it by taking an example of a
supermarket, as in a supermarket, all products that are purchased together are put together.
Types of Association Rule Learning
Association rule learning can be divided into three algorithms:
Apriori Algorithm
This algorithm uses frequent datasets to generate association rules. It is designed to work
on the databases that contain transactions. This algorithm uses a breadth-first search and
Hash Tree to calculate the itemset efficiently. It is mainly used for market basket analysis
and helps to understand the products that can be bought together. It can also be used in the
healthcare field to find drug reactions for patients.
13
Figure 1.5 Apriori Algorithm flowchart
Eclat Algorith
Eclat algorithm stands for Equivalence Class Transformation. This algorithm uses a depth-
first search technique to find frequent itemset in a transaction database. It performs faster
execution than Apriori Algorithm.
The F-P growth algorithm stands for Frequent Pattern, and it is the improved version of the
14
Apriori Algorithm. It represents the database in the form of a tree structure that is known
as a frequent pattern or tree. The purpose of this frequent tree is to extract the most frequent
patterns.
15
CHAPTER-4
ADVANTAGES AND DISADVANTAGES
Machine Learning can review large volumes of data and discover specific trends and
patterns that would not be apparent to humans. For instance, for an e-commerce website
like Amazon, it serves to understand the browsing behaviors and purchase histories of its
users to help cater to the right products, deals, and reminders relevant to them. It usesthe
results to reveal relevant advertisements to them.
With ML, you do not need to babysit your project every step of the way. Since it means
giving machines the ability to learn, it lets them make predictions and improve the
algorithms on their own. A common example of this is anti-virus software; they learn to
filter new threats as they are recognized. ML is also good at recognizing spam.
Machine Learning algorithms are good at handling data that are multi-dimensional and
multi-variety, and they can do this in dynamic or uncertain environments.
4. Continuous Improvement
As algorithms gain experience, they keep improving in accuracy and efficiency. This lets
them make better decisions. Say you need to make a weather forecast model. As the amount of
data you have keeps growing, your algorithms learn to make more accurate predictions
faster.
5. Wide Applications
You could be an e-tailer or a healthcare provider and make ML work for you. Where it
does apply, it holds the capability to help deliver a much more personal experience to
16
customers while also targeting the right customers.
6. Automation
Machine Learning is one of the driving forces behind automation, and it is cutting down
time and human workload. Automation can now be seen everywhere, and the complex
algorithm does the hard work for the user. Automation is more reliable, efficient, and quick.
With the help of machine learning, now advanced computers are being designed. Now this
advanced computer can handle several machine-learning models and complex algorithms.
However, automation is spreading faster in the industry but, a lot of research and
innovation are required in this field.
7. Scope of Improvement
Machine Learning is a field where things keep evolving. It gives many opportunities for
improvement and can become the leading technology in the future. A lot of research and
innovation is happening in this technology, which helps improve software and hardware.
Machine Learning is going to be used in the education sector extensively, and it will be
going to enhance the quality of education and student experience. It has emerged in China;
machine learning has improved student focus. In the e-commerce field, Machine Learning
studies your search feed and give suggestion based on them. Depending upon search and
browsing history, it pushes targeted advertisements and notifications to users.
This technology has a very wide range of applications. Machine learning plays a role in
almost every field, like hospitality, ed-tech, medicine, science, banking, and business. It
creates more opportunities.
You could be an e-tailer or a healthcare provider and make ML work for you. Where it
does apply, it holds the capability to help deliver a much more personal experience to
customers while also targeting the right customers.
17
Applications in machine learning
1. Image Recognition
Image recognition is one of the most common applications of machine learning. It is used
to identify objects, persons, places, digital images, etc. The popular use case of image
recognition and face detection is, Automatic friend tagging suggestion:
Facebook provides us a feature of auto friend tagging suggestion. Whenever we upload a
photo with our Facebook friends, then we automatically get a tagging suggestion with
name, and the technology behind this is machine learning's face detection and recognition
algorithm.
It is based on the Facebook project named "Deep Face," which is responsible for face
recognition and person identification in the picture.
2. Speech Recognition
While using Google, we get an option of "Search by voice," it comes under speech
recognition, and it is a popular application of machine learning.
Speech recognition is a process of converting voice instructions into text, and it is also
known as "Speech to text", or "Computer speech recognition." At present, machine
learning algorithms are widely used by various applications of speech recognition. Google
assistant, Siri, Cortana, and Alexa are using speech recognition technology to follow the
voice instructions.
3. Traffic prediction
If we want to visit a new place, we take help of Google Maps, which shows us the correct
path with the shortest route and predicts the traffic conditions.
It predicts the traffic conditions such as whether traffic is cleared, slow-moving, or heavily
congested with the help of two ways:
Real Time location of the vehicle form Google Map app and sensors Average time has
taken on past days at the same time.
18
4.2 Disadvantages of machine learning
In ML, we can choose the algorithms based on accurate results. For that, we must run the
results on every algorithm. The main problem occurs in the training and testing of data.
The data is huge, so sometimes removing errors becomes nearly impossible. These errors
can cause a headache to users. Since the data is huge, the errors take a lot of time to resolve.
2. Algorithm Selection
The selection of an algorithm in Machine Learning is still a manual job. We must run and
test our data in all the algorithms. After that only we can decide what algorithm, we want.
We choose them based on result accuracy. The process is very much time- consuming.
3. Data Acquisition
In ML, we constantly work on data. We take a huge amount of data for training and testing.
This process can sometimes cause data inconsistency. The reason is some data constantly
keep on updating. So, we must wait for the new data to arrive. If not, the old and new data
might give different results. That is not a good sign for an algorithm.
Many ML algorithms might take more time than you think. Even if it is the best algorithm,
it might sometimes surprise you. If your data is large and advanced, the system will take
time. This may sometimes cause the consumption of more CPU power. Even with GPUs
alongside, it sometimes becomes hectic. Also, the data might use more than the allotted
space.
5. Data management
The whole concept of machine learning is about identifying useful data. The outcome will
be incorrect if a credible data source is not provided. The quality of the data is also
significant. If the user or institution needs more quality data, wait for it. It will cause delays in
providing the output. So, machine learning significantly depends on the data and its
19
quality.
6. Resources
The data that machines process remains huge in quantity and differs greatly. Machines
require time so that their algorithm can adjust to the environment and learn it. Trials runs
are held to check the accuracy and reliability of the machine. It requires massive and
expensive resources and high-quality expertise to set up that quality of infrastructure. Trials
runs are costly as they would cost in terms of time and expenses.
7. Social Changes
Machine learning is bringing numerous social changes in society. The role of machine
learning-based technology in society has increased multifold. It is influencing the thought
process of society and creating unwanted problems in society. Character assassination and
sensitive details are disturbing the social fabric of society.
8. Privacy Concern
As we know that one of the pillars of machine learning is data. The collection of data has
raised the fundamental question of privacy. The way data is collected and used for
commercial purposes has always been a contentious issue. In India, the Supreme court of
India has declared privacy a fundamental right of Indians. Without the user's permission,
data cannot be collected, used, or stored. However, many cases have come up that big firms
collect the data without the user's knowledge and using it for their commercial gains.
20
CHAPTER-5
CASE STUDY
21
5.2. PROGRAM_CODE
import cv2
import numpy as np
yolo_config_path = "yolov3.cfg" # Path to YOLO configuration file
yolo_weights_path = "yolov3.weights" # Path to pre-trained YOLO weights
coco_labels_path = "coco.names" # Path to COCO labels file
# Load YOLO
net = cv2.dnn.readNetFromDarknet(yolo_config_path, yolo_weights_path)
layer_names = net.getLayerNames()
output_layers = [layer_names[i[0] - 1] for i in net.getUnconnectedOutLayers()]
# Load COCO class labels
with open(coco_labels_path, 'r') as f:
labels = f.read().strip().split("\n")
# Function to perform object detection
def detect_objects(frame):
height, width = frame.shape[:2]
blob = cv2.dnn.blobFromImage(frame, 1/255.0, (416, 416), swapRB=True, crop=False)
net.setInput(blob)
detections = net.forward(output_layers)
boxes, confidences, class_ids = [], [], []
for output in detections:
for detection in output:
scores = detection[5:]
class_id = np.argmax(scores)
confidence = scores[class_id]
if confidence > 0.5: # Minimum confidence threshold
# Calculate bounding box coordinates
box = detection[0:4] * np.array([width, height, width, height])
(centerX, centerY, box_width, box_height) = box.astype("int")
x = int(centerX - (box_width / 2))
y = int(centerY - (box_height / 2))
22
boxes.append([x, y, int(box_width), int(box_height)])
confidences.append(float(confidence))
class_ids.append(class_id)
# Apply non-maxima suppression
indices = cv2.dnn.NMSBoxes(boxes, confidences, score_threshold=0.5,
nms_threshold=0.3)
for i in indices.flatten():
x, y, w, h = boxes[i]
color = (0, 255, 0) # Bounding box color (green)
label = f"{labels[class_ids[i]]}: {int(confidences[i] * 100)}%"
cv2.rectangle(frame, (x, y), (x + w, y + h), color, 2)
cv2.putText(frame, label, (x, y - 10), cv2.FONT_HERSHEY_SIMPLEX, 0.5, color,
2)
return frame
# Video stream for real-time detection
cap = cv2.VideoCapture("test_video.mp4") # Replace with 0 for webcam
while True:
ret, frame = cap.read()
if not ret:
break
# Perform detection on the frame
frame = detect_objects(frame)
# Display the result
cv2.imshow("Object Detection", frame)
if cv2.waitKey(1) & 0xFF == ord('q'):
break
cap.release()
cv2.destroyAllWindows()
23
CONCLUSION
Machine learning is a field of artificial intelligence that deals with the design and
development of algorithms that can learn from and make predictions on data. The aim of
machine learning is to automate analytical model building and enable computers to learn
from data without being explicitly programmed to do so. Machine learning represents a
rapidly expanding frontier within computer science, permeating diverse fields of study and
finding widespread commercial applications. Its significance lies in its ability to address
complex problems that may be impractical or time-intensive for human solutions. At its
core, machine learning employs a spectrum of models to discern patterns within data,
subsequently enabling accurate predictions based on the identified patterns. This dynamic
and evolving field holds great promise for the future, contributing to advancements and
innovations across various domains.Machine learning is quickly growing field in computer
science. It has applications in nearly every other field of study and is already being
implemented commercially because machine learning can solve problems too difficult or
time consuming for humans to solve. To describe machine learning in general terms, a
variety models are used to learn patterns in data and make accurate predictions based on
the patterns it observes.Machine learning is a powerful tool for making predictions from
data. However, it is important to remember that machine learning is only as good as the
data that is used to train the algorithms. In order to make accurate predictions, it is
important to use high-quality data that is representative of the real word many applications are
being used and constructed using machine learning
24
REFERENCES
25