V Minor
V Minor
PROJECT REPORT
ON
Submitted To
Dr Vijay Kumar Dalla Sir
Submitted By
Premdeep (2022UGCM0016)
National Institute of Technology, Jamshedpur
CERTIFICATE
This is to certify that Mr. Premdeep (Reg. No: 2022UGCM016)
Bachelor of Engineering and Computational Mechanics Vth semester student
of National Institute of Technology, Jamshedpur, has done project work entitled
“Mnist Digit Classification” under the guidance of oursenior faculties during the
period of July 2024 to November 2024. They successfully completed the
project and during the period they were methodical and hardworking.
Premdeep
(2022UGCM0016)
This is to certify that the above statement made by the candidate is trueand
correct to the best of my knowledge and belief.
I would like to express our sincere gratitude to all those who contributed to the
completion of this research work titled “Mnist Digit Classification”. We are
deeply thankful to our supervisor Dr Vijay Kumar Dalla Sir whose guidance,
support, and valuable insights were instrumental throughout this research
project. Their expertise and encouragement have been invaluable.
I extend our appreciation to the participants of this study, whose cooperation
and willingness to participate were crucial in gathering meaningful data and
drawing insightful conclusions.
Lastly, I acknowledge the contributions of all the authors, researchers, and
institutions whose work and publications have enriched the literature and
served as a foundation for this study.
ABSTRACT
3.Experiments
3.1 MNIST Handwritten Digit Dataset
3.2 Model Hyperparameter Optimization
3.3 Learning Rate Attenuation and Classification Analysis
3.4 Model Optimization and Performance Evaluation
4 Conclusions
5 References
LIST OF FIGURE
The output of the last convolutional layer includes the extracted features, which
are flattened into a onedimensional tensor and then forwarded to two fully connected
(i.e., dense) layers. A Softmax layer is connected after the last fully connected layer
to predict the likelihood of each category for each sample. The class with the highest
probability is chosen as the sample’s predicted category.
In this study, three convolution layers are exploited for feature extraction (Fig.
2). The first layer is equipped with a kernel of 3 × 3 and a stride of (1, 1), while the
other two use a kernel of 6 × 6 and a stride of (2, 2) for two main purposes: 1)
speeding up the converging rate with larger kernel size in deeper layers, and 2) use
larger stride as an alternative to the max pooling layer for data dimension reduction.
After three convolution layers, each of which is accompanied with a batch
normalization and an activation layer, a total number of 1568 features are extracted
and fed to two dense layers for classification. Finally, the output layer has 10 nodes
for the 10 different number digit categories.
where y and y0 are the classifications of model prediction and the real classifications
(i.e., ground truths or labels) of samples respectively, and i 2 ^ is the sample indicator
corresponding to the dataset. The accuracy rate and losses are calculated separately
for each category, and these two indicators could describe the efficacy of the
proposed model more comprehensively.
3.Experiments
3.1 MNIST Handwritten Digit Dataset
MNIST database is one of the foremost classical imaging datasets within the field
of machine learning, and is broadly utilized for benchmarking in image
classification. The MNIST database contains 60,000 training samples and 10,000
testing samples (Table 1), each consisting of a 28*28 pixel grayscale image of a
handwritten Arabic digit. The number sample in each image is
normalized/standardized and centered. The 60,000 training samples are further
divided into 55,000 training dataset and
5,000 validation dataset.
4 Conclusions
This study elaborates the details of developing a light-weight DNN for
handwriting digital recognition/ classification using the MNIST dataset. The CNN-
based backbone architecture and the methodology of numerous hyperparameter
optimizations, including batch size, learning rate, etc., are explored in detail to
enhance the training and testing outcomes. The developed neural network model is
evaluated upon Keras/ Tensorflow framework, and the overall accuracy of the
developed model can reach 99.4% on the MNIST dataset, compatible with the results
from other state-of-the-art models.
Acknowledgement: We would like to thank the editors and anonymous reviewers for
their thoughtful comments and efforts towards improving our manuscript.
Funding Statement: The authors received no specific funding for this study.
Conflicts of Interest: The authors declare that they have no conflicts of interest to
report regarding the present study.
5.References
1. Dataset from Kaggle
https://www.kaggle.com/datasets/sachinshrivastv/simple-mnist
2.MNIST Handwritten Digit Classification Based on Convolutional Neural Network with
Hyperparameter Optimization
https://www.researchgate.net/publication/369265604_MNIST_Handwritten_Di
git_Classification_Based_on_Convolutional_Neural_Network_with_Hyperpara
meter_Optimization