Automatic Classification of Mechanical Components of Engines Using Deep Learning Techniques
Automatic Classification of Mechanical Components of Engines Using Deep Learning Techniques
ISSN No:-2456-2165
Abstract:- Mechanical parts of engines help to reduce transmitted motion while the friction between the thread and
friction and carry weight for linear or rotating motion. compression in nuts and bolts work together to form the
Modern engines are complex systems with structural fastener [4]. Assembling and disassembling of mechanical
elements, mechanisms, and mechanical parts. The components is a routine and important process in industries
building blocks of the engines are joined together using using machines [5].
several mechanical components that are similar in shape
and size. During the assembly and disassembly of these Most mechanical engine contains tens or hundreds of
complex engines, the mechanical components get mixed mechanical components. The classification of mechanical
up. The traditional classification techniques for components can be defined as a Fine Grained Visual
components are laborious with high costs. Existing Categorization (FGVC) problem. A significant number of
research for classifying mechanical components uses these mechanical components needs to be identified and
algorithms that work based on shape descriptors and classified [6]. However, some of these mechanical
geometric similarity thereby resulting in low accuracies. components are similar in shapes and sizes thereby making
Hence, there is a need to develop an automatic the manual extraction of distinguishing features difficult [7].
classification technique with high accuracy. This study Also, mechanical components with unrecognizable part
classified four mechanical components (bearing, nut, number or without part number makes the manual
gear, and bolt) using four deep learning models classification costly and time consuming for technicians [8].
(AlexNet, DenseNet-121, ResNet-50, and SqueezeNet). In Hence, an automatic classification technique would reduce
the result, Densenet-121 achieved the highest costs and save time [9].
performance at an accuracy of 98.3%, sensitivity of
95.8%, specificity of 98.5%, and Area under Curve Computer vision [1] is the technology and science that
(AUC) of 98.5%. focuses on the theory behind artificial systems to extract
information from images using an automated approach. It
Keywords: Mechanical Components, Deep Learning, has a great potential in the sorting, inspection, classification
Convolutional Neural Network, Engine, Transfer Learning. and quality control of mechanical parts during assembling,
manufacturing and disassembling stages [10]. Computer
I. INTRODUCTION vision has been combined with machine learning techniques
for achieving automatic image classification. The major
Mechanical design involves the knowledge of constraint of machine learning is that it cannot extract
numerous machine elements that could be assembled differentiating features from the training set of data.
together using mechanical components [1]. Mechanical However, This limitation has been remedied by the use of
components ensure the reliability and safety of engines [2]. deep learning technique [11].
Basic mechanical components such as screw, bearing, gear,
washer, bolt, and nut can be used for connecting mechanical Deep Learning (DL) is a branch of machine learning
building blocks into complex systems in the mechanical that use algorithms for processing information [11]. It is
industry [3]. Bearing and gear are structural components, the implemented using the architecture of neural network [12].
basic purpose of gear is to change the speed of direction for Various images has been successfully classified using deep
ResNet
Residual neural network (ResNet) was developed in
2015 by Kaiming et al. [21]. The network was developed by
piling residual blocks on top of each other. It was realized
by hopping connections on two or three layers containing
batch normalization and Rectified Linear Unit (ReLU)
between the architectures [22]. Compared to other CNN
models, ResNet’s training ability is better and the
computations are lighter. The available ResNet models are
ResNet-18, ResNet-34, ResNet-50, ResNet-101 and ResNet-
152 [23]. ResNet-50 won ILSVRC in year 2015 with 3.6%
error rate. Its architecture contains 48 convolution layers,
one maxpool layer and one average pool layer [24].
SqueezeNet
SqueezeNet was developed at the University of
California, Stanford University, Berkeley and DeepScale in
year 2016 [25] by researchers. The architecture of
SqueezeNet achieved an AlexNet-level precision on
ImageNet with less than 5 MB parameters and less
computational time [26]. The architecture contains two
convolutional layers, eight fire modules and one softmax
layer [26].
DenseNet
Densely connected convolutional networks (DenseNet)
was developed by Zhuang Liu, Gao Huang and their team in
year 2017 [27]. Each DenseNet layer has a direct contact
with the original input image and it does not learn redundant
feature maps [28]. The available variants of DenseNet are
DenseNet-121, DenseNet-160, and DenseNet-201.
DenseNet-121 is a 24-layer convolutional neural network
containing five convolution layers, five pooling layers, three
Start
Define Transformations
Train Model
Has NO
Network
Converged
YES
Save Model
Stop
The remaining part of this paper is organized as follows; Section 2 presents the literature survey. Section 3 contains the
proposed methodology. Section 4 reports the discussion of the result and finally, Section 5 presents the conclusion and future
scope.
Evaluation Metrics
Evaluation metrics are indicators for assessing the
performance of an experiment [40]. In this work, sensitivity,
accuracy, specificity and Area under Curve (AUC) were
selected as the quantitative evaluation metrics. The Fig 4 (b) Confusion Matrix for AlexNet Model
accuracy, sensitivity, and specificity were calculated using
Equations (1), (2), and (3) correspondingly.
𝑇𝑃+𝑇𝑁
𝐴𝑐𝑐𝑢𝑟𝑎𝑐𝑦 = (1)
𝑇𝑃+𝑇𝑁+𝐹𝑃+𝐹𝑁
𝑇𝑃
𝑆𝑒𝑛𝑠𝑖𝑡𝑖𝑣𝑖𝑡𝑦 = (2)
𝑇𝑃+𝐹𝑁
𝑇𝑁
𝑆𝑝𝑒𝑐𝑖𝑓𝑖𝑐𝑖𝑡𝑦 = (3)
𝑇𝑁+𝐹𝑃
Performance Analysis
This section presents the result obtained during the
training and the testing of the CNN models. Figures,
(4),(5),(6),and (7) shows the results obtained.
Fig 5 (b) Confusion Matrix for DenseNet-121 Model Fig 6 (b) Confusion Matrix for ResNet-50 Model
Fig 5 (c) ROC Plot for DenseNet-121 Model Fig 6 (c) ROC Plot for ResNet-50 Model
Fig 6 (a) Training Plot for ResNet-50 Model Fig 7 (a) Training Plot for SqueezeNet Model
Fig 7 (b) Confusion Matrix for SqueezeNet Model Fig 7 (c) ROC Plot for SqueezeNet Model
Table 1 The Result of Accuracy, Loss, and the Elapsed Time at 15th Epoch when Training the CNN Models
Validation accuracy at
15 th epoch (%) 97.893 96.975 96.179 96.504
Training accuracy at
15 th epoch (%) 98.520 98.849 97.867 97.533
validation loss at 15 th
epoch 0.00026 0.00169 0.00243 0.00129
Training loss at 15 th
epoch 0.00014 0.00112 0.00110 0.00104
Training-Time
(hh:mm:ss) 02:18:37 42:08:03 18:49:28 02:31:28
From Table 1, we can observe that all the models had Result of Testing the CNN Models
high training accuracies indicating that the models learnt the This section presents the result of testing the CNN
features of the images correctly. DenseNet-121 model had models in classifying the test dataset. It was observed from
the highest training accuracy of 98.849% while SqueezeNet the confusion matrixes that some of the mechanical
model had the least training accuracy of 97.533%. AlexNet components were misclassified, this indicated similarities
model had the highest validation accuracy of 97.893% while among mechanical components. From the confusion matrix
the ResNet-50 model had the least validation accuracy of for each CNN model, the evaluation metric for each
96.179%. It was observed that the validation accuracy for all mechanical component was calculated using equations (1),
the models were lesser than the training accuracy; this (2), and (3) correspondingly. The performance of each
indicated that the models did not over-fit and they CNN model was determined by calculating average
generalized well on the validation dataset. Considering the accuracy, average sensitivity, average specificity, and
elapsed time for the training, SqueezeNet model took the average AUC. The result obtained is as shown in Table 2.
least time, while DenseNet-121 took the longest time.
From Table 2, the test performances for all the models were good. DenseNet-121 model had the highest performance at
98.3% accuracy, 95.8% sensitivity, and 98.5% specificity. AlexNet had the lowest accuracy of 94.5%. sensitivity of 92.8%, and
the highest AUC of 93.5% while squeezeNet has the lowest AUC of 96.0%. AlexNet has the lowest specificity of 96.2%. ResNet-
50 had the highest AUC of 99.0%. Considering the time elapsed for the testing, SqueezeNet model took the least test time, while
DenseNet took the longest time.
From Table 3, we can observe that AlexNet model has The classification performance for all the considered
the highest number of parameters while SqueezeNet model CNN models are good, this indicates that deep learning
has the lowest number of parameters. ResNet-50 model has techniques are good image classifiers. DenseNet-121 has the
the deepest architecture while squeezeNet model has the highest performance in terms of accuracy, sensitivity, and
shallowest architecture. All the models were trained and specificity. However, it came second in the AUC value. This
tested under the same conditions. Comparing the model suggests that the use of more than one evaluation metric is
architecture and their performances. We observed that important in comparing and validating the performance of
DenseNet-121 model took the longest time for the training deep learning models as classifiers. From this work, it can
and testing while squeezeNet model took the shortest be established that under the same training and testing
training and testing time. This indicated that network conditions; the performance of CNN models differs due to
architecture could affect the training and testing time for differnces in their network architecture. It can also be
CNN models. deduced that the performance of CNN models in image
classification has improved over the years through the
variation of network architectures.