0% found this document useful (0 votes)
4 views12 pages

Deep Convolutional Neural Network For Weld Defect Classification in Radiographic Images

This paper presents a deep convolutional neural network (CNN) model based on ResNet50 for classifying weld defects in radiographic images, specifically targeting cracks, pores, non-penetration, and no defects. The model achieved high accuracies of 98.75%, 90.255%, and 75.83% on three different datasets, demonstrating its effectiveness even with low-quality images. The research emphasizes the importance of advanced computational techniques in improving quality control processes in the welding industry.

Uploaded by

aadityadevadeth
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
4 views12 pages

Deep Convolutional Neural Network For Weld Defect Classification in Radiographic Images

This paper presents a deep convolutional neural network (CNN) model based on ResNet50 for classifying weld defects in radiographic images, specifically targeting cracks, pores, non-penetration, and no defects. The model achieved high accuracies of 98.75%, 90.255%, and 75.83% on three different datasets, demonstrating its effectiveness even with low-quality images. The research emphasizes the importance of advanced computational techniques in improving quality control processes in the welding industry.

Uploaded by

aadityadevadeth
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 12

See discussions, stats, and author profiles for this publication at: https://www.researchgate.

net/publication/380282202

Deep convolutional neural network for weld defect classification in


radiographic images

Article in Heliyon · May 2024


DOI: 10.1016/j.heliyon.2024.e30590

CITATIONS READS

4 162

8 authors, including:

Dayana Palma Ramírez Bárbara Dora Ross Veitía


Universidad de Oriente Centro Nacional de Electromagnetismo Aplicado, Cuba
3 PUBLICATIONS 10 CITATIONS 4 PUBLICATIONS 13 CITATIONS

SEE PROFILE SEE PROFILE

Alejandro Espinel Hernández Angel Sánchez Roca


Universidad de Oriente Universidad de Oriente
9 PUBLICATIONS 42 CITATIONS 53 PUBLICATIONS 409 CITATIONS

SEE PROFILE SEE PROFILE

All content following this page was uploaded by José R. Nuñez Alvarez on 04 May 2024.

The user has requested enhancement of the downloaded file.


Heliyon 10 (2024) e30590

Contents lists available at ScienceDirect

Heliyon
journal homepage: www.cell.com/heliyon

Research article

Deep convolutional neural network for weld defect classification


in radiographic images
Dayana Palma-Ramírez a, Bárbara D. Ross-Veitía b, Pablo Font-Ariosa c,
Alejandro Espinel-Hernández d, Angel Sanchez-Roca e, Hipólito Carvajal-Fals f, José
R. Nuñez-Alvarez g, *, Hernan Hernández-Herrera h, **
a
Postgraduate Program Doctorate in Applied Computer Engineering School of Computer Engineering. University of Valparaiso. Valparaiso, Chile
b
Production Engineering Doctorate Postgraduate Program Federal Technological University of Paraná (UTFPR) - Ponta Grossa Campus. PR, Brazil
c
Defectoscopy and Welding Technical Services Company, Road O’Burke km. 2½ Pastorita, Cienfuegos, Cuba
d
National Center for Applied Electromagnetism (CNEA), Universidad de Oriente, Ave. de Las Américas s/n, 90100, Santiago de Cuba, Cuba
e
Intranox SL Pol. La Portalada C/ Circunde, 23 26006, Logroño, La Rioja, Spain
f
Pesquisador Visitante. Departamento de Engenharia de Manufatura e Materiais. Universidade Estadual de Campinas. SP, Brazil
g
Energy Department, Universidad de la Costa, (CUC), Calle 58 # 55-66, Barranquilla, 080002, Colombia
h
Faculty of Engineering, Universidad Simón Bolívar, Carrera 59 #59-132, Barranquilla, 080002, Colombia

A R T I C L E I N F O A B S T R A C T

Keywords: The quality of welds is critical to the safety of structures in construction, so early detection of
Radiographic testing irregularities is crucial. Advances in machine vision inspection technologies, such as deep
Classification learning models, have improved the detection of weld defects. This paper presents a new CNN
Weld defects
model based on ResNet50 to classify four types of weld defects in radiographic images: crack,
CNNs
Transfer learning
pore, non-penetration, and no defect. Stratified cross-validation, data augmentation, and regu­
larization were used to improve generalization and avoid over-fitting. The model was tested on
three datasets, RIAWELC, GDXray, and a private dataset of low image quality, obtaining an ac­
curacy of 98.75 %, 90.255 %, and 75.83 %, respectively. The model proposed in this paper
achieves high accuracies on different datasets and constitutes a valuable tool to improve the
efficiency and effectiveness of quality control processes in the welding industry. Moreover,
experimental tests show that the proposed approach performs well on even low-resolution
images.

1. Introduction

Welding is widely used in various modern industries, including energy, building construction, ships, automobiles, aircraft, pet­
rochemicals, food, nuclear power, and electronics [1–4]. To ensure the reliability and safety of built structures, it is crucial to have an
efficient quality control process, including detecting and classifying defects.

* Corresponding author.
** Corresponding author.
E-mail addresses: [email protected] (D. Palma-Ramírez), [email protected] (B.D. Ross-Veitía), [email protected] (P. Font-
Ariosa), [email protected] (A. Espinel-Hernández), [email protected] (A. Sanchez-Roca), [email protected] (H. Carvajal-Fals),
[email protected] (J.R. Nuñez-Alvarez), [email protected] (H. Hernández-Herrera).

https://doi.org/10.1016/j.heliyon.2024.e30590
Received 18 January 2024; Received in revised form 29 April 2024; Accepted 30 April 2024
Available online 1 May 2024
2405-8440/© 2024 Published by Elsevier Ltd. This is an open access article under the CC BY-NC-ND license
(http://creativecommons.org/licenses/by-nc-nd/4.0/).
D. Palma-Ramírez et al. Heliyon 10 (2024) e30590

In this regard, industrial radiography is a non-destructive testing (NDT) technique [5–7] that allows the inspection of welds and the
determination of the presence of internal faults, as well as leaving a permanent documentary record [8,9]. Certified examiners test
welded parts under visual inspection [10,11]. The evaluation criteria may vary depending on the inspector, so the result affects
effectiveness and accuracy. At the same time, defect detection and classification become difficult due to contrast changes and noise in
radiographic images.
The increase in industrial production brings with it the need to test a large volume of samples in a short period, so the fatigue of the
worker and the subjectivity of the evaluations make this process unreliable [12]. Due to this, new advanced computational techniques
are necessary to improve the quality and efficiency of the welded joint inspection process [13–17].
Several authors have focused their studies on applying artificial intelligence (AI) techniques for welding inspection from radio­
graphic images and defect classification [18]. In this sense, T. W. Liao et al. [19] proposes using statistical classifiers to extract features
from noisy radiographic images of cast aluminum with defects. In contrast, other authors have used fuzzy clustering techniques [20]
and fuzzy expert systems for defect classification [21]. Research by K. Carvajal et al. and L. Yang et al. [22,23] proposes pattern
recognition using neural networks. However, all these previous methods are based on extracting features chosen by groups of experts.
Otherwise, S. Wang et al. [24] uses deep neural networks (DNN) to classify five weld defects: pore, cracks, lack of fusion, pene­
tration, and slag inclusion, achieving 91.36 % classification accuracy but with a sample number of only 220 images. In a similar study
[25], proposed a combination of binary classification and flaw detection on steel surfaces with 98 % accuracy when combining both
techniques but with low accuracy for measuring crack defects.
In recent years, the excellent results achieved by convolutional neural networks for image classification have been demonstrated
[26–29]. This type of deep learning model has been widely used [30,31] because of its advantages in bringing together feature
extraction and classification results in a single structure. A study by Khumaidi et al. [32] classified ten classes in 752 radiographic
images, achieving 89 % accuracy using transfer learning techniques in a CNN mode. On the other hand, S. Perri et al. [33] obtained
95.8 % accuracy in classifying four types of weld defects using a dataset of 120 images from a webcam.
A recurring problem in the academic community applying deep learning in non-destructive testing is the need for more public,
varied, high volume, and high-quality data collections that allow the training, testing, and validation of the models obtained for defect
classification [34]. The GDXray collection is currently available [35], which has in its Welds category only 68 images, so additional
work of cropping and manual annotation of the images is needed to train a convolutional neural network (CNN). The WDXI collection
[36] has 13,766 images of seven different defect types but is private.
Recently, Totino et al. [37] published a public dataset suitable for weld defect classification containing 24407 radiographic images
divided into four defects: crack, pore, lack of penetration, and no fault. This dataset provides the scientific community with a working
tool to train neural network models for weld defect classification and identification. In addition, when combined with the transfer
learning technique in deep learning models, this dataset can further enhance the model’s ability to adapt to different scenarios and
improve the accuracy of defect classification.
In various machine learning applications, using transfer learning for feature extraction is an effective technique [38–46]. This
technique involves using a pre-trained model to extract meaningful features from the input images, which can improve the model’s
generalization ability and reduce the risk of over-fitting [47,48].
Transfer learning in deep learning models allows for avoiding overfitting, solving the problem of using small datasets, and
extracting the features of the images to be classified [38]. Several pre-trained networks such as VGG16 [49], ResNet50 [45], Incep­
tionResNetV2 [50], DenseNet [51], AlexNet [52], and others have served as feature extractors in classification problems [53–55].
These networks were trained using ImageNet [56], making them suitable for classifying 1000 object classes. Still, they do not consider
welding defect classes, so it is necessary to use a set of images to train the model according to the problem to be solved.
In this sense, this paper aims to present a new CNN model based on ResNet50, designed to classify defects in radiographic images.
Techniques such as stratified cross-validation, data augmentation, and regularization will be selected to improve model generalization
and avoid over-fitting. This approach will represent a valuable tool to enhance the efficiency of quality control procedures in the
welding industry, even when confronted with poor-quality images.

2. Materials and methods

Defects in welding can significantly impact the strength and integrity of joints. This research addresses four types of defects: crack,
pore, lack of penetration, and no defect. These discontinuities are highly prevalent in welds and were chosen for analysis based on the
criteria established by experts in Cuban industries.
Crack-type defects are cracks or fractures in the weld or the heat-affected zone. They can be microscopic or visible to the naked eye.
Cracks can weaken the structure and propagate under service loads, reducing the strength and service life of the material. Cracks are
critical and must be eliminated or rigorously controlled. However, detecting them with non-destructive techniques (NDT) can be
challenging due to their size and internal location [57–60].
Pore-type defects refer to small cavities or gas bubbles trapped in the weld. They can be spherical or elongated. Porosity reduces the
strength and tightness of the weld. It can affect the integrity of the joint. Although not as critical as cracks, porosity should be
minimized. NDT techniques can detect porosity, but their sensitivity varies [61,62].
On the other hand, the non-penetration type defect occurs when the weld does not fully penetrate the base material. There may be a
gap between the joined parts. The lack of penetration reduces the strength of the joint and can lead to premature failure. Although
there are no apparent defects in the weld on the no-defect samples, NDT inspection is still essential to verify the internal quality [63,
64].

2
D. Palma-Ramírez et al. Heliyon 10 (2024) e30590

2.1. Data base

The radiographic image database selected to train the weld defect classification model was RIAWELC [37]. This set has 24,407
radiographic images of size 224 x 224 pixels divided into four types of weld defects: crack, pore, lack of penetration, and no defect,
Fig. 1.
For training the neural network, images were selected and balanced, with 1600 samples with 400 images in each class. Class
balancing allows for improved model performance, avoids majority class bias, and provides a more accurate evaluation, making it a
helpful technique in the classification task [65–67]. The proposed model was trained and validated using a K-fold stratified
cross-validation technique [68,69] with K = 5, which allows the dataset to be split into five smaller subsets and used to train and
evaluate the model in five rounds. This ensures a more robust and reliable evaluation of the model by using different combinations of
training and validation data in each game.
As a test set, 400 new images from the RIAEWELC database were employed and evenly distributed among the four classes for
classification. The generalization assessment of the model was conducted using two image datasets. First, the GDXray dataset [35] was
employed. This dataset comprises 68 images in the welding category, necessitating manual cropping and annotation, Fig. 2. The
images were resized to 224x224x3 pixels while preserving the three-color channels in PNG format. A total of 100 images were ob­
tained, Table 1.
Additionally, a private dataset consisting of radiographic images of welds was utilized. These images constitute the historical
archive of the Defectoscopy and Welding Technical Services Company (CENEX) in Cienfuegos, Cuba. Collected over more than 10
years, these images were digitized under non-uniform conditions. Challenges associated with this dataset include low contrast,
inconsistent gray distribution, noise, and uneven illumination, Fig. 3.
The images were resized to 224x224x3, retaining all three color channels. To prepare input for the proposed neural network, the
region of interest was extracted from the original images, which had dimensions of 9280x6944x3. A total of 600 images were obtained
for this dataset, and certified specialists performed labeling from CENEX.
To carry out the experiments, a Google Colab instance with an Nvidia Tesla T4 graphics processor was used to accelerate the
processing of the Python code and to take advantage of the Keras and TensorFlow machine learning libraries in the training and
evaluation of the model. This platform was used due to its ease of use and availability of computational resources, allowing the ex­
periments to run efficiently and without additional costs. This allowed a higher accuracy of the model and a decrease in training time
compared to traditional methods.

2.2. ResNet50

ResNet50 is a very effective neural network architecture for image classification and other computer vision problems [28,70–74].
Being pre-trained on large image datasets, ResNet50 can provide high performance without the need to train a neural network from
scratch; it is further used to save training time and to aid the effective generalization of the models built by learning from visual
patterns and features of the image set it was initially trained on. The use of ResNet50 as a feature extractor in a deep convolutional
network for the classification of four types of weld defects in radiographic images is proposed in this paper.

2.3. CNN model

In constructing the proposed CNN, taking ResNet50 as the base model, the last classification layer was removed and replaced by
four new fully connected layers. The first layer added is a 2D Global Average Pooling layer, which inputs the output of the last
convolutional layer of the pre-entered ResNet50 network. Next, a dense layer of 512 neurons with a ReLU activation function is used to
reduce dimensionality and learn more abstract and complex features from the images. To avoid overfitting, a Dropout layer was
applied at a rate of 50 % after the dense layer. Dropout randomly "deactivates" neurons during training, which helps to regularize the
model and reduce dependency on specific features, Fig. 4.
Finally, a last dense layer was added with four neurons, corresponding to the four classes of weld defects to be classified. This layer
uses a softmax activation function, which assigns probabilities to each class and allows the classification to be performed.
In addition to the architectural modifications mentioned above, the input images were subjected to pre-processing before being fed
to the modified CNN. Image pre-processing was performed to ensure compatibility with the ResNet50 base model and improve the
input data quality. Image pre-processing involves several steps, Fig. 5. First, the images were converted into tensors, a numerical

Fig. 1. Images of defect types in the RIAWELC dataset [37].

3
D. Palma-Ramírez et al. Heliyon 10 (2024) e30590

Fig. 2. Images of defect types in the GDXray dataset [35].

Table 1
Comparison of weld defects in the three data collections: RIAWELC, GDXray, and Private Dataset.
Weld defect types Number of images Sample images Size of images

RIAWELC Crack 1600 224 x 224


Porosity
Lack of Penetration
No defect

GDXray 100

Private Dataset 600

Fig. 3. Sample images of porosity defects and lack of penetration in the private image set.

representation that the neural network can process. The images were in their original format of 224x224 pixels with three color
channels (red, green, and blue), so no further size adjustment was necessary.
Normalization was then applied to the image tensors. This involved dividing each tensor value by the standard deviation of the
values in the dataset. This normalization helped to scale the pixel values and ensured that the features learned by the model were more
stable and consistent during training. Once the images were converted into tensors and normalized, further pre-processing steps were
performed, such as data augmentation, which consisted of applying random transformations such as rotation, resizing, cropping,

4
D. Palma-Ramírez et al. Heliyon 10 (2024) e30590

Fig. 4. Neural network structure.

Fig. 5. Pre-processing of input images to the CNN network, feature extraction, and classification.

zooming, and horizontal flipping. These transformations generated additional versions of the images during training to improve the
model’s generalizability and reduce the risk of overfitting. Thus, the initial dataset of 1600 images was expanded to 8000 images.
The modified CNN could extract relevant features and learn from the pre-processed data during training by pre-processing the
images in this way, Fig. 5. These pre-processing steps ensured that the images were in a suitable format, with the dimensions
224x224x3, and contained information relevant to the classification task of weld defect detection classification in radiographic images.
Once the classification layers were modified, the previous layers were frozen to avoid changing the general features already learned
by the network. This way, a base model was obtained with 1051140 trainable parameters out of 24638852. These layers were then re-
trained and adapted to the specific task using the images selected from the RIAWELC dataset. This allowed the network features to be
matched to the study of classifying the weld defects in the radiographic images. Table 2 reflects the hyperparameters fitted for this
model.
For fine-tuning, the Adam optimizer was used with a learning rate of 0.01 in 50 training epochs with a batch size of 32 and with the
epoch steps calculated as a function of the number of images to adjust the number of iterations required to train the model. Also, a
dropout rate of 0.5 was used, a regularization technique that randomly deactivates some of the neurons during training to avoid
overdependence on specific features in the training data [75,76]. After fitting the adapted layers, we trained the whole model,
including the previously frozen layers, with a low learning rate of 0.0001. This way, a model with improved capability for classifying
weld defects in radiographic images was achieved using transfer learning for feature extraction and fine-tuning for task-specific
adaptation.

Table 2
Hyperparameters used in the training phase of the model.
Parameters Values

Batch Size 32
Optimizer Adam
Epochs 50
Learning rate 0.01
Loss function Categorical Cross-entropy
Data Augmentation Yes
Number of Classes 4

5
D. Palma-Ramírez et al. Heliyon 10 (2024) e30590

2.4. Evaluation of the model

Metrics commonly used to evaluate a neural network model are.

1. Accuracy: As shown in expression (1), it represents the proportion of correctly classified samples.
(TP + TN)
Accuracy = (1)
(TP + TN + FP + FN)

2. Recall: Represented in expression (2), it indicates the proportion of positive samples the model correctly identifies.
TP
Recall = (2)
(TP + FN)

3. Precision: Expressed in expression (3), it signifies the proportion of samples classified as positive that are genuinely positive.
TP
Precision = (3)
(TP + FP)

4. F1-Score: Expressed as expression (4), it combines precision and recall into a single measure.
2(Pr ecision ∗ Recall)
F1 − Score = (4)
(Precision + Recall)

Where, TP = True Positive; TN = True Negative; FP = False Positive and FN = False Negative.

3. Results and discussion

Table 3 presents the results obtained during the training, validation, and testing phase on the RIAWELC image dataset. This table
compiles the metrics and values obtained at each evaluation stage, providing an overview of the model’s performance in accuracy, loss,
and other relevant metrics. The results in this table reflect the model’s performance in classifying images within the RIAWELC dataset
and serve as a benchmark for evaluating the model’s effectiveness on the specific task.
The achieved accuracy in the testing phase, with a value of 98.75 %, expression (1), demonstrates the model’s ability to perform
precise classification on the evaluated dataset. The precision and sensitivity values obtained in the test set, also at 98.5 %, expressions
(2) and (3), indicate that the neural network correctly identifies true positives, i.e., welding defects.
The F1-Score is a measure that combines precision and sensitivity, providing an overall model performance measure. In this
research work, the obtained F1-Score was 99.13 %, expression (4), indicating a balance between the accuracy and sensitivity of the
model and surpassing those reported in previous research. This is crucial to ensure the reliability and effectiveness of the model in
defect detection and classification.
Although there are differences in architectures, amount of training data, and validation, it is essential to note that the results of this
research exceed those reported by Totino et al. [37], who used the same RIAWELC dataset and obtained 93.3 % accuracy. The superior
performance of the proposed model can be attributed to several factors. First, the structure selected to build the model has proven
effective in weld defect classification. Additionally, the model’s performance was improved by optimizing the hyperparameters.
Fig. 6 shows the accuracy curves for the training and validation sets. The training curve reaches an accuracy of 98.60 %, while the

Table 3
Values of the metrics obtained in the model’s training,
validation, and testing on the RIAWELC dataset.
Metrics Values

Train Accuracy (1) 98.60 %


Train Loss 0.145
Validation Accuracy (1) 98.13 %
Validation Loss 0.169
Test Accuracy (1) 98.75 %
Test Loss 0.1397
Test Precision (3) 98.75 %
Test Recall (2) 98.75 %
F1-Score Test (4) 99.13 %

6
D. Palma-Ramírez et al. Heliyon 10 (2024) e30590

validation curve reaches 98.13 % at epoch 23, where they begin to converge. These values indicate that the model fits the data well and
has learned enough from it. Furthermore, it is observed that accuracy improves rapidly during the first epochs, suggesting effective
learning of the model.
This rapid initial improvement indicates the model’s generalizability and ability to learn representative patterns. It is important to
note that the difference between the accuracy values of the training and validation sets is minimal, which can be attributed to loss
during training. The similarity in accuracy between the two sets also supports the model’s reliability in classifying weld defects.
The observed consistency in the model metrics, where both training and validation accuracy converge to a similar value, suggests
that the model is not experiencing overfitting. Overfitting occurs when a model fits the training data too well and needs help
generalizing to new data. When a model shows a significant gap between training and validation accuracy, it indicates overfitting. This
means that the model has memorized the training data instead of learning general patterns and features that can be applied to new
data. However, in this case, by observing a close convergence between the training and validation metrics, it is inferred that the model
has learned effectively and has managed to generalize the knowledge acquired during training to previously unseen data.
Fig. 7 shows the model’s loss behavior in the training and validation phases. A gradual decrease of the loss values is observed for
both sets. The loss reached during training is 0.145, while in the validation set, it is 0.169. The curves converge at epoch 23 and do not
show a significant decrease after that, showing a similar behavior as in Fig. 6. As the number of training epochs increases, the loss
approaches 0, indicating that the model effectively learns and reduces the discrepancy between predictions and accurate labels.
The normalized confusion matrix is used to evaluate the performance of a classification model. This matrix represents the rela­
tionship between the actual classes and the classes predicted by the model. The normalized confusion matrix shows the frequency of
correct and incorrect model predictions for each class, normalized by the total number of instances in that class. For the analysis of the
model’s performance on the RIAWELC test set, the normalized confusion matrix is shown in Fig. 8. The main diagonal of the
normalized confusion matrix indicates the proportion of correctly classified images for each class. In this case, it can be observed that
three classes (crack, pore, and no defect) have a value of 1.0 on the main diagonal, which means that the model correctly classified 100
% of these images. However, it is noted that a small percentage of images from the crack class were incorrectly classified (5 % error) as
lack of penetration, which may indicate a weakness of the model in recognizing specific patterns or features of that class.
From a mechanical point of view, classifying cracks as non-penetration may be less critical if the cracks in question are superficial
and do not compromise structural integrity in strength or durability. In cases where the cracks do not penetrate deeply into the material
and do not affect its ability to withstand significant loads, the impact of the classification error may be less. However, it is essential to
note that accurate assessment depends on the specific nature of the cracks and the loading conditions applied to the structure.
However, evaluating the model’s performance on different datasets is crucial to determine its ability to accurately generalize and
classify radiographic images of welds that were not seen during training. Table 4 presents the results of using this CNN model on
different test sets, including RIAWELC, GDXray, and our private dataset. Analyzing such samples by the model enables the verification
of its ability to adapt to images with diverse characteristics and assesses its generalizability. This additional evaluation of diverse
datasets will contribute to a deeper understanding of the model’s performance under realistic conditions and strengthen its practical
utility in defect classification within radiographic images of welds.
This analysis facilitates the evaluation of the model’s performance in classifying the four types of weld defects across various
datasets. From the series of RIAWELC radiographic images, 400 new images not used for training or validation were chosen; an ac­
curacy of 98.75 % was achieved. The GDXray set requires manual cropping and annotation work, and from these, 100 evenly balanced
images were selected, achieving an accuracy of 90.25 %. Finally, 600 radiographic images with low quality, low contrast, and uneven
illumination were used. These images were divided into equal parts, and an accuracy of 75.83 % was obtained, which validates its use
in industrial environments where there is no practical technique for digitizing radiographic plates. The proposed model demonstrates
its use for data sets with high-quality and high-density images and other datasets where images are not uniformly digitized.
In the testing phase, the proposed model reaches the lowest accuracy on its dataset, where the quality of the images is poor, but
obtains an accuracy of 75.83 %, which can be improved with other pre-processing techniques. Class balancing eliminated the majority
class bias so the model could be classified well. The highest accuracy value was obtained when evaluating the model performance on
the RIAWELC set with 98.75 %. Assessing the different datasets allowed us to ensure the model’s generalization and balance to a
specific set of images. The results on the RIAWELC database show an improvement concerning those obtained by D. Mery et al. [34] on
the same dataset using the SqueezNet network, getting a 93.3 % test accuracy. However, these authors use more data, which may
influence the model’s performance.

4. Conclusions

This paper proposed a new CNN model based on the ResNet50 architecture to classify four types of weld defects in radiographic
images (pore, crack, lack of penetration, and no defect). Regularization-stratified cross-validation and data augmentation techniques
were applied to prevent network overfitting and increase generalizability. Performance was tested on the GDXray test set with a test
accuracy of 90.25 %, on the RIAWELC set with 98.75 %, and on a private dataset with low-quality images with 75.83 %. The model
demonstrated its generalizability by adapting to different image sets. It was evidenced that the proposed neural network, when trained
with high-density images in the RIAWELC dataset, can be used to classify welding defects in different types of radiographic image
quality of welded parts. It is possible to use this network to classify weld defects in noisy and low-contrast image sets.
Given the results, future adjustments to the model are suggested to improve classification and increase overall accuracy. This may
involve refining the algorithms, expanding the training data, or exploring alternative approaches to address pattern detection in the
"crack" class and improve classification in noisy, low-contrast images. Therefore, future model tuning is suggested for better

7
D. Palma-Ramírez et al. Heliyon 10 (2024) e30590

Fig. 6. Accuracy curves during training and validation of the proposed model.

Fig. 7. Loss behavior curves during training and validation of the proposed model.

Fig. 8. Normalized confusion matrix for the evaluation of the model in the RIAWELC set in the test phase.

8
D. Palma-Ramírez et al. Heliyon 10 (2024) e30590

Table 4
Model performance on each test data set.
GDXray RIAWELC Private Dataset

Number of images 100 400 600


Size of images 224 x 224 224 x 224 224 x 224
Test Accuracy 90.25 % 98.75 % 75.83 %

classification and greater accuracy. The image dataset should be increased in quantity and variability to optimize the tuning of the
neural network hyperparameters and achieve more accurate results, particularly for classifying low-contrast and noisy images.

Funding

None.

Data availability statement

The authors confirm that no extra supporting data are available for this article.
No additional information is available for this paper.

CRediT authorship contribution statement

Dayana Palma-Ramírez: Validation, Supervision, Project administration, Methodology, Investigation, Conceptualization.


Bárbara D. Ross-Veitía: Investigation, Funding acquisition, Formal analysis, Data curation. Pablo Font-Ariosa: Validation, Super­
vision, Resources, Project administration, Methodology. Alejandro Espinel-Hernández: Writing – original draft, Visualization,
Validation, Supervision, Software, Resources, Investigation, Data curation, Conceptualization. Angel Sanchez-Roca: Supervision,
Software, Resources, Funding acquisition, Formal analysis, Data curation, Conceptualization. Hipólito Carvajal-Fals: Supervision,
Resources, Project administration, Investigation, Funding acquisition, Formal analysis. José R. Nuñez-Alvarez: Writing – review &
editing, Writing – original draft, Visualization, Validation, Formal analysis, Conceptualization. Hernan Hernández-Herrera: Writing
– review & editing, Writing – original draft, Visualization, Validation, Methodology.

Declaration of competing interest

The authors declare that they have no known competing financial interests or personal relationships that could have appeared to
influence the work reported in this paper.

References

[1] T.J. McPheron, R.M. Stwalley, Engineering Challenges Associated with Welding Field Repairs, Engineering Principles-Welding and Residual Stresses, 2022,
https://doi.org/10.5772/intechopen.104263.
[2] M. Pérez de la Parte, et al., Effect of zinc coating on delay nugget formation in dissimilar DP600-AISI304 welded joints obtained by the resistance spot welding
process, Int. J. Adv. Des. Manuf. Technol. 120 (2022) 1877–1887, https://doi.org/10.1007/s00170-022-08849-2.
[3] D. Varshney, K. Kumar, Application and use of different aluminium alloys with respect to workability, strength and welding parameter optimization, Ain Shams
Eng. J. 12 (1) (2021) 1143–1152, https://doi.org/10.1016/j.asej.2020.05.013.
[4] Q. Wang, et al., A tutorial on deep learning-based data analytics in manufacturing through a welding case study, J. Manuf. Process. 63 (2021) 2–13, https://doi.
org/10.1016/j.jmapro.2020.04.044.
[5] J. Deepak, et al., Non-destructive testing (NDT) techniques for low carbon steel welded joints: a review and experimental study, Mater. Today: Proc. 44 (5)
(2021) 3732–3737, https://doi.org/10.1016/j.matpr.2020.11.578.
[6] S.K. Dwivedi, M. Vishwakarma, A. Soni, Advances and researches on non destructive testing: a review, Mater. Today: Proc. 5 (2) (2018) 3690–3698, https://doi.
org/10.1016/j.matpr.2017.11.620.
[7] M. Shaloo, et al., A review of non-destructive testing (NDT) techniques for defect detection: application to fusion welding and future Wire arc additive
manufacturing processes, Materials 15 (10) (2022) 3697, https://doi.org/10.3390/ma15103697.
[8] S. Eckel, et al., Radiographic film system classification and noise characterisation by a camera-based digitisation procedure, NDT E Int. 111 (2020) 102241,
https://doi.org/10.1016/j.ndteint.2020.102241.
[9] J. Szusta, et al., Effect of welding process parameters on the strength of dissimilar joints of S355 and Strenx 700 steels used in the Manufacture of Agricultural
Machinery, Materials 16 (21) (2023) 6963, https://doi.org/10.3390/ma16216963.
[10] W. Hou, et al., Review on computer aided weld defect detection from radiography images, Appl. Sci. 10 (5) (2020) 1878, https://doi.org/10.3390/
app10051878.
[11] E. Yahaghi, M. Mirzapour, A. Movafeghi, Comparison of traditional and adaptive multi-scale products thresholding for enhancing the radiographs of welded
object, Eur. Phys. J. Plus 136 (7) (2021) 744, https://doi.org/10.1140/epjp/s13360-021-01733-0.
[12] M. Prunella, et al., Deep learning for automatic vision-based recognition of industrial surface defects: a survey, IEEE Access 11 (2023) 43370–43423, https://
doi.org/10.1109/ACCESS.2023.3271748.
[13] W. Dai, et al., Deep learning assisted vision inspection of resistance spot welds, J. Manuf. Process. 62 (2021) 262–274, https://doi.org/10.1016/j.
jmapro.2020.12.015.
[14] M.K. Ferguson, et al., Detection and segmentation of manufacturing defects with convolutional neural networks and transfer learning, Smart and sustainable
manufacturing systems 2 (1) (2018) 20180033, https://doi.org/10.48550/arXiv.1808.02518.
[15] Y. Han, J. Fan, X. Yang, A structured light vision sensor for on-line weld bead measurement and weld quality inspection, Int. J. Adv. Des. Manuf. Technol. 106
(2020) 2065–2078, https://doi.org/10.1007/s00170-019-04450-2.

9
D. Palma-Ramírez et al. Heliyon 10 (2024) e30590

[16] D. Wang, et al., Deep network-assisted quality inspection of laser welding on power Battery, Sensors 23 (21) (2023) 8894, https://doi.org/10.3390/s23218894.
[17] D.P. Ramírez, et al., Pore segmentation in industrial radiographic images using adaptive thresholding and Morphological analysis, Trends in Agricultural and
Environmental Sciences (2023). https://orcid.org/0000-0002-6485-9231.
[18] D.M. Hermosilla, et al., Shallow convolutional network excel for classifying motor imagery EEG in BCI applications, IEEE Access 9 (2021) 98275–98286,
https://doi.org/10.1109/ACCESS.2021.3091399.
[19] D. Mery, et al., Pattern recognition in the automatic inspection of aluminium castings, Insight-Non-Destructive Testing and Condition Monitoring 45 (7) (2023)
475–483, https://doi.org/10.1784/insi.45.7.475.54452.
[20] T. W. Liao, D.-M. Li, Y.-M. Li, Detection of welding flaws from radiographic images with fuzzy clustering methods, Fuzzy Set Syst., 108 (2) (199) 145-158.
https://doi.org/10.1016/S0165-0114(97)00307-2.
[21] T.W. Liao, Classification of welding flaw types with fuzzy expert systems, Expert Syst. Appl. 25 (1) (2003) 101–111, https://doi.org/10.1016/S0957-4174(03)
00010-1.
[22] R.R. da Silva, et al., Pattern recognition of weld defects detected by radiographic test, NDT E Int. 37 (6) (2004) 461–470, https://doi.org/10.1016/j.
ndteint.2003.12.004.
[23] K. Carvajal, et al., Neural network method for failure detection with skewed class distribution, Insight-Non-Destructive Testing and Condition Monitoring 46 (7)
(2004) 399–402, https://doi.org/10.1784/insi.46.7.399.55578.
[24] L. Yang, et al., Inspection of welding defect based on multi-feature fusion and a convolutional network, J. Nondestr. Eval. 40 (2021) 1–11, https://doi.org/
10.1007/s10921-021-00823-4.
[25] S. Wang, Automatic detection and classification of steel surface defect using deep convolutional neural networks, Metals 11 (3) (2021) 388, https://doi.org/
10.3390/met11030388.
[26] D. Say, et al., Automated categorization of multiclass welding defects using the x-ray image augmentation and convolutional neural network, Sensors 23 (14)
(2023) 6422, https://doi.org/10.3390/s23146422.
[27] Ross Veitía, Deep Learning for Quality Prediction in Dissimilar Spot Welding DP600-Aisi304, Using a Convolutional Neural Network and Infrared Image
Processing, 2020, https://doi.org/10.46354/i3m.2020.emss.057.
[28] R. Zhang, et al., Research on an ultrasonic detection method for weld defects based on neural network architecture search, Measurement 221 (2023) 113483,
https://doi.org/10.1016/j.measurement.2023.113483.
[29] Z. Li, F. Liu, W. Yang, S. Peng, J. Zhou, A survey of convolutional neural networks: analysis, applications, and Prospects, IEEE Transact. Neural Networks Learn.
Syst. 33 (12) (2022) 6999–7019, https://doi.org/10.1109/TNNLS.2021.3084827.
[30] K. Lee, et al., Review on the recent welding research with application of CNN-based deep learning part II: model evaluation and visualizations, Journal of
Welding and Joining 39 (1) (2021) 20–26, https://doi.org/10.5781/JWJ.2021.39.1.2.
[31] R.V. Patil, Y.P. Reddy, Multiform weld joint flaws detection and classification by sagacious artificial neural network technique, Int. J. Adv. Des. Manuf. Technol.
125 (1) (2023) 913–943, https://doi.org/10.1007/s00170-022-10719-w.
[32] S. Kumaresan, et al., Deep learning based Simple CNN weld defects classification using optimization technique, Russ. J. Nondestr. Test. 58 (6) (2022) 499–509,
https://doi.org/10.1134/S1061830922060109.
[33] A. Singh, K. Raj, T. Kumar, S. Verma, A.M. Roy, Deep learning-based Cost-effective and Responsive Robot for autism Treatment, Drones 7 (2) (2023) 81, https://
doi.org/10.3390/drones7020081.
[34] S. Perri, et al., Welding defects classification through a convolutional neural network, Manufacturing Letters 35 (2023) 29–32, https://doi.org/10.1016/j.
mfglet.2022.11.006.
[35] D. Mery, et al., The database of X-ray images for nondestructive testing, J. Nondestr. Eval. 34 (4) (2015) 42, https://doi.org/10.1007/s10921-015-0315-7.
[36] W. Guo, H. Qu, L. Liang, WDXI: the dataset of X-ray image for weld defects, in: 2018 14th International Conference on Natural Computation, Fuzzy Systems and
Knowledge Discovery (ICNC-FSKD), 2018, pp. 1051–1055, https://doi.org/10.1109/FSKD.2018.8686975.
[37] B. Totino, F. Spagnolo, S. Perri, RIAWELC: a Novel dataset of radiographic images for automatic weld defects classification, International Journal of Electrical
and Computer Engineering Research 3 (1) (2023) 13–17, https://doi.org/10.53375/ijecer.2023.320.
[38] S. Kumaresan, et al., Transfer learning with CNN for classification of weld defect, IEEE Access 9 (2021) 95097–95108, https://doi.org/10.1109/
ACCESS.2021.3093487.
[39] H.G. Hernandez-Palma, et al., Technological tools based on artificial intelligence in the sugar industry: a Bibliometric analysis and future Perspectives for energy
efficiency, LADEE 4 (2) (2023) 49–64, https://doi.org/10.17981/ladee.04.02.2023.4.
[40] S. Kumaresan, et al., Deep learning-based weld defect classification using VGG16 transfer learning adaptive fine-tuning, Int. J. Interact. Des. Manuf. 17 (6)
(2023) 2999–3010, https://doi.org/10.1007/s12008-023-01327-3.
[41] M. Hussain, J.J. Bird, D.R. Faria, A study on CNN transfer learning for image classification, Advances in Computational Intelligence Systems, Adv. Intell. Syst.
Comput. 840 (2019), https://doi.org/10.1007/978-3-319-97982-3_16. Springer, Cham.
[42] W. Jiao, et al., End-to-end prediction of weld penetration: a deep learning and transfer learning based method, J. Manuf. Process. 63 (2021) 191–197, https://
doi.org/10.1016/j.jmapro.2020.01.044.
[43] D.D. Kumar, et al., Semi-supervised transfer learning-based automatic weld defect detection and visual inspection, Eng. Struct. 292 (2023) 116580, https://doi.
org/10.1016/j.engstruct.2023.116580.
[44] H. Pan, et al., A new image recognition and classification method combining transfer learning algorithm and mobilenet model for welding defects, IEEE Access 8
(2020) 119951–119960, https://doi.org/10.1109/ACCESS.2020.3005450.
[45] A.M. Roy, R. Boseand J. Bhaduri, A fast accurate fine-grain object detection model based on YOLOv4 deep neural network, Neural Comput. Appl. 34 (5) (2022)
3895–3921, https://doi.org/10.1007/s00521-021-06651-x.
[46] A.M. Roy, J. Bhaduri, DenseSPH-YOLOv5: an automated damage detection model based on DenseNet and Swin-Transformer prediction head-enabled YOLOv5
with attention mechanism, Adv. Eng. Inf. 56 (2023) 102007, https://doi.org/10.1016/j.aei.2023.102007.
[47] D.V. Rayudu, J.F. Roseline, Accurate Weather Forecasting for Rainfall prediction using artificial neural network compared with deep learning neural network,
in: 2023 International Conference on Artificial Intelligence and Knowledge Discovery in Concurrent Engineering (ICECONF), 2023, pp. 1–6, https://doi.org/
10.1109/ICECONF57129.2023.10084252. Chennai, India.
[48] X. Wang, X. Yu, Understanding the effect of transfer learning on the automatic welding defect detection, NDT E Int. 134 (2023) 102784, https://doi.org/
10.1016/j.ndteint.2022.102784.
[49] J. Nuñez, et al., Design of a fuzzy controller for a hybrid generation system, IOP Conf. Ser. Mater. Sci. Eng. 844 (2020) 012017, https://doi.org/10.1088/1757-
899X/844/1/012017.
[50] B. Jiang, S. Chen, B. Wang, B. Luo, MGLNN: Semi-supervised learning via Multiple Graph Cooperative learning neural networks, Neural Network. 153 (2022)
204–214, https://doi.org/10.1016/j.neunet.2022.05.024.
[51] Z. Li, Y. Li, Y. Liu, P. Wang, R. Lu, H.B. Gooi, Deep learning based densely connected network for load Forecasting, IEEE Trans. Power Syst. 36 (4) (2021)
2829–2840, https://doi.org/10.1109/TPWRS.2020.3048359.
[52] A. Singh, L. Bruzzone, Mono- and Dual-Regulated Contractive-Expansive-Contractive deep convolutional networks for classification of Multispectral Remote
sensing images, Geosci. Rem. Sens. Lett. IEEE 19 (2022) 5513605, https://doi.org/10.1109/LGRS.2022.3211861, 1-5.
[53] B. Liu, et al., Weld Defect Images Classification with Vgg16-Based Neural Network, International Forum on Digital TV and Wireless Multimedia
Communications, vol. 815, Springer, Singapore, 2018, https://doi.org/10.1007/978-981-10-8108-8_20.
[54] L. Mohanasundari, Performance analysis of weld image classification using modified Resnet CNN architecture, Turkish Journal of Computer and Mathematics
Education (TURCOMAT) 12 (2) (2021) 2260–2266. https://turcomat.org/index.php/turkbilmat/article/view/1943.
[55] V.A. Golodov, A.A. Mittseva, Weld segmentation and defect detection in radiographic images of Pipe welds, in: 2019 International Russian Automation
Conference (RusAutoCon), Sochi, Russia, 2019, pp. 1–6, https://doi.org/10.1109/RUSAUTOCON.2019.8867734.

10
D. Palma-Ramírez et al. Heliyon 10 (2024) e30590

[56] J. Deng, W. Dong, R. Socher, L.-J. Li, Kai Li, Li Fei-Fei, ImageNet: a large-scale hierarchical image database, in: 2009 IEEE Conference on Computer Vision and
Pattern Recognition, 2009, pp. 248–255, https://doi.org/10.1109/CVPR.2009.5206848. Miami, FL, USA.
[57] D. Chauveau, Review of NDT and process monitoring techniques useable to produce high-quality parts by welding or additive manufacturing, Weld. World 62
(5) (2018) 1097–1118, https://doi.org/10.1007/s40194-018-0609-3.
[58] J. Rao, et al., Non-destructive testing of metal-based additively manufactured parts and processes: a review, Virtual Phys. Prototyp. 18 (1) (2023) e2266658,
https://doi.org/10.1080/17452759.2023.2266658.
[59] D. Priyasudana, et al., Double side friction stir welding effect on mechanical properties and corrosion rate of aluminum alloy AA6061, Heliyon 9 (2) (2023)
e13366, https://doi.org/10.1016/j.heliyon.2023.e13366.
[60] M.N. Ilman, Microstructure and mechanical properties of friction stir spot welded AA5052-H112 aluminum alloy, Heliyon 7 (2) (2021) e06009, https://doi.org/
10.1016/j.heliyon.2021.e06009.
[61] S. Shin, C. Jin, J. Yuand S. Rhee, Real-time detection of weld defects for automated welding process base on deep neural network, Metals 10 (3) (2020)
389–2020, https://doi.org/10.3390/met10030389.
[62] D.-Y. Kim, et al., Weld fatigue behavior of gas metal arc welded steel sheets based on porosity and gap size, Int. J. Adv. Des. Manuf. Technol. 124 (3) (2023)
1141–1153, https://doi.org/10.1007/s00170-022-10567-8.
[63] T. Fujii, N. Ogasawara, K. Tohgo, Y. Shimamura, Monte Carlo simulation of stress corrosion cracking in welded metal with surface defects and life estimation,
Int. J. Mech. Sci. 270 (2024) 109079, https://doi.org/10.1016/j.ijmecsci.2024.109079.
[64] Y. Li, et al., Detection model of invisible weld defects using magneto-optical imaging induced by rotating magnetic field, in: 2019 IEEE 15th International
Conference on Automation Science and Engineering (CASE), 2019, pp. 1–5, https://doi.org/10.1109/COASE.2019.8843280. Vancouver, BC, Canada.
[65] H.M. Rai, K. Chatterjee, Hybrid CNN-LSTM deep learning model and ensemble technique for automatic detection of myocardial infarction using big ECG data,
Appl. Intell. 52 (5) (2022) 5366–5384, https://doi.org/10.1007/s10489-021-02696-6.
[66] Z. Zhu, K. Lin, A.K. Jain, J. Zhou, Transfer learning in deep Reinforcement learning: a survey, IEEE Trans. Pattern Anal. Mach. Intell. 45 11 (2023)
13344–13362, https://doi.org/10.1109/TPAMI.2023.3292075.
[67] W. Wang, G. Wen, Z. Zheng, Design of deep learning Mixed Language short Text Sentiment classification system based on CNN algorithm, in: 2022 IEEE 2nd
International Conference on Mobile Networks and Wireless Communications (ICMNWC), Tumkur, 2022, pp. 1–5, https://doi.org/10.1109/
ICMNWC56175.2022.10031786. Karnataka, India.
[68] K. Pal, B.V. Patel, Data classification with K-fold cross validation and Holdout accuracy estimation methods with 5 different machine learning techniques, in:
2020 Fourth International Conference on Computing Methodologies and Communication (ICCMC), 2020, pp. 83–87, https://doi.org/10.1109/
ICCMC48092.2020.ICCMC-00016. Erode, India.
[69] S. Szeghalmy, A. Fazekas, A comparative study of the use of stratified cross-validation and distribution-balanced stratified cross-validation in imbalanced
learning, Sensors 23 (4) (2023) 2333, https://doi.org/10.3390/s23042333.
[70] J.-R. Lee, K.-W. Ng, Y.-J. Yoong, Face and facial expressions recognition system for blind people using ResNet50 architecture and CNN, Journal of Informatics
and Web Engineering 2 (2) (2023) 284–298, https://doi.org/10.33093/jiwe.2023.2.2.20.
[71] S. Mascarenhas, M. Agarwal, A comparison between VGG16, VGG19 and ResNet50 architecture frameworks for Image Classification, in: 2021 International
Conference on Disruptive Technologies for Multi-Disciplinary Research and Applications (CENTCON), 2021, pp. 96–99, https://doi.org/10.1109/
CENTCON52345.2021.9687944. Bengaluru, India.
[72] S. Tofigh, M.O. Ahmad, M.N.S. Swamy, A low-Complexity modified ThiNet algorithm for Pruning convolutional neural networks, IEEE Signal Process. Lett. 29
(2022) 1012–1016, https://doi.org/10.1109/LSP.2022.3164328.
[73] X. Huang, High resolution Remote sensing image classification based on deep transfer learning and multi feature network, IEEE Access 11 (2023)
110075–110085, https://doi.org/10.1109/ACCESS.2023.3320792.
[74] D. Milanés-Hermosilla, et al., Monte Carlo dropout for uncertainty estimation and motor imagery classification, Sensors 21 (21) (2021) 7241, https://doi.org/
10.3390/s21217241.
[75] P. Dileep, D. Das, P.K. Bora, Dense layer dropout based CNN architecture for automatic Modulation classification, in: 2020 National Conference on
Communications (NCC), Kharagpur, India, 2020, pp. 1–5, https://doi.org/10.1109/NCC48643.2020.9055989.
[76] Y. Liu, et al., Guided dropout: Improving deep networks without increased computation, Intelligent Automation & Soft Computing 36 (3) (2023) 2519–2528,
https://doi.org/10.32604/iasc.2023.033286.

11

View publication stats

You might also like