0% found this document useful (0 votes)
56 views32 pages

Zhao 2020

This paper evaluates deep learning algorithms for machinery fault diagnosis using various datasets and parameters. It collects public datasets, tests models with different data preprocessing and augmentation, and analyzes issues like imbalance and interpretability. Code is released for open benchmarking.

Uploaded by

Equipe GM
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
56 views32 pages

Zhao 2020

This paper evaluates deep learning algorithms for machinery fault diagnosis using various datasets and parameters. It collects public datasets, tests models with different data preprocessing and augmentation, and analyzes issues like imbalance and interpretability. Code is released for open benchmarking.

Uploaded by

Equipe GM
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 32

ISA Transactions xxx (xxxx) xxx

Contents lists available at ScienceDirect

ISA Transactions
journal homepage: www.elsevier.com/locate/isatrans

Research article

Deep learning algorithms for rotating machinery intelligent diagnosis:


An open source benchmark study

Zhibin Zhao, Tianfu Li, Jingyao Wu, Chuang Sun, Shibin Wang, Ruqiang Yan ,
Xuefeng Chen
School of Mechanical Engineering, Xi’an Jiaotong University, Xi’an, China

article info a b s t r a c t

Article history: Rotating machinery intelligent diagnosis based on deep learning (DL) has gone through tremendous
Received 25 January 2020 progress, which can help reduce costly breakdowns. However, different datasets and hyper-parameters
Received in revised form 30 July 2020 are recommended to be used, and few open source codes are publicly available, resulting in unfair
Accepted 7 August 2020
comparisons and ineffective improvement. To address these issues, we perform a comprehensive
Available online xxxx
evaluation of four models, including multi-layer perception (MLP), auto-encoder (AE), convolutional
Keywords: neural network (CNN), and recurrent neural network (RNN), with seven datasets to provide a
Deep learning benchmark study. We first gather nine publicly available datasets and give a comprehensive benchmark
Machinery intelligent diagnosis study of DL-based models with two data split strategies, five input formats, three normalization
Open source codes methods, and four augmentation methods. Second, we integrate the whole evaluation codes into a
Benchmark study
code library and release it to the public for better comparisons. Third, we use specific-designed cases
to point out the existing issues, including class imbalance, generalization ability, interpretability, few-
shot learning, and model selection. Finally, we release a unified code framework for comparing and
testing models fairly and quickly, emphasize the importance of open source codes, provide the baseline
accuracy (a lower bound), and discuss existing issues in this field. The code library is available at:
https://github.com/ZhaoZhibin/DL-based-Intelligent-Diagnosis-Benchmark.
© 2020 Published by Elsevier Ltd on behalf of ISA.

1. Introduction empirical. Thus, how to perform diagnosis more precisely and


efficiently is still a challenging problem.
Rotating machinery, as key mechanical equipment in the mod- DL, as a booming data mining technique, has swept many
ern industry, is chronically running in a complex environment fields including computer vision (CV) [4,5], natural language pro-
with elevated temperature, fatigue, and heavy load. Generated cessing (NLP) [6–8], and other fields [9,10]. In 2006, the concept
faults might cause severe accidents, resulting in enormous eco- of DL was first introduced via proposing deep belief network
nomic loss and casualties. Intelligent diagnosis, as a key ingre- (DBN) [11]. In 2013, MIT Technology Review ranked the DL tech-
dient of prognostics health management (PHM), which is one of nology as the top ten breakthrough technologies [12]. In 2015,
the most essential systems in a wide range of rotating machinery, a review [13] published in nature stated that DL allows com-
such as helicopter, aero-engine, wind turbine, and high-speed putational models composed of multiple processing layers to
train, is designed to detect faults effectively. Traditional intel- learn data representations with multiple levels of the abstraction.
ligent diagnosis methods mainly consist of feature extraction Due to its strong representation learning ability, DL is well-
using signal processing methods [1,2] and fault classification [3] suited to data analysis and classification. Therefore, in the field
using machine learning approaches, which have made consider- of intelligent diagnosis, many researchers have applied DL-based
able progress. However, facing with heterogeneous massive data,
techniques, such as multi-layer perception (MLP), auto-encoder
feature extraction methods and mapping abilities from signals to
(AE), convolutional neural network (CNN), DBN, and recurrent
conditions that are designed and chosen by experts, to a great
neural network (RNN) to boost the performance. However, differ-
extent depending on prior knowledge, are time-consuming and
ent researchers often recommended to use different inputs (such
as time domain input, frequency domain input, time–frequency
∗ Corresponding author.
domain input, etc.) and set different hyper-parameters (such
E-mail addresses: [email protected] (Z. Zhao),
as the learning rate, the batch size, the network architecture,
[email protected] (T. Li), [email protected] (J. Wu),
[email protected] (C. Sun), [email protected] (S. Wang), etc.). Unfortunately, a few authors made their codes available
[email protected] (R. Yan), [email protected] (X. Chen). for comparisons, resulting in unfair comparisons and ineffective

https://doi.org/10.1016/j.isatra.2020.08.010
0019-0578/© 2020 Published by Elsevier Ltd on behalf of ISA.

Please cite this article as: Z. Zhao, T. Li, J. Wu et al., Deep learning algorithms for rotating machinery intelligent diagnosis: An open source benchmark study. ISA Transactions
(2020), https://doi.org/10.1016/j.isatra.2020.08.010.
2 Z. Zhao, T. Li, J. Wu et al. / ISA Transactions xxx (xxxx) xxx

improvement. To address this problem, it is crucial to evalu-


ate and compare different DL-based intelligent diagnosis algo-
rithms to provide the benchmark study and open source codes,
thereby helping further studies to propose more persuasive and
appropriate algorithms.
For comprehensive performance comparisons, it is necessary
to collect different datasets in a library and evaluate the perfor-
mance of algorithms for different datasets on a unified platform.
In addition, one common issue in intelligent diagnosis is data
split, and researchers often use the random split strategy. This
strategy is dangerous since if the preparation process exists any
overlap, the evaluation of classification algorithms will have test
leakage [14]. As for industrial data, they are rarely random and Fig. 1. The relationship between the number of published papers and publica-
are always sequential (they might contain trends in the time tion years covering the last six years (as of April 2020). The basic descriptor is
‘‘TI = ((deep OR autoencoder OR convolutional network* OR neural network*)
domain), and it is more appropriate to split data according to time AND (fault OR condition monitoring OR health management OR intelligent
sequences (we simply call it order split). Conversely, if we ran- diagnosis))’’.
domly split the data, it might be possible for diagnosis algorithms
to record the future patterns, which might cause another pitfall
with test leakage.
(2) Benchmark accuracy and existing issues. We evaluate var-
To address these problems, in this paper, we collect nine
ious DL-based intelligent diagnosis algorithms for seven
publicly available datasets and discuss whether it is suitable for
datasets and provide the benchmark accuracy to make
intelligent diagnosis. After that, we evaluate DL-based intelli-
future studies more comparable. Based on the benchmark
gent diagnosis algorithms from different perspectives including
study, we highlight some evaluation results which are very
the data preparation for all datasets and the whole evaluation
important for comparing or testing new models. We also
framework with different input formats, normalization methods,
discuss the existing issues, including class imbalance, gen-
data split ways, augmentation methods, and DL-based models.
eralization ability, interpretability, few-shot learning, and
Based on the benchmark study, we highlight some evaluation
model selection.
results which are very important for comparing or testing new
(3) Open source codes. To emphasize the reproducibility of
models. First, not all datasets are suitable for comparing the
DL-based intelligent diagnosis algorithms, we release the
classification effectiveness of the proposed methods since basic
code library for the better comparisons. Meanwhile, it is
models can achieve very high accuracy on these datasets, like
a unified intelligent diagnosis library, which retains an
CWRU and XJTU-SY. Second, the frequency domain input can
extended interface for everyone to load their own datasets
achieve the highest accuracy in all datasets, so researchers should
and models to carry out new studies.
first try to use the frequency domain as the input. Third, it is
not necessary for CNN models to get the best results in all cases, The outlines of this paper are listed as follows: In Section 2,
and we also should consider the overfitting problem. Fourth, we give a brief review of the recent development of DL-based
when the accuracy of datasets is not very high, data augmen- intelligent diagnosis algorithms. Then, Sections 3–9 discuss the
tation methods improve the performance of models, especially evaluation algorithms, datasets, data preprocessing, data aug-
for the time domain input. Thus, more effective data augmen- mentation, data split, evaluation methodologies, and evaluation
tation methods need to be investigated. Finally, in some cases, results, respectively. After that, Section 10 makes some further
it may be more suitable for splitting the datasets according to discussions and the results, followed by conclusions in Section 11.
time sequences (order split) since the random split may pro-
vide a virtually high accuracy. We also release a code library to 2. Brief review
evaluate DL-based intelligent diagnosis algorithms and provide
the benchmark accuracy (a lower bound) to avoid useless im- Recently, DL has become a promising method in a large scope
provement. Meanwhile, we use specific-designed cases to discuss of fields, and a huge amount of papers related to DL have been
existing issues, including class imbalance, generalization ability, published since 2012. This paper mainly focuses on a benchmark
interpretability, few-shot learning, and model selection. Through study of intelligent diagnosis, rather than providing a comprehen-
these works, we aim to allow comparisons fairer and quicker, sive review on DL for other fields. Some famous DL researchers
emphasize the importance of open source codes, and provide
have published more professional references, and readers could
deep discussions of existing issues. To the best of our knowledge,
refer to [13,15].
this is the first work to comprehensively perform the bench-
Due to the efforts of many researchers, DL has become one
mark study and release the code library to the public. The code
of the most popular data-driven methods to perform intelligent
library is available at: https://github.com/ZhaoZhibin/DL-based-
diagnosis. In general, DL-based methods can extract represen-
Intelligent-Diagnosis-Benchmark.
tative features adaptively without any manual intervention and
The main contributions of this paper can be summarized as
can achieve a higher accuracy than traditional machine learning
three aspects:
algorithms in most of the tasks when the dataset is large enough.
(1) Various datasets and data preparing . We gather nine We conducted a literature search using Web of Science with a
publicly available datasets and give a detailed discussion database called web of science core collection. As shown in Fig. 1,
about its adaptability. For data preparation, we first dis- we can observe that the number of published papers related to
cuss different kinds of input formats and normalization DL-based intelligent algorithms increases year by year.
methods, and then we perform different data augmenta- Another interesting observation is that many review papers on
tion approaches to clarify that they have not been fully this topic have been published in the recent four years. Therefore,
investigated. We also discuss the way of data split and state in this paper, we first summarize the main contents of different
that it may be more appropriate to split data according to review papers, allowing readers who just enter this field to find
time sequences. suitable review papers quickly.

Please cite this article as: Z. Zhao, T. Li, J. Wu et al., Deep learning algorithms for rotating machinery intelligent diagnosis: An open source benchmark study. ISA Transactions
(2020), https://doi.org/10.1016/j.isatra.2020.08.010.
Z. Zhao, T. Li, J. Wu et al. / ISA Transactions xxx (xxxx) xxx 3

In bearing fault diagnosis, Li et al. [16] provided a systematic by [41,42]. For imbalanced learning, generation adversarial net-
review of fuzzy formalisms including combination with other work (GAN) was used to combine with AE models to generate
machine learning algorithms. Hoang et al. [17] provided a com- new labeled samples in [43–45]. In [46], a model called deep
prehensive review of three popular DL algorithms (AE, DBN, and Laplacian AE (DLapAE) was proposed by introducing the Laplacian
CNN) for bearing fault diagnosis. Zhang et al. [18] systemati- regularization to improve the generalization performance. For
cally reviewed the machine learning and DL-based algorithms for transfer learning, the pretrained and fine-tuned approach was
bearing fault diagnosis and also provided a comparison of the applied to AE models to realize the knowledge transfer in [47,48].
classification accuracy of CWRU with different DL-based methods. Domain adaptation was also used to transfer the knowledge
Hamadache et al. [19] reviewed different fault modes of rolling learned by AE models to the target domain in [49,50].
element bearings and described various health indexes for PHM. In CNN models, for model improvement, different input types,
Meanwhile, it also provided a survey of artificial intelligence (AI) such as time–frequency images [51], vibration spectrum im-
methods for PHM including shallow learning and DL.
ages [52], infrared thermal images [53], and two-dimensional
In rotating machinery intelligent diagnosis, Ali et al. [20] pro-
images [54], were used as the inputs of CNN models. Multiple
vided a review of AI-based methods using acoustic emission data
wavelet regularizations [55], data augmentation methods [56],
for rotating machinery condition monitoring. Liu et al. [21] re-
and information fusion technology [57] were also applied to
viewed Al-based approaches including k-nearest neighbors
improve the performance of CNN models. Hand-crafted features
(KNN), support vector machine (SVM), artificial neural networks
(ANN), Naive Bayes, and DL for fault diagnosis of rotating ma- were combined with CNN features to boost the performance
chinery. Wei et al. [22] summarized early fault diagnosis of in [58]. For imbalanced learning, GAN was also used to combine
gears, bearings, and rotors through signal processing methods with CNN models to generate new labeled samples in [59,60].
(adaptive decomposition methods, wavelet transform, and sparse Focal loss, which can deal with severe imbalanced problems,
decomposition) and AI-based methods (KNN, neural network, and was used by [61] to allow CNN models to learn discriminative
SVM). features. For transfer learning, the pretrained and fine-tuned
In machinery condition monitoring, Zhao et al. [23] and Duan approach was used to leverage the prior knowledge from the
et al. [24] reviewed diagnosis and prognosis of mechanical equip- source task in [62–64]. Domain adaptation methods were also
ment based on DL algorithms such as DBN and CNN. Zhang applied to allow CNN models to learn transferable features in [65,
et al. [25] reviewed computational intelligent approaches in- 66]. In addition, layer-wise relevance propagation was also used
cluding ANN, evolutionary algorithms, fuzzy logic, and SVM for to understand how CNN models learn to distinguish different
machinery fault diagnosis. Zhao et al. [26] reviewed data-driven patterns [67].
machine health monitoring through DL methods (AE, DBN, CNN, Beyond that, complex wavelet packet energy moment en-
and RNN) and provided the data and codes (in Keras) about an tropy [68] and the gray wolf optimizer algorithm [69] were
experimental study. Lei et al. [27] presented a systematical review combined with an enhanced deep gated recurrent unit to improve
to cover the development of intelligent diagnosis following the the security of rotating machinery. Joint distribution adaptation
progress of machine learning and DL models and offer a future was embedded into LSTM to realize learning transferable features
perspective called transfer learning theories. in [70]. DBN models were also modified in [71–73] to improve
In addition, Nasiri et al. [28] surveyed the state-of-the-art the diagnosis performance of rotating machinery. Deep reinforce-
AI-based approaches for fracture mechanics and provided the ment learning was also used in intelligent fault diagnosis for
accuracy comparisons achieved by different machine learning rotating machinery in [74,75]. Meanwhile, a deep graph convo-
algorithms for mechanical fault detection. Tian et al. [29] sur-
lutional network (DGCN) was first applied to rolling bearing fault
veyed different modes of traction induction motor fault and their
diagnosis based on acoustic signals [76].
diagnosis algorithms including model-based methods and AI-
Although a large body of DL-based methods and many related
based methods. Khan et al. [30] provided a comprehensive re-
reviews have been published in the field of intelligent diagno-
view of AI for system health management and emphasized the
sis, few studies thoroughly evaluate various DL-based intelligent
trend of DL-based methods with limitations and benefits. Stetco
et al. [31] reviewed machine learning approaches applied to diagnosis algorithms for most of the publicly available datasets,
wind turbine condition monitoring and made a discussion of provide the benchmark accuracy, and release the code library
the possibility for future research. Ellefsen et al. [32] reviewed for complete evaluation procedures. For example, a simple code
four well-established DL algorithms including AE, CNN, DBN, and written in Keras was published in [26], which is not compre-
long short-term memory network (LSTM) for PHM applications hensive enough for different datasets and models. The accuracy
and discussed challenges for the future studies, especially in the comparisons were provided in [18,28] according to existing pa-
field of PHM in autonomous ships. AI-based algorithms (tradi- pers, but they were not comprehensive enough due to different
tional machine learning algorithms and DL-based approaches) configurations. Therefore, this paper is intended to make up for
and applications (smart sensors, intelligent manufacturing, PHM, this gap and emphasize the importance of open source codes and
and cyber–physical systems) were reviewed in [33–36] for smart the benchmark study.
manufacturing and manufacturing diagnosis.
Due to the fact that there are already many review papers
3. Evaluation algorithms
covering DL-based rotating machinery intelligent diagnosis pub-
lished before 2020, we further review most of related papers
published in 2020 and summarize their main contributions to fill It is impossible to cover all the published models since there is
the void. currently no open source community in this field. Therefore, we
In AE models, for model improvement, AE models were com- switch to test the performance of four models (including MLP,
bined with some other data preprocessing methods, such as AE, CNN, and RNN) via embedding some advanced techniques.
singular value decomposition [37] and nonlinear frequency spec- It should be noted that DBN is also another commonly used DL
trum [38]. The ensemble learning strategy was also used to boost model in intelligent diagnosis, but we do not add it to this code
the performance of AE models in [39,40]. Meanwhile, the semi- library due to that the fact the training way of DBN is much
supervised learning methods were also embedded into AE models different from those four models.

Please cite this article as: Z. Zhao, T. Li, J. Wu et al., Deep learning algorithms for rotating machinery intelligent diagnosis: An open source benchmark study. ISA Transactions
(2020), https://doi.org/10.1016/j.isatra.2020.08.010.
4 Z. Zhao, T. Li, J. Wu et al. / ISA Transactions xxx (xxxx) xxx

shown in Fig. 3 change adaptively. Specifically, the network struc-


tures of DAE and SAE are the same with AE, and their differences
lie in the loss function and inputs. DAE takes an input corrupted
with noise and is trained to reconstruct the clean version of the
input; SAE uses a MSE loss regularized with a sparsity constraint
(the Kullback–Leibler divergence is often used) to train the AE
model. During the training of AE and its derivatives, the encoder
and decoder are trained jointly to get the low-dimension features
of data. After that, the encoder and classifier are trained jointly
using the softmax cross-entropy loss for the classification task.
The details of AE and its derivatives are shown in Fig. 3. In Fig. 3,
the MSE loss means the mean square error loss defined in (3),
Fig. 2. The structure of MLP.
Conv means the convolutional layer, ConvT means the transposed
convolutional (e.g. inverse convolution) layer, and the KLP loss
means the Kullback–Leibler divergence loss.
3.1. MLP
3.3. CNN
MLP [77], which was a fully connected network with multiple
hidden layers, was proposed in 1987 as the prototype of ANN. CNN [80] was first proposed in 1997 and the proposed net-
With such a simple structure, MLP can complete some easy work was also called LeNet. CNN is a specialized kind of neural
classification tasks such as MNIST. However, as the task becomes network for processing data that has a known grid-like topology.
more complex, it would be hard to train MLP because of the Sparse interactions, parameter sharing, and equivalent represen-
huge amount of parameters. MLP with five fully connected layers tations are realized with convolution and pooling operations on
and five batch normalization layers is used in this paper for one CNN. In 2012, AlexNet [4] won the title in the ImageNet compe-
dimension (1D) input data. The structure and parameters of this tition by far surpassing the second place, and CNN has attracted
model are shown in Fig. 2. Besides, in Fig. 2, FC means the fully wide attention. Besides, in 2016, ResNet [81] was proposed and
connected layer, BN means the Batch Normalization layer, and CE its classification accuracy exceeded the human baseline. In this
loss means the softmax cross-entropy loss. paper, we design 5 layers 1D CNN and 2D CNN for 1D input
data and 2D input data, respectively, and also adapt three well
3.2. AE known CNN models (LeNet, ResNet18, and AlexNet) for two types
of the input data. The details of CNN and its derivatives are
AE was first proposed in 2006, as a method for dimensionality shown in Fig. 4. In Fig. 4, MaxPool means the Max Pooling layer,
reduction. It can reduce the dimensionality of the input data AdaptiveMaxPool means the Adaptive Max Pooling layer, and
while retaining most of the information. AE consists of an encoder Dropout means the Dropout layer.
and a decoder, which tries to reconstruct the input from the out- As shown in Fig. 4, CNN mainly contains three kinds of layers,
put of an encoder, and the reconstruction error is used as a loss including the convolutional layer, the maxpooling layer, and the
function. In detail, an encoder takes x as an input and transforms classifier layer. Convolutional and maxpooling layers are utilized
it into a hidden representation h which can be formulated as: to perform feature learning, and the classifier layer classifies
learned features into different classes. For an input x, the con-
h = φ (W · x + b) (1) volutional layer can be defined as a multiplication with a filter
φ (·) denotes the nonlinear activation function (in this paper, we kernel w , and the final feature map after the nonlinear activation
use ReLU), and W and b represent the weight and bias required can be formulated as:
to learn, respectively. After that, a decoder generates the output hlk = φ (wkl ∗ x + blk ) (4)
x′ from the hidden representation h, which can be formulated as:
where ∗ denotes the convolution operator. hlk , wkl , and blk rep-
resent the obtained feature map, the weight and bias of kth
x′ = φ (W ′ · h + b′ ) (2) convolutional kernel of lth layer, respectively.
The maxpooling layer is set behind the convolution layer
where W ′ and b′ represent the weight and bias required to learn,
to extract the most significant local information in each fea-
respectively. For traditional AE, the mean squared error (MSE)
ture map and to reduce the dimension of obtained features. The
loss is often used as a loss function, shown as follows:
maxpooling layer can be defined as:
N
1 ∑
LMSE = ∥xi − x′i ∥22 (3) zkl = down(hlk ; s) (5)
N
i=1
where down(·) denotes the down-sampling function of the max-
where xi and x′i are the ith sample and its approximation, respec- pooling layer, zkl is the output feature map of the maxpooling
tively, and N is the number of data samples. In practice, we often layer, and s is the pooling size.
stack multiple layers in both encoder and decoder to produce After a number of stacked convolutional and pooling layers,
better learning results. the extracted high-level features of input data are input into the
Subsequently, various derivatives of AE were proposed by classifier layer. In this paper, a fully connected layer is used to
different researchers, such as denoising auto-encoder (DAE) [78] map features into different classes.
and sparse auto-encoder (SAE) [79]. In this paper, we design a
deep AE and its derivatives for 1D input data (using the fully 3.4. RNN
connected layer) and two dimension (2D) input data (using the
convolutional layer), respectively. Considering different features RNN can describe the temporal dynamic behavior and is very
of neural networks, the structures and hyper-parameters of them suitable for dealing with time series. However, RNN often exists

Please cite this article as: Z. Zhao, T. Li, J. Wu et al., Deep learning algorithms for rotating machinery intelligent diagnosis: An open source benchmark study. ISA Transactions
(2020), https://doi.org/10.1016/j.isatra.2020.08.010.
Z. Zhao, T. Li, J. Wu et al. / ISA Transactions xxx (xxxx) xxx 5

Fig. 3. The structure of AE and its derivatives.

the gradient vanishing and exploding problems during the train-


ing procedure. To overcome these problems, LSTM was proposed
in 1997 [82] for processing continual input streams and has
σi (ht −1 , xt ) = σ (Wi · [ht −1 , xt ] + bi ), i = 1, 2, 3 (7)
made great success in various fields. Bi-directional LSTM (BiLSTM) where Wi and bi represent the weight and bias, respectively. σ (·)
can capture bidirectional dependencies over long distances and denotes the sigmoid function. Similarly, the tanh layer replaces
learn to remember and forget information selectively. We utilize the sigmoid function with the tanh function. In addition, the
BiLSTM as the representation of RNN to deal with two types of output hidden state ht can be calculate as follows:
input data (1D and 2D) for the classification task. The details of
1D BiLSTM and 2D BiLSTM are shown in Fig. 5. Besides, in Fig. 5, ht = σ3 (ht −1 , xt ) ⊗ tanh(Ct ) (8)
Transpose means transposing the channel and feature dimensions Many repeating cells are linked together to form a LSTM block,
of the input data, and BiLSTM Block means the BiLSTM layer. designed for capturing both long-term and short-term dependen-
The structure of a LSTM cell is shown in Fig. 6, including the cies. The BiLSTM layer is the combining of forward and backward
forget gate layer σ1 , the input gate layer σ2 , the output gate layer LSTM blocks, in which information is bidirectionally transmitted.
σ3 , and the tanh layer. Firstly, the hidden state of a last cell ht −1 For each input, information of the whole time series can be used
and the current input xt are fed into the forget gate layer to decide simultaneously.
whether we should forget the last cell state Ct −1 . Secondly, ht −1
and xt are fed into the input gate layer and the tanh layer to 4. Datasets
decide values we want to update. Thirdly, ht −1 and xt are fed into
the output layer to decide what we should export for the last cell. In the field of intelligent diagnosis, publicly available datasets
Based on the structure shown in Fig. 6, the output cell state Ct can have not been investigated in depth. Actually, for comprehen-
be calculated as follows: sive performance comparisons, it is important to gather different
Ct = σ1 (ht −1 , xt ) ⊗ Ct −1 ⊕ σ2 (ht −1 , xt ) ⊗ tanh(ht −1 , xt ) (6) kinds of representative datasets. We collected nine commonly
used datasets which all have specific labels and explanations
where ⊗ and ⊕ denote the element-wise multiplication and ad- in addition to the PHM 2012 bearing dataset and Intelligent
dition, respectively. Meanwhile, σi (ht −1 , xt ) is defined as follows: Maintenance Systems (IMS) bearing dataset, so PHM 2012 and

Please cite this article as: Z. Zhao, T. Li, J. Wu et al., Deep learning algorithms for rotating machinery intelligent diagnosis: An open source benchmark study. ISA Transactions
(2020), https://doi.org/10.1016/j.isatra.2020.08.010.
6 Z. Zhao, T. Li, J. Wu et al. / ISA Transactions xxx (xxxx) xxx

Fig. 4. The structure of CNN and its derivatives.

Fig. 5. The structure of 1D BiLSTM and 2D BiLSTM.

Please cite this article as: Z. Zhao, T. Li, J. Wu et al., Deep learning algorithms for rotating machinery intelligent diagnosis: An open source benchmark study. ISA Transactions
(2020), https://doi.org/10.1016/j.isatra.2020.08.010.
Z. Zhao, T. Li, J. Wu et al. / ISA Transactions xxx (xxxx) xxx 7

Table 2
Detailed description of MFPT datasets.
Fault Mode Description
Health State Fault-free bearing working at 270 lbs
Outer ring 1 Outer ring fault bearing working at 25 lbs
Outer ring 2 Outer ring fault bearing working at 50 lbs
Outer ring 3 Outer ring fault bearing working at 100 lbs
Outer ring 4 Outer ring fault bearing working at 150 lbs
Outer ring 5 Outer ring fault bearing working at 200 lbs
Outer ring 6 Outer ring fault bearing working at 250 lbs
Outer ring 7 Outer ring fault bearing working at 300 lbs
Outer ring 1 Inner ring fault bearing working at 0 lbs
Inner ring 2 Inner ring fault bearing working at 50 lbs
Inner ring 3 Inner ring fault bearing working at 100 lbs
Inner ring 4 Inner ring fault bearing working at 150 lbs
Inner ring 5 Inner ring fault bearing working at 200 lbs
Inner ring 6 Inner ring fault bearing working at 250 lbs
Fig. 6. The structure of a LSTM cell. Inner ring 7 Inner ring fault bearing working at 300 lbs

Table 1
Detailed description of CWRU datasets.
4.3. PU bearing dataset
Fault Mode Description
Health State the normal bearing at 1791 rpm and 0 HP
Inner ring 1 0.007 inch inner ring fault at 1797 rpm and 0 HP Paderborn University (PU) datasets were provided by the
Inner ring 2 0.014 inch inner ring fault at 1797 rpm and 0 HP Paderborn University Bearing Data Center [85,86], and PU
Inner ring 3 0.021 inch inner ring fault at 1797 rpm and 0 HP
datasets consisted of 32 sets of current signals and vibration
Rolling Element 1 0.007 inch rolling element fault at 1797 rpm and 0 HP
Rolling Element 2 0.014 inch rolling element fault at 1797 rpm and 0 HP signals. As shown in Table 3, bearings were divided into: (1) six
Rolling Element 3 0.021 inch rolling element fault at 1797 rpm and 0 HP undamaged bearings; (2) twelve artificially damaged bearings;
Outer ring 1 0.007 inch outer ring fault at 1797rpm and 0 HP (3) fourteen bearings with real damages caused by accelerated
Outer ring 2 0.014 inch outer ring fault at 1797rpm and 0 HP
Outer ring 3 0.021 inch outer ring fault at 1797rpm and 0 HP
lifetime tests. Each dataset was collected under four working
conditions as shown in Table 4.
In this paper, since using all the data would cause huge com-
putational time, we only used the data collected from real dam-
IMS are not suitable for fault classification that requires labels. To
aged bearings (including KA04, KA15, KA16, KA22, KA30, KB23,
sum up, this paper uses seven datasets to verify the performance
KB24, KB27, KI14, KI16, KI17, KI18, and KI22) under the working
of models introduced in Section 3. The description of all these
datasets is listed as follows. condition N15_M07_F10 to carry out the performance verifica-
tion. Since KI04 was the same as KI14 completely shown in
4.1. CWRU bearing dataset Table 3, we deleted KI04 and the total number of classes was
thirteen. Besides, only vibration signals were used for testing the
Case Western Reserve University (CWRU) datasets were pro- models.
vided by the Case Western Reserve University Bearing Data Cen-
ter [83]. Vibration signals were collected at 12 kHz or 48 kHz for
normal bearings and damaged bearings with single-point defects 4.4. UoC gear fault dataset
under four different motor loads. Within each working condition,
single-point faults were introduced with fault diameters of 0.007, University of Connecticut (UoC) gear fault datasets were pro-
0.014, and 0.021 inches on the rolling element, the inner ring, vided by the University of Connecticut [87], and UoC datasets
and the outer ring, respectively. In this paper, we used the data were collected at 20 kHz. In this dataset, nine different gear
collected from the drive end, and the sampling frequency was
conditions were introduced to the pinions on the input shaft, in-
equivalent to 12 kHz. In Table 1, one healthy bearing and three
cluding healthy condition, missing tooth, root crack, spalling, and
fault modes, including the inner ring fault, the rolling element
fault, and the outer ring fault, were classified into ten categories chipping tip with 5 different levels of severity. All the collected
(one health state and 9 fault states) according to different fault datasets were used and classified into nine categories (one health
sizes. state and eight fault states) to test the performance.

4.2. MFPT bearing dataset


4.5. XJTU-SY bearing dataset
Machinery Failure Prevention Technology (MFPT) datasets
were provided by Society for Machinery Failure Prevention Tech- XJTU-SY bearing datasets were provided by the Institute of
nology [84]. MFPT datasets consisted of three bearing datasets: Design Science and Basic Component at Xi’an Jiaotong University
(1) a baseline dataset sampled at 97 656 Hz for six seconds in and the Changxing Sumyoung Technology Co. [88,89]. XJTU-SY
each file; (2) seven outer ring fault datasets sampled at 48 828 Hz datasets consisted of fifteen bearings run-to-failure data under
for three seconds in each file; (3) seven inner ring fault datasets
three different working conditions. The dataset was collected at
sampled at 48 828 Hz for three seconds in each file; (4) some
2.56 kHz. A total of 32 768 data points were recorded, and the
other datasets which were not used in this paper (more detailed
information can be referred to the website of MFPT datasets [84]). sampling period was equal to one minute. The details of bearing
In Table 2, one health state bearing and two fault bearings lifetime and fault elements were shown in Table 5. In this paper,
including the inner ring fault and the rolling element fault were we used all the data described in Table 6 and the total number
classified into 15 categories (one health state and 14 fault states) of classes was fifteen. It should be noticed that we used collected
according to different loads. data at the end of run-to-failure experiments.

Please cite this article as: Z. Zhao, T. Li, J. Wu et al., Deep learning algorithms for rotating machinery intelligent diagnosis: An open source benchmark study. ISA Transactions
(2020), https://doi.org/10.1016/j.isatra.2020.08.010.
8 Z. Zhao, T. Li, J. Wu et al. / ISA Transactions xxx (xxxx) xxx

Table 3
Detailed description of PU datasets (S: single damage; R: repetitive damage; M: multiple damage)
Bearing Code Fault Mode Description Bearing Code Fault Mode Description
K001 Health state Run-in 50 h before KI07 Artificial inner ring Made by electric
test fault (Level 2) engraver
K002 Health state Run-in 19 h before KI08 Artificial inner ring Made by electric
test fault (Level 2) engraver
K003 Health state Run-in 1 h before test KA04 Outer ring damage Caused by fatigue and
(single point + S + pitting
Level 1)
K004 Health state Run-in 5 h before test KA15 Outer ring damage Caused by plastic
(single point + S + deform and
Level 1) indentation
K005 Health state Run-in 10 h before KA16 Outer ring damage Caused by fatigue and
test (single point + R + pitting
Level 2)
K006 Health state Run-in 16 h before KA22 Outer ring damage Caused by fatigue and
test (single point + S + pitting
Level 1)
KA01 Artificial outer ring Made by EDM KA30 Outer ring damage Caused by plastic
fault (Level 1) (distributed + R + deform and
Level 1) indentation
KA03 Artificial outer ring Made by electric KB23 Outer ring and inner Caused by fatigue and
fault (Level 2) engraver ring damage (single pitting
point + M + Level 2)
KA05 Artificial outer ring Made by electric KB24 Outer ring and inner Caused by fatigue and
fault (Level 1) engraver ring damage pitting
(distributed + M +
Level 3)
KA06 Artificial outer ring Made by electric KB27 Outer ring and inner Caused by plastic
fault (Level 2) engraver ring damage deform and
(distributed + M + indentation
Level 1)
KA07 Artificial outer ring Made by drilling KI04 Inner ring damage Caused by fatigue and
fault (Level 1) (single point + M + pitting
Level 1)
KA08 Artificial outer ring Made by drilling KI14 Inner ring damage Caused by fatigue and
fault (Level 2) (single point + M + pitting
Level 1)
KA09 Artificial outer ring Made by drilling KI16 Inner ring damage Caused by fatigue and
fault (Level 2) (single point + S + pitting
Level 3)
KI01 Artificial inner ring Made by EDM KI17 Inner ring damage Caused by fatigue and
fault (Level 1) (single point + R + pitting
Level 1)
KI03 Artificial inner ring Made by electric KI18 Inner ring damage Caused by fatigue and
fault (Level 1) engraver (single point + S + pitting
Level 2)
KI05 Artificial inner ring Made by electric KI21 Inner ring damage Caused by fatigue and
fault (Level 1) engraver (single point + S + pitting
Level 1)

configuration (RS-LC) set to be 20 Hz - 0 V and 30 Hz - 2 V


Table 4 shown in Table 6. The total number of classes was equal to twenty
Four working conditions of PU datasets. according to Table 6 under different working conditions. Within
No. Rotating Load torque Radial force Name of setting
each file, there were eight rows of vibration signals, and we used
speed (rpm) (Nm) (N)
the second row of vibration signals.
0 1500 0.7 1000 N15_M07_F10
1 900 0.7 1000 N09_M07_F10
2 1500 0.1 1000 N15_M01_F10
4.7. JNU bearing dataset
3 1500 0.7 400 N15_M07_F04

Jiangnan University (JNU) bearing datasets were provided by


Jiangnan University [92,93]. JNU datasets consisted of three bear-
4.6. SEU gearbox dataset
ing vibration datasets with different rotating speeds, and the data
Southeast University (SEU) gearbox datasets were provided by were collected at 50 kHz. As shown in Table 7, JNU datasets con-
Southeast University [90,91]. SEU datasets contained two sub- tained one health state and three fault modes which include inner
datasets, including a bearing dataset and a gear dataset, which ring fault, outer ring fault, and rolling element fault. Therefore,
were both acquired on drivetrain dynamic simulator (DDS). There the total number of classes was equal to twelve according to
were two kinds of working conditions with rotating speed-load different working conditions.

Please cite this article as: Z. Zhao, T. Li, J. Wu et al., Deep learning algorithms for rotating machinery intelligent diagnosis: An open source benchmark study. ISA Transactions
(2020), https://doi.org/10.1016/j.isatra.2020.08.010.
Z. Zhao, T. Li, J. Wu et al. / ISA Transactions xxx (xxxx) xxx 9

Table 5
Detailed description of XJTU-SY datasets.
Condition File Lifetime Fault element
Bearing 1_1 2h 3 min Outer ring
Bearing 1_2 2h 41 min Outer ring
Speed: 35 Hz
Bearing 1_3 2h 38 min Outer ring
Load: 12 kN
Bearing 1_4 2h 2 min Cage
Bearing 1_5 52 min Inner ring and Outer ring
Bearing 2_1 8h 11 min Inner ring
Speed: 37.5 Bearing 2_2 2h 41 min Outer ring
Hz Load: 11 Bearing 2_3 8h 53 min Cage
kN Bearing 2_4 42 min Outer ring
Bearing 2_5 5h 39 min Outer ring
Bearing 3_1 42 h 18 min Outer ring
Bearing 3_2 41 h 36 min Inner ring, Rolling element, Cage, and Outer ring
Speed: 40 Hz
Bearing 3_3 6 h 11 min Inner ring
Load: 10 kN
Bearing 3_4 25 h 15 min Inner ring
Bearing 3_5 1 h 54 min Outer ring

Table 6 paper, the effects of five input types and three normalization
Detailed description of SEU datasets. methods on the performance of DL models are discussed.
Fault Mode RS-LC Fault Mode RS-LC
Health Gear 20 Hz–0 V Health Bearing 20 Hz–0 V 5.1. Input types
Health Gear 30 Hz–2 V Health Bearing 30 Hz–2 V
Chipped Tooth 20 Hz–0 V Inner ring 20 Hz–0 V
Chipped Tooth 30 Hz–2 V Inner ring 30 Hz–2 V Many researchers use signal processing methods to map the
Missing Tooth 20 Hz–0 V Outer ring 20 Hz–0 V time series to different domains to boost the performance. How-
Missing Tooth 30 Hz–2 V Outer ring 30 Hz–2 V
ever, which input type is more suitable for intelligent diagnosis is
Root Fault 20 Hz–0 V Inner + Outer ring 20 Hz–0 V
Root Fault 30 Hz–2 V Inner + Outer ring 30 Hz–2 V still an open question. In this paper, the effects of different input
Surface Fault 20 Hz–0 V Rolling Element 20 Hz–0 V types on model performance are discussed.
Surface Fault 30 Hz–2 V Rolling Element 30 Hz–2 V
5.1.1. Time domain input
For the time domain input, vibration signals are directly used
4.8. PHM 2012 bearing dataset as the input without data preprocessing. In this paper, the length
of each sample is 1024 and the total number of samples can be
PHM 2012 bearing datasets were used for PHM IEEE 2012 Data obtained from Eq. (9). After generating samples, we take 80% of
Challenge [94,95]. In PHM 2012 datasets, seventeen run-to-failure total samples as the training set and 20% of total samples as the
datasets were provided including six training sets and eleven testing set.
testing sets. Three different loads were considered. Vibration and L
temperature signals were gathered during all those experiments. N = floor( ) (9)
Since no label on the types of failures was given, it was not used 1024
in this paper. where L is the length of each signal, N is the total samples, and
floor means rounding towards minus infinity.
4.9. IMS bearing dataset
5.1.2. Frequency domain input
IMS bearing datasets were generated by the NSF I/UCR Center For the frequency domain input, FFT is used to transform
for Intelligent Maintenance Systems [96]. IMS datasets were made each sample xi from the time domain into the frequency domain
up of three bearing datasets, and each of them contained vibra- shown in Eq. (10). After this operation, the length of data will be
tion signals of four bearings installed on the different locations. At halved and the new sample can be expressed as:
the end of the run-to-failure experiment, a defect occurred on one
of the bearings. The failure occurred in the different locations of xFFT
i = FFT(xi ) (10)
bearings. It is inappropriate to classify these failures simply using
where the operator FFT(·) represents transforming xi into the
three classes, so IMS datasets were not evaluated in this paper.
frequency domain and taking the first half of the result.
5. Data prepreocessing
5.1.3. Time–frequency domain input
The type of input data and the way of normalization have a For the time–frequency domain input, short-time Fourier
great impact on the performance of DL models. Types of input transform (STFT) is applied to each sample xi to obtain the
data determine the difficulty of feature extraction, and normal- time–frequency representation shown in Eq. (11). The Hanning
ization methods determine the difficulty of calculation. So, in this window is used and the window length is 64. After this operation,

Table 7
Detailed description of JNU datasets.
Fault Mode Rotating Fault Mode Rotating Fault Mode Rotating
Speed Speed Speed
Health State 600 rpm Health State 800 rpm Health State 1000 rpm
Inner ring 600 rpm Inner ring 800 rpm Inner ring 1000 rpm
Outer ring 600 rpm Outer ring 800 rpm Outer ring 1000 rpm
Rolling Element 600 rpm Rolling Element 800 rpm Rolling Element 1000 rpm

Please cite this article as: Z. Zhao, T. Li, J. Wu et al., Deep learning algorithms for rotating machinery intelligent diagnosis: An open source benchmark study. ISA Transactions
(2020), https://doi.org/10.1016/j.isatra.2020.08.010.
10 Z. Zhao, T. Li, J. Wu et al. / ISA Transactions xxx (xxxx) xxx

the time–frequency representation (a 33 × 33 image) will be Z-score Normalization: This normalization method can be
generated as: implemented by

xSTFT = STFT(xi ), i = 1, 2, . . . , N (11) xi − xmean


i xnormalize
i
−3
= i
, i = 1, 2, . . . , N (16)
xstd
i
where the operator SFFT(·) represents transforming xi into the
time–frequency domain. where xmean
i is the mean value of xi , and xstd
i is the standard
deviation of xi .
5.1.4. Wavelet domain input
For the wavelet domain input, continuous wavelet transform 6. Data augmentation
(CWT) is applied to each sample xi to obtain the wavelet do-
main representation shown in Eq. (12). Because CWT is time- Data augmentation is important to make the training datasets
consuming, the length of each sample xi is set to 100. After this more diverse and to alleviate the learning difficulties caused
operation, the wavelet coefficients (an 100 × 100 image) will be by small sample problems. However, data augmentation for in-
obtained as: telligent diagnosis has not been investigated in depth. The key
challenge for data augmentation is to create the label-corrected
xCWT
i = CWT(xi ), i = 1, 2, . . . , N (12) samples from existing samples, and this procedure mainly de-
where the operator CWT(·) represents transforming xi into the pends on the domain knowledge. However, it is difficult to deter-
wavelet domain. mine whether the generated samples are label-corrected. So, this
paper provides some data augmentation techniques to increase
the concerns of other scholars. In addition, these data augmenta-
5.1.5. Slicing image input
tion strategies are only a simple test and their applications still
For slicing image input, each sample xi is reshaped into a
need to be studied in depth.
32 × 32 image. After this operation, the new sample can be
denoted as:
6.1. One dimension input augmentation
Reshape
xi = Reshape(xi ), i = 1, 2, . . . , N (13)
RandomAddGaussian: this strategy randomly adds Gaussian
where the operator Reshape(·) represents reshaping xi into a
noise into the input signal formulated as follows:
32 × 32 image.
However, above data preprocessing methods have some prob- x := x + n (17)
lems for training AE models and CNN models in the following two
aspects: (1) if AE models input a large 2D signal, it will lead the where x is the 1D input signal, and n is generated by Gaussian
decoder to have difficulty in the reconstruction procedure and distribution N (0, 0.01).
RandomScale: this strategy randomly multiplies the input
the reconstruction error is very large; (2) if CNN models input
signal with a random factor which is formulated as follows:
a small 2D signal, it will make CNN unable to extract appropriate
features. x := β ∗ x (18)
Therefore, we have made a compromise on the data size ob-
tained by the above data preprocessing methods. The sizes of the where x is the 1D input signal, and β is a scaler following the
time domain and the frequency domain input are unchanged as distribution N (1, 0.01).
shown in Eqs. (9) and (10). For the AE class, sizes of all 2D inputs RandomStretch: this strategy resamples the signal into a ran-
are adjusted to 32 × 32, while for CNN models, sizes of signals dom proportion and ensures the equal length by nulling and
after CWT, STFT, and slicing image are adjusted to 300 × 300, truncating.
330 × 330, and 320 × 320, respectively. It should be noted that RandomCrop: this strategy randomly covers partial signals
input sizes of CNN models can be different since we use the which is formulated as follows:
AdaptiveMaxPooling layer to adapt different input sizes. x := mask ∗ x (19)

5.2. Normalization where x is the 1D input signal, and mask is the binary sequence
whose subsequence of random position is zero. In this paper, the
length of the subsequence is equal to 10.
Input normalization is the basic step in data preparation,
which can facilitate the subsequent data processing and accel-
erate the convergence of DL models. Therefore, we discuss the 6.2. Two dimension input augmentation
effects of three normalization methods on the performance of DL
models. RandomScale: this strategy randomly multiplies the input
Maximum–Minimum Normalization: This normalization signal with a random factor which is formulated as follows:
method can be implemented by x := β ∗ x (20)
xi − xmin where x is the 2D input signal, and β is a scaler following the
xinormalize−1 = i
, i = 1, 2, . . . , N (14)
xmax
i − xmin
i distribution N (1, 0.01).
RandomCrop: this strategy randomly covers partial signals,
where xi is the input sample, xmin
i is the minimum value in xi , and
which is formulated as follows:
xmax
i is the maximum value in xi .
[-1-1] Normalization: This normalization method can be im- x := mask ∗ x (21)
plemented by
where x is the 2D input signal, and mask is the binary sequence
xi − xmin whose subsequence of random position is zero. In this paper, the
xinormalize−2 = −1 + 2 ∗ i
, i = 1, 2, . . . , N (15)
xmax − xmin length of the subsequence is equal to 20.
i i

Please cite this article as: Z. Zhao, T. Li, J. Wu et al., Deep learning algorithms for rotating machinery intelligent diagnosis: An open source benchmark study. ISA Transactions
(2020), https://doi.org/10.1016/j.isatra.2020.08.010.
Z. Zhao, T. Li, J. Wu et al. / ISA Transactions xxx (xxxx) xxx 11

Fig. 7. Random data splitting strategy with preprocessing without overlap. Fig. 8. Another condition with the training and testing sets split as the first
step.

7. Data split

One common practice of data split in intelligent diagnosis is


the random split strategy, and the diagram of this strategy is
shown in Fig. 7. From this diagram, we stress the preprocessing
step without overlap due to the fact that if the sample prepa-
ration process exists any overlap, the evaluation of classification
algorithms may have test leakage (if users split the training set
and the testing set from the beginning of the preprocessing step,
then they can use any processing to simultaneously deal with the
training and testing sets, as shown in Fig. 8).
The formal way is that the training set is further split into
the training set and the validation set for the model selection.
Fig. 7 shows the condition of 4-fold cross-validation, and we often
use the average accuracy of 4-fold cross-validation to represent
the generalization accuracy, if there is no testing set. In this
paper, for testing convenience and time saving, we only use 1-
fold validation and use the last epoch accuracy to represent the
testing accuracy (we also list the maximum accuracy in the whole
epochs for comparisons). It is worth noting that some papers use
the maximum accuracy of the validation set, and this strategy is Fig. 9. Data split according to time sequences.
also dangerous because the validation set is used to select the
parameters accidentally.
For industrial data, they are rarely random and are always
accuracy to evaluate the performance of algorithms. The overall
sequential (they might contain trends or other temporal corre-
accuracy is defined as the number of correctly classified samples
lation). Therefore, it is more appropriate to split data according
divided by the total number of samples. The average accuracy is
to time sequences (order split). The diagram of data split strategy
defined as the average classification accuracy of each category.
according to time sequences is shown in Fig. 9. From this diagram,
Since the performance of DL-based intelligent diagnosis algo-
it can be observed that we split the training and testing sets with
rithms fluctuates during the training process, to obtain reliable
the time phase instead of splitting the data randomly. In addition,
results and show the best overall accuracy that the model can
Fig. 9 also shows the condition of 4-fold cross-validation with
achieve, we repeat each experiment five times. Four indicators
time. In the following study, we will compare the results of this
strategy with the random split strategy. are used to assess the performance of models, including the mean
and maximum values of the overall accuracy obtained by the
8. Evaluation methodology last epoch (the accuracy in the last epoch can represent the real
accuracy without any test leakage), and the mean and maximum
8.1. Evaluation metrics values of the maximal overall accuracy. For simplicity, they can
be denoted as Last-Mean, Last-Max, Best-Mean, and Best-Max.
It is a rather challenging task to evaluate the performance of
intelligent diagnosis algorithms with suitable evaluation metrics. 8.2. Experimental setting
It has three standard evaluation metrics, which have been widely
used, including the overall accuracy, the average accuracy, and In the preparation stage, we use two strategies, including
the confusion matrix. In this paper, we only use the overall random split and order split to divide the dataset into training

Please cite this article as: Z. Zhao, T. Li, J. Wu et al., Deep learning algorithms for rotating machinery intelligent diagnosis: An open source benchmark study. ISA Transactions
(2020), https://doi.org/10.1016/j.isatra.2020.08.010.
12 Z. Zhao, T. Li, J. Wu et al. / ISA Transactions xxx (xxxx) xxx

and testing sets. For random split, a sliding window is used to D. PU dataset
truncate the vibration signal without any overlap and each data The results of PU dataset are shown in Appendix from Ta-
sample contains 1024 points. After the preparation, we randomly bles A.10 to A.12. It is shown that the accuracy of CNN models is
take 80% of samples as the training set and 20% of samples as generally higher than that of AE models. Besides, the accuracy is
the testing set. For order split, we take the former 80% of time worse when using the wavelet domain as the input, while using
series as the training set and the last 20% as the testing set. FFT and STFT to process the signal allows models to achieve better
Then, in two time series, a sliding window is used to truncate the accuracy. Using Z-score normalization enables AE models and
vibration signal without any overlap, and each sample contains CNN models to achieve higher accuracy. Data augmentation does
1024 points. not help AE models improve the accuracy, while it can increase
In order to verify how input types, data normalization meth- the accuracy of CNN models. Similarly, the order split would
ods, and data split methods affect the performance of models, we heavily reduce the accuracy.
set up three configurations of experiments (shown in Table 8,
E. SEU dataset
Table 9, and Table 10.). During model training, we use Adam
The results of SEU dataset are shown in Appendix from
as the optimizer. The learning rate and the batch size of each
Tables A.13 to A.15. We can observe that when using the time
experiment are set to 0.001 and 64, respectively. Each model
domain or wavelet domain as the input, models would achieve
is trained for 100 epochs, and during the training procedure,
worse accuracy. However, using FFT to process the signal allows
model training and model testing are alternated. In addition, all models to achieve better accuracy and the accuracy of AE models
the experiments are executed under Windows 10 and Pytorch is even higher than that of CNN models. Using Z-score normal-
1.1 through running on a computer with an Intel Core i7-9700K, ization allows AE models and CNN models to achieve higher
GeForce RTX 2080Ti, and 16G RAM. accuracy. Data augmentation can improve the accuracy of both
CNN and AE models. In this case, the order split would slightly
9. Evaluation results reduce the accuracy.

In this section, we will first discuss the experimental results F. UoC dataset
of different datasets in depth. After that, the results of datasets, The results of UoC dataset are shown in Appendix from
input types, models, input normalization, data augmentation, and Tables A.16 to A.18. We can observe that most models do not
data splitting will be summarized separately. Complete results perform well in this case, and among them, the performance
are shown in Appendix and each accuracy which is larger than of AlexNet is relatively worse. Besides, using FFT to process the
95% are bold. signal allows models to achieve better accuracy, and the accuracy
of AE models is higher than that of CNN models. AE models and
9.1. Detailed analysis of different datasets CNN models with Z-score normalization would achieve higher ac-
curacy. Data augmentation can help different models improve the
final accuracy. The order split would heavily reduce the accuracy.
A. CWRU dataset
The results of CWRU dataset are shown in Appendix from
G. XJTU-SY dataset
Tables A.1 to A.3. From those results, we can observe that the The results of XJTU-SY dataset are shown in Appendix from
accuracy of CNN models is generally higher than that of AE mod- Tables A.19 to A.21. We can observe that most models perform
els. In addition, using FFT and STFT to process the signal allows well in this dataset. Besides, we can find that using FFT and
models to achieve better accuracy among five kinds of input. CNN STFT to process the signal allows models to achieve the better
models with Z-score normalization can get better accuracy while accuracy, and the accuracy of CNN models is higher than that of
using -1-1 normalization allows AE models to achieve higher ac- AE models, generally. AE models and CNN models with Z-score
curacy. Using data augmentation does not improve the accuracy normalization would achieve higher accuracy. Data augmentation
of AE models, but it can improve the accuracy of CNN models. The can help different models improve the final accuracy. The order
order split would slightly reduce the accuracy. split would quietly reduce the accuracy.

B. JNU dataset 9.2. Results of datasets


The results of JNU dataset are shown in Appendix from
Tables A.4 to A.6. As can be seen from those tables, using FFT to It can be seen from the results that with the exception of
process the raw signal allows models to achieve better accuracy the UoC dataset, the accuracy of both AE and CNN models on
among five types of input. CNN models with Z-score normal- those datasets exceeds 95%. In addition, the accuracy of CWRU,
ization can get better accuracy while using -1-1 normalization SEU and XJTU-SY datasets can reach to 100%. The accuracy of
enables AE models to achieve higher accuracy. Using data aug- UoC is much lower than others in all conditions. Besides, the
mentation can improve the accuracy of CNN models and AE diagnostic difficulty of seven datasets can be ranked according to
the number of diagnostic accuracy exceeding 95% in each dataset.
models. The order split would highly reduce the accuracy.
As shown in Fig. 10, we can split the datasets into four levels of
difficulty.
C. MFPT dataset
The results of MFPT dataset are shown in Appendix from 9.3. Results of input types
Tables A.7 to A.9. We can observe that models with time or
wavelet domain as the input would have the worse performance. In all datasets, the frequency domain input can always achieve
However, using FFT to process the signal allows models to achieve the highest accuracy followed by the time–frequency domain
better accuracy, and the accuracy of AE models is even higher input since in the frequency domain, the noise is spread over the
than CNN models in this dataset. CNN models with Z-score nor- full frequency band and the fault information is much easier to
malization can get better accuracy while using -1-1 normalization be distinguished than that in the time domain. According to the
enables the AE models to achieve higher accuracy. Using data computational load of CWT, we use the short length of samples to
augmentation can improve the accuracy of CNN models and AE perform CWT and then upsample the wavelet coefficients. These
models. The order split would heavily reduce the accuracy. steps may degrade the classification accuracy of CWT.

Please cite this article as: Z. Zhao, T. Li, J. Wu et al., Deep learning algorithms for rotating machinery intelligent diagnosis: An open source benchmark study. ISA Transactions
(2020), https://doi.org/10.1016/j.isatra.2020.08.010.
Z. Zhao, T. Li, J. Wu et al. / ISA Transactions xxx (xxxx) xxx 13

Table 8
Experiment setup 1.

of different data normalization methods also depends on the used


models and datasets. In general, Z-score normalization can make
models achieve better accuracy.

9.6. Results of data augmentation

We can conclude that when the accuracy of datasets is al-


ready high enough, data augmentation methods may slightly
degrade the performance because models have already fitted
original datasets well. More augmentation methods may change
the distribution of the original data and make the learning process
harder. However, when the accuracy of datasets is not very high,
data augmentation methods improve the performance of models,
Fig. 10. The level of dataset difficulty. especially for the time domain input. Therefore, researchers can
design other various data augmentation methods for their specific
inputs.
9.4. Results of models
9.7. Results of data splitting
From the results, we can observe that models, especially
ResNet18, can achieve the best accuracy in most of datasets When the datasets are easy to handle (CWRU and XJTU-SY),
including CWRU, JNU, PU, SEU, and XJTU-SY. However, for MFPT
the results between random split and order split are quite similar.
and UoC, models belonging to AE can perform better than other
However, the accuracy of some datasets (PU and UoC) decreases
models. This phenomenon may be caused by the size of datasets
sharply when using order split. What we should pay more at-
and the overfitting problem. Therefore, not every dataset can get
tention to is that whether randomly splitting these datasets has
better results using a more complex model.
the risk of test leakage. It may be more suitable for splitting the
9.5. Results of data normalization datasets according to time sequences to verify the performance.
According to the above discussion, we summarize the follow-
It is hard to conclude which data normalization method is the ing conclusion coming from the evaluation results. First, not all
best one, and from the results, we can observe that the accuracy datasets are suitable for comparing the classification effectiveness

Please cite this article as: Z. Zhao, T. Li, J. Wu et al., Deep learning algorithms for rotating machinery intelligent diagnosis: An open source benchmark study. ISA Transactions
(2020), https://doi.org/10.1016/j.isatra.2020.08.010.
14 Z. Zhao, T. Li, J. Wu et al. / ISA Transactions xxx (xxxx) xxx

Table 9
Experiment setup 2.

of the proposed methods since basic models can achieve very the following five issues including class imbalance, generalization
high accuracy on these datasets, like CWRU and XJTU-SY. Second, ability, interpretability, few-shot learning, and model selection
the frequency domain input can achieve the highest accuracy in using experimental cases.
all datasets, so researchers should first try to use the frequency
domain as the input. Third, it is not necessary for CNN models to
10.1. Class imbalance
get the best results in all cases, and we also should consider the
overfitting problem. Fourth, when the accuracy of datasets is not
very high, data augmentation methods improve the performance Most of measured signals are in the normal state, and only
of models, especially for the time domain input. Thus, more a few of them are in the fault state, which means that fault
effective data augmentation methods need to be investigated. modes often have different probabilities of happening. Therefore,
Finally, in some cases, it may be more suitable for splitting the the class imbalance issue will occur when using intelligent algo-
datasets according to time sequences (order split) since random rithms in real applications. Recently, although some researchers
split may provide virtually high accuracy. We also release a code have published some related papers using traditional imbalanced
library to evaluate DL-based intelligent diagnosis algorithms and learning methods [97] or GAN [98] to solve this problem, these
provide the benchmark accuracy (a lower bound) to avoid use- studies are far from enough. In this paper, PU Bearing Datasets
less improvement. Meanwhile, we use specific-designed cases to are used to simulate the class imbalance issue. In this experiment,
discuss existing issues, including class imbalance, generalization we adopt ResNet18 as the experimental model and only use two
ability, interpretability, few-shot learning, and model selection. kinds of input types (the time domain input and the frequency
Through these works, we aim to allow comparisons fairer and domain input). Besides, data augmentation methods are used and
quicker, emphasize the importance of open source codes, and the normalization method is the Z-score normalization, while the
provide deep discussions of existing issues. To the best of our dataset is randomly split. Three groups of datasets with different
knowledge, this is the first work to comprehensively perform the imbalance ratios are constructed, which are shown in Table 11.
benchmark study and release the code library to the public. As shown in Table 11, three datasets (Group1, Group2, and
Group3) are constituted with different imbalanced ratios. Group1
10. Discussion is a balanced dataset, and there is no imbalance for each state.
In real applications, it is almost impossible to let the number of
Although intelligent diagnosis algorithms can achieve high data samples be the same. We reduce the training samples of
classification accuracy in many datasets, there are still many some fault modes in Group1 to construct Group2, and then the
issues that need to be discussed. In this paper, we further discuss imbalanced classification is simulated. In Group3, the imbalanced

Please cite this article as: Z. Zhao, T. Li, J. Wu et al., Deep learning algorithms for rotating machinery intelligent diagnosis: An open source benchmark study. ISA Transactions
(2020), https://doi.org/10.1016/j.isatra.2020.08.010.
Z. Zhao, T. Li, J. Wu et al. / ISA Transactions xxx (xxxx) xxx 15

Table 10
Experiment setup 3.

Table 11 thresholding-based methods are applied in the test phase to ad-


Number of samples in three groups of imbalanced datasets.
just the decision threshold of the classifier. Besides, cost-sensitive
Fault mode Training samples Testing samples
learning methods assign different weights to different classes
Group1 Group2 Group3 Group1/2/3 to avoid the suppression of categories with a small number of
KA04 125 125 125 125 samples. In the field of intelligent diagnosis, other methods based
KA15 125 75 50 125
on physical meanings and fault attention need to be explored.
KA16 125 75 50 125
KA22 125 75 50 125
KA30 125 37 25 125
10.2. Generalization ability
KB23 125 37 25 125
KB24 125 37 25 125
KB27 125 25 6 125 Many existing intelligent algorithms perform very well on one
KI14 125 25 6 125
working condition, but the diagnostic performance tends to drop
KI16 125 25 6 125
KI17 125 12 2 125 significantly on another working condition, and here, we call
KI18 125 12 2 125 it the generalization problem. Recently, many researchers have
KI21 125 12 2 125 used algorithms based on transfer learning strategies to solve this
problem, and a comparative study with open source codes was
performed in [100]. To illustrate the weak generalization abil-
ratio between fault modes increases further. Group2 can be con- ity of the intelligent diagnosis algorithms, experiments are also
sidered as a moderately imbalanced dataset, while Group3 can be carried out on the PU bearing dataset. Experiments use the data
considered as a highly imbalanced dataset. under three working conditions (N15_M07_F10, N09_M07_F10,
Experimental results are shown in Fig. 11, and it can be N15_M01_F10). In these experiments, data under one working
observed that the overall accuracy in Group3 is much lower condition is used to train models, and data under another working
than that of Group1, which indicates that the class imbalance condition is used to test the performance. A total of six groups are
will greatly degrade the performance of models. To address the performed, and the detailed information is shown in Table 12.
problem of class imbalance, data-level methods and classifier- The experimental results are shown in Fig. 12. It can be con-
level methods can be used [99]. Oversampling and undersam- cluded that in most cases, intelligent diagnosis algorithms trained
pling methods are the most commonly used data-level meth- on one working condition cannot perform well on another work-
ods, and some methods for generating samples based on GAN ing condition, which means the generalization ability of algo-
have also been studied recently. For the classifier-level methods, rithms is insufficient. In general, we expect that our algorithms

Please cite this article as: Z. Zhao, T. Li, J. Wu et al., Deep learning algorithms for rotating machinery intelligent diagnosis: An open source benchmark study. ISA Transactions
(2020), https://doi.org/10.1016/j.isatra.2020.08.010.
16 Z. Zhao, T. Li, J. Wu et al. / ISA Transactions xxx (xxxx) xxx

Fig. 12. Experimental results of working conditions transfer. (a) time domain
Fig. 11. Experimental results of three groups of datasets. (a) time domain input, input, and (b) frequency domain input.
and (b) frequency domain input.

Table 12 Table 13
Training data and testing data for each experiment. The bearing code and the number of samples used in each experiment.
Group Data for training Data for testing Group Bearing code Training samples Testing samples
Group1 N15_M07_F10 N09_M07_F10 KA03 200 50
Group2 N15_M07_F10 N15_M01_F10 Group1
KA06 200 50
Group3 N09_M07_F10 N15_M07_F10
KA08 200 50
Group4 N09_M07_F10 N15_M01_F10 Group2
KA09 200 50
Group5 N15_M01_F10 N15_M07_F10
Group6 N15_M01_F10 N09_M07_F10 KI07 200 50
Group3
KI08 200 50

could adapt to the changes in working conditions or measure- and the working condition at the time of acquisition are the
ment situations since these changes occur frequently in real ap- same between two classes, theoretically, methods should not be
plications. Therefore, studies are still required on how to transfer able to achieve such high accuracy. These expected results are
the trained algorithms to different working conditions effectively. exactly contrary to those of the experiment, which shows that
Two excellent review papers [101,102] and other applica- models only learn the discrimination of different collection points
tions [103,104] published recently pointed out several poten- and do not learn how to extract the essential characteristics of
tial research directions which could be considered and studied fault signals. Therefore, it is very important to figure out whether
further to improve the generalization ability. models can learn essential fault characteristics or just classify the
different conditions of collected signals.
10.3. Interpretability According to the development of interpretability, we might
be able to study the interpretability of DL-based models from
Although intelligent diagnosis algorithms can achieve high the following aspects: (1) visualize the results of neurons to an-
diagnostic accuracy in their tasks, the interpretability of these alyze the attention points of models [108]; (2) add physical con-
models is often insufficient and these black-box models would
straints to the loss function [109] to meet specific needs of fault
generate high-risk results [105], which greatly limits their ap-
feature extraction; (3) add prior knowledge to network struc-
plications. Actually, some papers in intelligent diagnosis have
tures and convolutions [110] or unroll the existing optimization
noted this problem and attempted to propose some interpretable
algorithms [111] to extract corresponding fault features.
models [106,107].
To point out that intelligent diagnostic algorithms lack in-
terpretability, we perform three sets of experiments on the PU 10.4. Few-shot learning
bearing dataset, and the datasets are shown in Table 13. In each
set of experiments, we use two different data, which have the In intelligent diagnosis, the amount of data is far from big data
same fault pattern and are acquired under the same condition. because of the preciousness of fault data and the high cost of fault
The results, in which intelligent algorithms can get high ac- simulation experiments, especially for the key components. To
curacy in each set of experiments, are shown in Fig. 13. Never- manifest the influence of the sample number on the classification
theless, for each binary classification task, since the fault mode accuracy, we use the PU bearing dataset to design a few-shot

Please cite this article as: Z. Zhao, T. Li, J. Wu et al., Deep learning algorithms for rotating machinery intelligent diagnosis: An open source benchmark study. ISA Transactions
(2020), https://doi.org/10.1016/j.isatra.2020.08.010.
Z. Zhao, T. Li, J. Wu et al. / ISA Transactions xxx (xxxx) xxx 17

Fig. 13. Experimental results of three groups. (a) time domain input, and (b) Fig. 14. Experimental results of different few-shot training patterns. (a) time
frequency domain input. domain input, and (b) frequency domain input.

training pattern with six groups of different sample numbers in trial and error cost multiplied by the number of trial and error
each class. can easily reach a huge cost. Besides, reducing this cost is also
Results of the time domain input and the frequency domain the partial purpose of this benchmark study which provides some
input are shown in Fig. 14. It is shown that with the decrease guidelines to choose a baseline model.
of the sample number, the accuracy decreases sharply. As shown Actually, there is another way called neural architecture search
in Fig. 14, for the time domain input, the Best-Max accuracy (NAS) [113] to avoid the huge cost of trial and error. NAS can al-
decreases from 91.46% to 20.39% as the sample number decreases low designing a neural network automatically through searching
from 100 to 1. Meanwhile, the Best-Max accuracy decreases from for a specific network based on a specific dataset. Limited search
97.73% to 29.67% as the sample number decreases from 100 to 1 space of the network is first constructed according to the physical
with the frequency domain input. prior. After that, a neural network matching a specific dataset is
Although the accuracy can be increased after using FFT, it sampled from the search space through reinforcement learning,
is still too low to be accepted when the number of samples is the evolutionary algorithm or the gradient strategy. Besides, the
extremely small. It is necessary to develop methods based on whole construction process does not require manual participa-
few-shot learning to cope with application scenarios with limited tion, which greatly reduces the cost of building a neural network
samples.
and allows us to focus on specific engineering applications.
Many DL-based few-shot learning models have been proposed
in recent years [112], most of these methods adopt a meta-
learning paradigm by training networks with a large number of 11. Conclusion
tasks, which means that the big data in other related fields is
necessary for these methods. In the field of fault diagnosis, there In this paper, we collect nine publicly available datasets to
is no relevant data with such a big size available, so methods
evaluate the performance of MLP, AE, CNN, and RNN models. This
embedding with physical mechanisms are required to address
work mainly focuses on evaluating DL-based intelligent diagnosis
this problem effectively.
algorithms from different perspectives and providing the bench-
mark accuracy (a lower bound) to avoid useless improvement.
10.5. Model selection
In addition, we release a code library for other researchers to
For intelligent diagnosis, designing a neural network is not the test the performance of their own DL-based intelligent diagnosis
final goal, and our task is to apply the model to real industrial models of these datasets. We hope that the evaluation results
applications while designing a neural network is only a small part and the code library could promote a better understanding of
of our task. However, to achieve a good effect, we have to spend DL-based models and provide a unified framework for generating
considerable time and energy on designing the corresponding more effective models. For further studies, we will focus on five
networks. Because building a neural network is an iterative pro- listed issues (class imbalance, generalization ability, interpretabil-
cess consisting of repeated trial and error, and the performance ity, few-shot learning, and model selection) to propose more
of models should be fed back to us to adjust models. The single customized works.

Please cite this article as: Z. Zhao, T. Li, J. Wu et al., Deep learning algorithms for rotating machinery intelligent diagnosis: An open source benchmark study. ISA Transactions
(2020), https://doi.org/10.1016/j.isatra.2020.08.010.
18 Z. Zhao, T. Li, J. Wu et al. / ISA Transactions xxx (xxxx) xxx

Table A.1
CWRU: Results with random split and data augmentation.
Nor Input Loc AE DAE SAE MLP CNN LeNet AlexNet ResNet18 LSTM
Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max
Last 54.71 66.67 54.33 64.37 67.13 69.73 52.95 67.43 98.47 99.62 96.4 99.23 91.11 96.17 88.28 98.47 98.47 98.85
1
Best 71.42 73.95 78.24 80.84 72.95 76.25 73.49 75.48 98.93 100 99.08 99.62 91.11 96.17 100 100 98.47 98.85
Last 100 100 100 100 99.92 100 100 100 99.89 100 100 100 100 100 99.08 100 100 100
2
Best 100 100 100 100 100 100 100 100 99.92 100 100 100 100 100 100 100 100 100
A Last 86.95 89.89 82.33 84.27 86.02 91.39 – – 95.28 98.88 98.59 99.33 97.89 98.73 98.08 99.52 98.11 98.43
3
Best 90.95 92.51 86.14 88.39 90.01 91.76 – – 97.03 99.18 99.17 99.33 97.89 98.73 99.66 99.85 98.11 98.43
Last 100 100 100 100 100 100 – – 98.7 99.23 98.7 100 99.69 100 100 100 99.85 100
4
Best 100 100 100 100 100 100 – – 99.46 99.62 100 100 99.69 100 100 100 99.85 100
Last 94.71 97.32 92.19 94.64 94.10 97.32 – – 86.74 95.02 74.94 89.27 46.36 94.25 90.19 100 84.29 89.27
5
Best 96.86 98.47 95.94 97.32 96.09 98.85 – – 94.86 96.93 89.81 90.8 46.36 94.25 100 100 84.29 89.27
Last 74.79 77.01 80.31 83.14 76.09 77.78 76.02 78.93 99.45 100 98.08 99.23 98.31 99.23 99.46 100 98.85 100
1
Best 78.85 80.46 82.68 83.52 77.78 78.54 76.71 79.69 99.67 100 98.83 99.88 98.54 99.62 99.46 100 99.92 100
Last 99.92 100 100 100 99.54 100 99.92 100 99.39 100 100 100 99.92 100 99.92 100 99.85 100
2
Best 100 100 100 100 100 100 99.92 100 99.39 100 100 100 99.92 100 99.92 100 100 100
B Last 85.89 90.64 85.27 88.01 87.58 89.14 – – 98.94 99.22 98.84 99.07 98.6 98.84 99.17 99.78 93.71 94.76
3
Best 88.83 92.13 89.01 92.13 90.7 92.13 – – 99.12 99.29 99.15 99.36 98.71 99.03 99.6 99.95 95.36 95.88
Last 100 100 100 100 100 100 – – 97.62 99.62 99 99.62 97.85 100 99.92 100 99.77 100
4
Best 100 100 100 100 100 100 – – 98.93 100 99.23 100 98.31 100 99.92 100 100 100
Last 93.1 95.79 94.56 96.55 89.65 95.02 – – 86.9 93.87 86.05 88.51 91.03 97.32 97.09 99.23 91.65 93.87
5
Best 95.33 96.17 97.01 98.85 93.95 97.32 – – 89.46 95.64 86.67 90.04 92.27 97.32 97.62 100 93.33 94.64
Last 66.13 72.03 74.87 77.78 68.81 74.71 69.88 74.71 99.16 100 99.15 99.23 99.46 100 99.16 100 99.16 99.62
1
Best 70.8 72.8 77.32 79.69 71.8 74.71 71.95 74.71 100 100 99.69 100 99.92 100 100 100 100 100
Last 100 100 100 100 100 100 100 100 99.85 100 100 100 100 100 100 100 100 100
2
Best 100 100 100 100 100 100 100 100 100 100 100 100 100 100 100 100 100 100
C Last 88.49 91.39 88.97 90.64 88.39 92.13 – – 99.17 99.37 99.07 99.25 98.69 98.92 99.59 99.81 98.3 98.58
3
Best 91.35 93.63 91.25 92.51 90.94 93.63 – – 99.43 99.55 99.23 99.37 99.13 99.25 99.86 99.89 98.86 99.07
Last 100 100 100 100 100 100 – – 99.85 100 97.85 100 100 100 100 100 99.85 100
4
Best 100 100 100 100 100 100 – – 100 100 99.92 100 100 100 100 100 100 100
Last 95.25 98.85 94.86 97.7 97.09 98.47 – – 92.03 93.87 87.2 91.57 87.2 98.85 91.19 100 92.26 96.17
5
Best 97.93 99.62 96.47 98.47 98.62 100 – – 94.41 95.02 92.11 93.1 98.08 99.23 100 100 96.01 97.7

*A is the max–min normalization; B is the -1-1 normalization; C is the Z-score normalization.


1 is the time domain input; 2 is the frequency domain input; 3 is the wavelet domain input; 4 is the time domain sample after STFT; 5 is the time domain sample
reshape to a 2D matrix
In the same input, the first line are the results of last epoch, and the second line are is the results of the best epoch.

Table A.2
CWRU: Results with random split and without data augmentation.
Nor Input Loc AE DAE SAE MLP CNN LeNet AlexNet ResNet18 LSTM
Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max
Last 60.08 65.9 54.48 68.97 69.58 71.26 64.29 70.11 99.54 99.62 99.46 100 95.79 98.08 99.77 100 99.92 100
1
Best 70.96 73.56 71.34 76.63 71.57 73.18 74.41 76.63 99.85 100 99.77 100 97.4 98.47 100 100 100 100
Last 100 100 100 100 100 100 100 100 99.85 100 100 100 100 100 99.77 100 99.92 100
2
Best 100 100 100 100 100 100 100 100 100 100 100 100 100 100 100 100 100 100
A Last 85.14 91.01 85.33 89.89 89.45 91.76 – – 98.85 99.07 98.61 99.18 82.17 98.77 96.6 98.66 97.71 98.66
3
Best 88.08 91.39 87.08 91.39 91.01 93.63 – – 99.35 99.4 99.16 99.18 82.59 98.84 99.55 99.63 98.67 98.84
Last 100 100 100 100 100 100 – – 99.31 100 99.69 100 100 100 100 100 100 100
4
Best 100 100 100 100 100 100 – – 99.46 100 99.92 100 100 100 100 100 100 100
Last 92.95 95.79 90.11 96.55 85.29 96.17 – – 85.52 93.87 79.62 85.44 63.68 95.79 96.55 100 86.82 91.19
5
Best 93.87 96.93 93.41 98.47 94.94 97.7 – – 94.64 96.55 88.81 90.04 67.2 97.32 100 100 90.96 93.1
Last 69.96 71.65 67.28 73.18 70.8 73.56 71.1 73.18 99.4 99.62 99.07 100 97.7 98.85 99.51 100 99.78 100
1
Best 74.48 76.63 71.11 73.95 73.18 76.25 75.26 78.54 99.62 100 99.56 100 98.58 99.62 100 100 100 100
Last 99.62 100 99.92 100 100 100 100 100 99.95 100 100 100 99.89 100 97.37 100 100 100
2
Best 100 100 100 100 100 100 100 100 100 100 100 100 100 100 100 100 100 100
B Last 89.45 91.39 85.39 89.51 90.38 93.63 – – 98.87 99.29 97.79 98.99 98.58 99.14 99.28 99.74 98.27 98.81
3
Best 90.32 92.13 88.14 90.26 91.2 94.01 – – 99.21 99.4 99.15 99.33 99.04 99.33 99.58 99.74 98.82 98.96
Last 90.65 100 100 100 100 100 – – 99.4 100 99.45 100 99.78 100 100 100 99.78 100
4
Best 100 100 100 100 100 100 – – 99.73 100 99.78 100 100 100 100 100 100 100
Last 93.03 97.7 94.33 96.17 95.25 96.55 – – 91.57 96.93 85.77 91.95 83.2 96.93 95.57 99.62 85.11 92.34
5
(continued on next page)

Please cite this article as: Z. Zhao, T. Li, J. Wu et al., Deep learning algorithms for rotating machinery intelligent diagnosis: An open source benchmark study. ISA Transactions
(2020), https://doi.org/10.1016/j.isatra.2020.08.010.
Z. Zhao, T. Li, J. Wu et al. / ISA Transactions xxx (xxxx) xxx 19

Table A.2 (continued).


Nor Input Loc AE DAE SAE MLP CNN LeNet AlexNet ResNet18 LSTM
Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max
Best 96.09 97.7 96.63 97.32 96.32 97.7 – – 95.46 96.93 89.82 94.25 86.75 97.32 100 100 94.25 95.4
Last 64.75 67.43 66.9 69.73 62.68 69.35 62.53 63.98 99.62 100 98.7 99.62 99 100 99.46 100 99.77 100
1
Best 68.28 70.5 69.81 71.26 67.36 70.88 68.43 71.26 99.7 100 99.85 100 99.69 100 100 100 100 100
Last 100 100 100 100 100 100 100 100 100 100 100 100 100 100 100 100 100 100
2
Best 100 100 100 100 100 100 100 100 100 100 100 100 100 100 100 100 100 100
C Last 88.91 91.76 88.46 89.51 90.56 92.88 – – 99.17 99.52 98.57 99.22 98.6 98.88 99.36 99.78 98.56 98.7
3
Best 89.44 92.88 88.99 90.64 92.06 94.38 – – 99.46 99.55 99.35 99.52 99.14 99.4 99.87 99.89 99.05 99.14
Last 100 100 100 100 100 100 – – 99.69 100 99.77 100 100 100 100 100 100 100
4
Best 100 100 100 100 100 100 – – 99.85 100 99.77 100 100 100 100 100 100 100
Last 95.4 97.32 90.12 99.62 93.79 97.32 – – 93.79 94.64 91.95 93.49 90.8 95.4 95.17 100 91.26 96.17
5
Best 96.47 98.47 93.94 100 97.09 98.08 – – 94.94 95.4 93.03 94.25 97.62 98.47 100 100 95.79 97.32

Table A.3
CWRU: Results with order split and data augmentation.
Nor Input Loc AE DAE SAE MLP CNN LeNet AlexNet ResNet18 LSTM
Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max
Last 58.71 70.08 55.84 69.7 64.55 68.18 65.23 73.11 95.99 98.11 97.04 98.86 91.21 95.83 82.65 99.62 97.27 98.48
1
Best 71.67 72.73 74.77 76.89 71.29 73.48 69.39 78.41 98.18 100 97.57 99.24 92.5 96.97 98.41 99.62 98.1 98.86
Last 100 100 100 100 100 100 100 100 99.54 100 99.85 100 99.92 100 98.79 100 99.85 100
2
Best 100 100 100 100 100 100 100 100 99.7 100 100 100 100 100 98.79 100 99.85 100
A Last 72.69 78.1 73.9 76.64 76.64 79.2 – – 87.15 88.32 96.67 97.54 95.12 96.76 93.16 99.29 95.17 96.76
3
Best 76.82 80.29 77.07 80.66 79.5 80.66 – – 89.2 90.51 96.86 97.85 96.34 96.8 98.66 99.29 95.81 97.54
Last 99.92 100 99.77 100 100 100 – – 97.73 98.11 99.47 100 97.27 100 100 100 99.01 100
4
Best 100 100 100 100 100 100 – – 98.79 99.24 99.62 100 97.35 100 100 100 99.85 100
Last 92.35 97.73 92.35 98.11 92.2 96.21 – – 78.56 95.45 83.18 90.53 93.47 96.59 100 100 87.35 90.15
5
Best 96.74 98.11 96.06 99.62 95.07 99.62 – – 94.85 96.59 85.08 92.8 94.98 96.59 100 100 88.26 92.42
Last 70.46 72.73 76.36 78.79 70.91 74.24 72.24 73.48 98.7 99.24 96.81 98.86 97.08 97.73 99.24 100 98.41 99.24
1
Best 74.54 76.89 78.86 79.92 74.85 76.52 72.84 73.48 98.97 99.26 97.24 98.86 97.62 98.11 99.57 100 99.09 99.24
Last 100 100 99.85 100 100 100 100 100 98.81 99.62 100 100 99.56 100 97.73 100 100 100
2
Best 100 100 100 100 100 100 100 100 99.13 100 100 100 99.56 100 99.95 100 100 100
B Last 76.95 79.93 74.21 80.29 76.83 78.83 – – 97.04 97.92 96.68 97.65 96.88 97.65 98.19 98.92 86.72 88.32
3
Best 78.83 80.29 77.31 80.29 79.5 80.66 – – 97.23 98.1 96.8 97.8 97.29 98.29 98.53 98.92 88.69 89.78
Last 99.92 100 99.92 100 98.94 100 – – 98.05 98.48 99.24 99.62 97.24 100 99.57 100 98.94 99.62
4
Best 100 100 100 100 100 100 – – 98.37 98.48 99.57 100 99.3 100 99.57 100 99.92 100
Last 94.32 97.73 92.35 98.11 91.21 95.08 – – 88.53 96.59 92.64 94.32 93.02 97.73 96.91 100 89.32 94.32
5
Best 97.12 98.86 95.53 98.86 95.83 96.97 – – 88.8 96.59 93.51 95.83 93.35 97.73 96.91 100 92.88 94.32
Last 64.24 68.94 73.79 77.27 65.46 67.05 66.16 68.18 98.7 99.62 97.98 98.86 98.17 99.24 97.98 100 98.56 98.86
1
Best 67.73 70.83 75.15 77.27 67.12 68.94 67.17 72.35 99.02 99.62 98.23 99.24 98.29 99.24 100 100 98.79 99.24
Last 100 100 100 100 100 100 100 100 99.3 100 99.94 100 99.94 100 99.68 100 100 100
2
Best 100 100 100 100 100 100 100 100 99.78 100 99.94 100 100 100 100 100 100 100
C Last 78.29 79.93 75.79 78.47 77.07 81.75 – – 96.54 97.99 97.05 97.77 95.45 98.21 98.93 99.55 96.44 97.02
3
Best 81.63 84.31 80.17 83.21 80.84 84.67 – – 97.61 98.36 97.29 98.18 96.32 98.21 99.28 99.55 96.69 97.32
Last 99.92 100 100 100 100 100 – – 98.97 99.62 99.56 100 99.77 100 100 100 99.77 100
4
Best 100 100 100 100 100 100 – – 99.08 99.62 99.68 100 99.77 100 100 100 99.92 100
Last 93.71 97.35 87.5 93.18 94.09 98.11 – – 93.34 97.35 93.56 95.08 89.01 93.94 98.18 100 93.64 95.83
5
Best 96.21 97.73 92.42 97.35 96.74 98.11 – – 93.56 97.35 94.07 95.83 89.92 95.45 98.41 100 94.52 95.83

Table A.4
JNU: Results with random split and data augmentation.
Nor Input Loc AE DAE SAE MLP CNN LeNet AlexNet ResNet18 LSTM
Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max
Last 46.74 49.03 46.01 50.51 43.8 45.9 46.97 49.32 80.57 83.28 73.23 77.25 82.97 83.9 77.14 91.47 82.14 83.62
1
Best 52.6 54.72 51.19 53.58 53.41 54.66 52.48 53.81 82.28 85.32 77.23 79.18 83.86 85.32 93.07 93.69 83.2 84.7
Last 95.93 96.3 95.2 95.73 63.47 95.51 95.89 96.76 93.83 94.48 94.59 95.45 95.82 96.19 95.61 96.19 95.02 95.56
2
Best 96.62 96.93 95.58 96.13 64.52 96.64 97.21 97.38 94.36 94.99 95.86 96.02 96.23 96.7 96.76 96.99 95.36 95.56
A Last 33.11 36.2 30.47 31.76 33.99 37.92 – – 43.5 45.81 44.93 46.56 37.04 44.87 52.28 55.19 42.93 44.17
3
Best 37.42 41.42 36.48 36.92 38.98 40.92 – – 46.66 48.25 48.66 49.28 37.59 44.87 55.51 57.16 43.55 45.27
Last 78.63 80.34 78.92 80.63 77.55 82.91 – – 69.69 73.22 66.84 69.52 79.77 81.77 87.12 90.88 71.11 73.22
4
Best 82.91 83.76 82.68 84.05 81.99 86.32 – – 73.16 75.21 70.66 72.36 80.46 84.05 91.17 92.02 72.71 74.07
Last 54.81 57.83 50.26 54.7 45.24 51.57 – – 39.94 43.3 43.65 50.71 50.71 55.27 75.44 80.91 54.3 56.7
5
(continued on next page)

Please cite this article as: Z. Zhao, T. Li, J. Wu et al., Deep learning algorithms for rotating machinery intelligent diagnosis: An open source benchmark study. ISA Transactions
(2020), https://doi.org/10.1016/j.isatra.2020.08.010.
20 Z. Zhao, T. Li, J. Wu et al. / ISA Transactions xxx (xxxx) xxx

Table A.4 (continued).


Nor Input Loc AE DAE SAE MLP CNN LeNet AlexNet ResNet18 LSTM
Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max
Best 57.44 60.11 55.96 57.26 47.98 54.99 – – 48.15 50.43 49.12 52.14 52.31 57.83 81.94 83.76 56.53 60.4
Last 60.36 61.6 16.67 16.67 61.14 62 61.97 63.54 84.96 86.41 78.2 80.49 87.2 88.4 92.43 93.63 83.24 84.36
1
Best 61.93 63.14 16.67 16.67 62.26 62.8 62.88 64.11 86.2 86.86 78.55 80.89 87.99 88.51 92.83 93.63 84.46 84.93
Last 94.97 96.59 94.63 96.08 95.69 96.76 92.73 96.53 93.94 94.6 94.11 95.22 95.56 95.79 95.73 95.96 94.73 95.34
2
Best 96.7 96.93 96.59 97.1 96.66 96.76 94.31 97.04 94.38 94.88 94.91 95.73 95.71 96.25 95.92 96.13 95.64 95.9
B Last 33.33 34.65 33 34.87 33.96 37.09 – – 46.69 47.67 46.82 48.03 48.63 52.83 51.37 52.44 37.79 38.37
3
Best 39 39.81 38.42 40.2 39.09 40.64 – – 47.5 50.53 47.42 48.64 48.78 53.3 52.58 53.39 43.07 43.75
Last 76.69 77.78 80.46 81.48 77.49 79.77 – – 65.93 72.08 65.81 71.51 76.47 78.06 85.58 90.03 68.03 73.22
4
Best 81.03 82.62 84.44 86.89 82.56 84.9 – – 67.41 76.64 68.49 71.51 77.89 83.48 87.29 92.31 73.67 75.21
Last 48.43 56.98 48.03 52.14 39.03 44.73 – – 43.87 47.86 44.21 46.15 56.52 58.69 77.72 79.49 55.67 57.26
5
Best 55.16 62.11 52.71 56.7 44.33 51.28 – – 44.96 48.15 45.58 47.01 57.78 61.25 78.83 80.63 60.28 60.68
Last 65.75 66.78 52.23 62.17 67.14 68.15 67.91 69 89.78 90.39 85.71 86.29 90.89 91.7 94.15 94.88 87.65 88.57
1
Best 66.66 67.35 52.96 62.74 67.71 68.71 68.98 69.28 91.24 91.47 86.73 87.88 91.99 92.72 95.68 95.85 89.2 89.65
Last 96.35 96.7 96.34 96.7 96.22 96.36 96.58 97.16 93.94 94.65 95.03 95.73 96.02 96.59 95.64 96.53 94.9 95.45
2
Best 97.01 97.21 97.19 97.38 96.92 97.1 97.14 97.27 95.05 95.62 95.94 96.25 96.64 96.99 96.59 96.87 95.62 95.85
C Last 34.14 37.2 34.14 36.81 32.89 35.37 – – 48.69 50.25 48.19 52.64 50.41 52.14 52.34 53.77 42.24 43.62
3
Best 38.92 41.03 37.88 39.98 38.99 40.09 – – 52.21 53.22 51.16 52.64 53.13 54.08 55.29 55.6 46.91 48.34
Last 77.09 78.06 80.91 84.33 77.32 80.91 – – 69.69 73.5 69.86 71.23 81.37 82.91 91.28 93.45 73.22 75.5
4
Best 82.74 83.76 84.16 86.04 83.65 84.62 – – 74.76 76.64 72.42 74.07 86.44 87.46 93.11 93.73 76.75 77.49
Last 63.13 66.1 61.2 64.1 62.45 66.67 – – 54.3 58.4 50.14 51.57 70.09 73.5 84.73 87.18 61.71 64.67
5
Best 65.98 68.38 65.81 68.66 67.12 69.8 – – 58.4 60.68 55.67 58.4 75.21 76.35 86.38 87.46 65.93 67.52

*A is the max–min normalization; B is the -1-1 normalization; C is the Z-score normalization.


1 is the time domain input, 2 is the frequency domain input; 3 is the wavelet domain input; 4 is the time domain sample after STFT; 5 is the time domain sample
reshape to a 2D matrix.

Table A.5
JNU: Results with random split and without data augmentation.
Nor Input Loc AE DAE SAE MLP CNN LeNet AlexNet ResNet18 LSTM
Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max
Last 41.64 45.11 39.06 42.78 38.39 41.92 40.6 44.94 80.79 82.31 63.06 69.4 77.57 79.12 87.51 90.96 76.08 77.42
1
Best 44.68 45.96 43.12 44.2 45.83 48.75 45.17 46.25 83.56 84.81 71.96 72.81 80.19 81.51 90.26 91.24 78.88 81.8
Last 95.77 96.47 95.51 97.04 80.21 96.3 96.52 97.21 92.88 94.25 95.05 95.62 95.37 95.9 96.49 97.38 95.23 95.79
2
Best 96.78 96.99 96.84 97.1 80.79 96.99 97.44 97.78 94.01 94.6 95.49 95.96 96.55 96.76 97.4 97.72 95.93 96.53
A Last 32.96 34.81 29.17 29.93 35.01 37.26 – – 43.7 48.11 40.85 47.23 44.6 47.25 45.43 51.69 40.74 44.34
3
Best 37.06 38.59 36.8 38.76 38.51 40.53 – – 47.16 51.66 47.36 49.67 46.64 49.42 52.81 56.22 45.7 47.5
Last 81.31 86.04 81.14 87.18 82.45 84.62 – – 69.74 72.36 67.24 68.95 78.58 81.2 90.19 91.74 71.32 76.07
4
Best 83.76 87.18 85.76 88.32 83.88 85.47 – – 72.31 75.5 69.83 72.08 83.25 84.9 91.77 92.59 74.74 76.64
Last 47.52 56.41 50.14 56.13 45.01 50.14 – – 42.85 46.44 42.48 45.87 52.25 54.42 72.62 79.49 53.72 56.41
5
Best 52.48 58.69 54.59 58.4 48.2 53.56 – – 48.66 49.57 48.35 52.14 56.52 58.69 80.69 82.91 59.42 60.68
Last 50.4 52.1 16.67 16.67 51.09 52.84 51.87 53.53 81.02 83.67 70.67 72.24 81.64 83.45 86.97 88.74 77.93 78.67
1
Best 53.31 54.1 16.67 16.67 53.29 54.21 55.49 55.75 82.31 83.67 72.41 73.15 82.94 84.41 89.94 90.39 78.43 80.03
Last 95.58 96.87 94.29 96.42 95.65 96.19 88.35 94.77 92.08 94.25 95.38 96.42 95.31 96.08 96.81 97.38 94.73 95.73
2
Best 96.96 97.44 97.11 97.27 96.91 97.38 97.18 97.61 93.13 94.54 95.86 96.42 95.43 96.08 97.51 97.67 94.98 95.73
B Last 33.73 36.76 35.28 36.92 35.46 37.98 – – 46.72 48.31 47.58 48.5 50.17 51.22 51.6 52.66 41.3 42.34
3
Best 38.16 39.76 39.05 39.98 39.29 40.48 – – 47.7 49.67 50.51 51.55 50.82 51.44 56.01 56.38 41.75 43.23
Last 84.1 90.03 81.82 86.32 79.03 81.2 – – 72.08 72.65 68.26 70.94 76.18 78.06 85.13 90.31 69.29 70.66
4
Best 85.47 90.6 83.76 87.75 80.23 82.34 – – 72.93 73.79 70.09 72.36 77.55 78.92 91.51 92.02 69.92 72.08
Last 54.64 60.11 52.76 59.83 43.82 45.58 – – 44.3 44.44 46.38 48.43 54.87 61.25 78.52 80.34 54.13 56.98
5
Best 56.58 61.82 54.7 61.54 45.24 49 – – 50.14 52.71 51.28 52.99 56.47 61.25 79.54 81.48 55.84 57.55
Last 55.22 56.43 45.39 47.72 55.09 55.86 53.66 54.72 87.29 88 77.53 78.84 86.65 88.79 92.67 93.57 87.55 87.88
1
Best 57.18 57.74 47.48 48.41 56.88 57.74 56.22 56.71 88.33 89.19 79.92 82.82 88.77 89.19 94.43 94.88 88.71 89.25
Last 96.19 96.53 96.24 96.81 95.79 96.25 96.03 96.47 93.72 94.71 95.14 95.51 95.45 96.08 96.1 97.16 95.26 95.79
2
Best 96.8 96.99 97.34 97.38 96.9 97.04 97.19 97.33 94.34 95.05 95.53 95.96 96.1 96.53 97.53 97.9 95.92 96.13
C Last 35.82 37.03 34.59 36.92 33.34 38.03 – – 47.91 49.42 46.34 47.84 42.51 50.08 49.11 52.36 42.67 42.98
3
Best 38.95 40.26 37.91 39.64 38.52 41.31 – – 52.77 53.63 49.21 49.97 44.43 53.33 54.55 55.22 46.91 47.14
Last 79.48 85.75 81.58 82.34 84.14 84.9 – – 71.62 73.5 69.57 72.36 84.05 86.04 91.4 93.73 73.96 74.93
4
Best 83.26 86.89 83.19 83.48 85.57 86.61 – – 73.16 74.93 71.62 74.07 86.27 88.32 93.62 94.87 77.09 78.92
Last 62.89 65.24 60.87 65.24 64.96 66.95 – – 54.87 58.97 50.88 54.7 69.86 72.36 85.93 88.03 62.16 65.81
5
Best 64.6 67.81 63.44 66.67 66.86 69.23 – – 57.95 59.54 55.67 58.4 73.05 74.36 86.72 88.32 65.53 66.95

Please cite this article as: Z. Zhao, T. Li, J. Wu et al., Deep learning algorithms for rotating machinery intelligent diagnosis: An open source benchmark study. ISA Transactions
(2020), https://doi.org/10.1016/j.isatra.2020.08.010.
Z. Zhao, T. Li, J. Wu et al. / ISA Transactions xxx (xxxx) xxx 21

Table A.6
JNU: Results with order split and data augmentation.
Nor Input Loc AE DAE SAE MLP CNN LeNet AlexNet ResNet18 LSTM
Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max
Last 42.16 43.88 42.67 45.18 44.72 48.64 46.76 49.77 79.91 83.62 72.99 77.55 81.54 82.43 90.84 93.59 82.22 85.09
1
Best 52.07 52.27 49.67 51.3 51.63 54.25 51.68 54.25 85.33 86.05 77.61 78.46 81.54 82.43 93.11 93.59 84.2 85.71
Last 94.92 95.46 94.66 95.92 48.11 95.29 95.95 96.2 93.84 94.56 94.47 94.73 96.1 96.54 95.83 96.2 95.28 95.63
2
Best 96.03 96.09 95.16 96.26 48.37 95.98 96.8 97.05 94.81 95.07 95.42 95.58 96.1 96.54 97.1 97.34 95.99 96.2
A Last 32.66 35.66 29.53 30.84 32.6 34.39 – – 42.73 47.81 46.23 48.56 36.18 43.47 52.2 53.65 42.51 44.96
3
Best 36.93 38.82 36.05 37.94 37.83 38.44 – – 48.73 49.42 49.84 50.75 37.87 43.47 54.94 55.68 47.75 49.03
Last 75.28 79.55 76.19 77.31 75.41 76.75 – – 68.57 72.55 69.25 70.31 79.41 82.35 81.57 89.08 74.37 75.35
4
Best 80.11 84.31 80.56 82.35 79.77 82.07 – – 71.82 75.07 71.54 72.27 82.84 84.59 90.31 90.76 76.89 77.87
Last 53.36 55.18 49.36 57.98 42.52 50.98 – – 38.38 45.38 45.72 49.3 54.23 56.02 63.7 80.11 50.59 56.02
5
Best 57.42 61.62 55.4 60.5 48.96 52.38 – – 49.41 51.26 50.7 51.54 56.75 58.82 80.78 84.31 57.42 58.26
Last 60.38 61.62 37.86 54.37 59.24 60.94 62.01 63.15 83.89 86.34 77.22 79.59 86.24 88.1 92.8 93.65 75.64 83.73
1
Best 61.32 61.85 39.07 55.27 60.48 61.39 62.01 63.15 85.63 86.85 77.22 79.59 86.95 88.76 93.11 93.65 76.66 84.86
Last 95.5 96.09 95.12 95.86 95.09 95.92 94.43 96.15 93.32 93.82 93.59 94.27 95.77 96.43 95.63 96.71 94.55 94.9
2
Best 96.75 97.05 96.37 96.6 96.45 96.54 94.8 96.15 93.58 94.33 93.83 95.12 95.99 96.77 96.18 96.71 95.62 95.75
B Last 32.75 35.16 32.6 34.39 32.78 34.33 – – 47.91 48.64 47.3 49 51.27 53.88 52.05 53.38 40.85 41.82
3
Best 38.26 40.27 38.25 39.66 38.47 38.99 – – 48.29 49.22 48.04 49 51.42 53.88 52.98 53.49 45.28 46.15
Last 72.1 81.51 75.24 77.87 73.78 77.59 – – 68.13 69.75 67.45 71.15 74.23 75.91 85.49 86.27 69.75 71.71
4
Best 74.96 85.99 77.93 81.23 77.81 82.07 – – 68.6 70.03 67.45 71.15 74.96 78.15 87.06 89.64 73.45 76.47
Last 50.76 52.94 51.99 54.62 43.64 50.42 – – 45.27 47.06 47.23 50.14 55.24 56.86 76.41 81.23 54.34 56.86
5
Best 57.48 58.82 56.52 57.98 47 54.06 – – 45.6 47.62 47.96 50.98 56.59 60.22 78.99 81.23 58.88 59.94
Last 65.76 67.35 60.07 62.81 65.53 67.74 66.79 68.54 87.94 90.99 85.48 86.28 90.7 91.33 94.06 95.18 87.52 88.21
1
Best 66.39 67.35 61.04 63.15 66.59 67.74 67.35 68.54 88.96 90.99 85.61 86.28 91.32 92.12 94.27 95.18 87.78 88.55
Last 95.74 96.6 96.01 96.77 96.27 96.54 96.12 96.54 93.59 94.22 94.72 95.58 95.83 97.05 95.34 96.43 94.62 95.35
2
Best 96.71 96.88 96.87 97.22 96.69 96.88 96.34 96.54 93.81 95.01 94.72 95.58 96.08 97.05 95.52 96.43 94.84 95.35
C Last 33.84 36 32.95 34 33.25 35.39 – – 48.94 50.28 49.1 50.55 44.06 51.88 51.17 55.84 43.03 44.63
3
Best 38.43 39.93 37.11 38.44 37.42 39.93 – – 49.22 50.28 49.29 50.55 44.38 52.77 52.65 55.84 43.42 44.63
Last 78.43 82.07 75.35 77.59 79.66 81.23 – – 70.12 71.71 69.28 72.55 80.67 82.07 87 90.76 72.66 75.35
4
Best 81.96 85.43 77.65 80.39 83.92 84.31 – – 70.73 71.71 69.94 72.55 81.4 84.03 90.26 90.76 73.73 78.43
Last 64.76 69.19 60.05 63.87 62.13 64.15 – – 55.97 58.54 53.08 55.18 67.85 70.03 66.38 87.39 61.4 63.59
5
Best 69.02 72.83 67.34 72.83 65.72 71.43 – – 56.49 58.54 53.38 55.18 68.01 70.03 66.72 87.39 62.19 63.59

Table A.7
MFPT: Results with random split and data augmentation.
Nor Input Loc AE DAE SAE MLP CNN LeNet AlexNet ResNet18 LSTM
Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max
Last 35.84 39.22 35.07 40.97 35.92 38.45 30.63 36.12 86.72 87.57 65.9 67.77 77.16 79.61 93.86 96.5 76.82 78.64
1
Best 41.83 45.83 41.09 42.72 41.59 42.91 40.29 41.36 89.36 91.07 69.63 71.46 78.6 80.97 95.92 96.89 78.52 79.61
Last 94.21 94.76 94.76 95.92 94.6 95.92 94.84 95.15 85.63 88.54 91.81 93.4 92.04 92.43 91.8 93.4 91.88 92.82
2
Best 95.53 95.92 95.84 96.12 95.61 95.92 96.23 96.5 88.39 89.51 92.97 93.4 92.97 93.59 94.6 95.34 93.09 93.59
A Last 23.8 26.62 23.91 27.76 26.62 29.66 – – 37.98 40.3 38.63 41.83 41.25 43.92 46.96 47.91 40 41.44
3
Best 28.03 28.9 27.68 29.85 30.34 33.84 – – 42.36 43.35 42.32 43.73 43.58 45.25 48.86 49.81 43.39 43.92
Last 88.47 90.68 87.23 88.74 88.35 90.68 – – 70.21 75.15 77.09 78.45 88.23 90.29 91.26 95.15 80.97 83.3
4
Best 90.52 92.04 90.56 92.62 90.45 92.04 – – 75.03 76.12 79.3 80.78 89.55 90.87 95.34 95.92 84.54 84.85
Last 46.83 52.23 48.86 51.26 46.25 49.71 – – 50.22 52.43 50.29 53.2 55.54 63.69 85.9 91.26 54.18 58.06
5
Best 51.18 52.23 53.75 55.73 50.02 54.76 – – 54.33 56.12 53.51 55.34 62.06 66.21 90.25 91.26 57.9 61.36
Last 47.03 49.71 46.56 48.93 47.53 48.54 49.79 51.26 87.26 89.13 67.3 68.54 84.97 87.57 92.97 95.73 77.09 79.03
1
Best 50.21 52.04 49.09 50.1 50.41 51.46 50.18 51.46 87.62 90.12 68.54 69.9 85.67 87.96 93.94 96.7 78.87 79.42
Last 93.59 96.31 93.44 95.92 92.15 95.15 94.14 94.95 85.59 86.8 91.19 92.62 91.49 92.23 92.08 93.4 92.47 93.2
2
Best 95.73 96.31 96.19 96.89 95.84 96.12 94.52 94.95 86.02 87.38 91.58 92.63 92.08 92.82 92.47 93.98 93.79 94.37
B Last 22.7 25.67 23.8 27.19 26.96 29.47 – – 39.39 41.06 38.59 41.63 42.09 44.11 46.92 48.86 39.58 40.87
3
Best 27.23 28.52 27.87 29.85 30.38 31.56 – – 39.54 41.63 39.65 41.63 42.17 44.11 47.87 48.86 43.12 43.92
Last 87.42 88.93 86.06 88.93 87.96 90.29 – – 72.04 73.59 72.04 76.7 86.79 87.96 91.57 95.53 79.46 80.97
4
Best 89.17 90.49 88.31 90.1 89.83 91.84 – – 73.01 74.56 73.94 77.86 87.1 87.96 95.07 95.53 82.21 83.11
Last 48.97 53.79 50.91 55.73 42.83 47.18 – – 50.8 53.59 51.41 54.17 62.18 65.05 89.01 90.68 57.17 58.64
5
Best 53.75 56.89 52.97 58.06 47.22 53.4 – – 51.26 54.56 52.07 54.17 62.56 65.44 89.32 90.68 60.08 61.17
Last 49.71 51.07 49.36 52.62 48.47 51.84 47.69 51.07 88.85 90.29 69.98 72.43 91.38 92.43 93.86 97.67 82.21 83.5
1
Best 50.72 52.62 50.52 53.2 49.9 52.82 49.47 51.07 90.76 91.84 70.14 73.2 92.54 93.4 96.04 97.67 84.55 87.38
Last 94.56 95.53 95.03 95.53 95.49 96.5 95.15 96.12 86.41 88.35 91.03 92.43 91.34 92.23 90.87 91.65 91.61 93.01
2
Best 96.16 96.5 96.58 96.89 96.62 97.09 95.88 97.28 87.85 88.93 91.77 92.62 92.54 93.4 92.2 93.01 92.54 93.4
C
(continued on next page)

Please cite this article as: Z. Zhao, T. Li, J. Wu et al., Deep learning algorithms for rotating machinery intelligent diagnosis: An open source benchmark study. ISA Transactions
(2020), https://doi.org/10.1016/j.isatra.2020.08.010.
22 Z. Zhao, T. Li, J. Wu et al. / ISA Transactions xxx (xxxx) xxx

Table A.7 (continued).


Nor Input Loc AE DAE SAE MLP CNN LeNet AlexNet ResNet18 LSTM
Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max
Last 26.01 30.42 24.4 27.95 28.03 30.23 – – 40 41.44 39.62 41.44 40.41 41.25 45.02 48.67 40.19 42.4
3
Best 28.82 32.51 27.95 29.09 30.27 31.75 – – 41.52 43.16 41.18 42.78 42.47 44.11 48.29 49.43 41.79 42.4
Last 88.23 89.71 87.5 88.74 88.58 90.68 – – 75.07 78.64 77.94 79.61 88.04 90.29 94.83 96.12 79.81 82.91
4
Best 90.21 91.46 89.83 91.07 90.91 92.23 – – 76.04 78.83 78.87 80 89.98 91.46 95.38 96.5 83.57 91.84
Last 56.04 61.36 51.54 57.09 50.33 57.48 – – 54.6 60 54.33 55.92 60.51 71.07 91.53 93.4 61.71 64.08
5
Best 60.47 66.41 55.11 64.47 54.45 61.36 – – 57.01 60 55.69 57.48 63.42 71.07 93.2 93.98 63.26 64.08

*A is the max–min normalization; B is the -1-1 normalization; C is the Z-score normalization.


1 is the time domain input, 2 is the frequency domain input; 3 is the wavelet domain input; 4 is the time domain sample after STFT; 5 is the time domain sample
reshape to a 2D matrix.

Table A.8
MFPT: Results with random split and without data augmentation.
Nor Input Loc AE DAE SAE MLP CNN LeNet AlexNet ResNet18 LSTM
Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max
Last 20.93 27.38 17.59 22.33 22.91 26.02 22.49 25.05 83.22 86.8 62.25 62.91 76.08 78.64 78.18 85.44 69.48 71.07
1
Best 28.54 32.04 26.52 28.35 27.81 30.29 27.77 28.74 87.07 87.96 65.24 67.18 78.37 82.33 86.88 87.96 72.58 75.15
Last 94.95 95.92 94.14 95.15 94.76 95.34 94.33 95.53 79.96 84.27 93.75 94.76 92.04 93.4 92.27 95.92 93.09 93.79
2
Best 96.23 96.5 96 96.7 96.27 96.89 95.96 96.5 82.68 84.66 93.86 94.95 93.36 94.17 95.85 96.31 94.06 94.76
A Last 26.12 29.09 23.08 24.52 25.59 28.14 – – 38.1 40.11 40.76 41.44 29.46 41.06 44.22 48.67 40.49 43.16
3
Best 28.14 29.66 27.49 28.71 29.17 30.42 – – 42.02 43.73 43.16 44.11 30.53 43.35 49.05 50.38 43.35 46.01
Last 89.01 90.49 89.79 91.46 88.74 90.1 – – 73.12 76.31 78.1 79.03 88.39 90.1 94.87 95.34 80.43 83.69
4
Best 89.98 91.26 91.26 92.23 89.9 91.26 – – 74.99 76.89 79.34 80.97 90.41 91.65 95.42 95.73 83.5 86.02
Last 47.49 51.07 48.23 53.01 46.6 51.26 – – 50.84 53.98 50.87 54.76 52.2 65.44 88.31 89.9 53.98 58.25
5
Best 49.98 54.76 51.42 57.09 48.82 55.53 – – 53.67 57.09 54.68 55.92 53.32 65.44 90.37 91.07 59.11 61.36
Last 25.32 27.96 22.33 23.3 25.52 28.16 25.4 27.18 85.2 86.99 63.57 66.21 80.62 82.91 86.99 87.96 65.94 72.04
1
Best 28.58 30.1 26.17 27.18 28.89 30.87 25.4 27.18 85.2 86.99 64.04 67.57 80.93 82.91 87.11 87.96 66.18 72.04
Last 92.93 95.34 92.19 94.95 93.67 95.92 91.96 94.76 80.04 80.97 93.09 93.4 90.99 92.23 94.68 95.92 93.2 94.56
2
Best 95.61 95.73 95.42 95.53 96.12 96.7 95.54 96.12 81.13 82.91 94.1 94.37 93.09 93.98 96.23 96.7 94.06 94.56
B Last 26.47 29.85 25.59 26.43 27.07 30.8 – – 39.24 40.87 38.86 40.11 39.92 42.59 44.22 46.58 39.54 40.11
3
Best 28.94 30.42 29.13 30.04 30 34.22 – – 39.92 42.59 39.58 41.44 41.1 42.59 44.83 47.53 40.44 41.63
Last 87.88 89.51 86.83 89.9 88.39 90.29 – – 71.84 76.5 77.59 79.22 86.56 87.77 94.02 95.34 80.58 82.52
4
Best 89.01 90.49 89.63 92.04 89.79 90.87 – – 75.15 77.09 78.95 79.81 88.85 89.71 95.69 96.31 82.41 83.11
Last 50.06 57.67 50.41 57.86 45.94 50.1 – – 50.14 52.04 51.11 54.76 56.66 61.36 78.52 90.49 56.89 59.61
5
Best 53.94 59.03 55.42 60.58 48.58 51.84 – – 51.65 53.4 51.92 54.76 57.82 64.27 83.57 90.49 57.57 60.39
Last 24.19 26.6 23.26 24.85 22.95 24.08 21.79 22.72 86.21 87.57 61.36 64.08 84.19 85.24 86.6 91.84 76.27 77.48
1
Best 26.68 28.35 25.98 26.99 26.6 27.96 24.47 26.41 86.8 88.56 61.75 64.08 84.82 86.8 88.74 91.84 77.09 79.22
Last 94.99 95.34 94.49 95.92 94.37 95.15 94.37 94.76 80.39 82.33 91.84 92.62 90.72 92.82 94.56 95.34 92.15 93.01
2
Best 96.35 96.89 96.58 97.48 96.12 96.89 96.04 96.31 81.51 82.52 92.43 93.4 92.82 93.59 95.49 95.92 93.9 94.37
C Last 22.17 25.67 25.89 28.14 27.76 31.18 – – 40.26 42.21 40.19 41.06 39.73 41.63 40.84 49.05 39.92 40.68
3
Best 28.1 32.13 28.25 29.28 30.8 32.7 – – 40.87 43.35 40.34 41.06 40.45 42.97 41.37 49.81 41.02 42.78
Last 89.28 89.71 88.35 89.9 88.97 91.07 – – 83.16 91.26 86.45 87.57 88.04 89.71 89.79 91.46 87.34 89.71
4
Best 89.71 90.29 89.05 90.1 90.95 92.43 – – 84.33 91.46 86.6 87.57 91.22 92.43 90.14 91.46 88.43 90.29
Last 54.1 57.86 53.55 60.58 50.6 53.59 – – 58.52 62.52 54.21 56.12 63.07 66.41 91.84 92.62 60.93 62.14
5
Best 57.48 60.97 56.31 62.33 53.59 56.7 – – 59.34 63.69 55.46 56.31 64 69.51 92.39 92.82 61.83 63.11

Table A.9
MFPT: Results with order split and data augmentation.
Nor Input Loc AE DAE SAE MLP CNN LeNet AlexNet ResNet18 LSTM
Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max
Last 35.28 40.69 36.39 41.27 35.43 41.07 31.09 37.62 86.03 90.02 69.64 73.9 80.27 83.3 92.32 93.86 79.08 81.57
1
Best 42.65 44.72 43.8 45.68 44.3 46.64 43.72 47.22 89.83 90.4 72.78 74.66 81.42 85.03 94.74 95.59 81.84 83.3
Last 93.59 95.2 95.47 96.55 95.01 95.39 95.55 96.16 87.03 88.1 91.63 93.09 92.55 93.09 92.4 94.63 91.94 93.09
2
Best 96.08 96.35 96.43 97.31 96.12 96.74 96.51 96.74 88.98 89.44 93.17 94.05 93.55 94.05 94.63 94.82 93.66 95.01
A Last 26.07 26.82 26.52 28.12 28.34 33.33 – – 39.89 41.53 39.96 41.34 34.45 43.76 44.95 45.81 39.63 42.27
3
Best 29.31 29.98 29.79 31.28 30.43 33.33 – – 43.35 44.69 44.1 45.62 35.42 46.18 48.46 49.16 42.42 44.13
Last 89.25 91.94 90.06 91.55 88.79 90.6 – – 77.7 80.81 78.31 81.77 90.75 91.75 90.6 96.55 80.69 81.96
4
Best 91.71 92.51 92.28 93.86 91.52 92.71 – – 78.77 81.96 80.69 82.92 92.29 92.71 97.2 97.7 85.14 86.37
Last 52.55 57.2 51.32 53.74 51.28 55.66 – – 50.86 55.47 53.24 55.47 65.87 69.1 74.82 89.44 60.61 63.53
5
Best 55.47 58.16 55.01 57.58 55.36 58.35 – – 57.16 57.77 56.89 57.97 68.18 73.51 92.28 92.9 63.45 66.6
(continued on next page)

Please cite this article as: Z. Zhao, T. Li, J. Wu et al., Deep learning algorithms for rotating machinery intelligent diagnosis: An open source benchmark study. ISA Transactions
(2020), https://doi.org/10.1016/j.isatra.2020.08.010.
Z. Zhao, T. Li, J. Wu et al. / ISA Transactions xxx (xxxx) xxx 23

Table A.9 (continued).


Nor Input Loc AE DAE SAE MLP CNN LeNet AlexNet ResNet18 LSTM
Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max
Last 52.97 54.89 50.29 52.02 51.55 52.98 51.87 54.89 89.44 90.98 72.5 74.47 87.39 89.44 93.16 95.39 80.08 81.19
1
Best 54.01 55.09 51.59 53.55 54.43 56.81 53.4 56.24 90.64 91.94 73.61 75.43 88.27 89.44 94.46 96.55 81.84 83.3
Last 91.82 93.86 94.74 95.2 91.79 94.05 93.38 95.39 85.58 88.48 91.99 93.86 92.35 93.47 92.44 94.05 92.09 93.28
2
Best 95.28 95.97 95.7 96.16 95.7 96.16 95.15 96.74 87.69 89.64 92.8 94.24 92.87 93.67 94.07 95.2 93.59 93.86
B Last 27.71 30.35 27.11 30.73 27.41 29.8 – – 40.83 41.9 40.99 42.64 41.61 43.95 45.23 46.74 40.19 41.53
3
Best 31.47 32.77 30.99 33.15 30.13 32.03 – – 42.13 43.76 43.11 45.44 42.62 44.88 46.6 48.42 42.83 43.58
Last 88.02 90.21 86.14 89.25 88.25 90.6 – – 76.87 79.46 74.4 79.85 88.13 89.44 92.65 96.55 82.23 84.26
4
Best 90.29 92.51 89.1 91.94 90.21 91.75 – – 77.71 79.65 77.06 80.42 89.55 91.17 93.25 97.5 84.57 85.41
Last 51.94 55.47 51.78 55.28 51.25 56.24 – – 56.72 61.23 55.85 57.58 66.05 69.87 77.68 92.71 61.61 65.07
5
Best 57.12 62.76 54.32 56.05 53.9 57.97 – – 58.09 61.42 57.68 61.61 67.92 72.74 92.1 93.47 64.18 66.6
Last 49.75 51.06 52.67 55.28 51.63 52.98 51.06 52.78 90.71 92.13 74.51 75.24 91.63 94.24 94.05 96.16 87.75 89.44
1
Best 52.55 53.74 53.86 56.62 53.74 54.51 51.29 52.98 92.17 94.05 74.89 76.58 92.4 94.43 94.7 96.16 88.48 89.44
Last 95.63 96.55 95.47 96.74 95.05 95.78 95.51 96.16 87.22 88.87 92.28 93.28 92.32 93.67 91.98 93.67 92.86 93.47
2
Best 96.74 97.12 96.51 96.74 96.36 96.74 95.85 96.35 88.02 88.87 92.78 93.47 92.63 93.67 92.86 93.67 93.13 93.47
C Last 28.31 31.1 30.1 31.1 29.53 32.03 – – 42.16 51.29 46.97 50.95 48.8 54.19 50.98 55.59 46.47 50.78
3
Best 31.32 33.71 30.76 31.47 33.26 35.75 – – 44.53 52.1 47.38 51.65 48.98 55.06 51.82 55.59 46.86 51.37
Last 90.36 92.51 89.25 90.98 89.71 92.13 – – 78.23 79.46 77.7 81 90.17 92.9 95.24 97.5 81.15 84.07
4
Best 92.93 93.86 92.71 95.01 92.94 94.63 – – 80.42 81 78.54 81.19 90.29 92.9 95.66 97.89 83.03 84.64
Last 57.62 61.42 54.51 57.2 51.32 60.08 – – 60.38 62 58.35 60.08 69.29 76.01 68.91 95.59 65.91 67.56
5
Best 62.3 67.56 58.54 60.46 58.93 62.38 – – 62.42 63.34 59.04 61.61 69.75 76.01 94.63 95.59 66.68 69.48

Table A.10
PU: Results with random split and data augmentation.
Nor Input Loc AE DAE SAE MLP CNN LeNet AlexNet ResNet18 LSTM
Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max
Last 47.22 53.15 44.91 53 48.79 53.3 46.18 49.16 79.02 81.87 68.76 72.66 82.46 84.33 89.74 92.17 75.02 76.04
1
Best 55.08 59.14 55.21 58.06 52.69 54.53 51.68 57.16 80.68 85.87 71.43 75.12 83.1 84.33 90.85 94.01 75.6 76.34
Last 97.45 97.85 97.27 97.85 97.64 98.31 98.31 99.08 95.18 96.16 94.19 96.16 96.53 98.31 97.14 97.54 94.99 95.39
2
Best 98.34 98.46 98.43 98.77 98.52 98.77 98.53 99.23 95.58 96.62 94.68 97.54 96.84 98.31 97.66 98.46 95.11 95.55
A Last 30.08 33.58 29.74 31.63 31.42 32.23 – – 36.37 39.88 33.52 38.08 33.52 35.23 41.77 47.68 37.39 38.68
3
Best 33.34 36.58 33.67 34.63 34.36 36.13 – – 37.24 40.63 34.27 38.83 34.69 36.73 43.93 48.28 37.9 39.28
Last 94.22 95.7 95.51 96.31 93.98 96.31 – – 82.52 85.71 82.3 84.18 94.26 95.55 93.03 97.7 84.55 86.94
4
Best 96.59 96.93 96.96 97.39 95.79 97.08 – – 83.01 87.86 82.8 84.49 95.08 96.31 93.68 98.82 86.7 88.48
Last 56.99 68.36 63.96 68.05 59.36 62.67 – – 45.71 53.46 44.09 49.16 7.83 7.83 91.37 94.47 48.63 52.38
5
Best 66.76 73.73 68.26 72.35 65.19 70.51 – – 48.83 56.61 47.59 51.31 7.83 7.83 94.1 95.55 53.67 59.29
Last 65.95 66.97 62.13 64.06 65.79 67.74 70.72 72.2 83.44 86.18 72.26 72.96 86.61 88.94 91.58 93.24 75.88 77.11
1
Best 67.61 68.97 63.59 65.9 66.62 68.82 71.24 72.66 84.55 86.33 73.18 74.5 87.5 90.02 92.32 94.01 78.49 79.42
Last 97.06 98.31 96.26 97.24 96.49 98 97.69 98.46 93.7 95.55 95.02 96.31 97.05 97.54 90.91 98 94.66 95.39
2
Best 98.28 98.46 98.08 98.46 97.83 98.16 97.97 98.46 95.15 95.7 95.42 96.62 97.24 97.54 97.76 98.16 96.31 96.47
B Last 29.41 32.53 31.86 35.83 33.91 37.03 – – 37.6 40.63 38.86 40.33 33.4 35.98 45.04 47.98 39.76 41.68
3
Best 32.65 34.93 34.2 38.53 37.33 38.38 – – 39.52 43.18 39.46 41.53 34.06 37.63 45.43 47.98 42.82 44.53
Last 91.94 94.32 92.06 93.39 93.42 95.08 – – 77.17 85.71 76.19 79.42 91.06 92.78 95.14 99.23 82 84.64
4
Best 93.78 96.47 94.42 96.31 95.03 96.01 – – 77.64 85.71 78.96 81.26 91.83 93.55 99.02 99.39 87.1 89.25
Last 60.01 66.67 56.91 66.82 62.88 68.82 – – 33.79 55.61 50.38 53.92 49.09 61.6 90.23 95.85 57.33 59.75
5
Best 66.36 72.04 63.59 72.04 65.97 70.05 – – 34.87 57.3 51.58 54.85 49.09 61.6 94.87 95.85 60.09 60.83
Last 69.83 71.58 64.98 68.05 70.72 71.43 71.09 72.96 85.96 88.63 78.92 80.03 90.57 92.17 93.92 95.24 80.31 82.03
1
Best 70.97 74.04 66.73 69.59 71.31 72.04 72.19 72.96 86.79 88.63 78.99 80.18 90.66 92.17 94.44 96.01 81.26 83.26
Last 97.45 98 97.7 98.62 97.42 97.54 97.94 98.92 95.3 96.31 95.33 95.85 97.3 98.16 97.17 98 95.02 95.55
2
Best 98.25 98.62 98.5 98.62 98.34 98.62 98.22 98.92 95.55 96.47 95.85 96.93 97.73 98.16 97.64 98.77 95.36 96.16
C Last 25.52 33.58 29.95 32.23 33.01 34.63 – – 39.1 41.68 39.61 41.53 37 38.98 41.86 47.08 40.21 41.68
3
Best 26.84 33.73 32.77 34.03 36.04 38.53 – – 40.45 41.68 39.82 42.58 38.02 41.23 45.91 47.08 41.29 41.83
Last 94.96 96.01 93.64 95.39 94.69 97.54 – – 82.34 85.41 79.85 82.03 91.95 93.55 92.87 98.16 85.1 86.64
4
Best 96.07 96.47 95.02 96.31 96.44 97.54 – – 83.41 86.64 81.04 82.64 92.47 94.62 93.18 98.92 86.33 86.94
Last 59.11 65.75 56.62 66.36 62.15 69.59 – – 38.55 64.36 53.09 55.61 28.45 62.37 91.55 96.62 59.51 60.83
5
Best 64.27 66.05 61.26 70.51 67.77 72.96 – – 39.11 65.13 53.58 55.61 29.31 62.37 95.85 96.77 59.85 61.44

*A is the max–min normalization; B is the -1-1 normalization; C is the Z-score normalization.


1 is the time domain input, 2 is the frequency domain input; 3 is the wavelet domain input; 4 is the time domain sample after STFT; 5 is the time domain sample
reshape to a 2D matrix .

Please cite this article as: Z. Zhao, T. Li, J. Wu et al., Deep learning algorithms for rotating machinery intelligent diagnosis: An open source benchmark study. ISA Transactions
(2020), https://doi.org/10.1016/j.isatra.2020.08.010.
24 Z. Zhao, T. Li, J. Wu et al. / ISA Transactions xxx (xxxx) xxx

Table A.11
PU: Results with random split and without data augmentation.
Nor Input Loc AE DAE SAE MLP CNN LeNet AlexNet ResNet18 LSTM
Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max
Last 37.7 44.19 29.59 34.33 38.31 43.13 50.39 56.84 78.96 84.95 71.49 74.96 82.74 84.49 81.85 85.41 73.24 77.11
1
Best 46.85 48.5 39.36 40.89 45.63 47.25 55.95 60.22 79.97 84.95 73.92 75.27 83.23 85.87 85.29 95.24 74.62 77.27
Last 74.62 93.38 91.91 93.76 92.66 93.2 97.91 98.46 90.57 91.86 95.85 97.08 95.18 96.77 98.77 99.85 93.86 95.55
2
Best 75.48 93.88 92.86 93.95 93.5 94.01 98.43 98.92 90.72 91.86 96.13 97.54 95.57 97.85 99.48 99.85 94.29 96.01
A Last 31.03 33.58 31.06 33.28 32.86 37.03 – – 46.3 49.36 47.1 48.59 43.36 54.62 47.94 55.64 46.65 48.56
3
Best 34.12 37.18 33.58 35.83 35.2 38.23 – – 48.04 51.26 48.26 49.37 43.5 55.07 53.56 55.8 47.34 48.56
Last 89.31 90.01 89.83 90.89 89.45 90.57 – – 77.63 86.94 81.38 82.03 92.35 95.7 95.48 98.46 78.06 86.64
4
Best 90.72 91.14 90.94 91.64 89.98 91.2 – – 85.74 87.71 81.81 82.49 94.01 95.7 96.99 98.62 84.73 88.17
Last 56.09 59.74 56.02 59.3 53.06 60.61 – – 51.21 54.22 49.65 54.84 7.83 7.83 89.59 94.47 55.36 58.37
5
Best 59.51 61.67 58.5 61.99 58.54 67.23 – – 55.21 55.91 55.24 57.76 7.83 7.83 94.99 95.39 58.37 59.91
Last 55.01 55.68 31.37 38.14 53.08 55.43 70.2 72.81 84.06 85.56 72.04 73.73 87.04 87.71 89.68 93.86 72.41 76.65
1
Best 55.98 56.55 31.85 39.26 54.79 56.18 71.27 72.81 84.67 86.48 73.49 75.73 87.65 88.79 90.87 95.39 75.67 78.34
Last 90.79 92.7 89.63 93.26 92.52 93.95 80.92 97.85 88.63 92.17 96.44 97.39 95.73 97.24 97.94 99.54 93.67 95.39
2
Best 94.06 94.32 93.84 94.19 94.16 94.51 95.15 98.92 88.83 92.17 96.59 97.39 96.41 97.24 99.32 99.69 94.87 95.7
B Last 32.47 34.78 32.71 38.68 35.35 37.93 – – 50.56 52.14 50.27 51.56 53.39 53.87 53.64 55.46 48.34 48.76
3
Best 34.54 36.13 34.75 40.18 36.7 39.13 – – 51.13 52.29 50.77 51.56 53.76 55.07 54.32 56.22 48.81 49.24
Last 87.88 90.32 87.45 87.83 86.22 90.14 – – 83.76 88.33 76.01 81.57 91.92 92.93 81.97 98.62 81.87 86.48
4
Best 88.71 90.32 88.4 89.58 88.26 90.45 – – 84.16 88.63 77.6 83.41 92.11 93.55 98.22 99.08 82.43 86.48
Last 57.7 60.61 58.94 61.17 58.69 60.86 – – 51.36 51.46 51.77 52.69 51.77 68.2 94.77 95.08 55.54 57.45
5
Best 60.43 61.92 62.66 67.79 62.26 66.42 – – 53.97 54.38 55.49 56.84 54.47 72.04 95.11 95.85 59.69 61.44
Last 59.39 60.36 51.25 54.18 57.15 58.49 71.77 72.81 83.41 87.1 77.94 80.18 91.43 92.78 91.92 95.55 80.18 81.57
1
Best 60.18 62.3 52 54.43 57.81 58.93 72.53 73.12 85.53 87.1 79.02 80.18 91.98 92.78 92.75 97.08 80.83 82.64
Last 93.12 93.95 92.8 93.51 93.17 94.38 97.85 98.46 89.43 92.63 95.61 96.77 97.3 98.92 99.63 99.69 94.07 94.78
2
Best 94.07 94.26 94.24 94.57 94.13 94.38 98.52 98.77 89.43 92.63 95.79 96.77 97.82 98.92 99.63 99.69 94.72 96.01
C Last 28.04 32.23 24.71 29.39 34.45 37.03 – – 51.58 52.05 50.88 51.84 54.5 55.91 55.21 56.36 48.96 49.6
3
Best 31.12 33.88 29 33.43 37.15 40.93 – – 51.99 52.91 51.36 52.33 54.86 55.91 55.4 57.24 49.4 49.93
Last 90.04 90.89 89.96 91.51 83.43 90.82 – – 82.8 86.33 79.17 82.03 92.17 95.39 97.79 98.62 85.38 86.94
4
Best 90.4 91.07 90.26 91.51 84.14 91.07 – – 83.07 86.64 79.45 82.03 93.24 95.39 97.94 98.62 86.54 87.71
Last 56.34 59.68 47.64 62.48 58.04 62.98 – – 45.16 57.3 51.43 53.76 7.83 7.83 89.28 96.62 59.63 61.14
5
Best 60.26 65.04 49.74 63.17 60.12 64.04 – – 50.31 59.45 56.84 58.68 8.35 10.45 96.44 96.93 62.18 63.13

Table A.12
PU: Results with order split and data augmentation.
Nor Input Loc AE DAE SAE MLP CNN LeNet AlexNet ResNet18 LSTM
Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max
Last 48.11 53.3 49.86 55.45 48.33 52.69 51.06 55.3 80.83 82.8 64.24 71.43 83.9 84.95 85.07 93.7 71.64 73.89
1
Best 54.75 56.68 54.75 57.14 55.08 57.6 54.41 62.37 82.22 85.6 67.96 71.43 84.73 86.79 90.38 93.7 72.01 74.96
Last 97.63 98.46 97.63 98.92 97.67 98.46 89.54 92.31 66.15 72.31 77.54 84.62 65.54 75.38 88 92.31 76.92 80
2
Best 99.14 99.23 98.89 98.92 98.77 98.92 95.69 98.46 69.23 72.31 82.77 87.69 69.23 75.38 91.08 93.85 82.77 86.15
A Last 29.96 32.2 28.42 31.61 29.69 31.61 – – 32.59 37.52 32.94 36.48 25.61 31.02 41.54 44.46 34.5 35.89
3
Best 32.79 35.01 32.61 34.56 33.38 35.3 – – 35.3 38.4 34.53 37.37 25.61 31.02 42.63 46.23 35.63 37.52
Last 94.44 95.55 93.7 94.62 94.32 95.24 – – 78.74 83.56 77.42 78.65 93.24 95.24 95.45 98.92 83.96 88.33
4
Best 96.9 97.39 96.37 97.08 95.82 96.62 – – 80.18 83.56 78.83 81.72 93.52 95.55 97.97 98.92 87.07 88.33
Last 61.2 66.97 57.7 62.98 58 65.59 – – 42.58 49.31 45.16 51.61 7.83 7.83 85.9 94.47 45.53 51.15
5
Best 64.52 68.51 62.43 69.89 62.03 67.9 – – 45.68 50.84 48.82 51.61 7.83 7.83 87.19 95.39 50.51 51.92
Last 65.5 67.9 63.1 65.28 66.82 69.59 69.65 71.58 79.66 84.64 70.2 72.04 86.95 88.33 91.58 93.39 74.38 75.88
1
Best 67.34 69.12 64.33 67.13 68.36 69.89 70.08 71.89 80.74 85.25 71.15 74.5 87.62 88.33 92.44 94.47 76.4 77.11
Last 97.2 98.46 98.19 98.46 96.59 98 75.69 89.23 58.15 67.69 79.08 84.62 66.77 70.77 72.31 80 78.15 83.08
2
Best 98.95 99.23 99.05 99.23 98.59 98.92 92.31 93.85 68.31 73.85 82.15 86.15 68.62 72.31 88.92 90.77 82.16 84.62
B Last 31.82 33.68 29.87 32.35 31.28 31.76 – – 37.52 39 35.66 36.93 31.34 34.86 42.51 45.2 36.1 38.85
3
Best 34.65 36.48 33.03 35.45 34.5 35.45 – – 37.84 39.44 36.87 38.7 32.26 34.86 43.4 45.2 39.44 41.36
Last 92.1 94.62 91.58 92.78 92.23 93.24 – – 80.15 84.02 72.72 81.72 91.52 92.47 90.38 98 83.84 86.18
4
Best 94.13 97.24 94.56 95.85 94.22 95.39 – – 81.66 84.02 74.5 82.95 91.74 92.63 97.02 98.31 86.54 87.71
Last 61.32 66.21 60.46 63.13 56.93 66.67 – – 48.54 52.23 48.05 50.54 30.87 67.74 88.88 95.55 50.63 53.46
5
Best 64.88 67.74 64.64 66.36 67.96 71.43 – – 48.6 52.38 48.76 50.54 30.87 67.74 94.84 95.55 54.5 57.6
(continued on next page)

Please cite this article as: Z. Zhao, T. Li, J. Wu et al., Deep learning algorithms for rotating machinery intelligent diagnosis: An open source benchmark study. ISA Transactions
(2020), https://doi.org/10.1016/j.isatra.2020.08.010.
Z. Zhao, T. Li, J. Wu et al. / ISA Transactions xxx (xxxx) xxx 25

Table A.12 (continued).


Nor Input Loc AE DAE SAE MLP CNN LeNet AlexNet ResNet18 LSTM
Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max
Last 69.4 71.43 65.83 66.97 69.31 71.58 72.93 75.73 84.73 86.64 75.7 77.57 91.15 91.71 93.67 95.39 78.43 78.96
1
Best 70.6 71.89 67.4 69.43 70.75 72.96 73.46 75.73 86.18 88.33 76.4 78.03 91.8 93.7 95.02 96.16 80.06 81.87
Last 98.62 99.23 99.05 99.39 98.53 99.23 88.61 90.77 64 66.15 78.77 86.15 77.23 80 85.54 89.23 78.15 81.54
2
Best 99.14 99.54 99.36 99.39 99.36 99.54 93.54 93.85 74.15 76.92 82.15 86.15 81.85 84.62 92.92 95.38 84.62 86.15
C Last 27.5 31.76 30.07 33.68 31.7 32.79 – – 37.93 40.62 36.1 37.37 34.71 37.08 44.19 46.82 38.08 38.85
3
Best 30.46 34.27 32.41 35.01 34.92 36.63 – – 38.85 40.62 37.19 38.85 35.36 37.37 45.73 46.82 38.76 40.92
Last 94.35 95.7 93.95 95.85 94.75 95.55 – – 82.43 85.56 81.07 82.49 92.2 94.32 94.65 98.92 84.7 85.56
4
Best 96.44 97.08 95.79 96.62 96.31 97.24 – – 82.67 85.56 81.78 82.95 93.15 94.32 98.34 98.92 85.41 88.63
Last 61.97 72.81 60.89 63.29 61.78 69.12 – – 44.49 58.53 50.26 52.84 35.51 63.29 96.5 97.24 54.5 55.61
5
Best 67.1 75.88 64.45 69.89 66.21 73.43 – – 46.48 58.53 50.78 52.84 37.26 63.29 96.65 97.24 55.85 56.37

Table A.13
SEU: Results with random split and data augmentation.
Nor Input Loc AE DAE SAE MLP CNN LeNet AlexNet ResNet18 LSTM
Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max
Last 50.25 53.19 45.1 52.7 46.91 52.7 60.64 65.69 88.24 94.36 83.19 86.27 53.43 75.74 95.73 99.75 87.38 91.67
1
Best 61.72 66.18 62.55 64.22 65.1 68.14 63.87 66.91 90.98 95.59 85.25 87.75 55.15 78.43 97.15 99.75 87.93 92.65
Last 97.16 97.79 97.45 98.28 97.16 98.04 97.35 97.55 97.3 98.04 96.81 98.53 96.03 96.81 97.25 97.55 96.02 96.81
2
Best 98.14 98.53 98.38 98.77 98.48 98.77 97.79 98.28 97.55 98.53 97.3 98.53 97.4 98.04 97.7 98.53 96.5 97.55
A Last 50.39 54.57 53.94 57.69 55.24 63.46 – – 52.94 59.62 45.34 54.81 64.95 66.59 59.52 65.14 54.69 55.77
3
Best 52.4 57.45 58.99 62.02 60.39 66.35 – – 53.7 59.62 45.48 54.81 66.02 67.55 61.88 66.59 56.07 57.69
Last 97.5 99.02 97.99 99.02 98.09 98.77 – – 84.41 88.73 84.31 89.46 78.04 98.04 96.22 100 88.48 90.44
4
Best 98.77 99.51 99.07 99.51 98.72 99.26 – – 86.81 90.2 85 90.2 79.12 98.77 99.66 100 89.95 91.67
Last 52.89 86.76 66.32 83.58 56.37 79.9 – – 20.24 43.38 34.61 43.14 4.9 4.9 67.11 100 43.38 46.08
5
Best 54.22 88.48 69.27 88.73 61.42 81.13 – – 21.92 50.52 42.7 49.75 4.9 4.9 82.94 100 44.98 47.06
Last 77.55 79.9 79.85 82.84 78.24 80.39 82.6 83.82 92.84 95.83 86.08 89.71 89.26 94.36 99.46 99.75 90.35 91.91
1
Best 79.61 82.6 81.08 83.09 78.97 80.39 83.48 85.78 93.14 96.81 87.79 91.91 90.78 94.36 99.66 100 91.03 93.63
Last 95.59 97.3 95.74 97.06 94.9 97.06 96.47 97.79 97.84 98.28 97.25 97.55 96.96 97.55 96.17 97.79 96.86 97.55
2
Best 97.7 98.04 97.89 98.28 97.84 98.04 96.96 97.79 98.09 98.77 97.44 97.99 97.25 97.79 96.76 98.28 97.3 98.28
B Last 54.85 56.97 54.9 60.58 56.2 61.06 – – 52.74 59.62 50.34 54.09 62.31 65.38 60.67 66.59 56.63 60.1
3
Best 58.89 62.26 57.5 61.78 59.57 64.18 – – 54.47 60.1 51.68 56.01 63.85 66.35 62.65 71.88 57.98 62.5
Last 97.4 98.28 95.93 98.53 97.45 98.04 – – 87.25 89.71 86.32 89.71 96.32 98.53 99.26 100 88.38 91.42
4
Best 98.92 99.51 97.45 98.77 98.87 99.51 – – 88.68 90.69 87.06 91.42 96.86 98.53 99.8 100 89.26 91.42
Last 86.08 88.48 81.23 86.52 73.43 87.75 – – 14.02 50.49 43.48 47.55 59.41 85.05 97.16 99.51 46.77 49.02
5
Best 88.87 92.16 87.3 91.91 78.48 90.44 – – 14.12 50.49 45.34 47.55 61.42 90.44 97.4 100 47.6 50.49
Last 84.31 87.25 83.72 86.03 83.33 90.44 87.79 91.18 95.83 97.3 93.33 94.85 99.02 99.51 100 100 96.86 97.55
1
Best 85.78 87.25 85.15 87.25 85.1 90.44 88.38 91.18 96.27 97.3 93.97 96.32 99.26 100 100 100 97.3 98.28
Last 97.74 98.04 97.79 98.28 97.84 98.53 97.84 98.28 98.82 99.26 97.6 98.04 98.18 99.51 97.84 98.53 97.5 98.28
2
Best 98.58 98.77 99.02 99.02 98.67 98.77 98.15 99.02 99.12 99.51 97.99 99.26 98.67 100 98.09 99.02 97.7 98.53
C Last 51.68 54.09 49.42 53.85 56.4 59.86 – – 57.93 63.94 56.97 59.86 63.12 65.62 71.88 74.76 60.29 63.94
3
Best 54.9 58.65 52.02 55.53 59.71 61.54 – – 61.59 63.94 57.74 61.54 64.09 66.59 72.89 74.76 61.06 63.94
Last 96.96 99.26 98.63 98.77 97.55 99.02 – – 90 92.16 87.4 90.2 96.81 99.02 99.7 100 90.05 91.18
4
Best 98.48 99.75 99.46 99.75 99.26 100 – – 90.34 92.65 88.38 90.2 97.85 99.02 99.8 100 90.98 91.91
Last 89.95 92.16 88.04 89.95 86.47 92.16 – – 31.72 53.92 51.23 56.13 63.04 80.64 99.26 100 52.84 56.13
5
Best 93.19 96.57 91.22 94.36 92.26 94.12 – – 32.5 54.66 52.3 56.13 63.63 81.13 99.41 100 55.07 58.18

*A is the max–min normalization; B is the -1-1 normalization; C is the Z-score normalization.


1 is the time domain input, 2 is the frequency domain input; 3 is the wavelet domain input; 4 is the time domain sample after STFT; 5 is the time domain sample
reshape to a 2D matrix.

Table A.14
SEU: Results with random split and without data augmentation.
Nor Input Loc AE DAE SAE MLP CNN LeNet AlexNet ResNet18 LSTM
Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max
Last 40.13 44.85 31.52 34.56 36.42 42.16 37.45 44.61 88.09 91.42 83.24 87.01 69.93 72.55 98.04 99.02 79.22 81.37
1
Best 44.67 48.53 35.54 37.99 43.78 45.59 42.94 48.28 88.38 92.16 83.58 87.5 72.57 77.77 98.28 99.51 80.78 82.35
Last 96.71 97.55 97.35 98.04 97.11 97.55 97.6 98.28 96.86 99.26 98.09 98.53 97.2 97.55 99.85 100 97.35 97.79
2
Best 98.38 98.53 98.58 98.77 98.43 98.77 97.99 98.28 97.4 99.75 98.43 99.26 97.5 98.28 99.9 100 97.94 98.77
A Last 57.6 63.46 57.45 60.1 54.18 60.34 – – 48.51 56.97 50.53 54.57 62.62 65.87 52.16 62.98 54.28 56.73
3
Best 61.54 65.38 60.53 62.74 60.14 62.5 – – 53.89 57.69 52.39 54.57 63.16 66.35 56.3 72.6 55.91 57.93
Last 98.58 99.02 98.18 98.77 98.09 98.77 – – 85.69 90.69 88.24 90.2 96.42 97.06 99.85 100 88.78 90.2
4
Best 98.97 99.51 99.21 99.75 98.77 99.51 – – 86.52 90.69 89 91.42 97.35 98.77 99.9 100 89.81 90.69
Last 64.12 86.03 69.8 86.03 80.59 89.46 – – 14.31 42.65 33.97 39.46 4.9 4.9 98.72 100 41.42 48.04
5
Best 66.76 87.5 75.05 89.95 83.14 89.71 – – 14.46 42.65 36.52 44.85 5.05 5.15 98.77 100 46.08 50.49
(continued on next page)

Please cite this article as: Z. Zhao, T. Li, J. Wu et al., Deep learning algorithms for rotating machinery intelligent diagnosis: An open source benchmark study. ISA Transactions
(2020), https://doi.org/10.1016/j.isatra.2020.08.010.
26 Z. Zhao, T. Li, J. Wu et al. / ISA Transactions xxx (xxxx) xxx

Table A.14 (continued).


Nor Input Loc AE DAE SAE MLP CNN LeNet AlexNet ResNet18 LSTM
Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max
Last 44.97 48.28 36.13 39.71 43.22 44.61 43.87 47.06 87.35 89.22 81.62 84.31 85.25 97.06 97.89 99.26 79.66 82.35
1
Best 49.02 52.7 37.94 41.42 47.96 50.98 44.27 47.55 87.99 89.95 82.65 84.31 85.78 97.55 98.77 99.26 80.64 82.35
Last 92.61 97.55 78.58 91.91 79.7 97.79 93.87 97.3 98.73 99.51 97.89 98.28 97.5 98.28 99.66 100 97.5 98.04
2
Best 98.16 98.28 98.08 98.77 98.28 98.53 94.85 98.04 98.92 99.51 98.23 99.02 97.6 98.28 99.66 100 97.89 98.28
B Last 55.89 61.06 57.11 58.89 59.05 65.62 – – 56.39 59.62 48.75 51.2 62.36 65.87 59.71 66.59 55.14 58.41
3
Best 58.97 64.42 58.75 60.34 60.38 65.87 – – 58.46 59.62 51.64 55.77 63.99 66.35 64.81 72.6 57.07 61.06
Last 94.08 99.02 97.01 98.28 97.74 99.26 – – 89.22 91.67 79.22 90.69 96.13 96.57 99.85 100 86.37 88.24
4
Best 98.49 99.02 98.23 99.26 98.43 99.26 – – 89.66 91.67 88.68 91.91 96.71 98.28 99.85 100 87.4 90.93
Last 85.7 92.4 86.52 90.44 83.68 90.93 – – 13.38 47.3 45.64 49.51 14.8 54.41 99.46 99.75 45.1 46.08
5
Best 86.97 92.4 88.33 90.93 85.15 92.16 – – 13.38 47.3 46.82 51.23 14.85 54.41 99.61 100 46.81 50.25
Last 46.03 48.77 34.9 39.46 48.23 49.75 43.53 48.04 88.14 92.16 87.7 91.67 96.91 98.53 99.56 100 91.81 93.38
1
Best 50.88 53.43 40.25 42.4 50.34 52.21 45.44 50.74 91.62 92.89 88.43 93.14 97.74 100 99.56 100 92.74 93.38
Last 98.14 99.02 97.89 99.02 97.79 98.28 97.84 98.28 98.38 99.51 99.16 99.75 99.65 100 100 100 98.33 99.26
2
Best 98.87 99.02 99.36 99.51 98.87 99.26 97.99 98.53 98.58 99.51 99.31 99.75 99.75 100 100 100 98.63 99.51
C Last 43.37 61.3 40.67 50.72 54.71 57.69 – – 59.61 61.54 55.86 58.17 61.54 63.46 69.33 72.12 59.23 63.7
3
Best 45.05 62.5 42.26 52.16 57.16 59.13 – – 63.3 64.42 57.45 58.65 62.69 64.9 70.92 75.96 60.63 63.7
Last 98.87 99.51 98.53 99.26 98.58 99.51 – – 89.56 91.18 87.55 88.73 95.39 96.81 100 100 90.15 92.65
4
Best 99.02 99.51 98.68 99.51 99.12 99.51 – – 90.88 92.65 88.09 90.2 95.83 97.79 100 100 91.96 94.36
Last 91.32 93.63 91.96 94.85 90.83 92.65 – – 4.9 4.9 47.01 53.43 51.23 87.5 99.9 100 53.04 55.15
5
Best 92.4 94.61 93.04 96.32 91.77 93.14 – – 4.95 5.15 48.97 55.64 51.28 87.5 99.95 100 53.83 57.11

Table A.15
SEU: Results with order split and data augmentation.
Nor Input Loc AE DAE SAE MLP CNN LeNet AlexNet ResNet18 LSTM
Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max
Last 38.81 45.95 31.81 34.76 39.9 49.52 38.05 46.67 83.24 90.71 67.05 74.76 67.62 73.1 95.53 98.33 77.76 79.76
1
Best 44.48 47.86 48.43 52.14 46.33 49.52 41.24 46.67 84.05 90.71 67.38 74.76 69.48 75.1 96.05 99.05 79.24 81.9
Last 99.24 99.76 99 100 98.71 99.52 99.38 99.76 98.48 99.76 98.24 99.52 98.62 99.05 98.38 99.52 99.48 99.76
2
Best 100 100 100 100 100 100 99.57 99.76 98.86 99.76 98.33 99.52 98.9 99.52 98.81 99.76 99.52 100
A Last 55.57 59.76 54.38 55.71 55.43 64.05 – – 51.48 58.33 50.28 61.9 57.29 60 53 64.76 53.62 55
3
Best 59.47 62.14 60.9 62.14 60.1 64.05 – – 55.1 59.05 55 61.9 59.52 63.57 54.52 68.57 55.72 60.24
Last 88 90.48 87.62 92.38 89.62 95 – – 80.62 87.38 84.05 86.19 97.08 97.86 98.81 99.29 88.62 90.24
4
Best 93.48 96.43 94.48 96.67 94.09 96.67 – – 82.71 88.33 84.95 87.62 97.86 98.57 98.91 99.29 90.62 92.38
Last 47.24 63.33 60.05 73.33 50.29 53.57 – – 13.62 27.62 34.67 41.43 5 5 94.86 98.1 30.62 38.57
5
Best 51.19 65.95 66.33 76.19 55.53 65 – – 13.76 27.62 38.91 41.67 5 5 97.1 98.57 32.57 41.43
Last 99 99.52 95.33 99.29 96.57 99.05 54.52 58.33 88.95 92.38 64.81 69.05 89.43 94.76 99.19 99.76 85.59 86.67
1
Best 99.86 100 99.9 100 100 100 56.24 62.62 91.14 93.33 67.05 70 91.19 96.9 99.43 99.76 86.61 88.1
Last 50.14 57.14 52.29 59.52 53.81 54.29 98.33 99.76 97.67 98.81 99.28 99.76 98.28 99.52 95.67 99.29 98.93 99.76
2
Best 53.95 57.14 57.38 60.95 59.48 61.43 99.48 99.76 98.29 99.29 99.52 99.76 99 99.52 96.57 99.76 99.52 99.76
B Last 91.95 92.86 89.43 93.81 88.76 92.86 – – 50.43 53.81 54.29 56.43 60.81 63.81 64.62 67.86 52.98 56.19
3
Best 95.43 96.43 94 97.62 93.38 95.71 – – 53.19 56.9 55 59.05 62.29 65.95 65.86 69.29 56.38 58.62
Last 54.76 57.86 53.81 58.33 41.14 55.24 – – 83.87 87.38 84.57 86.67 95.48 96.67 93.76 99.05 85.36 86.43
4
Best 59.9 64.52 60 63.57 46.91 61.67 – – 85.42 87.38 85.33 86.9 96.33 98.57 94.09 99.52 86.49 88.33
Last 99 99.52 95.33 99.29 96.57 99.05 – – 44.29 44.29 37.86 44.52 64.52 68.81 97.33 98.57 39.94 42.38
5
Best 99.86 100 99.9 100 100 100 – – 46.67 46.67 39.27 44.52 68.65 76.9 97.62 98.57 40.48 42.38
Last 56.52 59.29 63.24 68.1 59.19 60.95 56.29 58.57 93.52 95.24 82.95 85.48 97.24 99.29 97.29 99.29 93.29 94.76
1
Best 59.91 64.05 66 68.33 62.19 62.86 58.81 63.57 94.71 97.14 82.95 85.48 98.1 99.29 98.05 99.29 93.52 95
Last 99.52 100 99.52 99.76 99.1 99.29 99.24 99.52 99.14 99.52 99 99.52 99.38 99.76 99.09 99.76 99 99.29
2
Best 99.95 100 99.95 100 99.9 100 99.33 99.76 99.48 100 99.28 99.52 99.48 100 99.81 100 99.29 99.76
C Last 46.48 48.81 51.14 56.43 53.24 55.95 – – 56.24 58.33 54.24 55.24 62.38 63.57 62.53 69.05 57.71 62.14
3
Best 50.38 55.24 53.29 60.24 56.81 58.57 – – 59 62.62 56.14 58.81 62.95 64.76 65.48 73.57 58.76 62.14
Last 86.14 91.19 91.67 94.29 88.62 95.95 – – 83.95 91.43 84.67 86.67 95.67 97.86 99.09 99.52 88.86 91.67
4
Best 92.71 97.38 94.33 95.48 94.05 98.1 – – 86.33 91.43 86.05 88.57 97.48 98.57 99.19 99.76 91.81 93.81
Last 55.33 58.81 55 59.05 55.28 62.62 – – 22.48 53.57 42.91 48.1 31 77.14 97.95 99.52 45.28 48.1
5
Best 58.09 59.52 60.71 65.48 62.47 65.24 – – 23.29 53.57 44.29 48.81 33.09 82.14 98.52 99.76 45.76 48.1

Please cite this article as: Z. Zhao, T. Li, J. Wu et al., Deep learning algorithms for rotating machinery intelligent diagnosis: An open source benchmark study. ISA Transactions
(2020), https://doi.org/10.1016/j.isatra.2020.08.010.
Z. Zhao, T. Li, J. Wu et al. / ISA Transactions xxx (xxxx) xxx 27

Table A.16
UoC: Results with random split and data augmentation.
Nor Input Loc AE DAE SAE MLP CNN LeNet AlexNet ResNet18 LSTM
Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max
Last 27.09 29.22 27.49 29.83 25.57 28.01 26.88 28.92 41.98 48.25 31.93 33.49 11.11 11.11 66.18 76.26 33.12 34.09
1
Best 29.92 31.35 31.66 32.57 28.46 29.38 30.2 31.2 47.55 51.45 33.94 35.46 11.45 12.79 76.8 78.39 36.89 37.9
Last 92.94 94.06 90.78 91.63 92.76 93.46 92.66 93.46 68.49 71.08 79.97 81.58 70.62 75.49 88.16 89.65 79.21 80.67
2
Best 94.7 95.13 93.3 93.76 94.55 95.13 94.91 95.13 69.53 71.08 84.78 86.45 73.3 78.69 90.59 91.93 80.58 81.74
A Last 15.25 19.44 19.11 24.04 21.39 24.78 – – 24.57 31.6 34.21 38.13 11.13 11.13 62.34 63.8 37.48 40.36
3
Best 17.57 23.15 20.65 24.04 23.95 26.71 – – 26.44 32.05 36.26 40.36 11.16 11.28 67.3 69.14 38.9 42.43
Last 52.42 57.69 50.71 55.1 51.14 53.88 – – 31.32 32.42 34.19 36.99 45.91 47.34 73.52 74.43 34.4 35.77
4
Best 55.98 60.27 53 56.47 54.43 56.16 – – 34.43 35.01 37.57 39.12 48.04 50.53 79.18 80.67 37.26 37.75
Last 35.37 49.01 34.03 45.97 37.26 46.27 – – 21.92 23.59 24.35 26.03 11.11 11.11 78.51 83.56 20.12 22.53
5
Best 37.56 49.01 36.83 49.01 39.39 48.1 – – 25.69 27.85 26.88 29.07 11.11 11.11 85.93 87.82 24.99 26.64
Last 26.61 27.55 29.98 31.81 28.01 29.68 27.67 28.92 46.18 47.79 32.17 34.09 30.11 40.79 69.16 72.15 35.13 37.29
1
Best 29.47 31.2 31.81 32.72 30.14 32.12 30.44 31.96 47.34 48.86 34.19 35.01 31.29 42.01 76.84 77.78 36.71 38.81
Last 79.87 84.32 80.67 85.54 81.4 87.37 81.19 90.11 62.89 67.28 78.72 82.34 68.77 74.89 83.87 88.43 77.62 78.84
2
Best 88.05 89.19 90.02 91.63 89.53 91.63 89.5 90.87 67.91 68.65 81.19 82.34 72.33 76.56 89.8 90.87 80.7 82.19
B Last 12.85 19.73 17.86 21.81 21.19 23.15 – – 27.18 33.09 36.29 38.72 13.21 21.51 65.19 67.66 37.62 40.65
3
Best 13.5 22.4 19.97 24.18 24.6 25.67 – – 28.66 34.12 38.9 40.5 14.22 24.48 67.89 68.99 39.91 42.73
Last 47.76 54.03 44.87 47.03 48.01 55.4 – – 30.93 32.42 34.4 36.38 33 48.25 75.04 79.15 34 36.07
4
Best 52.48 56.16 48.07 52.36 52.06 56.47 – – 34.58 35.01 38.39 39.73 35.13 52.51 79.63 80.97 37.99 39.88
Last 39.7 51.45 41.95 52.97 35.59 44.9 – – 16.96 22.68 27 29.22 11.11 11.11 71.38 87.21 20.7 22.68
5
Best 40.82 52.82 44.2 53.58 39.36 47.03 – – 19.15 28.46 28.31 29.22 11.17 11.42 86.03 88.13 26.12 27.7
Last 26.27 30.14 28.1 28.46 25.15 28.16 27.67 32.12 42.98 46.42 29.86 32.57 34.95 46.58 67.09 73.52 37.63 39.73
1
Best 29.13 30.59 30.26 32.88 27.64 30.29 29.77 32.12 45.9 50.68 32.69 35.31 35.92 46.58 76.16 77.32 40.67 42.92
Last 93.12 94.22 94.09 94.82 94.49 95.59 94.12 94.98 70.23 71.84 85.87 86.45 83.68 85.69 86.91 89.35 82.41 84.47
2
Best 94.95 95.28 95.22 95.89 95.22 95.74 95.68 96.19 73.55 74.73 87.73 88.74 86.27 87.21 89.2 90.26 84.05 85.84
C Last 12.88 19.88 18.1 21.66 20.03 21.81 – – 34.69 38.87 38.6 39.17 11.13 11.13 61.45 65.58 36.94 40.5
3
Best 13.47 22.85 19.38 23.15 23.26 24.93 – – 35.99 40.5 41.25 42.43 11.22 11.42 66.08 67.95 39.97 42.58
Last 52.05 53.42 53.45 57.99 52.91 59.06 – – 30.59 31.2 36.1 40.03 45.75 51.75 75.95 77.78 37.05 38.05
4
Best 55.8 57.08 55.56 59.06 57.57 59.97 – – 34.8 35.46 39.36 41.7 47.73 52.82 80.31 81.28 39.63 40.18
Last 44.57 49.01 44.9 48.25 49.25 53.58 – – 16.29 26.64 26 27.09 11.11 11.11 71.29 88.74 22.59 23.29
5
Best 47.03 49.47 47.64 51.29 51.9 56.16 – – 18.06 28.77 27.52 29.22 11.11 11.11 87.34 88.74 27.95 30.29

*A is the max–min normalization; B is the -1-1 normalization; C is the Z-score normalization.


1 is the time domain input, 2 is the frequency domain input; 3 is the wavelet domain input; 4 is the time domain sample after STFT; 5 is the time domain sample
reshape to a 2D matrix.

Table A.17
UoC: Results with random split and without data augmentation.
Nor Input Loc AE DAE SAE MLP CNN LeNet AlexNet ResNet18 LSTM
Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max
Last 21.55 25.27 22.13 25.57 21.89 25.11 24.41 27.55 36.23 42.92 28.07 31.05 11.11 11.11 72.12 88.28 29.71 32.42
1
Best 25.39 27.7 26.7 27.7 24.47 25.11 27.37 28.46 44.08 48.55 31.96 34.4 11.26 11.87 88.04 89.5 34.43 36.38
Last 92.53 94.67 91.45 93.15 92.15 94.06 91.69 92.69 65.36 69.25 83.38 84.93 75.04 78.54 88.13 90.72 80.82 82.34
2
Best 93.94 94.67 92.95 93.61 93.59 94.37 94.07 94.52 67.03 69.25 85.33 86.76 76.29 78.69 91.9 92.39 82.5 84.02
A Last 18.22 21.51 18.22 21.36 18.37 19.44 – – 27.15 29.38 31.54 33.68 11.13 11.13 63.02 66.17 35.7 42.14
3
Best 21.72 23.29 18.66 21.81 21.87 22.85 – – 29.26 31.9 35.34 38.72 11.13 11.13 67.36 69.44 39.17 44.66
Last 55.4 57.99 51.78 54.49 53.52 58.75 – – 31.6 32.12 34.15 37.29 44.23 52.21 75.89 78.69 35.68 36.99
4
Best 58.3 59.67 54.98 56.62 55.65 59.36 – – 34.1 35.01 37.93 39.57 46.91 54.19 79.15 80.06 38.33 39.57
Last 35.43 50.23 30.23 32.57 36.38 42.92 – – 18.2 25.57 22.16 24.05 11.11 11.11 71.14 85.39 19.39 21.31
5
Best 37.75 50.84 32.12 33.33 38.69 45.97 – – 19.27 27.7 27.06 28.01 11.17 11.42 86.24 87.98 23.87 25.88
Last 22.98 23.59 24.2 28.16 23.68 24.96 24.41 26.64 39.73 43.07 27.64 31.05 36.04 42.31 81.74 87.37 28.52 30.75
1
Best 26.39 27.4 26.94 28.61 26.06 28.77 27.76 29.07 43.74 45.97 32.21 33.79 38.93 42.31 87.06 87.98 33.3 34.09
Last 78.17 84.78 68.16 78.08 77.11 82.34 73.85 78.84 64.32 67.58 81.43 83.71 78.26 81.58 89.93 91.63 80.37 82.8
2
Best 85.75 88.13 86.03 88.43 85.91 86.91 86.48 88.28 65.72 68.8 84.05 85.69 80.12 83.56 92.33 93.3 82.71 84.17
B Last 13 20.47 20.18 21.96 20.77 22.85 – – 29.94 31.01 32.17 36.2 11.13 11.13 64.01 67.51 36.86 39.76
3
Best 13.68 22.7 22.46 24.63 23.65 26.11 – – 31.75 32.94 34.95 39.91 11.51 12.31 68.25 69.14 39.35 41.99
Last 49.07 51.75 53.21 55.56 50.5 54.79 – – 30.69 31.51 32.6 34.09 39.63 48.4 67.52 76.41 35.95 37.29
4
Best 51.14 53.73 55.8 59.06 53.52 56.01 – – 34.49 35.62 37.35 39.57 42.31 49.62 79.6 81.74 39.21 40.64
Last 42.59 53.42 47.36 50.68 40.25 46.73 – – 19.66 23.74 24.99 26.79 15.19 22.98 71.63 81.28 22.04 23.59
5
Best 44.14 54.03 49.1 51.6 41.43 47.95 – – 22.04 27.09 27.55 28.31 16.47 24.96 87.12 88.13 27.37 28.01
(continued on next page)

Please cite this article as: Z. Zhao, T. Li, J. Wu et al., Deep learning algorithms for rotating machinery intelligent diagnosis: An open source benchmark study. ISA Transactions
(2020), https://doi.org/10.1016/j.isatra.2020.08.010.
28 Z. Zhao, T. Li, J. Wu et al. / ISA Transactions xxx (xxxx) xxx

Table A.17 (continued).


Nor Input Loc AE DAE SAE MLP CNN LeNet AlexNet ResNet18 LSTM
Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max
Last 22.68 24.66 26.24 29.07 22.46 25.11 26.12 27.85 37.66 42.92 28.22 32.88 28.89 45.21 83.23 87.82 34.21 37.44
1
Best 26.15 27.4 29.22 31.35 26.4 27.25 28.53 29.38 41.34 44.44 31.78 35.62 30.59 45.81 88.34 89.35 38.26 40.33
Last 93.27 93.91 93.45 94.52 93.58 94.52 93.97 95.74 70.47 76.41 85.54 86.61 82.89 84.02 89.92 92.54 85.36 86.45
2
Best 94.58 95.13 95.58 96.04 94.89 95.74 95.4 95.74 72.51 77.93 86.64 87.82 84.29 85.69 92.91 93.46 87.15 87.67
C Last 12.74 20.92 13.55 18.55 22.28 25.96 – – 30.45 36.2 37.42 38.87 17.24 41.84 58.22 63.35 40.09 41.1
3
Best 13.11 23 14.64 21.96 24.76 27.6 – – 31.78 36.2 40.12 41.99 17.72 43.92 65.67 67.21 42.67 44.36
Last 52.54 54.03 49.96 57.99 55.59 58.45 – – 31.2 32.57 34.06 35.46 44.02 45.97 75.19 78.69 35.65 36.99
4
Best 54.1 55.86 52.39 58.14 57.63 59.51 – – 35.53 36.83 37.41 38.51 47.76 50.99 79.54 80.97 39.51 40.18
Last 46.45 49.32 42.83 48.25 49.22 60.88 – – 13.36 22.37 25.54 28.16 11.11 11.11 71.72 77.78 22.4 24.35
5
Best 47.12 49.77 44.05 48.25 53.79 60.88 – – 14.37 25.27 27.86 30.14 11.11 11.11 86.67 87.98 26.64 27.55

Table A.18
UoC: Results with order split and data augmentation.
Nor Input Loc AE DAE SAE MLP CNN LeNet AlexNet ResNet18 LSTM
Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max
Last 25.75 27.09 26.3 29.53 24.72 26.03 25.72 27.7 28.83 32.72 29.95 31.05 11.11 11.11 38.96 43.99 34.79 35.46
1
Best 28.55 31.05 29.53 31.81 28 28.46 29.65 31.05 33.18 34.09 32.11 32.88 11.41 12.63 44.23 44.75 36.99 38.51
Last 62.74 64.54 63.23 67.88 64.02 65.3 65.36 70.02 52.24 53.27 53.06 54.79 53.36 55.86 53.82 56.32 52.48 55.56
2
Best 67.88 70.47 68.31 68.95 68.52 71.08 70.04 71.99 55.86 57.99 58.02 59.82 56.99 57.84 61.67 63.32 55.98 57.53
A Last 17.9 22.37 15.79 19.56 21.72 23.26 – – 24.89 27.7 26.43 29.48 13.04 20.74 43.41 45.63 27.23 28.3
3
Best 20.27 25.19 16.62 20.74 23.59 25.48 – – 26.81 28 29.07 30.96 13.18 21.48 49.04 50.67 29.54 31.7
Last 35.89 37.75 37.9 40.18 36.28 37.44 – – 30.65 34.55 30.93 33.18 36.47 38.81 43.8 45.66 31.38 33.33
4
Best 39.82 42.16 40.43 41.55 39.42 42.01 – – 36.32 37.6 36.29 38.2 39.12 39.73 48.46 50.23 35.74 37.44
Last 21.52 32.57 22.8 28.31 24.05 27.85 – – 18.6 26.33 24.32 26.79 11.11 11.11 34.61 39.88 21.58 23.14
5
Best 24.63 35.31 24.78 32.57 26.91 31.51 – – 20.64 27.09 29.01 30.59 11.11 11.11 41.89 45.36 25.69 26.64
Last 24.87 26.64 24.9 28.16 26.33 28.31 25.42 27.7 29.65 32.88 30.07 31.35 24.48 32.12 37.87 41.7 34.25 35.01
1
Best 27.89 28.92 27.52 28.31 28.92 30.44 28.71 29.68 33.76 37.6 32.94 35.01 26.76 33.18 43.38 44.6 36.74 37.29
Last 52.27 59.21 55.95 60.43 46.76 53.58 57.26 63.47 48.8 50.99 54.7 56.16 49.77 54.03 51.81 55.1 50.99 52.51
2
Best 62.41 64.08 65.36 67.88 62.31 64.99 65.9 68.49 53.42 56.01 57.2 58.9 53.36 57.23 59.97 61.64 54.7 55.1
B Last 13.95 20.59 14.72 20.59 19.23 21.48 – – 22.55 24.3 24.03 25.33 12.56 17.78 46.19 48.44 20.89 21.61
3
Best 14.73 23.7 16.12 23.7 22.04 23.7 – – 23.76 24.3 27.73 29.33 13.21 20.89 49.07 50.07 23.75 23.99
Last 32.94 35.77 36.01 39.73 32.88 34.55 – – 31.54 33.03 29.01 30.14 37.02 38.96 40.21 49.16 31.35 33.49
4
Best 36.71 40.33 39.18 41.7 38.23 39.57 – – 37.2 38.05 35.4 36.99 41.25 43.23 49.22 49.92 35.62 36.07
Last 26.79 28.77 19.72 27.85 26.36 29.22 – – 18.32 24.81 25.48 28.46 17.47 21.61 30.75 40.03 23.01 24.96
5
Best 31.17 35.01 23.01 31.51 30.05 32.57 – – 20.97 28.31 29.41 31.2 20.4 24.51 41.61 43.53 27.27 28.92
Last 23.62 27.25 24.47 27.09 23.93 26.48 23.93 25.11 31.2 32.88 29.68 32.72 29.96 31.81 38.63 45.66 36.44 38.2
1
Best 27.08 29.38 27.2 28.46 27.37 29.22 28.13 28.77 35.68 37.9 31.48 34.4 32.33 34.25 44.63 45.66 38.39 39.12
Last 63.18 67.28 62.97 66.97 62.59 66.21 62.59 64.84 51.29 54.03 57.9 59.21 54.46 57.69 53.36 56.16 50.9 51.75
2
Best 68.25 70.62 67.78 70.47 68.51 72.15 69.89 70.78 55.68 56.62 61.8 62.71 58.66 60.12 59.21 61.04 56.89 58.14
C Last 12.8 19.56 15.82 20.89 20.89 24.15 – – 23.53 28.44 26.49 28.89 23.11 34.52 39.73 48.59 25.96 27.26
3
Best 13.39 21.63 16.68 21.48 22.67 25.33 – – 26.16 30.67 30.28 31.11 24.62 35.11 47.32 48.89 29.63 30.81
Last 37.17 40.33 37.09 39.27 37.28 38.51 – – 31.05 33.03 30.78 32.42 28.34 35.92 43.35 49.62 30.93 32.72
4
Best 40.62 43.68 40.76 42.62 41.48 43.84 – – 35.83 36.68 36.13 37.14 31.29 37.29 49.86 52.51 35.59 36.68
Last 28.17 32.88 28.1 32.27 28.11 35.77 – – 11.11 11.11 24.29 27.55 11.11 11.11 31.84 37.6 23.65 26.03
5
Best 32.04 34.86 30.59 34.86 31.93 36.99 – – 11.96 13.09 28.71 30.44 11.14 11.26 39.64 41.25 28.31 29.38

Table A.19
XJTU-SY: Results with random split and data augmentation.
Nor Input Loc AE DAE SAE MLP CNN LeNet AlexNet ResNet18 LSTM
Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max
Last 68.59 74.22 71.98 80.99 70.62 78.12 66.72 73.7 94.64 99.74 96.09 99.22 84.12 98.18 88.39 99.74 96.82 97.92
1
Best 74.17 76.3 79.84 83.07 75.62 80.21 77.24 80.21 99.95 100 98.6 99.22 86.36 98.18 99.79 100 98.91 99.48
Last 100 100 100 100 100 100 100 100 96.82 100 90.89 100 6.51 6.51 100 100 95.83 100
2
Best 100 100 100 100 100 100 100 100 100 100 100 100 6.72 6.77 100 100 100 100
A Last 37.81 48.7 30.78 40.89 38.85 47.14 – – 71.2 73.44 65.05 71.88 6.51 6.51 74.38 92.19 67.45 70.57
3
Best 39.01 49.48 34.9 47.66 41.15 48.18 – – 75.47 76.56 75.52 76.3 7.08 7.81 91.83 92.71 72.55 74.22
Last 99.95 100 99.84 100 99.69 100 – – 98.65 99.22 98.33 99.48 99.58 100 100 100 99.17 99.74
4
Best 100 100 100 100 100 100 – – 98.75 99.22 99.48 99.74 100 100 100 100 99.79 100
Last 42.97 90.62 68.02 90.89 59.17 89.84 – – 6.51 6.51 60.16 74.48 6.51 6.51 100 100 71.41 76.56
5
Best 56.04 92.71 72.71 93.75 65.68 91.93 – – 6.72 6.77 74.58 76.82 6.72 6.77 100 100 77.97 80.99
(continued on next page)

Please cite this article as: Z. Zhao, T. Li, J. Wu et al., Deep learning algorithms for rotating machinery intelligent diagnosis: An open source benchmark study. ISA Transactions
(2020), https://doi.org/10.1016/j.isatra.2020.08.010.
Z. Zhao, T. Li, J. Wu et al. / ISA Transactions xxx (xxxx) xxx 29

Table A.19 (continued).


Nor Input Loc AE DAE SAE MLP CNN LeNet AlexNet ResNet18 LSTM
Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max
Last 86.67 88.54 88.9 90.36 85.57 88.54 87.08 89.06 99.53 100 97.55 98.7 97.03 98.44 98.96 99.74 96.35 98.18
1
Best 89.17 90.89 90.52 91.93 88.23 89.58 89.27 91.41 100 100 98.75 98.96 98.86 99.74 100 100 98.23 99.22
Last 11.93 21.88 46.25 57.03 12.39 16.93 13.18 17.45 61.56 99.22 93.54 100 6.51 6.51 68.75 100 100 100
2
Best 41.3 68.49 55.42 63.8 42.13 67.97 31.56 40.62 99.48 99.74 100 100 6.72 6.77 100 100 100 100
B Last 28.91 45.31 37.03 47.66 44.43 49.74 – – 74.22 77.34 72.87 77.34 79.63 83.85 89.79 91.15 70.89 73.7
3
Best 31.3 46.35 39.69 51.82 47.45 53.65 – – 77.19 79.69 76.92 77.86 82.34 84.11 91.93 92.19 74.32 76.3
Last 99.9 100 100 100 99.84 100 – – 98.39 99.22 86.62 99.74 99.32 100 99.9 100 98.65 100
4
Best 100 100 100 100 100 100 – – 98.65 99.48 99.32 99.74 100 100 100 100 99.9 100
Last 87.08 90.89 80.78 90.89 72.4 86.72 – – 38.65 67.97 72.5 75.52 40.16 90.89 100 100 76.09 80.21
5
Best 88.39 93.23 84.89 94.27 82.13 87.24 – – 42.13 76.56 76.56 78.91 40.52 91.41 100 100 80.21 83.33
Last 88.07 90.36 90.36 91.93 86.41 88.54 88.07 89.58 97.6 100 97.61 99.22 98.86 99.48 99.22 100 98.23 98.7
1
Best 89.01 91.15 91.3 92.19 89.27 90.89 90.36 91.41 99.9 100 98.7 100 99.79 100 100 100 99.01 99.22
Last 100 100 100 100 100 100 99.95 100 99.84 100 99.84 100 99.95 100 100 100 100 100
2
Best 100 100 100 100 100 100 100 100 100 100 100 100 100 100 100 100 100 100
C Last 10.89 28.39 16.04 40.89 39.74 51.82 – – 74.17 78.39 75.05 76.82 83.44 84.64 91.93 92.71 70.89 72.4
3
Best 11.98 33.07 16.2 40.89 42.34 53.39 – – 77.4 79.95 77.03 78.39 85.89 86.46 92.76 93.23 75.1 76.82
Last 100 100 99.9 100 100 100 – – 96.88 98.96 98.6 98.96 99.79 100 100 100 98.96 99.74
4
Best 100 100 100 100 100 100 – – 98.8 99.22 99.32 99.74 100 100 100 100 100 100
Last 90 93.75 88.23 94.79 80.57 91.15 – – 34.64 78.39 69.79 75.26 88.85 94.79 92.14 100 80.94 85.42
5
Best 94.11 96.61 89.64 97.14 88.8 95.31 – – 36.25 80.47 79.53 82.03 91.25 95.83 100 100 85.36 87.24

*A is the max–min normalization; B is the -1-1 normalization; C is the Z-score normalization.


1 is the time domain input, 2 is the frequency domain input; 3 is the wavelet domain input; 4 is the time domain sample after STFT; 5 is the time domain sample
reshape to a 2D matrix.

Table A.20
XJTU-SY: Results with random split and without data augmentation.
Nor Input Loc AE DAE SAE MLP CNN LeNet AlexNet ResNet18 LSTM
Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max
Last 36.04 48.18 55.21 57.29 48.85 52.6 60 62.24 99.58 100 93.8 96.35 73.44 93.23 94.69 100 94.95 97.14
1
Best 56.98 58.85 57.55 59.64 54.95 57.03 61.72 63.02 99.79 100 97.4 98.7 75.16 95.83 99.95 100 97.87 98.18
Last 100 100 100 100 100 100 100 100 99.64 99.74 100 100 6.51 6.51 100 100 100 100
2
Best 100 100 100 100 100 100 100 100 99.79 100 100 100 6.72 6.77 100 100 100 100
A Last 39.64 48.96 36.2 43.75 37.92 41.41 – – 70.36 76.56 71.46 75.52 43.85 72.4 82.08 90.62 69.53 72.14
3
Best 42.14 50.52 39.43 46.88 41.51 42.97 – – 77.18 77.6 76.09 79.43 46.67 76.56 91.72 92.45 73.02 75
Last 100 100 100 100 99.9 100 – – 98.34 98.7 99.27 100 99.74 100 99.06 100 98.91 99.74
4
Best 100 100 100 100 99.9 100 – – 98.49 98.96 99.53 100 100 100 100 100 99.84 100
Last 54.43 92.19 36.93 86.72 63.28 92.97 – – 6.51 6.51 52.97 66.15 6.51 6.51 78.54 100 59.9 75.78
5
Best 54.95 92.71 38.65 87.24 63.8 93.49 – – 7.66 10.94 74.53 79.69 6.82 7.03 100 100 77.61 79.69
Last 57.87 60.16 57.24 59.11 57.97 59.9 60 61.2 95.68 100 93.91 96.61 96.88 98.18 99.79 100 95.83 98.44
1
Best 60.37 62.5 58.33 60.16 59.85 60.68 61.82 63.54 99.84 100 97.97 98.96 98.6 99.22 100 100 98.02 98.44
Last 68.18 100 70.89 84.64 62.66 100 62.66 100 99.84 100 100 100 6.51 6.51 100 100 99.95 100
2
Best 100 100 85.68 89.06 100 100 100 100 99.95 100 100 100 6.77 6.77 100 100 100 100
B Last 20.47 42.71 24.58 43.23 44.95 50.78 – – 75.78 77.6 72.97 75.26 79.89 83.85 90.52 91.93 70.26 72.14
3
Best 21.98 45.31 26.98 46.35 47.34 52.6 – – 78.28 80.21 75.21 76.56 81.93 84.38 92.14 92.71 74.38 75.78
Last 99.64 100 100 100 100 100 – – 98.39 98.96 98.44 98.96 99.48 100 100 100 96.15 100
4
Best 99.79 100 100 100 100 100 – – 98.6 99.22 99.01 99.22 100 100 100 100 100 100
Last 86.77 89.84 50.78 89.06 68.13 87.24 – – 31.93 76.56 72.19 75.78 37.5 91.15 98.91 100 77.5 80.99
5
Best 88.75 93.23 51.46 90.36 85.47 93.49 – – 35.05 78.65 76.25 80.21 38.91 92.97 100 100 80.42 82.55
Last 62.29 63.02 62.19 64.84 61.67 64.84 61.1 64.84 99.22 99.74 96.93 98.18 98.54 99.48 99.95 100 97.03 98.44
1
Best 64.16 65.62 63.75 65.62 65.42 66.15 62.66 65.1 99.48 100 98.7 99.22 99.12 99.74 100 100 98.86 99.22
Last 99.9 100 100 100 100 100 100 100 99.74 100 100 100 100 100 100 100 100 100
2
Best 100 100 100 100 100 100 100 100 99.9 100 100 100 100 100 100 100 100 100
C Last 17.76 36.46 17.97 40.1 35.78 41.15 – – 77.13 81.77 74.27 76.04 82.76 84.9 91.88 92.97 72.08 76.3
3
Best 19.58 40.62 19.01 43.49 38.23 42.71 – – 78.65 81.77 77.5 78.39 85.37 87.24 93.39 94.53 75 76.3
Last 100 100 100 100 100 100 – – 98.34 99.22 98.6 99.22 99.74 100 99.84 100 99.48 100
4
Best 100 100 100 100 100 100 – – 98.75 99.48 98.86 99.48 100 100 100 100 100 100
Last 93.96 97.66 85.52 96.35 90.26 93.75 – – 34.84 77.86 76.77 79.43 89.79 94.79 98.75 99.74 82.61 85.16
5
Best 94.9 98.7 86.3 98.18 92.34 94.79 – – 35.62 79.95 80.16 81.51 91.77 95.31 100 100 85.62 87.76

Please cite this article as: Z. Zhao, T. Li, J. Wu et al., Deep learning algorithms for rotating machinery intelligent diagnosis: An open source benchmark study. ISA Transactions
(2020), https://doi.org/10.1016/j.isatra.2020.08.010.
30 Z. Zhao, T. Li, J. Wu et al. / ISA Transactions xxx (xxxx) xxx

Table A.21
XJTU-SY: Results with order split and data augmentation.
Nor Input Loc AE DAE SAE MLP CNN LeNet AlexNet ResNet18 LSTM
Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max
Last 62.97 73.59 68.36 76.41 60.46 71.54 68.41 75.38 98.97 99.74 96.31 96.67 91.9 95.13 89.54 96.41 96.26 97.95
1
Best 72.67 76.92 81.08 83.33 70.92 72.05 77.64 80.26 99.84 100 98.1 98.97 94.2 96.15 99.69 100 98 98.97
Last 100 100 100 100 100 100 73.33 73.33 73.02 73.33 72.2 73.33 6.67 6.67 73.33 73.33 71.38 73.33
2
Best 100 100 100 100 100 100 73.33 73.33 73.33 73.33 73.33 73.33 6.67 6.67 73.33 73.33 73.33 73.33
A Last 26.05 45.64 29.95 40.77 41.95 46.67 – – 72.3 76.15 75.28 78.21 14.72 46.92 73.8 88.46 72.77 74.62
3
Best 27.64 47.44 33.13 43.08 46.72 51.28 – – 76.77 78.21 78.26 79.74 15.03 46.92 90.51 91.79 77.38 80
Last 99.95 100 99.84 100 99.9 100 – – 97.38 98.97 97.49 99.49 99.64 100 100 100 98.15 99.23
4
Best 100 100 100 100 100 100 – – 98.26 98.97 99.08 99.74 100 100 100 100 99.59 99.74
Last 58.36 92.56 54.05 88.46 78.1 89.74 – – 19.44 70.51 53.28 75.9 6.67 6.67 95.18 99.74 71.79 78.97
5
Best 60.56 93.59 55.39 89.23 82.31 95.13 – – 21.33 79.74 74 76.92 6.67 6.67 100 100 78.05 79.49
Last 82.67 85.9 84.26 86.92 82.05 83.59 85.13 87.44 98.87 99.74 97.59 97.95 97.54 98.46 97.44 99.49 96.51 97.69
1
Best 84.41 85.9 87.03 87.44 85.03 85.9 87.54 88.21 99.79 100 98.46 99.23 98.51 99.49 99.84 100 97.9 98.46
Last 14 36.15 29.23 32.82 20.82 47.44 11.39 18.46 54.2 72.05 73.28 73.33 6.67 6.67 52.77 73.33 73.33 73.33
2
Best 42.92 58.21 52.31 63.59 32.51 47.44 33.95 38.97 73.28 73.33 73.33 73.33 6.67 6.67 73.33 73.33 73.33 73.33
B Last 42.87 47.44 33.54 47.18 45.59 47.69 – – 75.74 78.46 76.41 79.23 81.64 82.56 87.49 88.97 75.13 76.67
3
Best 45.9 50.77 36.05 48.72 49.85 52.31 – – 78.62 79.23 78.67 79.74 85.28 85.9 90.62 91.28 77.79 79.23
Last 100 100 98.51 100 99.59 100 – – 96.87 98.46 96.1 98.72 81.23 100 100 100 98.62 99.49
4
Best 100 100 98.82 100 100 100 – – 98.41 98.97 98.51 99.23 81.33 100 100 100 99.44 99.74
Last 89.33 92.56 55.28 91.79 66.67 94.1 – – 26.72 76.67 70.72 74.36 35.39 84.62 99.33 100 72.31 75.64
5
Best 91.18 93.59 56.46 93.85 72.36 94.87 – – 27.44 76.67 76.72 79.23 38.05 89.74 100 100 78.15 82.05
Last 81.44 83.85 85.38 87.44 81.08 83.59 83.9 85.13 96.67 99.74 98.05 98.97 98.82 98.97 98.92 100 97.29 98.21
1
Best 83.59 86.15 87.44 89.23 84.26 85.64 85.9 87.18 99.95 100 99.33 99.74 99.64 99.74 100 100 99.18 99.49
Last 99.95 100 100 100 100 100 73.33 73.33 73.33 73.33 73.23 73.33 73.33 73.33 73.33 73.33 73.33 73.33
2
Best 100 100 100 100 100 100 73.33 73.33 73.33 73.33 73.33 73.33 73.33 73.33 73.33 73.33 73.33 73.33
C Last 6.67 6.67 6.67 6.67 49.69 57.44 – – 75.64 76.92 76.97 81.54 83.69 85.9 90.46 91.54 75.49 76.92
3
Best 6.67 6.67 6.67 6.67 51.85 61.54 – – 78.05 80.26 78.82 81.54 85.95 88.46 93.18 93.59 78.87 80
Last 99.95 100 99.79 100 99.95 100 – – 98.26 98.97 97.59 98.46 99.95 100 100 100 98.87 99.74
4
Best 100 100 100 100 100 100 – – 98.77 98.97 98.15 98.72 100 100 100 100 99.44 99.74
Last 74.05 93.08 69.95 92.56 90.1 93.85 – – 57.95 73.85 77.38 81.28 90.51 94.62 90.67 100 79.23 82.82
5
Best 76.77 95.13 73.33 96.92 94.56 98.21 – – 64.46 82.31 81.28 82.31 93.69 97.69 100 100 85.79 88.72

Declaration of competing interest [8] Young T, Hazarika D, Poria S, Cambria E. Recent trends in deep
learning based natural language processing. IEEE Comput Intell Mag
2018;13:55–75.
The authors declare that they have no known competing finan-
[9] Feng Q, Zhao X, Fan D, Cai B, Liu Y, Ren Y. Resilience design method
cial interests or personal relationships that could have appeared based on meta-structure: A case study of offshore wind farm. Reliab Eng
to influence the work reported in this paper. Syst Saf 2019;186:232–44.
[10] Li D, Liu Y, Huang D. Development of semi-supervised multiple-output
soft-sensors with co-training and tri-training mpls and mrvm. Chemometr
Acknowledgment
Intell Lab Syst 2020;199:103970.
[11] Hinton GE, Salakhutdinov RR. Reducing the dimensionality of data with
This work was supported by Natural Science Foundation of neural networks. Science 2006;313:504–7.
China (No. 51835009, No. 51705398). [12] MIT Technology Review. 10 breakthrough technologies 2013. 2019,
https://www.technologyreview.com/lists/technologies/2013/ [Accessed on
August 2019].
Appendix. Evaluation results [13] LeCun Y, Bengio Y, Hinton G. Deep learning. Nature 2015;521:436.
[14] Riley P. Three pitfalls to avoid in machine learning. 2019.
See Tables A.1–A.21. [15] Goodfellow I, Bengio Y, Courville A. Deep learning. MIT Press; 2016.
[16] Li C, d. Oliveira JLV, Lozada MC, Cabrera D, Sanchez V, Zurita G. A
systematic review of fuzzy formalisms for bearing fault diagnosis. IEEE
References Trans Fuzzy Syst 2018.
[17] Hoang D-T, Kang H-J. A survey on deep learning based bearing fault
[1] Zhao Z, Wu S, Qiao B, Wang S, Chen X. Enhanced sparse period- diagnosis. Neurocomputing 2019;335:327–35.
group lasso for bearing fault diagnosis. IEEE Trans Ind Electron [18] Zhang S, Zhang S, Wang B, Habetler TG. Deep learning algorithms
2018;66:2143–53. for bearing fault diagnosticsx-a comprehensive review. IEEE Access
[2] Wang S, Chen X, Tong C, Zhao Z. Matching synchrosqueezing wavelet 2020;8:29857–81.
transform and application to aeroengine vibration monitoring. IEEE Trans [19] Hamadache M, Jung JH, Park J, Youn BD. A comprehensive review of
Instrum Meas 2016;66:360–72. artificial intelligence-based approaches for rolling element bearing phm:
[3] Sun C, Ma M, Zhao Z, Chen X. Sparse deep stacking network for fault shallow and deep learning. JMST Adv 2019;1:125–51.
diagnosis of motor. IEEE Trans Ind Inf 2018;14:3261–70. [20] Ali YH, Ali SM, Rahman RA, Hamzah RIR. Acoustic emission and artifi-
[4] Krizhevsky A, Sutskever I, Hinton GE. Imagenet classification with cial intelligent methods in condition monitoring of rotating machine–a
deep convolutional neural networks. In: Advances in neural information review. In: National conference for postgraduate research.
processing systems, p. 1097–105. [21] Liu R, Yang B, Zio E, Chen X. Artificial intelligence for fault diagnosis of
[5] Farabet C, Couprie C, Najman L, LeCun Y. Learning hierarchical features rotating machinery: A review. Mech Syst Signal Process 2018;108:33–47.
for scene labeling. IEEE Trans Pattern Anal Mach Intell 2012;35:1915–29. [22] Wei Y, Li Y, Xu M, Huang W. A review of early fault diagnosis approaches
[6] Hirschberg J, Manning CD. Advances in natural language processing. and their applications in rotating machinery. Entropy 2019;21:409.
Science 2015;349:261–6. [23] Zhao G, Zhang G, Ge Q, Liu X. Research advances in fault diagnosis
[7] Sun S, Luo C, Chen J. A review of natural language processing techniques and prognostic based on deep learning. In: 2016 Prognostics and system
for opinion mining systems. Inf Fusion 2017;36:10–25. health management conference. IEEE; p. 1–6.

Please cite this article as: Z. Zhao, T. Li, J. Wu et al., Deep learning algorithms for rotating machinery intelligent diagnosis: An open source benchmark study. ISA Transactions
(2020), https://doi.org/10.1016/j.isatra.2020.08.010.
Z. Zhao, T. Li, J. Wu et al. / ISA Transactions xxx (xxxx) xxx 31

[24] Duan L, Xie M, Wang J, Bai T. Deep learning enabled intelligent [52] Khodja AY, Guersi N, Saadi MN, Boutasseta N. Rolling element bear-
fault diagnosis: Overview and applications. J Intell Fuzzy Systems ing fault diagnosis for rotating machinery using vibration spectrum
2018;35:5771–84. imaging and convolutional neural networks. Int J Adv Manuf Technol
[25] Zhang W, Jia M-P, Zhu L, Yan X-A. Comprehensive overview on compu- 2020;106:1737–51.
tational intelligence techniques for machinery condition monitoring and [53] Li Y, Du X, Wan F, Wang X, Yu H. Rotating machinery fault diagnosis
fault diagnosis. Chin J Mech Eng 2017;30:782–95. based on convolutional neural network and infrared thermal imaging.
[26] Zhao R, Yan R, Chen Z, Mao K, Wang P, Gao RX. Deep learning and Chin J Aeronaut 2020;33:427–38.
its applications to machine health monitoring. Mech Syst Signal Process [54] Zhang J, Sun Y, Guo L, Gao H, Hong X, Song H. A new bearing fault
2019;115:213–37. diagnosis method based on modified convolutional neural networks. Chin
[27] Lei Y, Yang B, Jiang X, Jia F, Li N, Nandi AK. Applications of machine J Aeronaut 2020;33:439–47.
learning to machine fault diagnosis: A review and roadmap. Mech Syst [55] Zhao M, Tang B, Deng L, Pecht M. Multiple wavelet regularized deep
Signal Process 2020;138:106587. residual networks for fault diagnosis. Measurement 2020;152.
[28] Nasiri S, Khosravani MR, Weinberg K. Fracture mechanics and mechanical [56] Li X, Zhang W, Ding Q, Sun J-Q. Intelligent rotating machinery fault
fault detection by artificial intelligence methods: A review. Eng Fail Anal diagnosis based on deep learning using data augmentation. J Intell Manuf
2017;81:270–93. 2020;31:433–52.
[29] Tian Y, Guo D, Zhang K, Jia L, Qiao H, Tang H. A review of fault diagnosis [57] Tang T, Hu T, Chen M, Lin R, Chen G. A deep convolutional neural
for traction induction motor. In: 2018 37th Chinese control conference. network approach with information fusion for bearing fault diagnosis
2018, p. 5763–8. under different working conditions. Proc Inst Mech Eng C 2020.
[30] Khan S, Yairi T. A review on the application of deep learning in system [58] Xue Y, Dou D, Yang J. Multi-fault diagnosis of rotating machinery
health management. Mech Syst Signal Process 2018;107:241–65. based on deep convolution neural network and support vector machine.
[31] Stetco A, Dinmohammadi F, Zhao X, Robu V, Flynn D, Barnes M, et Measurement 2020;156.
al. Machine learning methods for wind turbine condition monitoring: A [59] Verstraete DB, Lope Droguett E, Meruane V, Modarres M, Ferrada A.
review. Renew Energy 2018. Deep semi-supervised generative adversarial fault diagnostics of rolling
[32] Ellefsen AL, Æsøy V, Ushakov S, Zhang H. A comprehensive survey element bearings. Struct Health Monitor 2020;19:390–411.
of prognostics and health management based on deep learning for [60] Zhang W, Li X, Jia X-D, Ma H, Luo Z, Li X. Machinery fault diagno-
autonomous ships. IEEE Trans Reliab 2019;68:720–40. sis with imbalanced data using deep generative adversarial networks.
[33] Ademujimi TT, Brundage MP, Prabhu VV. A review of current machine Measurement 2020;152.
learning techniques used in manufacturing diagnosis. In: IFIP inter- [61] Li T, Zhao Z, Sun C, Yan R, Chen X. Adaptive channel weighted cnn with
national conference on advances in production management systems. multi-sensor fusion for condition monitoring of helicopter transmission
Springer; p. 407–15. system. IEEE Sens J 2020.
[34] Chang C-W, Lee H-W, Liu C-H. A review of artificial intelligence
[62] Xu G, Liu M, Jiang Z, Shen W, Huang C. Online fault diagnosis method
algorithms used for smart machine tools. Inventions 2018;3:41.
based on transfer convolutional neural networks. IEEE Trans Instrum
[35] Wang J, Ma Y, Zhang L, Gao RX, Wu D. Deep learning for smart Meas 2020;69:509–20.
manufacturing: Methods and applications. J Manuf Syst 2018;48:144–56.
[63] Mao W, Ding L, Tian S, Liang X. Online detection for bearing incipient
[36] Sharp M, Ak R, Hedber Jr T. A survey of the advancing use and
fault based on deep transfer learning. Measurement 2020;152.
development of machine learning in smart manufacturing. J Manuf Syst
[64] Chen Z, Gryllias K, Li W. Intelligent fault diagnosis for rotary machinery
2018;48:170–9.
using transferable convolutional neural network. IEEE Trans Ind Inf
[37] Mao W, Chen J, Liang X, Zhang X. A new online detection approach for
2020;16:339–49.
rolling bearing incipient fault via self-adaptive deep feature matching.
[65] Li Q, Tang B, Deng L, Wu Y, Wang Y. Deep balanced domain adaptation
IEEE Trans Instrum Meas 2020;69:443–56.
neural networks for fault diagnosis of planetary gearboxes with limited
[38] Chen L, Zhang Z, Cao J, Wang X. A novel method of combining non-
labeled data. Measurement 2020;156.
linear frequency spectrum and deep learning for complex system fault
[66] Jiao J, Zhao M, Lin J. Unsupervised adversarial adaptation network for
diagnosis. Measurement 2020;151.
intelligent fault diagnosis. IEEE Trans Ind Electron 2019.
[39] Zhang Y, Li X, Gao L, Chen W, Li P. Intelligent fault diagnosis of
[67] Grezmak J, Zhang J, Wang P, Loparo KA, Gao RX. Interpretable convo-
rotating machinery using a new ensemble deep auto-encoder method.
lutional neural network through layer-wise relevance propagation for
Measurement 2020;151.
machine fault diagnosis. IEEE Sensors J 2020;20:3172–81.
[40] Kong X, Mao G, Wang Q, Ma H, Yang W. A multi-ensemble method
based on deep auto-encoders for fault diagnosis of rolling bearings. [68] Haidong S, Junsheng C, Hongkai J, Yu Y, Zhantao W. Enhanced deep gated
Measurement 2020;151. recurrent unit and complex wavelet packet energy moment entropy for
early fault prognosis of bearing. Knowl-Based Syst 2020;188.
[41] Jiang N, Hu X, Li N. Graphical temporal semi-supervised deep learning-
based principal fault localization in wind turbine systems. Proc. Inst. [69] Zhao K, Jiang H, Li X, Wang R. An optimal deep sparse autoencoder with
Mech. Eng. I 2020. gated recurrent unit for rolling bearing fault diagnosis. Meas Sci Technol
[42] Li X, Li J, Qu Y, He D. Semi-supervised gear fault diagnosis using raw 2020;31.
vibration signal based on deep learning. Chin J Aeronaut 2020;33:418–26. [70] Wu Z, Jiang H, Zhao K, Li X. An adaptive deep transfer learning method
[43] Xiong X, Jiang H, Li X, Niu M. A wasserstein gradient-penalty generative for bearing fault diagnosis. Measurement 2020;151.
adversarial network with deep auto-encoder for bearing intelligent fault [71] Ma Y, Jia X, Bai H, Wang G, Liu G, Guo C. A new fault diagnosis
diagnosis. Meas Sci Technol 2020;31. method using deep belief network and compressive sensing. J Vibroeng
[44] Zhou F, Yang S, Fujita H, Chen D, Wen C. Deep learning fault diag- 2020;22:83–97.
nosis method based on global optimization GAN for unbalanced data. [72] Yan L-P, Dong X-Z, Wang T, Gao Q, Tan C-Q, Zeng D-T, et al. A fault
Knowl-Based Syst 2020;187. diagnosis method for gas turbines based on improved data preprocessing
[45] Guo Q, Li Y, Song Y, Wang D, Chen W. Intelligent fault diagnosis method and an optimization deep belief network. Meas Sci Technol 2020;31.
based on full 1-d convolutional generative adversarial network. IEEE Trans [73] Yu K, Lin TR, Tan J. A bearing fault and severity diagnostic technique
Ind Inf 2020;16:2044–53. using adaptive deep belief networks and Dempster–Shafer theory. Struct
[46] Zhao X, Jia M, Lin M. Deep Laplacian Auto-encoder and its application Health Monitor 2020;19:240–61.
into imbalanced fault diagnosis of rotating machinery. Measurement [74] Ding Y, Ma L, Ma J, Suo M, Tao L, Cheng Y, et al. Intelligent fault
2020;152. diagnosis for rotating machinery using deep q-network based health
[47] Zhiyi H, Haidong S, Lin J, Junsheng C, Yu Y. Transfer fault diagno- state classification: A deep reinforcement learning approach. Adv Eng Inf
sis of bearing installed in different machines using enhanced deep 2019;42:100977.
auto-encoder. Measurement 2020a;152. [75] Dai W, Mo Z, Luo C, Jiang J, Zhang H, Miao Q. Fault diagnosis of
[48] Zhiyi H, Haidong S, Ping W, Lin JJ, Junsheng C, Yu Y. Deep transfer multi- rotating machinery based on deep reinforcement learning and reciprocal
wavelet auto-encoder for intelligent fault diagnosis of gearbox with few of smoothness index. IEEE Sens J 2020.
target training samples. Knowl-Based Syst 2020b;191. [76] Zhang D, Stewart E, Entezami M, Roberts C, Yu D. Intelligent acoustic-
[49] Li X, Jia X-D, Zhang W, Ma H, Luo Z, Li X. Intelligent cross-machine based fault diagnosis of roller bearings using a deep graph convolutional
fault diagnosis approach with deep auto-encoder and domain adaptation. network. Measurement 2020;156.
Neurocomputing 2020;383:235–47. [77] Rumelhart DE, Hinton GE, Williams RJ. Learning internal representations
[50] Mao W, Zhang D, Tian S, Tang J. Robust detection of bearing early fault by error propagation. Technical report, California Univ San Diego La Jolla
based on deep transfer learning. Electronics 2020;9. Inst for Cognitive Science; 1985.
[51] Zhang Y, Xing K, Bai R, Sun D, Meng Z. An enhanced convolutional neural [78] Vincent P, Larochelle H, Bengio Y, Manzagol P-A. Extracting and compos-
network for bearing fault diagnosis based on time–frequency image. ing robust features with denoising autoencoders. In: Proceedings of the
Measurement 2020;157. 25th international conference on Machine learning. ACM; p. 1096–103.

Please cite this article as: Z. Zhao, T. Li, J. Wu et al., Deep learning algorithms for rotating machinery intelligent diagnosis: An open source benchmark study. ISA Transactions
(2020), https://doi.org/10.1016/j.isatra.2020.08.010.
32 Z. Zhao, T. Li, J. Wu et al. / ISA Transactions xxx (xxxx) xxx

[79] Ranzato M, Poultney C, Chopra S, Cun YL. Efficient learning of sparse [96] Lee J, Qiu H, Yu G, Lin Ja. Bearing data set. Moffett Field, CA: IMS,
representations with an energy-based model. In: Advances in neural University of Cincinnati, NASA Ames Prognostics Data Repository, NASA
information processing systems. p. 1137–44. Ames Research Center; 2007, https://ti.arc.nasa.gov/tech/dash/groups/
[80] LeCun Y, Bengio Y. Convolutional networks for images, speech, and time pcoe/prognostic-data-repository/.
series. In: The handbook of brain theory and neural networks. vol. 3361. [97] Zhang Y, Li X, Gao L, Wang L, Wen L. Imbalanced data fault diagnosis of
1995, p. 1995. rotating machinery using synthetic oversampling and feature learning. J
[81] He K, Zhang X, Ren S, Sun J. Deep residual learning for image recognition. Manuf Syst 2018;48:34–50.
In: Proceedings of the IEEE conference on computer vision and pattern [98] Mao W, Liu Y, Ding L, Li Y. Imbalanced fault diagnosis of rolling bearing
recognition. p. 770–8. based on generative adversarial network: A comparative study. IEEE
[82] Hochreiter S, Schmidhuber J. Long short-term memory. Neural Comput Access 2019;7:9515–30.
1997;9:1735–80. [99] Buda M, Maki A, Mazurowski MA. A systematic study of the class
[83] Case western reserve university (CWRU) bearing data center. 2019, imbalance problem in convolutional neural networks. Neural Netw
[Online]. Available: https://csegroups.case.edu/bearingdatacenter/pages/ 2018;106:249–59.
download-data-file/. [Accessed September 2019]. [100] Zhao Z, Zhang Q, Yu X, Sun C, Wang S, Yan R, et al. Unsupervised
[84] Society for machinery failure prevention technology. 2019, [Online]. deep transfer learning for intelligent fault diagnosis: An open source and
Available: https://mfpt.org/fault-data-sets/. [Accessed September 2019]. comparative study. 2019, arXiv preprint arXiv:1912.12528.
[85] Lessmeier C, Kimotho JK, Zimmer D, Sextro W. KAt-datacenter, chair of [101] Zheng H, Wang R, Yang Y, Yin J, Li Y, Li Y, et al. Cross-domain
design and drive technology. Paderborn University; 2019, https://mb.uni- fault diagnosis using knowledge transfer strategy: a review. IEEE Access
paderborn.de/kat/forschung/datacenter/bearing-datacenter/ [Accessed on 2019;7:129260–90.
August 2019]. [102] Yan R, Shen F, Sun C, Chen X. Knowledge transfer for rotary machine
[86] Lessmeier C, Kimotho JK, Zimmer D, Sextro W. Condition monitoring fault diagnosis. IEEE Sens J 2019.
of bearing damage in electromechanical drive systems by using motor [103] Han T, Liu C, Yang W, Jiang D. Deep transfer network with joint
current signals of electric motors: A benchmark data set for data- distribution adaptation: A new intelligent fault diagnosis framework for
driven classification. In: Proceedings of the European conference of the industry application. ISA Trans 2019a.
prognostics and health management society. p. 05–8. [104] Han T, Liu C, Yang W, Jiang D. Learning transferable features in deep
[87] Cao P, Zhang S, Tang J. Gear fault data. 2019, [Online]. Available: https: convolutional neural networks for diagnosing unseen machine conditions.
//doi.org/10.6084/m9.figshare.6127874.v1. [Accessed September 2019]. ISA Trans 2019b;93:341–53.
[88] XJTU-SY bearing datasets. 2019, [Online]. Available:http://biaowang.tech/ [105] Rudin C. Stop explaining black box machine learning models for high
xjtu-sy-bearing-datasets/. [Accessed September 2019]. stakes decisions and use interpretable models instead. Nat Mach Intell
[89] Wang B, Lei Y, Li N, Li N. A hybrid prognostics approach for estimating 2019;1:206–15.
remaining useful life of rolling element bearings. IEEE Trans Reliab 2018. [106] Li X, Zhang W, Ding Q. Understanding and improving deep learning-based
[90] SEU gearbox datasets. 2019, [Online]. Available: https://github.com/ rolling bearing fault diagnosis with attention mechanism. Signal Process
cathysiyu/Mechanical-datasets. [Accessed September 2019]. 2019a;161:136–54.
[91] Shao S, McAleer S, Yan R, Baldi P. Highly accurate machine fault diagnosis [107] Li T, Zhao Z, Sun C, Cheng L, Chen X, Yan R, et al. Waveletkernelnet:
using deep transfer learning. IEEE Trans Ind Inf 2018;15:2446–55. An interpretable deep neural network for industrial intelligent diagnosis.
[92] Li K. School of mechanical engineering. Jiangnan University; 2019, 2019b, arXiv preprint arXiv:1911.07925.
http://mad-net.org:8765/explore.html?t=0.5831516555847212 [Accessed [108] Zeiler MD, Fergus R. Visualizing and understanding convolutional net-
on August 2019]. works. In: European conference on computer vision. Springer; p.
[93] Li K, Ping X, Wang H, Chen P, Cao Y. Sequential fuzzy diagnosis method 818–33.
for motor roller bearing in variable operating conditions based on [109] Tang M, Perazzi F, Djelouah A, Ben Ayed I, Schroers C, Boykov Y. On reg-
vibration analysis. Sensors 2013;13:8013–41. ularized losses for weakly-supervised cnn segmentation. In: Proceedings
[94] PHM IEEE 2012 data challenge. 2019, [Online]. Available: https: of the European conference on computer vision. p. 507–22.
//github.com/wkzs111/phm-ieee-2012-data-challenge-dataset. [Accessed [110] Ravanelli M, Bengio Y. Interpretable convolutional filters with sincnet.
September 2019]. 2018, arXiv preprint arXiv:1811.09725.
[95] Nectoux P, Gouriveau R, Medjaher K, Ramasso E, Chebel-Morello B, [111] Gregor K, LeCun Y. Learning fast approximations of sparse coding. In: Pro-
Zerhouni N, et al. Pronostia: An experimental platform for bearings accel- ceedings of the 27th international conference on international conference
erated degradation tests. In: IEEE international conference on prognostics on machine learning. Omnipress; p. 399–406.
and health management. IEEE Catalog Number: CPF12PHM-CDR. p. 1–8. [112] Wang Y, Yao Q, Kwok J, Ni LM. Generalizing from a few examples: A
survey on few-shot learning. 2019, arXiv:1904.05046.
[113] Elsken T, Metzen JH, Hutter F. Neural architecture search: A survey. J
Mach Learn Res 2019;20:1–21.

Please cite this article as: Z. Zhao, T. Li, J. Wu et al., Deep learning algorithms for rotating machinery intelligent diagnosis: An open source benchmark study. ISA Transactions
(2020), https://doi.org/10.1016/j.isatra.2020.08.010.

You might also like