Zhao 2020
Zhao 2020
ISA Transactions
journal homepage: www.elsevier.com/locate/isatrans
Research article
article info a b s t r a c t
Article history: Rotating machinery intelligent diagnosis based on deep learning (DL) has gone through tremendous
Received 25 January 2020 progress, which can help reduce costly breakdowns. However, different datasets and hyper-parameters
Received in revised form 30 July 2020 are recommended to be used, and few open source codes are publicly available, resulting in unfair
Accepted 7 August 2020
comparisons and ineffective improvement. To address these issues, we perform a comprehensive
Available online xxxx
evaluation of four models, including multi-layer perception (MLP), auto-encoder (AE), convolutional
Keywords: neural network (CNN), and recurrent neural network (RNN), with seven datasets to provide a
Deep learning benchmark study. We first gather nine publicly available datasets and give a comprehensive benchmark
Machinery intelligent diagnosis study of DL-based models with two data split strategies, five input formats, three normalization
Open source codes methods, and four augmentation methods. Second, we integrate the whole evaluation codes into a
Benchmark study
code library and release it to the public for better comparisons. Third, we use specific-designed cases
to point out the existing issues, including class imbalance, generalization ability, interpretability, few-
shot learning, and model selection. Finally, we release a unified code framework for comparing and
testing models fairly and quickly, emphasize the importance of open source codes, provide the baseline
accuracy (a lower bound), and discuss existing issues in this field. The code library is available at:
https://github.com/ZhaoZhibin/DL-based-Intelligent-Diagnosis-Benchmark.
© 2020 Published by Elsevier Ltd on behalf of ISA.
https://doi.org/10.1016/j.isatra.2020.08.010
0019-0578/© 2020 Published by Elsevier Ltd on behalf of ISA.
Please cite this article as: Z. Zhao, T. Li, J. Wu et al., Deep learning algorithms for rotating machinery intelligent diagnosis: An open source benchmark study. ISA Transactions
(2020), https://doi.org/10.1016/j.isatra.2020.08.010.
2 Z. Zhao, T. Li, J. Wu et al. / ISA Transactions xxx (xxxx) xxx
Please cite this article as: Z. Zhao, T. Li, J. Wu et al., Deep learning algorithms for rotating machinery intelligent diagnosis: An open source benchmark study. ISA Transactions
(2020), https://doi.org/10.1016/j.isatra.2020.08.010.
Z. Zhao, T. Li, J. Wu et al. / ISA Transactions xxx (xxxx) xxx 3
In bearing fault diagnosis, Li et al. [16] provided a systematic by [41,42]. For imbalanced learning, generation adversarial net-
review of fuzzy formalisms including combination with other work (GAN) was used to combine with AE models to generate
machine learning algorithms. Hoang et al. [17] provided a com- new labeled samples in [43–45]. In [46], a model called deep
prehensive review of three popular DL algorithms (AE, DBN, and Laplacian AE (DLapAE) was proposed by introducing the Laplacian
CNN) for bearing fault diagnosis. Zhang et al. [18] systemati- regularization to improve the generalization performance. For
cally reviewed the machine learning and DL-based algorithms for transfer learning, the pretrained and fine-tuned approach was
bearing fault diagnosis and also provided a comparison of the applied to AE models to realize the knowledge transfer in [47,48].
classification accuracy of CWRU with different DL-based methods. Domain adaptation was also used to transfer the knowledge
Hamadache et al. [19] reviewed different fault modes of rolling learned by AE models to the target domain in [49,50].
element bearings and described various health indexes for PHM. In CNN models, for model improvement, different input types,
Meanwhile, it also provided a survey of artificial intelligence (AI) such as time–frequency images [51], vibration spectrum im-
methods for PHM including shallow learning and DL.
ages [52], infrared thermal images [53], and two-dimensional
In rotating machinery intelligent diagnosis, Ali et al. [20] pro-
images [54], were used as the inputs of CNN models. Multiple
vided a review of AI-based methods using acoustic emission data
wavelet regularizations [55], data augmentation methods [56],
for rotating machinery condition monitoring. Liu et al. [21] re-
and information fusion technology [57] were also applied to
viewed Al-based approaches including k-nearest neighbors
improve the performance of CNN models. Hand-crafted features
(KNN), support vector machine (SVM), artificial neural networks
(ANN), Naive Bayes, and DL for fault diagnosis of rotating ma- were combined with CNN features to boost the performance
chinery. Wei et al. [22] summarized early fault diagnosis of in [58]. For imbalanced learning, GAN was also used to combine
gears, bearings, and rotors through signal processing methods with CNN models to generate new labeled samples in [59,60].
(adaptive decomposition methods, wavelet transform, and sparse Focal loss, which can deal with severe imbalanced problems,
decomposition) and AI-based methods (KNN, neural network, and was used by [61] to allow CNN models to learn discriminative
SVM). features. For transfer learning, the pretrained and fine-tuned
In machinery condition monitoring, Zhao et al. [23] and Duan approach was used to leverage the prior knowledge from the
et al. [24] reviewed diagnosis and prognosis of mechanical equip- source task in [62–64]. Domain adaptation methods were also
ment based on DL algorithms such as DBN and CNN. Zhang applied to allow CNN models to learn transferable features in [65,
et al. [25] reviewed computational intelligent approaches in- 66]. In addition, layer-wise relevance propagation was also used
cluding ANN, evolutionary algorithms, fuzzy logic, and SVM for to understand how CNN models learn to distinguish different
machinery fault diagnosis. Zhao et al. [26] reviewed data-driven patterns [67].
machine health monitoring through DL methods (AE, DBN, CNN, Beyond that, complex wavelet packet energy moment en-
and RNN) and provided the data and codes (in Keras) about an tropy [68] and the gray wolf optimizer algorithm [69] were
experimental study. Lei et al. [27] presented a systematical review combined with an enhanced deep gated recurrent unit to improve
to cover the development of intelligent diagnosis following the the security of rotating machinery. Joint distribution adaptation
progress of machine learning and DL models and offer a future was embedded into LSTM to realize learning transferable features
perspective called transfer learning theories. in [70]. DBN models were also modified in [71–73] to improve
In addition, Nasiri et al. [28] surveyed the state-of-the-art the diagnosis performance of rotating machinery. Deep reinforce-
AI-based approaches for fracture mechanics and provided the ment learning was also used in intelligent fault diagnosis for
accuracy comparisons achieved by different machine learning rotating machinery in [74,75]. Meanwhile, a deep graph convo-
algorithms for mechanical fault detection. Tian et al. [29] sur-
lutional network (DGCN) was first applied to rolling bearing fault
veyed different modes of traction induction motor fault and their
diagnosis based on acoustic signals [76].
diagnosis algorithms including model-based methods and AI-
Although a large body of DL-based methods and many related
based methods. Khan et al. [30] provided a comprehensive re-
reviews have been published in the field of intelligent diagno-
view of AI for system health management and emphasized the
sis, few studies thoroughly evaluate various DL-based intelligent
trend of DL-based methods with limitations and benefits. Stetco
et al. [31] reviewed machine learning approaches applied to diagnosis algorithms for most of the publicly available datasets,
wind turbine condition monitoring and made a discussion of provide the benchmark accuracy, and release the code library
the possibility for future research. Ellefsen et al. [32] reviewed for complete evaluation procedures. For example, a simple code
four well-established DL algorithms including AE, CNN, DBN, and written in Keras was published in [26], which is not compre-
long short-term memory network (LSTM) for PHM applications hensive enough for different datasets and models. The accuracy
and discussed challenges for the future studies, especially in the comparisons were provided in [18,28] according to existing pa-
field of PHM in autonomous ships. AI-based algorithms (tradi- pers, but they were not comprehensive enough due to different
tional machine learning algorithms and DL-based approaches) configurations. Therefore, this paper is intended to make up for
and applications (smart sensors, intelligent manufacturing, PHM, this gap and emphasize the importance of open source codes and
and cyber–physical systems) were reviewed in [33–36] for smart the benchmark study.
manufacturing and manufacturing diagnosis.
Due to the fact that there are already many review papers
3. Evaluation algorithms
covering DL-based rotating machinery intelligent diagnosis pub-
lished before 2020, we further review most of related papers
published in 2020 and summarize their main contributions to fill It is impossible to cover all the published models since there is
the void. currently no open source community in this field. Therefore, we
In AE models, for model improvement, AE models were com- switch to test the performance of four models (including MLP,
bined with some other data preprocessing methods, such as AE, CNN, and RNN) via embedding some advanced techniques.
singular value decomposition [37] and nonlinear frequency spec- It should be noted that DBN is also another commonly used DL
trum [38]. The ensemble learning strategy was also used to boost model in intelligent diagnosis, but we do not add it to this code
the performance of AE models in [39,40]. Meanwhile, the semi- library due to that the fact the training way of DBN is much
supervised learning methods were also embedded into AE models different from those four models.
Please cite this article as: Z. Zhao, T. Li, J. Wu et al., Deep learning algorithms for rotating machinery intelligent diagnosis: An open source benchmark study. ISA Transactions
(2020), https://doi.org/10.1016/j.isatra.2020.08.010.
4 Z. Zhao, T. Li, J. Wu et al. / ISA Transactions xxx (xxxx) xxx
Please cite this article as: Z. Zhao, T. Li, J. Wu et al., Deep learning algorithms for rotating machinery intelligent diagnosis: An open source benchmark study. ISA Transactions
(2020), https://doi.org/10.1016/j.isatra.2020.08.010.
Z. Zhao, T. Li, J. Wu et al. / ISA Transactions xxx (xxxx) xxx 5
Please cite this article as: Z. Zhao, T. Li, J. Wu et al., Deep learning algorithms for rotating machinery intelligent diagnosis: An open source benchmark study. ISA Transactions
(2020), https://doi.org/10.1016/j.isatra.2020.08.010.
6 Z. Zhao, T. Li, J. Wu et al. / ISA Transactions xxx (xxxx) xxx
Please cite this article as: Z. Zhao, T. Li, J. Wu et al., Deep learning algorithms for rotating machinery intelligent diagnosis: An open source benchmark study. ISA Transactions
(2020), https://doi.org/10.1016/j.isatra.2020.08.010.
Z. Zhao, T. Li, J. Wu et al. / ISA Transactions xxx (xxxx) xxx 7
Table 2
Detailed description of MFPT datasets.
Fault Mode Description
Health State Fault-free bearing working at 270 lbs
Outer ring 1 Outer ring fault bearing working at 25 lbs
Outer ring 2 Outer ring fault bearing working at 50 lbs
Outer ring 3 Outer ring fault bearing working at 100 lbs
Outer ring 4 Outer ring fault bearing working at 150 lbs
Outer ring 5 Outer ring fault bearing working at 200 lbs
Outer ring 6 Outer ring fault bearing working at 250 lbs
Outer ring 7 Outer ring fault bearing working at 300 lbs
Outer ring 1 Inner ring fault bearing working at 0 lbs
Inner ring 2 Inner ring fault bearing working at 50 lbs
Inner ring 3 Inner ring fault bearing working at 100 lbs
Inner ring 4 Inner ring fault bearing working at 150 lbs
Inner ring 5 Inner ring fault bearing working at 200 lbs
Inner ring 6 Inner ring fault bearing working at 250 lbs
Fig. 6. The structure of a LSTM cell. Inner ring 7 Inner ring fault bearing working at 300 lbs
Table 1
Detailed description of CWRU datasets.
4.3. PU bearing dataset
Fault Mode Description
Health State the normal bearing at 1791 rpm and 0 HP
Inner ring 1 0.007 inch inner ring fault at 1797 rpm and 0 HP Paderborn University (PU) datasets were provided by the
Inner ring 2 0.014 inch inner ring fault at 1797 rpm and 0 HP Paderborn University Bearing Data Center [85,86], and PU
Inner ring 3 0.021 inch inner ring fault at 1797 rpm and 0 HP
datasets consisted of 32 sets of current signals and vibration
Rolling Element 1 0.007 inch rolling element fault at 1797 rpm and 0 HP
Rolling Element 2 0.014 inch rolling element fault at 1797 rpm and 0 HP signals. As shown in Table 3, bearings were divided into: (1) six
Rolling Element 3 0.021 inch rolling element fault at 1797 rpm and 0 HP undamaged bearings; (2) twelve artificially damaged bearings;
Outer ring 1 0.007 inch outer ring fault at 1797rpm and 0 HP (3) fourteen bearings with real damages caused by accelerated
Outer ring 2 0.014 inch outer ring fault at 1797rpm and 0 HP
Outer ring 3 0.021 inch outer ring fault at 1797rpm and 0 HP
lifetime tests. Each dataset was collected under four working
conditions as shown in Table 4.
In this paper, since using all the data would cause huge com-
putational time, we only used the data collected from real dam-
IMS are not suitable for fault classification that requires labels. To
aged bearings (including KA04, KA15, KA16, KA22, KA30, KB23,
sum up, this paper uses seven datasets to verify the performance
KB24, KB27, KI14, KI16, KI17, KI18, and KI22) under the working
of models introduced in Section 3. The description of all these
datasets is listed as follows. condition N15_M07_F10 to carry out the performance verifica-
tion. Since KI04 was the same as KI14 completely shown in
4.1. CWRU bearing dataset Table 3, we deleted KI04 and the total number of classes was
thirteen. Besides, only vibration signals were used for testing the
Case Western Reserve University (CWRU) datasets were pro- models.
vided by the Case Western Reserve University Bearing Data Cen-
ter [83]. Vibration signals were collected at 12 kHz or 48 kHz for
normal bearings and damaged bearings with single-point defects 4.4. UoC gear fault dataset
under four different motor loads. Within each working condition,
single-point faults were introduced with fault diameters of 0.007, University of Connecticut (UoC) gear fault datasets were pro-
0.014, and 0.021 inches on the rolling element, the inner ring, vided by the University of Connecticut [87], and UoC datasets
and the outer ring, respectively. In this paper, we used the data were collected at 20 kHz. In this dataset, nine different gear
collected from the drive end, and the sampling frequency was
conditions were introduced to the pinions on the input shaft, in-
equivalent to 12 kHz. In Table 1, one healthy bearing and three
cluding healthy condition, missing tooth, root crack, spalling, and
fault modes, including the inner ring fault, the rolling element
fault, and the outer ring fault, were classified into ten categories chipping tip with 5 different levels of severity. All the collected
(one health state and 9 fault states) according to different fault datasets were used and classified into nine categories (one health
sizes. state and eight fault states) to test the performance.
Please cite this article as: Z. Zhao, T. Li, J. Wu et al., Deep learning algorithms for rotating machinery intelligent diagnosis: An open source benchmark study. ISA Transactions
(2020), https://doi.org/10.1016/j.isatra.2020.08.010.
8 Z. Zhao, T. Li, J. Wu et al. / ISA Transactions xxx (xxxx) xxx
Table 3
Detailed description of PU datasets (S: single damage; R: repetitive damage; M: multiple damage)
Bearing Code Fault Mode Description Bearing Code Fault Mode Description
K001 Health state Run-in 50 h before KI07 Artificial inner ring Made by electric
test fault (Level 2) engraver
K002 Health state Run-in 19 h before KI08 Artificial inner ring Made by electric
test fault (Level 2) engraver
K003 Health state Run-in 1 h before test KA04 Outer ring damage Caused by fatigue and
(single point + S + pitting
Level 1)
K004 Health state Run-in 5 h before test KA15 Outer ring damage Caused by plastic
(single point + S + deform and
Level 1) indentation
K005 Health state Run-in 10 h before KA16 Outer ring damage Caused by fatigue and
test (single point + R + pitting
Level 2)
K006 Health state Run-in 16 h before KA22 Outer ring damage Caused by fatigue and
test (single point + S + pitting
Level 1)
KA01 Artificial outer ring Made by EDM KA30 Outer ring damage Caused by plastic
fault (Level 1) (distributed + R + deform and
Level 1) indentation
KA03 Artificial outer ring Made by electric KB23 Outer ring and inner Caused by fatigue and
fault (Level 2) engraver ring damage (single pitting
point + M + Level 2)
KA05 Artificial outer ring Made by electric KB24 Outer ring and inner Caused by fatigue and
fault (Level 1) engraver ring damage pitting
(distributed + M +
Level 3)
KA06 Artificial outer ring Made by electric KB27 Outer ring and inner Caused by plastic
fault (Level 2) engraver ring damage deform and
(distributed + M + indentation
Level 1)
KA07 Artificial outer ring Made by drilling KI04 Inner ring damage Caused by fatigue and
fault (Level 1) (single point + M + pitting
Level 1)
KA08 Artificial outer ring Made by drilling KI14 Inner ring damage Caused by fatigue and
fault (Level 2) (single point + M + pitting
Level 1)
KA09 Artificial outer ring Made by drilling KI16 Inner ring damage Caused by fatigue and
fault (Level 2) (single point + S + pitting
Level 3)
KI01 Artificial inner ring Made by EDM KI17 Inner ring damage Caused by fatigue and
fault (Level 1) (single point + R + pitting
Level 1)
KI03 Artificial inner ring Made by electric KI18 Inner ring damage Caused by fatigue and
fault (Level 1) engraver (single point + S + pitting
Level 2)
KI05 Artificial inner ring Made by electric KI21 Inner ring damage Caused by fatigue and
fault (Level 1) engraver (single point + S + pitting
Level 1)
Please cite this article as: Z. Zhao, T. Li, J. Wu et al., Deep learning algorithms for rotating machinery intelligent diagnosis: An open source benchmark study. ISA Transactions
(2020), https://doi.org/10.1016/j.isatra.2020.08.010.
Z. Zhao, T. Li, J. Wu et al. / ISA Transactions xxx (xxxx) xxx 9
Table 5
Detailed description of XJTU-SY datasets.
Condition File Lifetime Fault element
Bearing 1_1 2h 3 min Outer ring
Bearing 1_2 2h 41 min Outer ring
Speed: 35 Hz
Bearing 1_3 2h 38 min Outer ring
Load: 12 kN
Bearing 1_4 2h 2 min Cage
Bearing 1_5 52 min Inner ring and Outer ring
Bearing 2_1 8h 11 min Inner ring
Speed: 37.5 Bearing 2_2 2h 41 min Outer ring
Hz Load: 11 Bearing 2_3 8h 53 min Cage
kN Bearing 2_4 42 min Outer ring
Bearing 2_5 5h 39 min Outer ring
Bearing 3_1 42 h 18 min Outer ring
Bearing 3_2 41 h 36 min Inner ring, Rolling element, Cage, and Outer ring
Speed: 40 Hz
Bearing 3_3 6 h 11 min Inner ring
Load: 10 kN
Bearing 3_4 25 h 15 min Inner ring
Bearing 3_5 1 h 54 min Outer ring
Table 6 paper, the effects of five input types and three normalization
Detailed description of SEU datasets. methods on the performance of DL models are discussed.
Fault Mode RS-LC Fault Mode RS-LC
Health Gear 20 Hz–0 V Health Bearing 20 Hz–0 V 5.1. Input types
Health Gear 30 Hz–2 V Health Bearing 30 Hz–2 V
Chipped Tooth 20 Hz–0 V Inner ring 20 Hz–0 V
Chipped Tooth 30 Hz–2 V Inner ring 30 Hz–2 V Many researchers use signal processing methods to map the
Missing Tooth 20 Hz–0 V Outer ring 20 Hz–0 V time series to different domains to boost the performance. How-
Missing Tooth 30 Hz–2 V Outer ring 30 Hz–2 V
ever, which input type is more suitable for intelligent diagnosis is
Root Fault 20 Hz–0 V Inner + Outer ring 20 Hz–0 V
Root Fault 30 Hz–2 V Inner + Outer ring 30 Hz–2 V still an open question. In this paper, the effects of different input
Surface Fault 20 Hz–0 V Rolling Element 20 Hz–0 V types on model performance are discussed.
Surface Fault 30 Hz–2 V Rolling Element 30 Hz–2 V
5.1.1. Time domain input
For the time domain input, vibration signals are directly used
4.8. PHM 2012 bearing dataset as the input without data preprocessing. In this paper, the length
of each sample is 1024 and the total number of samples can be
PHM 2012 bearing datasets were used for PHM IEEE 2012 Data obtained from Eq. (9). After generating samples, we take 80% of
Challenge [94,95]. In PHM 2012 datasets, seventeen run-to-failure total samples as the training set and 20% of total samples as the
datasets were provided including six training sets and eleven testing set.
testing sets. Three different loads were considered. Vibration and L
temperature signals were gathered during all those experiments. N = floor( ) (9)
Since no label on the types of failures was given, it was not used 1024
in this paper. where L is the length of each signal, N is the total samples, and
floor means rounding towards minus infinity.
4.9. IMS bearing dataset
5.1.2. Frequency domain input
IMS bearing datasets were generated by the NSF I/UCR Center For the frequency domain input, FFT is used to transform
for Intelligent Maintenance Systems [96]. IMS datasets were made each sample xi from the time domain into the frequency domain
up of three bearing datasets, and each of them contained vibra- shown in Eq. (10). After this operation, the length of data will be
tion signals of four bearings installed on the different locations. At halved and the new sample can be expressed as:
the end of the run-to-failure experiment, a defect occurred on one
of the bearings. The failure occurred in the different locations of xFFT
i = FFT(xi ) (10)
bearings. It is inappropriate to classify these failures simply using
where the operator FFT(·) represents transforming xi into the
three classes, so IMS datasets were not evaluated in this paper.
frequency domain and taking the first half of the result.
5. Data prepreocessing
5.1.3. Time–frequency domain input
The type of input data and the way of normalization have a For the time–frequency domain input, short-time Fourier
great impact on the performance of DL models. Types of input transform (STFT) is applied to each sample xi to obtain the
data determine the difficulty of feature extraction, and normal- time–frequency representation shown in Eq. (11). The Hanning
ization methods determine the difficulty of calculation. So, in this window is used and the window length is 64. After this operation,
Table 7
Detailed description of JNU datasets.
Fault Mode Rotating Fault Mode Rotating Fault Mode Rotating
Speed Speed Speed
Health State 600 rpm Health State 800 rpm Health State 1000 rpm
Inner ring 600 rpm Inner ring 800 rpm Inner ring 1000 rpm
Outer ring 600 rpm Outer ring 800 rpm Outer ring 1000 rpm
Rolling Element 600 rpm Rolling Element 800 rpm Rolling Element 1000 rpm
Please cite this article as: Z. Zhao, T. Li, J. Wu et al., Deep learning algorithms for rotating machinery intelligent diagnosis: An open source benchmark study. ISA Transactions
(2020), https://doi.org/10.1016/j.isatra.2020.08.010.
10 Z. Zhao, T. Li, J. Wu et al. / ISA Transactions xxx (xxxx) xxx
the time–frequency representation (a 33 × 33 image) will be Z-score Normalization: This normalization method can be
generated as: implemented by
5.2. Normalization where x is the 1D input signal, and mask is the binary sequence
whose subsequence of random position is zero. In this paper, the
length of the subsequence is equal to 10.
Input normalization is the basic step in data preparation,
which can facilitate the subsequent data processing and accel-
erate the convergence of DL models. Therefore, we discuss the 6.2. Two dimension input augmentation
effects of three normalization methods on the performance of DL
models. RandomScale: this strategy randomly multiplies the input
Maximum–Minimum Normalization: This normalization signal with a random factor which is formulated as follows:
method can be implemented by x := β ∗ x (20)
xi − xmin where x is the 2D input signal, and β is a scaler following the
xinormalize−1 = i
, i = 1, 2, . . . , N (14)
xmax
i − xmin
i distribution N (1, 0.01).
RandomCrop: this strategy randomly covers partial signals,
where xi is the input sample, xmin
i is the minimum value in xi , and
which is formulated as follows:
xmax
i is the maximum value in xi .
[-1-1] Normalization: This normalization method can be im- x := mask ∗ x (21)
plemented by
where x is the 2D input signal, and mask is the binary sequence
xi − xmin whose subsequence of random position is zero. In this paper, the
xinormalize−2 = −1 + 2 ∗ i
, i = 1, 2, . . . , N (15)
xmax − xmin length of the subsequence is equal to 20.
i i
Please cite this article as: Z. Zhao, T. Li, J. Wu et al., Deep learning algorithms for rotating machinery intelligent diagnosis: An open source benchmark study. ISA Transactions
(2020), https://doi.org/10.1016/j.isatra.2020.08.010.
Z. Zhao, T. Li, J. Wu et al. / ISA Transactions xxx (xxxx) xxx 11
Fig. 7. Random data splitting strategy with preprocessing without overlap. Fig. 8. Another condition with the training and testing sets split as the first
step.
7. Data split
Please cite this article as: Z. Zhao, T. Li, J. Wu et al., Deep learning algorithms for rotating machinery intelligent diagnosis: An open source benchmark study. ISA Transactions
(2020), https://doi.org/10.1016/j.isatra.2020.08.010.
12 Z. Zhao, T. Li, J. Wu et al. / ISA Transactions xxx (xxxx) xxx
and testing sets. For random split, a sliding window is used to D. PU dataset
truncate the vibration signal without any overlap and each data The results of PU dataset are shown in Appendix from Ta-
sample contains 1024 points. After the preparation, we randomly bles A.10 to A.12. It is shown that the accuracy of CNN models is
take 80% of samples as the training set and 20% of samples as generally higher than that of AE models. Besides, the accuracy is
the testing set. For order split, we take the former 80% of time worse when using the wavelet domain as the input, while using
series as the training set and the last 20% as the testing set. FFT and STFT to process the signal allows models to achieve better
Then, in two time series, a sliding window is used to truncate the accuracy. Using Z-score normalization enables AE models and
vibration signal without any overlap, and each sample contains CNN models to achieve higher accuracy. Data augmentation does
1024 points. not help AE models improve the accuracy, while it can increase
In order to verify how input types, data normalization meth- the accuracy of CNN models. Similarly, the order split would
ods, and data split methods affect the performance of models, we heavily reduce the accuracy.
set up three configurations of experiments (shown in Table 8,
E. SEU dataset
Table 9, and Table 10.). During model training, we use Adam
The results of SEU dataset are shown in Appendix from
as the optimizer. The learning rate and the batch size of each
Tables A.13 to A.15. We can observe that when using the time
experiment are set to 0.001 and 64, respectively. Each model
domain or wavelet domain as the input, models would achieve
is trained for 100 epochs, and during the training procedure,
worse accuracy. However, using FFT to process the signal allows
model training and model testing are alternated. In addition, all models to achieve better accuracy and the accuracy of AE models
the experiments are executed under Windows 10 and Pytorch is even higher than that of CNN models. Using Z-score normal-
1.1 through running on a computer with an Intel Core i7-9700K, ization allows AE models and CNN models to achieve higher
GeForce RTX 2080Ti, and 16G RAM. accuracy. Data augmentation can improve the accuracy of both
CNN and AE models. In this case, the order split would slightly
9. Evaluation results reduce the accuracy.
In this section, we will first discuss the experimental results F. UoC dataset
of different datasets in depth. After that, the results of datasets, The results of UoC dataset are shown in Appendix from
input types, models, input normalization, data augmentation, and Tables A.16 to A.18. We can observe that most models do not
data splitting will be summarized separately. Complete results perform well in this case, and among them, the performance
are shown in Appendix and each accuracy which is larger than of AlexNet is relatively worse. Besides, using FFT to process the
95% are bold. signal allows models to achieve better accuracy, and the accuracy
of AE models is higher than that of CNN models. AE models and
9.1. Detailed analysis of different datasets CNN models with Z-score normalization would achieve higher ac-
curacy. Data augmentation can help different models improve the
final accuracy. The order split would heavily reduce the accuracy.
A. CWRU dataset
The results of CWRU dataset are shown in Appendix from
G. XJTU-SY dataset
Tables A.1 to A.3. From those results, we can observe that the The results of XJTU-SY dataset are shown in Appendix from
accuracy of CNN models is generally higher than that of AE mod- Tables A.19 to A.21. We can observe that most models perform
els. In addition, using FFT and STFT to process the signal allows well in this dataset. Besides, we can find that using FFT and
models to achieve better accuracy among five kinds of input. CNN STFT to process the signal allows models to achieve the better
models with Z-score normalization can get better accuracy while accuracy, and the accuracy of CNN models is higher than that of
using -1-1 normalization allows AE models to achieve higher ac- AE models, generally. AE models and CNN models with Z-score
curacy. Using data augmentation does not improve the accuracy normalization would achieve higher accuracy. Data augmentation
of AE models, but it can improve the accuracy of CNN models. The can help different models improve the final accuracy. The order
order split would slightly reduce the accuracy. split would quietly reduce the accuracy.
Please cite this article as: Z. Zhao, T. Li, J. Wu et al., Deep learning algorithms for rotating machinery intelligent diagnosis: An open source benchmark study. ISA Transactions
(2020), https://doi.org/10.1016/j.isatra.2020.08.010.
Z. Zhao, T. Li, J. Wu et al. / ISA Transactions xxx (xxxx) xxx 13
Table 8
Experiment setup 1.
Please cite this article as: Z. Zhao, T. Li, J. Wu et al., Deep learning algorithms for rotating machinery intelligent diagnosis: An open source benchmark study. ISA Transactions
(2020), https://doi.org/10.1016/j.isatra.2020.08.010.
14 Z. Zhao, T. Li, J. Wu et al. / ISA Transactions xxx (xxxx) xxx
Table 9
Experiment setup 2.
of the proposed methods since basic models can achieve very the following five issues including class imbalance, generalization
high accuracy on these datasets, like CWRU and XJTU-SY. Second, ability, interpretability, few-shot learning, and model selection
the frequency domain input can achieve the highest accuracy in using experimental cases.
all datasets, so researchers should first try to use the frequency
domain as the input. Third, it is not necessary for CNN models to
10.1. Class imbalance
get the best results in all cases, and we also should consider the
overfitting problem. Fourth, when the accuracy of datasets is not
very high, data augmentation methods improve the performance Most of measured signals are in the normal state, and only
of models, especially for the time domain input. Thus, more a few of them are in the fault state, which means that fault
effective data augmentation methods need to be investigated. modes often have different probabilities of happening. Therefore,
Finally, in some cases, it may be more suitable for splitting the the class imbalance issue will occur when using intelligent algo-
datasets according to time sequences (order split) since random rithms in real applications. Recently, although some researchers
split may provide virtually high accuracy. We also release a code have published some related papers using traditional imbalanced
library to evaluate DL-based intelligent diagnosis algorithms and learning methods [97] or GAN [98] to solve this problem, these
provide the benchmark accuracy (a lower bound) to avoid use- studies are far from enough. In this paper, PU Bearing Datasets
less improvement. Meanwhile, we use specific-designed cases to are used to simulate the class imbalance issue. In this experiment,
discuss existing issues, including class imbalance, generalization we adopt ResNet18 as the experimental model and only use two
ability, interpretability, few-shot learning, and model selection. kinds of input types (the time domain input and the frequency
Through these works, we aim to allow comparisons fairer and domain input). Besides, data augmentation methods are used and
quicker, emphasize the importance of open source codes, and the normalization method is the Z-score normalization, while the
provide deep discussions of existing issues. To the best of our dataset is randomly split. Three groups of datasets with different
knowledge, this is the first work to comprehensively perform the imbalance ratios are constructed, which are shown in Table 11.
benchmark study and release the code library to the public. As shown in Table 11, three datasets (Group1, Group2, and
Group3) are constituted with different imbalanced ratios. Group1
10. Discussion is a balanced dataset, and there is no imbalance for each state.
In real applications, it is almost impossible to let the number of
Although intelligent diagnosis algorithms can achieve high data samples be the same. We reduce the training samples of
classification accuracy in many datasets, there are still many some fault modes in Group1 to construct Group2, and then the
issues that need to be discussed. In this paper, we further discuss imbalanced classification is simulated. In Group3, the imbalanced
Please cite this article as: Z. Zhao, T. Li, J. Wu et al., Deep learning algorithms for rotating machinery intelligent diagnosis: An open source benchmark study. ISA Transactions
(2020), https://doi.org/10.1016/j.isatra.2020.08.010.
Z. Zhao, T. Li, J. Wu et al. / ISA Transactions xxx (xxxx) xxx 15
Table 10
Experiment setup 3.
Please cite this article as: Z. Zhao, T. Li, J. Wu et al., Deep learning algorithms for rotating machinery intelligent diagnosis: An open source benchmark study. ISA Transactions
(2020), https://doi.org/10.1016/j.isatra.2020.08.010.
16 Z. Zhao, T. Li, J. Wu et al. / ISA Transactions xxx (xxxx) xxx
Fig. 12. Experimental results of working conditions transfer. (a) time domain
Fig. 11. Experimental results of three groups of datasets. (a) time domain input, input, and (b) frequency domain input.
and (b) frequency domain input.
Table 12 Table 13
Training data and testing data for each experiment. The bearing code and the number of samples used in each experiment.
Group Data for training Data for testing Group Bearing code Training samples Testing samples
Group1 N15_M07_F10 N09_M07_F10 KA03 200 50
Group2 N15_M07_F10 N15_M01_F10 Group1
KA06 200 50
Group3 N09_M07_F10 N15_M07_F10
KA08 200 50
Group4 N09_M07_F10 N15_M01_F10 Group2
KA09 200 50
Group5 N15_M01_F10 N15_M07_F10
Group6 N15_M01_F10 N09_M07_F10 KI07 200 50
Group3
KI08 200 50
could adapt to the changes in working conditions or measure- and the working condition at the time of acquisition are the
ment situations since these changes occur frequently in real ap- same between two classes, theoretically, methods should not be
plications. Therefore, studies are still required on how to transfer able to achieve such high accuracy. These expected results are
the trained algorithms to different working conditions effectively. exactly contrary to those of the experiment, which shows that
Two excellent review papers [101,102] and other applica- models only learn the discrimination of different collection points
tions [103,104] published recently pointed out several poten- and do not learn how to extract the essential characteristics of
tial research directions which could be considered and studied fault signals. Therefore, it is very important to figure out whether
further to improve the generalization ability. models can learn essential fault characteristics or just classify the
different conditions of collected signals.
10.3. Interpretability According to the development of interpretability, we might
be able to study the interpretability of DL-based models from
Although intelligent diagnosis algorithms can achieve high the following aspects: (1) visualize the results of neurons to an-
diagnostic accuracy in their tasks, the interpretability of these alyze the attention points of models [108]; (2) add physical con-
models is often insufficient and these black-box models would
straints to the loss function [109] to meet specific needs of fault
generate high-risk results [105], which greatly limits their ap-
feature extraction; (3) add prior knowledge to network struc-
plications. Actually, some papers in intelligent diagnosis have
tures and convolutions [110] or unroll the existing optimization
noted this problem and attempted to propose some interpretable
algorithms [111] to extract corresponding fault features.
models [106,107].
To point out that intelligent diagnostic algorithms lack in-
terpretability, we perform three sets of experiments on the PU 10.4. Few-shot learning
bearing dataset, and the datasets are shown in Table 13. In each
set of experiments, we use two different data, which have the In intelligent diagnosis, the amount of data is far from big data
same fault pattern and are acquired under the same condition. because of the preciousness of fault data and the high cost of fault
The results, in which intelligent algorithms can get high ac- simulation experiments, especially for the key components. To
curacy in each set of experiments, are shown in Fig. 13. Never- manifest the influence of the sample number on the classification
theless, for each binary classification task, since the fault mode accuracy, we use the PU bearing dataset to design a few-shot
Please cite this article as: Z. Zhao, T. Li, J. Wu et al., Deep learning algorithms for rotating machinery intelligent diagnosis: An open source benchmark study. ISA Transactions
(2020), https://doi.org/10.1016/j.isatra.2020.08.010.
Z. Zhao, T. Li, J. Wu et al. / ISA Transactions xxx (xxxx) xxx 17
Fig. 13. Experimental results of three groups. (a) time domain input, and (b) Fig. 14. Experimental results of different few-shot training patterns. (a) time
frequency domain input. domain input, and (b) frequency domain input.
training pattern with six groups of different sample numbers in trial and error cost multiplied by the number of trial and error
each class. can easily reach a huge cost. Besides, reducing this cost is also
Results of the time domain input and the frequency domain the partial purpose of this benchmark study which provides some
input are shown in Fig. 14. It is shown that with the decrease guidelines to choose a baseline model.
of the sample number, the accuracy decreases sharply. As shown Actually, there is another way called neural architecture search
in Fig. 14, for the time domain input, the Best-Max accuracy (NAS) [113] to avoid the huge cost of trial and error. NAS can al-
decreases from 91.46% to 20.39% as the sample number decreases low designing a neural network automatically through searching
from 100 to 1. Meanwhile, the Best-Max accuracy decreases from for a specific network based on a specific dataset. Limited search
97.73% to 29.67% as the sample number decreases from 100 to 1 space of the network is first constructed according to the physical
with the frequency domain input. prior. After that, a neural network matching a specific dataset is
Although the accuracy can be increased after using FFT, it sampled from the search space through reinforcement learning,
is still too low to be accepted when the number of samples is the evolutionary algorithm or the gradient strategy. Besides, the
extremely small. It is necessary to develop methods based on whole construction process does not require manual participa-
few-shot learning to cope with application scenarios with limited tion, which greatly reduces the cost of building a neural network
samples.
and allows us to focus on specific engineering applications.
Many DL-based few-shot learning models have been proposed
in recent years [112], most of these methods adopt a meta-
learning paradigm by training networks with a large number of 11. Conclusion
tasks, which means that the big data in other related fields is
necessary for these methods. In the field of fault diagnosis, there In this paper, we collect nine publicly available datasets to
is no relevant data with such a big size available, so methods
evaluate the performance of MLP, AE, CNN, and RNN models. This
embedding with physical mechanisms are required to address
work mainly focuses on evaluating DL-based intelligent diagnosis
this problem effectively.
algorithms from different perspectives and providing the bench-
mark accuracy (a lower bound) to avoid useless improvement.
10.5. Model selection
In addition, we release a code library for other researchers to
For intelligent diagnosis, designing a neural network is not the test the performance of their own DL-based intelligent diagnosis
final goal, and our task is to apply the model to real industrial models of these datasets. We hope that the evaluation results
applications while designing a neural network is only a small part and the code library could promote a better understanding of
of our task. However, to achieve a good effect, we have to spend DL-based models and provide a unified framework for generating
considerable time and energy on designing the corresponding more effective models. For further studies, we will focus on five
networks. Because building a neural network is an iterative pro- listed issues (class imbalance, generalization ability, interpretabil-
cess consisting of repeated trial and error, and the performance ity, few-shot learning, and model selection) to propose more
of models should be fed back to us to adjust models. The single customized works.
Please cite this article as: Z. Zhao, T. Li, J. Wu et al., Deep learning algorithms for rotating machinery intelligent diagnosis: An open source benchmark study. ISA Transactions
(2020), https://doi.org/10.1016/j.isatra.2020.08.010.
18 Z. Zhao, T. Li, J. Wu et al. / ISA Transactions xxx (xxxx) xxx
Table A.1
CWRU: Results with random split and data augmentation.
Nor Input Loc AE DAE SAE MLP CNN LeNet AlexNet ResNet18 LSTM
Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max
Last 54.71 66.67 54.33 64.37 67.13 69.73 52.95 67.43 98.47 99.62 96.4 99.23 91.11 96.17 88.28 98.47 98.47 98.85
1
Best 71.42 73.95 78.24 80.84 72.95 76.25 73.49 75.48 98.93 100 99.08 99.62 91.11 96.17 100 100 98.47 98.85
Last 100 100 100 100 99.92 100 100 100 99.89 100 100 100 100 100 99.08 100 100 100
2
Best 100 100 100 100 100 100 100 100 99.92 100 100 100 100 100 100 100 100 100
A Last 86.95 89.89 82.33 84.27 86.02 91.39 – – 95.28 98.88 98.59 99.33 97.89 98.73 98.08 99.52 98.11 98.43
3
Best 90.95 92.51 86.14 88.39 90.01 91.76 – – 97.03 99.18 99.17 99.33 97.89 98.73 99.66 99.85 98.11 98.43
Last 100 100 100 100 100 100 – – 98.7 99.23 98.7 100 99.69 100 100 100 99.85 100
4
Best 100 100 100 100 100 100 – – 99.46 99.62 100 100 99.69 100 100 100 99.85 100
Last 94.71 97.32 92.19 94.64 94.10 97.32 – – 86.74 95.02 74.94 89.27 46.36 94.25 90.19 100 84.29 89.27
5
Best 96.86 98.47 95.94 97.32 96.09 98.85 – – 94.86 96.93 89.81 90.8 46.36 94.25 100 100 84.29 89.27
Last 74.79 77.01 80.31 83.14 76.09 77.78 76.02 78.93 99.45 100 98.08 99.23 98.31 99.23 99.46 100 98.85 100
1
Best 78.85 80.46 82.68 83.52 77.78 78.54 76.71 79.69 99.67 100 98.83 99.88 98.54 99.62 99.46 100 99.92 100
Last 99.92 100 100 100 99.54 100 99.92 100 99.39 100 100 100 99.92 100 99.92 100 99.85 100
2
Best 100 100 100 100 100 100 99.92 100 99.39 100 100 100 99.92 100 99.92 100 100 100
B Last 85.89 90.64 85.27 88.01 87.58 89.14 – – 98.94 99.22 98.84 99.07 98.6 98.84 99.17 99.78 93.71 94.76
3
Best 88.83 92.13 89.01 92.13 90.7 92.13 – – 99.12 99.29 99.15 99.36 98.71 99.03 99.6 99.95 95.36 95.88
Last 100 100 100 100 100 100 – – 97.62 99.62 99 99.62 97.85 100 99.92 100 99.77 100
4
Best 100 100 100 100 100 100 – – 98.93 100 99.23 100 98.31 100 99.92 100 100 100
Last 93.1 95.79 94.56 96.55 89.65 95.02 – – 86.9 93.87 86.05 88.51 91.03 97.32 97.09 99.23 91.65 93.87
5
Best 95.33 96.17 97.01 98.85 93.95 97.32 – – 89.46 95.64 86.67 90.04 92.27 97.32 97.62 100 93.33 94.64
Last 66.13 72.03 74.87 77.78 68.81 74.71 69.88 74.71 99.16 100 99.15 99.23 99.46 100 99.16 100 99.16 99.62
1
Best 70.8 72.8 77.32 79.69 71.8 74.71 71.95 74.71 100 100 99.69 100 99.92 100 100 100 100 100
Last 100 100 100 100 100 100 100 100 99.85 100 100 100 100 100 100 100 100 100
2
Best 100 100 100 100 100 100 100 100 100 100 100 100 100 100 100 100 100 100
C Last 88.49 91.39 88.97 90.64 88.39 92.13 – – 99.17 99.37 99.07 99.25 98.69 98.92 99.59 99.81 98.3 98.58
3
Best 91.35 93.63 91.25 92.51 90.94 93.63 – – 99.43 99.55 99.23 99.37 99.13 99.25 99.86 99.89 98.86 99.07
Last 100 100 100 100 100 100 – – 99.85 100 97.85 100 100 100 100 100 99.85 100
4
Best 100 100 100 100 100 100 – – 100 100 99.92 100 100 100 100 100 100 100
Last 95.25 98.85 94.86 97.7 97.09 98.47 – – 92.03 93.87 87.2 91.57 87.2 98.85 91.19 100 92.26 96.17
5
Best 97.93 99.62 96.47 98.47 98.62 100 – – 94.41 95.02 92.11 93.1 98.08 99.23 100 100 96.01 97.7
Table A.2
CWRU: Results with random split and without data augmentation.
Nor Input Loc AE DAE SAE MLP CNN LeNet AlexNet ResNet18 LSTM
Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max
Last 60.08 65.9 54.48 68.97 69.58 71.26 64.29 70.11 99.54 99.62 99.46 100 95.79 98.08 99.77 100 99.92 100
1
Best 70.96 73.56 71.34 76.63 71.57 73.18 74.41 76.63 99.85 100 99.77 100 97.4 98.47 100 100 100 100
Last 100 100 100 100 100 100 100 100 99.85 100 100 100 100 100 99.77 100 99.92 100
2
Best 100 100 100 100 100 100 100 100 100 100 100 100 100 100 100 100 100 100
A Last 85.14 91.01 85.33 89.89 89.45 91.76 – – 98.85 99.07 98.61 99.18 82.17 98.77 96.6 98.66 97.71 98.66
3
Best 88.08 91.39 87.08 91.39 91.01 93.63 – – 99.35 99.4 99.16 99.18 82.59 98.84 99.55 99.63 98.67 98.84
Last 100 100 100 100 100 100 – – 99.31 100 99.69 100 100 100 100 100 100 100
4
Best 100 100 100 100 100 100 – – 99.46 100 99.92 100 100 100 100 100 100 100
Last 92.95 95.79 90.11 96.55 85.29 96.17 – – 85.52 93.87 79.62 85.44 63.68 95.79 96.55 100 86.82 91.19
5
Best 93.87 96.93 93.41 98.47 94.94 97.7 – – 94.64 96.55 88.81 90.04 67.2 97.32 100 100 90.96 93.1
Last 69.96 71.65 67.28 73.18 70.8 73.56 71.1 73.18 99.4 99.62 99.07 100 97.7 98.85 99.51 100 99.78 100
1
Best 74.48 76.63 71.11 73.95 73.18 76.25 75.26 78.54 99.62 100 99.56 100 98.58 99.62 100 100 100 100
Last 99.62 100 99.92 100 100 100 100 100 99.95 100 100 100 99.89 100 97.37 100 100 100
2
Best 100 100 100 100 100 100 100 100 100 100 100 100 100 100 100 100 100 100
B Last 89.45 91.39 85.39 89.51 90.38 93.63 – – 98.87 99.29 97.79 98.99 98.58 99.14 99.28 99.74 98.27 98.81
3
Best 90.32 92.13 88.14 90.26 91.2 94.01 – – 99.21 99.4 99.15 99.33 99.04 99.33 99.58 99.74 98.82 98.96
Last 90.65 100 100 100 100 100 – – 99.4 100 99.45 100 99.78 100 100 100 99.78 100
4
Best 100 100 100 100 100 100 – – 99.73 100 99.78 100 100 100 100 100 100 100
Last 93.03 97.7 94.33 96.17 95.25 96.55 – – 91.57 96.93 85.77 91.95 83.2 96.93 95.57 99.62 85.11 92.34
5
(continued on next page)
Please cite this article as: Z. Zhao, T. Li, J. Wu et al., Deep learning algorithms for rotating machinery intelligent diagnosis: An open source benchmark study. ISA Transactions
(2020), https://doi.org/10.1016/j.isatra.2020.08.010.
Z. Zhao, T. Li, J. Wu et al. / ISA Transactions xxx (xxxx) xxx 19
Table A.3
CWRU: Results with order split and data augmentation.
Nor Input Loc AE DAE SAE MLP CNN LeNet AlexNet ResNet18 LSTM
Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max
Last 58.71 70.08 55.84 69.7 64.55 68.18 65.23 73.11 95.99 98.11 97.04 98.86 91.21 95.83 82.65 99.62 97.27 98.48
1
Best 71.67 72.73 74.77 76.89 71.29 73.48 69.39 78.41 98.18 100 97.57 99.24 92.5 96.97 98.41 99.62 98.1 98.86
Last 100 100 100 100 100 100 100 100 99.54 100 99.85 100 99.92 100 98.79 100 99.85 100
2
Best 100 100 100 100 100 100 100 100 99.7 100 100 100 100 100 98.79 100 99.85 100
A Last 72.69 78.1 73.9 76.64 76.64 79.2 – – 87.15 88.32 96.67 97.54 95.12 96.76 93.16 99.29 95.17 96.76
3
Best 76.82 80.29 77.07 80.66 79.5 80.66 – – 89.2 90.51 96.86 97.85 96.34 96.8 98.66 99.29 95.81 97.54
Last 99.92 100 99.77 100 100 100 – – 97.73 98.11 99.47 100 97.27 100 100 100 99.01 100
4
Best 100 100 100 100 100 100 – – 98.79 99.24 99.62 100 97.35 100 100 100 99.85 100
Last 92.35 97.73 92.35 98.11 92.2 96.21 – – 78.56 95.45 83.18 90.53 93.47 96.59 100 100 87.35 90.15
5
Best 96.74 98.11 96.06 99.62 95.07 99.62 – – 94.85 96.59 85.08 92.8 94.98 96.59 100 100 88.26 92.42
Last 70.46 72.73 76.36 78.79 70.91 74.24 72.24 73.48 98.7 99.24 96.81 98.86 97.08 97.73 99.24 100 98.41 99.24
1
Best 74.54 76.89 78.86 79.92 74.85 76.52 72.84 73.48 98.97 99.26 97.24 98.86 97.62 98.11 99.57 100 99.09 99.24
Last 100 100 99.85 100 100 100 100 100 98.81 99.62 100 100 99.56 100 97.73 100 100 100
2
Best 100 100 100 100 100 100 100 100 99.13 100 100 100 99.56 100 99.95 100 100 100
B Last 76.95 79.93 74.21 80.29 76.83 78.83 – – 97.04 97.92 96.68 97.65 96.88 97.65 98.19 98.92 86.72 88.32
3
Best 78.83 80.29 77.31 80.29 79.5 80.66 – – 97.23 98.1 96.8 97.8 97.29 98.29 98.53 98.92 88.69 89.78
Last 99.92 100 99.92 100 98.94 100 – – 98.05 98.48 99.24 99.62 97.24 100 99.57 100 98.94 99.62
4
Best 100 100 100 100 100 100 – – 98.37 98.48 99.57 100 99.3 100 99.57 100 99.92 100
Last 94.32 97.73 92.35 98.11 91.21 95.08 – – 88.53 96.59 92.64 94.32 93.02 97.73 96.91 100 89.32 94.32
5
Best 97.12 98.86 95.53 98.86 95.83 96.97 – – 88.8 96.59 93.51 95.83 93.35 97.73 96.91 100 92.88 94.32
Last 64.24 68.94 73.79 77.27 65.46 67.05 66.16 68.18 98.7 99.62 97.98 98.86 98.17 99.24 97.98 100 98.56 98.86
1
Best 67.73 70.83 75.15 77.27 67.12 68.94 67.17 72.35 99.02 99.62 98.23 99.24 98.29 99.24 100 100 98.79 99.24
Last 100 100 100 100 100 100 100 100 99.3 100 99.94 100 99.94 100 99.68 100 100 100
2
Best 100 100 100 100 100 100 100 100 99.78 100 99.94 100 100 100 100 100 100 100
C Last 78.29 79.93 75.79 78.47 77.07 81.75 – – 96.54 97.99 97.05 97.77 95.45 98.21 98.93 99.55 96.44 97.02
3
Best 81.63 84.31 80.17 83.21 80.84 84.67 – – 97.61 98.36 97.29 98.18 96.32 98.21 99.28 99.55 96.69 97.32
Last 99.92 100 100 100 100 100 – – 98.97 99.62 99.56 100 99.77 100 100 100 99.77 100
4
Best 100 100 100 100 100 100 – – 99.08 99.62 99.68 100 99.77 100 100 100 99.92 100
Last 93.71 97.35 87.5 93.18 94.09 98.11 – – 93.34 97.35 93.56 95.08 89.01 93.94 98.18 100 93.64 95.83
5
Best 96.21 97.73 92.42 97.35 96.74 98.11 – – 93.56 97.35 94.07 95.83 89.92 95.45 98.41 100 94.52 95.83
Table A.4
JNU: Results with random split and data augmentation.
Nor Input Loc AE DAE SAE MLP CNN LeNet AlexNet ResNet18 LSTM
Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max
Last 46.74 49.03 46.01 50.51 43.8 45.9 46.97 49.32 80.57 83.28 73.23 77.25 82.97 83.9 77.14 91.47 82.14 83.62
1
Best 52.6 54.72 51.19 53.58 53.41 54.66 52.48 53.81 82.28 85.32 77.23 79.18 83.86 85.32 93.07 93.69 83.2 84.7
Last 95.93 96.3 95.2 95.73 63.47 95.51 95.89 96.76 93.83 94.48 94.59 95.45 95.82 96.19 95.61 96.19 95.02 95.56
2
Best 96.62 96.93 95.58 96.13 64.52 96.64 97.21 97.38 94.36 94.99 95.86 96.02 96.23 96.7 96.76 96.99 95.36 95.56
A Last 33.11 36.2 30.47 31.76 33.99 37.92 – – 43.5 45.81 44.93 46.56 37.04 44.87 52.28 55.19 42.93 44.17
3
Best 37.42 41.42 36.48 36.92 38.98 40.92 – – 46.66 48.25 48.66 49.28 37.59 44.87 55.51 57.16 43.55 45.27
Last 78.63 80.34 78.92 80.63 77.55 82.91 – – 69.69 73.22 66.84 69.52 79.77 81.77 87.12 90.88 71.11 73.22
4
Best 82.91 83.76 82.68 84.05 81.99 86.32 – – 73.16 75.21 70.66 72.36 80.46 84.05 91.17 92.02 72.71 74.07
Last 54.81 57.83 50.26 54.7 45.24 51.57 – – 39.94 43.3 43.65 50.71 50.71 55.27 75.44 80.91 54.3 56.7
5
(continued on next page)
Please cite this article as: Z. Zhao, T. Li, J. Wu et al., Deep learning algorithms for rotating machinery intelligent diagnosis: An open source benchmark study. ISA Transactions
(2020), https://doi.org/10.1016/j.isatra.2020.08.010.
20 Z. Zhao, T. Li, J. Wu et al. / ISA Transactions xxx (xxxx) xxx
Table A.5
JNU: Results with random split and without data augmentation.
Nor Input Loc AE DAE SAE MLP CNN LeNet AlexNet ResNet18 LSTM
Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max
Last 41.64 45.11 39.06 42.78 38.39 41.92 40.6 44.94 80.79 82.31 63.06 69.4 77.57 79.12 87.51 90.96 76.08 77.42
1
Best 44.68 45.96 43.12 44.2 45.83 48.75 45.17 46.25 83.56 84.81 71.96 72.81 80.19 81.51 90.26 91.24 78.88 81.8
Last 95.77 96.47 95.51 97.04 80.21 96.3 96.52 97.21 92.88 94.25 95.05 95.62 95.37 95.9 96.49 97.38 95.23 95.79
2
Best 96.78 96.99 96.84 97.1 80.79 96.99 97.44 97.78 94.01 94.6 95.49 95.96 96.55 96.76 97.4 97.72 95.93 96.53
A Last 32.96 34.81 29.17 29.93 35.01 37.26 – – 43.7 48.11 40.85 47.23 44.6 47.25 45.43 51.69 40.74 44.34
3
Best 37.06 38.59 36.8 38.76 38.51 40.53 – – 47.16 51.66 47.36 49.67 46.64 49.42 52.81 56.22 45.7 47.5
Last 81.31 86.04 81.14 87.18 82.45 84.62 – – 69.74 72.36 67.24 68.95 78.58 81.2 90.19 91.74 71.32 76.07
4
Best 83.76 87.18 85.76 88.32 83.88 85.47 – – 72.31 75.5 69.83 72.08 83.25 84.9 91.77 92.59 74.74 76.64
Last 47.52 56.41 50.14 56.13 45.01 50.14 – – 42.85 46.44 42.48 45.87 52.25 54.42 72.62 79.49 53.72 56.41
5
Best 52.48 58.69 54.59 58.4 48.2 53.56 – – 48.66 49.57 48.35 52.14 56.52 58.69 80.69 82.91 59.42 60.68
Last 50.4 52.1 16.67 16.67 51.09 52.84 51.87 53.53 81.02 83.67 70.67 72.24 81.64 83.45 86.97 88.74 77.93 78.67
1
Best 53.31 54.1 16.67 16.67 53.29 54.21 55.49 55.75 82.31 83.67 72.41 73.15 82.94 84.41 89.94 90.39 78.43 80.03
Last 95.58 96.87 94.29 96.42 95.65 96.19 88.35 94.77 92.08 94.25 95.38 96.42 95.31 96.08 96.81 97.38 94.73 95.73
2
Best 96.96 97.44 97.11 97.27 96.91 97.38 97.18 97.61 93.13 94.54 95.86 96.42 95.43 96.08 97.51 97.67 94.98 95.73
B Last 33.73 36.76 35.28 36.92 35.46 37.98 – – 46.72 48.31 47.58 48.5 50.17 51.22 51.6 52.66 41.3 42.34
3
Best 38.16 39.76 39.05 39.98 39.29 40.48 – – 47.7 49.67 50.51 51.55 50.82 51.44 56.01 56.38 41.75 43.23
Last 84.1 90.03 81.82 86.32 79.03 81.2 – – 72.08 72.65 68.26 70.94 76.18 78.06 85.13 90.31 69.29 70.66
4
Best 85.47 90.6 83.76 87.75 80.23 82.34 – – 72.93 73.79 70.09 72.36 77.55 78.92 91.51 92.02 69.92 72.08
Last 54.64 60.11 52.76 59.83 43.82 45.58 – – 44.3 44.44 46.38 48.43 54.87 61.25 78.52 80.34 54.13 56.98
5
Best 56.58 61.82 54.7 61.54 45.24 49 – – 50.14 52.71 51.28 52.99 56.47 61.25 79.54 81.48 55.84 57.55
Last 55.22 56.43 45.39 47.72 55.09 55.86 53.66 54.72 87.29 88 77.53 78.84 86.65 88.79 92.67 93.57 87.55 87.88
1
Best 57.18 57.74 47.48 48.41 56.88 57.74 56.22 56.71 88.33 89.19 79.92 82.82 88.77 89.19 94.43 94.88 88.71 89.25
Last 96.19 96.53 96.24 96.81 95.79 96.25 96.03 96.47 93.72 94.71 95.14 95.51 95.45 96.08 96.1 97.16 95.26 95.79
2
Best 96.8 96.99 97.34 97.38 96.9 97.04 97.19 97.33 94.34 95.05 95.53 95.96 96.1 96.53 97.53 97.9 95.92 96.13
C Last 35.82 37.03 34.59 36.92 33.34 38.03 – – 47.91 49.42 46.34 47.84 42.51 50.08 49.11 52.36 42.67 42.98
3
Best 38.95 40.26 37.91 39.64 38.52 41.31 – – 52.77 53.63 49.21 49.97 44.43 53.33 54.55 55.22 46.91 47.14
Last 79.48 85.75 81.58 82.34 84.14 84.9 – – 71.62 73.5 69.57 72.36 84.05 86.04 91.4 93.73 73.96 74.93
4
Best 83.26 86.89 83.19 83.48 85.57 86.61 – – 73.16 74.93 71.62 74.07 86.27 88.32 93.62 94.87 77.09 78.92
Last 62.89 65.24 60.87 65.24 64.96 66.95 – – 54.87 58.97 50.88 54.7 69.86 72.36 85.93 88.03 62.16 65.81
5
Best 64.6 67.81 63.44 66.67 66.86 69.23 – – 57.95 59.54 55.67 58.4 73.05 74.36 86.72 88.32 65.53 66.95
Please cite this article as: Z. Zhao, T. Li, J. Wu et al., Deep learning algorithms for rotating machinery intelligent diagnosis: An open source benchmark study. ISA Transactions
(2020), https://doi.org/10.1016/j.isatra.2020.08.010.
Z. Zhao, T. Li, J. Wu et al. / ISA Transactions xxx (xxxx) xxx 21
Table A.6
JNU: Results with order split and data augmentation.
Nor Input Loc AE DAE SAE MLP CNN LeNet AlexNet ResNet18 LSTM
Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max
Last 42.16 43.88 42.67 45.18 44.72 48.64 46.76 49.77 79.91 83.62 72.99 77.55 81.54 82.43 90.84 93.59 82.22 85.09
1
Best 52.07 52.27 49.67 51.3 51.63 54.25 51.68 54.25 85.33 86.05 77.61 78.46 81.54 82.43 93.11 93.59 84.2 85.71
Last 94.92 95.46 94.66 95.92 48.11 95.29 95.95 96.2 93.84 94.56 94.47 94.73 96.1 96.54 95.83 96.2 95.28 95.63
2
Best 96.03 96.09 95.16 96.26 48.37 95.98 96.8 97.05 94.81 95.07 95.42 95.58 96.1 96.54 97.1 97.34 95.99 96.2
A Last 32.66 35.66 29.53 30.84 32.6 34.39 – – 42.73 47.81 46.23 48.56 36.18 43.47 52.2 53.65 42.51 44.96
3
Best 36.93 38.82 36.05 37.94 37.83 38.44 – – 48.73 49.42 49.84 50.75 37.87 43.47 54.94 55.68 47.75 49.03
Last 75.28 79.55 76.19 77.31 75.41 76.75 – – 68.57 72.55 69.25 70.31 79.41 82.35 81.57 89.08 74.37 75.35
4
Best 80.11 84.31 80.56 82.35 79.77 82.07 – – 71.82 75.07 71.54 72.27 82.84 84.59 90.31 90.76 76.89 77.87
Last 53.36 55.18 49.36 57.98 42.52 50.98 – – 38.38 45.38 45.72 49.3 54.23 56.02 63.7 80.11 50.59 56.02
5
Best 57.42 61.62 55.4 60.5 48.96 52.38 – – 49.41 51.26 50.7 51.54 56.75 58.82 80.78 84.31 57.42 58.26
Last 60.38 61.62 37.86 54.37 59.24 60.94 62.01 63.15 83.89 86.34 77.22 79.59 86.24 88.1 92.8 93.65 75.64 83.73
1
Best 61.32 61.85 39.07 55.27 60.48 61.39 62.01 63.15 85.63 86.85 77.22 79.59 86.95 88.76 93.11 93.65 76.66 84.86
Last 95.5 96.09 95.12 95.86 95.09 95.92 94.43 96.15 93.32 93.82 93.59 94.27 95.77 96.43 95.63 96.71 94.55 94.9
2
Best 96.75 97.05 96.37 96.6 96.45 96.54 94.8 96.15 93.58 94.33 93.83 95.12 95.99 96.77 96.18 96.71 95.62 95.75
B Last 32.75 35.16 32.6 34.39 32.78 34.33 – – 47.91 48.64 47.3 49 51.27 53.88 52.05 53.38 40.85 41.82
3
Best 38.26 40.27 38.25 39.66 38.47 38.99 – – 48.29 49.22 48.04 49 51.42 53.88 52.98 53.49 45.28 46.15
Last 72.1 81.51 75.24 77.87 73.78 77.59 – – 68.13 69.75 67.45 71.15 74.23 75.91 85.49 86.27 69.75 71.71
4
Best 74.96 85.99 77.93 81.23 77.81 82.07 – – 68.6 70.03 67.45 71.15 74.96 78.15 87.06 89.64 73.45 76.47
Last 50.76 52.94 51.99 54.62 43.64 50.42 – – 45.27 47.06 47.23 50.14 55.24 56.86 76.41 81.23 54.34 56.86
5
Best 57.48 58.82 56.52 57.98 47 54.06 – – 45.6 47.62 47.96 50.98 56.59 60.22 78.99 81.23 58.88 59.94
Last 65.76 67.35 60.07 62.81 65.53 67.74 66.79 68.54 87.94 90.99 85.48 86.28 90.7 91.33 94.06 95.18 87.52 88.21
1
Best 66.39 67.35 61.04 63.15 66.59 67.74 67.35 68.54 88.96 90.99 85.61 86.28 91.32 92.12 94.27 95.18 87.78 88.55
Last 95.74 96.6 96.01 96.77 96.27 96.54 96.12 96.54 93.59 94.22 94.72 95.58 95.83 97.05 95.34 96.43 94.62 95.35
2
Best 96.71 96.88 96.87 97.22 96.69 96.88 96.34 96.54 93.81 95.01 94.72 95.58 96.08 97.05 95.52 96.43 94.84 95.35
C Last 33.84 36 32.95 34 33.25 35.39 – – 48.94 50.28 49.1 50.55 44.06 51.88 51.17 55.84 43.03 44.63
3
Best 38.43 39.93 37.11 38.44 37.42 39.93 – – 49.22 50.28 49.29 50.55 44.38 52.77 52.65 55.84 43.42 44.63
Last 78.43 82.07 75.35 77.59 79.66 81.23 – – 70.12 71.71 69.28 72.55 80.67 82.07 87 90.76 72.66 75.35
4
Best 81.96 85.43 77.65 80.39 83.92 84.31 – – 70.73 71.71 69.94 72.55 81.4 84.03 90.26 90.76 73.73 78.43
Last 64.76 69.19 60.05 63.87 62.13 64.15 – – 55.97 58.54 53.08 55.18 67.85 70.03 66.38 87.39 61.4 63.59
5
Best 69.02 72.83 67.34 72.83 65.72 71.43 – – 56.49 58.54 53.38 55.18 68.01 70.03 66.72 87.39 62.19 63.59
Table A.7
MFPT: Results with random split and data augmentation.
Nor Input Loc AE DAE SAE MLP CNN LeNet AlexNet ResNet18 LSTM
Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max
Last 35.84 39.22 35.07 40.97 35.92 38.45 30.63 36.12 86.72 87.57 65.9 67.77 77.16 79.61 93.86 96.5 76.82 78.64
1
Best 41.83 45.83 41.09 42.72 41.59 42.91 40.29 41.36 89.36 91.07 69.63 71.46 78.6 80.97 95.92 96.89 78.52 79.61
Last 94.21 94.76 94.76 95.92 94.6 95.92 94.84 95.15 85.63 88.54 91.81 93.4 92.04 92.43 91.8 93.4 91.88 92.82
2
Best 95.53 95.92 95.84 96.12 95.61 95.92 96.23 96.5 88.39 89.51 92.97 93.4 92.97 93.59 94.6 95.34 93.09 93.59
A Last 23.8 26.62 23.91 27.76 26.62 29.66 – – 37.98 40.3 38.63 41.83 41.25 43.92 46.96 47.91 40 41.44
3
Best 28.03 28.9 27.68 29.85 30.34 33.84 – – 42.36 43.35 42.32 43.73 43.58 45.25 48.86 49.81 43.39 43.92
Last 88.47 90.68 87.23 88.74 88.35 90.68 – – 70.21 75.15 77.09 78.45 88.23 90.29 91.26 95.15 80.97 83.3
4
Best 90.52 92.04 90.56 92.62 90.45 92.04 – – 75.03 76.12 79.3 80.78 89.55 90.87 95.34 95.92 84.54 84.85
Last 46.83 52.23 48.86 51.26 46.25 49.71 – – 50.22 52.43 50.29 53.2 55.54 63.69 85.9 91.26 54.18 58.06
5
Best 51.18 52.23 53.75 55.73 50.02 54.76 – – 54.33 56.12 53.51 55.34 62.06 66.21 90.25 91.26 57.9 61.36
Last 47.03 49.71 46.56 48.93 47.53 48.54 49.79 51.26 87.26 89.13 67.3 68.54 84.97 87.57 92.97 95.73 77.09 79.03
1
Best 50.21 52.04 49.09 50.1 50.41 51.46 50.18 51.46 87.62 90.12 68.54 69.9 85.67 87.96 93.94 96.7 78.87 79.42
Last 93.59 96.31 93.44 95.92 92.15 95.15 94.14 94.95 85.59 86.8 91.19 92.62 91.49 92.23 92.08 93.4 92.47 93.2
2
Best 95.73 96.31 96.19 96.89 95.84 96.12 94.52 94.95 86.02 87.38 91.58 92.63 92.08 92.82 92.47 93.98 93.79 94.37
B Last 22.7 25.67 23.8 27.19 26.96 29.47 – – 39.39 41.06 38.59 41.63 42.09 44.11 46.92 48.86 39.58 40.87
3
Best 27.23 28.52 27.87 29.85 30.38 31.56 – – 39.54 41.63 39.65 41.63 42.17 44.11 47.87 48.86 43.12 43.92
Last 87.42 88.93 86.06 88.93 87.96 90.29 – – 72.04 73.59 72.04 76.7 86.79 87.96 91.57 95.53 79.46 80.97
4
Best 89.17 90.49 88.31 90.1 89.83 91.84 – – 73.01 74.56 73.94 77.86 87.1 87.96 95.07 95.53 82.21 83.11
Last 48.97 53.79 50.91 55.73 42.83 47.18 – – 50.8 53.59 51.41 54.17 62.18 65.05 89.01 90.68 57.17 58.64
5
Best 53.75 56.89 52.97 58.06 47.22 53.4 – – 51.26 54.56 52.07 54.17 62.56 65.44 89.32 90.68 60.08 61.17
Last 49.71 51.07 49.36 52.62 48.47 51.84 47.69 51.07 88.85 90.29 69.98 72.43 91.38 92.43 93.86 97.67 82.21 83.5
1
Best 50.72 52.62 50.52 53.2 49.9 52.82 49.47 51.07 90.76 91.84 70.14 73.2 92.54 93.4 96.04 97.67 84.55 87.38
Last 94.56 95.53 95.03 95.53 95.49 96.5 95.15 96.12 86.41 88.35 91.03 92.43 91.34 92.23 90.87 91.65 91.61 93.01
2
Best 96.16 96.5 96.58 96.89 96.62 97.09 95.88 97.28 87.85 88.93 91.77 92.62 92.54 93.4 92.2 93.01 92.54 93.4
C
(continued on next page)
Please cite this article as: Z. Zhao, T. Li, J. Wu et al., Deep learning algorithms for rotating machinery intelligent diagnosis: An open source benchmark study. ISA Transactions
(2020), https://doi.org/10.1016/j.isatra.2020.08.010.
22 Z. Zhao, T. Li, J. Wu et al. / ISA Transactions xxx (xxxx) xxx
Table A.8
MFPT: Results with random split and without data augmentation.
Nor Input Loc AE DAE SAE MLP CNN LeNet AlexNet ResNet18 LSTM
Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max
Last 20.93 27.38 17.59 22.33 22.91 26.02 22.49 25.05 83.22 86.8 62.25 62.91 76.08 78.64 78.18 85.44 69.48 71.07
1
Best 28.54 32.04 26.52 28.35 27.81 30.29 27.77 28.74 87.07 87.96 65.24 67.18 78.37 82.33 86.88 87.96 72.58 75.15
Last 94.95 95.92 94.14 95.15 94.76 95.34 94.33 95.53 79.96 84.27 93.75 94.76 92.04 93.4 92.27 95.92 93.09 93.79
2
Best 96.23 96.5 96 96.7 96.27 96.89 95.96 96.5 82.68 84.66 93.86 94.95 93.36 94.17 95.85 96.31 94.06 94.76
A Last 26.12 29.09 23.08 24.52 25.59 28.14 – – 38.1 40.11 40.76 41.44 29.46 41.06 44.22 48.67 40.49 43.16
3
Best 28.14 29.66 27.49 28.71 29.17 30.42 – – 42.02 43.73 43.16 44.11 30.53 43.35 49.05 50.38 43.35 46.01
Last 89.01 90.49 89.79 91.46 88.74 90.1 – – 73.12 76.31 78.1 79.03 88.39 90.1 94.87 95.34 80.43 83.69
4
Best 89.98 91.26 91.26 92.23 89.9 91.26 – – 74.99 76.89 79.34 80.97 90.41 91.65 95.42 95.73 83.5 86.02
Last 47.49 51.07 48.23 53.01 46.6 51.26 – – 50.84 53.98 50.87 54.76 52.2 65.44 88.31 89.9 53.98 58.25
5
Best 49.98 54.76 51.42 57.09 48.82 55.53 – – 53.67 57.09 54.68 55.92 53.32 65.44 90.37 91.07 59.11 61.36
Last 25.32 27.96 22.33 23.3 25.52 28.16 25.4 27.18 85.2 86.99 63.57 66.21 80.62 82.91 86.99 87.96 65.94 72.04
1
Best 28.58 30.1 26.17 27.18 28.89 30.87 25.4 27.18 85.2 86.99 64.04 67.57 80.93 82.91 87.11 87.96 66.18 72.04
Last 92.93 95.34 92.19 94.95 93.67 95.92 91.96 94.76 80.04 80.97 93.09 93.4 90.99 92.23 94.68 95.92 93.2 94.56
2
Best 95.61 95.73 95.42 95.53 96.12 96.7 95.54 96.12 81.13 82.91 94.1 94.37 93.09 93.98 96.23 96.7 94.06 94.56
B Last 26.47 29.85 25.59 26.43 27.07 30.8 – – 39.24 40.87 38.86 40.11 39.92 42.59 44.22 46.58 39.54 40.11
3
Best 28.94 30.42 29.13 30.04 30 34.22 – – 39.92 42.59 39.58 41.44 41.1 42.59 44.83 47.53 40.44 41.63
Last 87.88 89.51 86.83 89.9 88.39 90.29 – – 71.84 76.5 77.59 79.22 86.56 87.77 94.02 95.34 80.58 82.52
4
Best 89.01 90.49 89.63 92.04 89.79 90.87 – – 75.15 77.09 78.95 79.81 88.85 89.71 95.69 96.31 82.41 83.11
Last 50.06 57.67 50.41 57.86 45.94 50.1 – – 50.14 52.04 51.11 54.76 56.66 61.36 78.52 90.49 56.89 59.61
5
Best 53.94 59.03 55.42 60.58 48.58 51.84 – – 51.65 53.4 51.92 54.76 57.82 64.27 83.57 90.49 57.57 60.39
Last 24.19 26.6 23.26 24.85 22.95 24.08 21.79 22.72 86.21 87.57 61.36 64.08 84.19 85.24 86.6 91.84 76.27 77.48
1
Best 26.68 28.35 25.98 26.99 26.6 27.96 24.47 26.41 86.8 88.56 61.75 64.08 84.82 86.8 88.74 91.84 77.09 79.22
Last 94.99 95.34 94.49 95.92 94.37 95.15 94.37 94.76 80.39 82.33 91.84 92.62 90.72 92.82 94.56 95.34 92.15 93.01
2
Best 96.35 96.89 96.58 97.48 96.12 96.89 96.04 96.31 81.51 82.52 92.43 93.4 92.82 93.59 95.49 95.92 93.9 94.37
C Last 22.17 25.67 25.89 28.14 27.76 31.18 – – 40.26 42.21 40.19 41.06 39.73 41.63 40.84 49.05 39.92 40.68
3
Best 28.1 32.13 28.25 29.28 30.8 32.7 – – 40.87 43.35 40.34 41.06 40.45 42.97 41.37 49.81 41.02 42.78
Last 89.28 89.71 88.35 89.9 88.97 91.07 – – 83.16 91.26 86.45 87.57 88.04 89.71 89.79 91.46 87.34 89.71
4
Best 89.71 90.29 89.05 90.1 90.95 92.43 – – 84.33 91.46 86.6 87.57 91.22 92.43 90.14 91.46 88.43 90.29
Last 54.1 57.86 53.55 60.58 50.6 53.59 – – 58.52 62.52 54.21 56.12 63.07 66.41 91.84 92.62 60.93 62.14
5
Best 57.48 60.97 56.31 62.33 53.59 56.7 – – 59.34 63.69 55.46 56.31 64 69.51 92.39 92.82 61.83 63.11
Table A.9
MFPT: Results with order split and data augmentation.
Nor Input Loc AE DAE SAE MLP CNN LeNet AlexNet ResNet18 LSTM
Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max
Last 35.28 40.69 36.39 41.27 35.43 41.07 31.09 37.62 86.03 90.02 69.64 73.9 80.27 83.3 92.32 93.86 79.08 81.57
1
Best 42.65 44.72 43.8 45.68 44.3 46.64 43.72 47.22 89.83 90.4 72.78 74.66 81.42 85.03 94.74 95.59 81.84 83.3
Last 93.59 95.2 95.47 96.55 95.01 95.39 95.55 96.16 87.03 88.1 91.63 93.09 92.55 93.09 92.4 94.63 91.94 93.09
2
Best 96.08 96.35 96.43 97.31 96.12 96.74 96.51 96.74 88.98 89.44 93.17 94.05 93.55 94.05 94.63 94.82 93.66 95.01
A Last 26.07 26.82 26.52 28.12 28.34 33.33 – – 39.89 41.53 39.96 41.34 34.45 43.76 44.95 45.81 39.63 42.27
3
Best 29.31 29.98 29.79 31.28 30.43 33.33 – – 43.35 44.69 44.1 45.62 35.42 46.18 48.46 49.16 42.42 44.13
Last 89.25 91.94 90.06 91.55 88.79 90.6 – – 77.7 80.81 78.31 81.77 90.75 91.75 90.6 96.55 80.69 81.96
4
Best 91.71 92.51 92.28 93.86 91.52 92.71 – – 78.77 81.96 80.69 82.92 92.29 92.71 97.2 97.7 85.14 86.37
Last 52.55 57.2 51.32 53.74 51.28 55.66 – – 50.86 55.47 53.24 55.47 65.87 69.1 74.82 89.44 60.61 63.53
5
Best 55.47 58.16 55.01 57.58 55.36 58.35 – – 57.16 57.77 56.89 57.97 68.18 73.51 92.28 92.9 63.45 66.6
(continued on next page)
Please cite this article as: Z. Zhao, T. Li, J. Wu et al., Deep learning algorithms for rotating machinery intelligent diagnosis: An open source benchmark study. ISA Transactions
(2020), https://doi.org/10.1016/j.isatra.2020.08.010.
Z. Zhao, T. Li, J. Wu et al. / ISA Transactions xxx (xxxx) xxx 23
Table A.10
PU: Results with random split and data augmentation.
Nor Input Loc AE DAE SAE MLP CNN LeNet AlexNet ResNet18 LSTM
Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max
Last 47.22 53.15 44.91 53 48.79 53.3 46.18 49.16 79.02 81.87 68.76 72.66 82.46 84.33 89.74 92.17 75.02 76.04
1
Best 55.08 59.14 55.21 58.06 52.69 54.53 51.68 57.16 80.68 85.87 71.43 75.12 83.1 84.33 90.85 94.01 75.6 76.34
Last 97.45 97.85 97.27 97.85 97.64 98.31 98.31 99.08 95.18 96.16 94.19 96.16 96.53 98.31 97.14 97.54 94.99 95.39
2
Best 98.34 98.46 98.43 98.77 98.52 98.77 98.53 99.23 95.58 96.62 94.68 97.54 96.84 98.31 97.66 98.46 95.11 95.55
A Last 30.08 33.58 29.74 31.63 31.42 32.23 – – 36.37 39.88 33.52 38.08 33.52 35.23 41.77 47.68 37.39 38.68
3
Best 33.34 36.58 33.67 34.63 34.36 36.13 – – 37.24 40.63 34.27 38.83 34.69 36.73 43.93 48.28 37.9 39.28
Last 94.22 95.7 95.51 96.31 93.98 96.31 – – 82.52 85.71 82.3 84.18 94.26 95.55 93.03 97.7 84.55 86.94
4
Best 96.59 96.93 96.96 97.39 95.79 97.08 – – 83.01 87.86 82.8 84.49 95.08 96.31 93.68 98.82 86.7 88.48
Last 56.99 68.36 63.96 68.05 59.36 62.67 – – 45.71 53.46 44.09 49.16 7.83 7.83 91.37 94.47 48.63 52.38
5
Best 66.76 73.73 68.26 72.35 65.19 70.51 – – 48.83 56.61 47.59 51.31 7.83 7.83 94.1 95.55 53.67 59.29
Last 65.95 66.97 62.13 64.06 65.79 67.74 70.72 72.2 83.44 86.18 72.26 72.96 86.61 88.94 91.58 93.24 75.88 77.11
1
Best 67.61 68.97 63.59 65.9 66.62 68.82 71.24 72.66 84.55 86.33 73.18 74.5 87.5 90.02 92.32 94.01 78.49 79.42
Last 97.06 98.31 96.26 97.24 96.49 98 97.69 98.46 93.7 95.55 95.02 96.31 97.05 97.54 90.91 98 94.66 95.39
2
Best 98.28 98.46 98.08 98.46 97.83 98.16 97.97 98.46 95.15 95.7 95.42 96.62 97.24 97.54 97.76 98.16 96.31 96.47
B Last 29.41 32.53 31.86 35.83 33.91 37.03 – – 37.6 40.63 38.86 40.33 33.4 35.98 45.04 47.98 39.76 41.68
3
Best 32.65 34.93 34.2 38.53 37.33 38.38 – – 39.52 43.18 39.46 41.53 34.06 37.63 45.43 47.98 42.82 44.53
Last 91.94 94.32 92.06 93.39 93.42 95.08 – – 77.17 85.71 76.19 79.42 91.06 92.78 95.14 99.23 82 84.64
4
Best 93.78 96.47 94.42 96.31 95.03 96.01 – – 77.64 85.71 78.96 81.26 91.83 93.55 99.02 99.39 87.1 89.25
Last 60.01 66.67 56.91 66.82 62.88 68.82 – – 33.79 55.61 50.38 53.92 49.09 61.6 90.23 95.85 57.33 59.75
5
Best 66.36 72.04 63.59 72.04 65.97 70.05 – – 34.87 57.3 51.58 54.85 49.09 61.6 94.87 95.85 60.09 60.83
Last 69.83 71.58 64.98 68.05 70.72 71.43 71.09 72.96 85.96 88.63 78.92 80.03 90.57 92.17 93.92 95.24 80.31 82.03
1
Best 70.97 74.04 66.73 69.59 71.31 72.04 72.19 72.96 86.79 88.63 78.99 80.18 90.66 92.17 94.44 96.01 81.26 83.26
Last 97.45 98 97.7 98.62 97.42 97.54 97.94 98.92 95.3 96.31 95.33 95.85 97.3 98.16 97.17 98 95.02 95.55
2
Best 98.25 98.62 98.5 98.62 98.34 98.62 98.22 98.92 95.55 96.47 95.85 96.93 97.73 98.16 97.64 98.77 95.36 96.16
C Last 25.52 33.58 29.95 32.23 33.01 34.63 – – 39.1 41.68 39.61 41.53 37 38.98 41.86 47.08 40.21 41.68
3
Best 26.84 33.73 32.77 34.03 36.04 38.53 – – 40.45 41.68 39.82 42.58 38.02 41.23 45.91 47.08 41.29 41.83
Last 94.96 96.01 93.64 95.39 94.69 97.54 – – 82.34 85.41 79.85 82.03 91.95 93.55 92.87 98.16 85.1 86.64
4
Best 96.07 96.47 95.02 96.31 96.44 97.54 – – 83.41 86.64 81.04 82.64 92.47 94.62 93.18 98.92 86.33 86.94
Last 59.11 65.75 56.62 66.36 62.15 69.59 – – 38.55 64.36 53.09 55.61 28.45 62.37 91.55 96.62 59.51 60.83
5
Best 64.27 66.05 61.26 70.51 67.77 72.96 – – 39.11 65.13 53.58 55.61 29.31 62.37 95.85 96.77 59.85 61.44
Please cite this article as: Z. Zhao, T. Li, J. Wu et al., Deep learning algorithms for rotating machinery intelligent diagnosis: An open source benchmark study. ISA Transactions
(2020), https://doi.org/10.1016/j.isatra.2020.08.010.
24 Z. Zhao, T. Li, J. Wu et al. / ISA Transactions xxx (xxxx) xxx
Table A.11
PU: Results with random split and without data augmentation.
Nor Input Loc AE DAE SAE MLP CNN LeNet AlexNet ResNet18 LSTM
Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max
Last 37.7 44.19 29.59 34.33 38.31 43.13 50.39 56.84 78.96 84.95 71.49 74.96 82.74 84.49 81.85 85.41 73.24 77.11
1
Best 46.85 48.5 39.36 40.89 45.63 47.25 55.95 60.22 79.97 84.95 73.92 75.27 83.23 85.87 85.29 95.24 74.62 77.27
Last 74.62 93.38 91.91 93.76 92.66 93.2 97.91 98.46 90.57 91.86 95.85 97.08 95.18 96.77 98.77 99.85 93.86 95.55
2
Best 75.48 93.88 92.86 93.95 93.5 94.01 98.43 98.92 90.72 91.86 96.13 97.54 95.57 97.85 99.48 99.85 94.29 96.01
A Last 31.03 33.58 31.06 33.28 32.86 37.03 – – 46.3 49.36 47.1 48.59 43.36 54.62 47.94 55.64 46.65 48.56
3
Best 34.12 37.18 33.58 35.83 35.2 38.23 – – 48.04 51.26 48.26 49.37 43.5 55.07 53.56 55.8 47.34 48.56
Last 89.31 90.01 89.83 90.89 89.45 90.57 – – 77.63 86.94 81.38 82.03 92.35 95.7 95.48 98.46 78.06 86.64
4
Best 90.72 91.14 90.94 91.64 89.98 91.2 – – 85.74 87.71 81.81 82.49 94.01 95.7 96.99 98.62 84.73 88.17
Last 56.09 59.74 56.02 59.3 53.06 60.61 – – 51.21 54.22 49.65 54.84 7.83 7.83 89.59 94.47 55.36 58.37
5
Best 59.51 61.67 58.5 61.99 58.54 67.23 – – 55.21 55.91 55.24 57.76 7.83 7.83 94.99 95.39 58.37 59.91
Last 55.01 55.68 31.37 38.14 53.08 55.43 70.2 72.81 84.06 85.56 72.04 73.73 87.04 87.71 89.68 93.86 72.41 76.65
1
Best 55.98 56.55 31.85 39.26 54.79 56.18 71.27 72.81 84.67 86.48 73.49 75.73 87.65 88.79 90.87 95.39 75.67 78.34
Last 90.79 92.7 89.63 93.26 92.52 93.95 80.92 97.85 88.63 92.17 96.44 97.39 95.73 97.24 97.94 99.54 93.67 95.39
2
Best 94.06 94.32 93.84 94.19 94.16 94.51 95.15 98.92 88.83 92.17 96.59 97.39 96.41 97.24 99.32 99.69 94.87 95.7
B Last 32.47 34.78 32.71 38.68 35.35 37.93 – – 50.56 52.14 50.27 51.56 53.39 53.87 53.64 55.46 48.34 48.76
3
Best 34.54 36.13 34.75 40.18 36.7 39.13 – – 51.13 52.29 50.77 51.56 53.76 55.07 54.32 56.22 48.81 49.24
Last 87.88 90.32 87.45 87.83 86.22 90.14 – – 83.76 88.33 76.01 81.57 91.92 92.93 81.97 98.62 81.87 86.48
4
Best 88.71 90.32 88.4 89.58 88.26 90.45 – – 84.16 88.63 77.6 83.41 92.11 93.55 98.22 99.08 82.43 86.48
Last 57.7 60.61 58.94 61.17 58.69 60.86 – – 51.36 51.46 51.77 52.69 51.77 68.2 94.77 95.08 55.54 57.45
5
Best 60.43 61.92 62.66 67.79 62.26 66.42 – – 53.97 54.38 55.49 56.84 54.47 72.04 95.11 95.85 59.69 61.44
Last 59.39 60.36 51.25 54.18 57.15 58.49 71.77 72.81 83.41 87.1 77.94 80.18 91.43 92.78 91.92 95.55 80.18 81.57
1
Best 60.18 62.3 52 54.43 57.81 58.93 72.53 73.12 85.53 87.1 79.02 80.18 91.98 92.78 92.75 97.08 80.83 82.64
Last 93.12 93.95 92.8 93.51 93.17 94.38 97.85 98.46 89.43 92.63 95.61 96.77 97.3 98.92 99.63 99.69 94.07 94.78
2
Best 94.07 94.26 94.24 94.57 94.13 94.38 98.52 98.77 89.43 92.63 95.79 96.77 97.82 98.92 99.63 99.69 94.72 96.01
C Last 28.04 32.23 24.71 29.39 34.45 37.03 – – 51.58 52.05 50.88 51.84 54.5 55.91 55.21 56.36 48.96 49.6
3
Best 31.12 33.88 29 33.43 37.15 40.93 – – 51.99 52.91 51.36 52.33 54.86 55.91 55.4 57.24 49.4 49.93
Last 90.04 90.89 89.96 91.51 83.43 90.82 – – 82.8 86.33 79.17 82.03 92.17 95.39 97.79 98.62 85.38 86.94
4
Best 90.4 91.07 90.26 91.51 84.14 91.07 – – 83.07 86.64 79.45 82.03 93.24 95.39 97.94 98.62 86.54 87.71
Last 56.34 59.68 47.64 62.48 58.04 62.98 – – 45.16 57.3 51.43 53.76 7.83 7.83 89.28 96.62 59.63 61.14
5
Best 60.26 65.04 49.74 63.17 60.12 64.04 – – 50.31 59.45 56.84 58.68 8.35 10.45 96.44 96.93 62.18 63.13
Table A.12
PU: Results with order split and data augmentation.
Nor Input Loc AE DAE SAE MLP CNN LeNet AlexNet ResNet18 LSTM
Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max
Last 48.11 53.3 49.86 55.45 48.33 52.69 51.06 55.3 80.83 82.8 64.24 71.43 83.9 84.95 85.07 93.7 71.64 73.89
1
Best 54.75 56.68 54.75 57.14 55.08 57.6 54.41 62.37 82.22 85.6 67.96 71.43 84.73 86.79 90.38 93.7 72.01 74.96
Last 97.63 98.46 97.63 98.92 97.67 98.46 89.54 92.31 66.15 72.31 77.54 84.62 65.54 75.38 88 92.31 76.92 80
2
Best 99.14 99.23 98.89 98.92 98.77 98.92 95.69 98.46 69.23 72.31 82.77 87.69 69.23 75.38 91.08 93.85 82.77 86.15
A Last 29.96 32.2 28.42 31.61 29.69 31.61 – – 32.59 37.52 32.94 36.48 25.61 31.02 41.54 44.46 34.5 35.89
3
Best 32.79 35.01 32.61 34.56 33.38 35.3 – – 35.3 38.4 34.53 37.37 25.61 31.02 42.63 46.23 35.63 37.52
Last 94.44 95.55 93.7 94.62 94.32 95.24 – – 78.74 83.56 77.42 78.65 93.24 95.24 95.45 98.92 83.96 88.33
4
Best 96.9 97.39 96.37 97.08 95.82 96.62 – – 80.18 83.56 78.83 81.72 93.52 95.55 97.97 98.92 87.07 88.33
Last 61.2 66.97 57.7 62.98 58 65.59 – – 42.58 49.31 45.16 51.61 7.83 7.83 85.9 94.47 45.53 51.15
5
Best 64.52 68.51 62.43 69.89 62.03 67.9 – – 45.68 50.84 48.82 51.61 7.83 7.83 87.19 95.39 50.51 51.92
Last 65.5 67.9 63.1 65.28 66.82 69.59 69.65 71.58 79.66 84.64 70.2 72.04 86.95 88.33 91.58 93.39 74.38 75.88
1
Best 67.34 69.12 64.33 67.13 68.36 69.89 70.08 71.89 80.74 85.25 71.15 74.5 87.62 88.33 92.44 94.47 76.4 77.11
Last 97.2 98.46 98.19 98.46 96.59 98 75.69 89.23 58.15 67.69 79.08 84.62 66.77 70.77 72.31 80 78.15 83.08
2
Best 98.95 99.23 99.05 99.23 98.59 98.92 92.31 93.85 68.31 73.85 82.15 86.15 68.62 72.31 88.92 90.77 82.16 84.62
B Last 31.82 33.68 29.87 32.35 31.28 31.76 – – 37.52 39 35.66 36.93 31.34 34.86 42.51 45.2 36.1 38.85
3
Best 34.65 36.48 33.03 35.45 34.5 35.45 – – 37.84 39.44 36.87 38.7 32.26 34.86 43.4 45.2 39.44 41.36
Last 92.1 94.62 91.58 92.78 92.23 93.24 – – 80.15 84.02 72.72 81.72 91.52 92.47 90.38 98 83.84 86.18
4
Best 94.13 97.24 94.56 95.85 94.22 95.39 – – 81.66 84.02 74.5 82.95 91.74 92.63 97.02 98.31 86.54 87.71
Last 61.32 66.21 60.46 63.13 56.93 66.67 – – 48.54 52.23 48.05 50.54 30.87 67.74 88.88 95.55 50.63 53.46
5
Best 64.88 67.74 64.64 66.36 67.96 71.43 – – 48.6 52.38 48.76 50.54 30.87 67.74 94.84 95.55 54.5 57.6
(continued on next page)
Please cite this article as: Z. Zhao, T. Li, J. Wu et al., Deep learning algorithms for rotating machinery intelligent diagnosis: An open source benchmark study. ISA Transactions
(2020), https://doi.org/10.1016/j.isatra.2020.08.010.
Z. Zhao, T. Li, J. Wu et al. / ISA Transactions xxx (xxxx) xxx 25
Table A.13
SEU: Results with random split and data augmentation.
Nor Input Loc AE DAE SAE MLP CNN LeNet AlexNet ResNet18 LSTM
Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max
Last 50.25 53.19 45.1 52.7 46.91 52.7 60.64 65.69 88.24 94.36 83.19 86.27 53.43 75.74 95.73 99.75 87.38 91.67
1
Best 61.72 66.18 62.55 64.22 65.1 68.14 63.87 66.91 90.98 95.59 85.25 87.75 55.15 78.43 97.15 99.75 87.93 92.65
Last 97.16 97.79 97.45 98.28 97.16 98.04 97.35 97.55 97.3 98.04 96.81 98.53 96.03 96.81 97.25 97.55 96.02 96.81
2
Best 98.14 98.53 98.38 98.77 98.48 98.77 97.79 98.28 97.55 98.53 97.3 98.53 97.4 98.04 97.7 98.53 96.5 97.55
A Last 50.39 54.57 53.94 57.69 55.24 63.46 – – 52.94 59.62 45.34 54.81 64.95 66.59 59.52 65.14 54.69 55.77
3
Best 52.4 57.45 58.99 62.02 60.39 66.35 – – 53.7 59.62 45.48 54.81 66.02 67.55 61.88 66.59 56.07 57.69
Last 97.5 99.02 97.99 99.02 98.09 98.77 – – 84.41 88.73 84.31 89.46 78.04 98.04 96.22 100 88.48 90.44
4
Best 98.77 99.51 99.07 99.51 98.72 99.26 – – 86.81 90.2 85 90.2 79.12 98.77 99.66 100 89.95 91.67
Last 52.89 86.76 66.32 83.58 56.37 79.9 – – 20.24 43.38 34.61 43.14 4.9 4.9 67.11 100 43.38 46.08
5
Best 54.22 88.48 69.27 88.73 61.42 81.13 – – 21.92 50.52 42.7 49.75 4.9 4.9 82.94 100 44.98 47.06
Last 77.55 79.9 79.85 82.84 78.24 80.39 82.6 83.82 92.84 95.83 86.08 89.71 89.26 94.36 99.46 99.75 90.35 91.91
1
Best 79.61 82.6 81.08 83.09 78.97 80.39 83.48 85.78 93.14 96.81 87.79 91.91 90.78 94.36 99.66 100 91.03 93.63
Last 95.59 97.3 95.74 97.06 94.9 97.06 96.47 97.79 97.84 98.28 97.25 97.55 96.96 97.55 96.17 97.79 96.86 97.55
2
Best 97.7 98.04 97.89 98.28 97.84 98.04 96.96 97.79 98.09 98.77 97.44 97.99 97.25 97.79 96.76 98.28 97.3 98.28
B Last 54.85 56.97 54.9 60.58 56.2 61.06 – – 52.74 59.62 50.34 54.09 62.31 65.38 60.67 66.59 56.63 60.1
3
Best 58.89 62.26 57.5 61.78 59.57 64.18 – – 54.47 60.1 51.68 56.01 63.85 66.35 62.65 71.88 57.98 62.5
Last 97.4 98.28 95.93 98.53 97.45 98.04 – – 87.25 89.71 86.32 89.71 96.32 98.53 99.26 100 88.38 91.42
4
Best 98.92 99.51 97.45 98.77 98.87 99.51 – – 88.68 90.69 87.06 91.42 96.86 98.53 99.8 100 89.26 91.42
Last 86.08 88.48 81.23 86.52 73.43 87.75 – – 14.02 50.49 43.48 47.55 59.41 85.05 97.16 99.51 46.77 49.02
5
Best 88.87 92.16 87.3 91.91 78.48 90.44 – – 14.12 50.49 45.34 47.55 61.42 90.44 97.4 100 47.6 50.49
Last 84.31 87.25 83.72 86.03 83.33 90.44 87.79 91.18 95.83 97.3 93.33 94.85 99.02 99.51 100 100 96.86 97.55
1
Best 85.78 87.25 85.15 87.25 85.1 90.44 88.38 91.18 96.27 97.3 93.97 96.32 99.26 100 100 100 97.3 98.28
Last 97.74 98.04 97.79 98.28 97.84 98.53 97.84 98.28 98.82 99.26 97.6 98.04 98.18 99.51 97.84 98.53 97.5 98.28
2
Best 98.58 98.77 99.02 99.02 98.67 98.77 98.15 99.02 99.12 99.51 97.99 99.26 98.67 100 98.09 99.02 97.7 98.53
C Last 51.68 54.09 49.42 53.85 56.4 59.86 – – 57.93 63.94 56.97 59.86 63.12 65.62 71.88 74.76 60.29 63.94
3
Best 54.9 58.65 52.02 55.53 59.71 61.54 – – 61.59 63.94 57.74 61.54 64.09 66.59 72.89 74.76 61.06 63.94
Last 96.96 99.26 98.63 98.77 97.55 99.02 – – 90 92.16 87.4 90.2 96.81 99.02 99.7 100 90.05 91.18
4
Best 98.48 99.75 99.46 99.75 99.26 100 – – 90.34 92.65 88.38 90.2 97.85 99.02 99.8 100 90.98 91.91
Last 89.95 92.16 88.04 89.95 86.47 92.16 – – 31.72 53.92 51.23 56.13 63.04 80.64 99.26 100 52.84 56.13
5
Best 93.19 96.57 91.22 94.36 92.26 94.12 – – 32.5 54.66 52.3 56.13 63.63 81.13 99.41 100 55.07 58.18
Table A.14
SEU: Results with random split and without data augmentation.
Nor Input Loc AE DAE SAE MLP CNN LeNet AlexNet ResNet18 LSTM
Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max
Last 40.13 44.85 31.52 34.56 36.42 42.16 37.45 44.61 88.09 91.42 83.24 87.01 69.93 72.55 98.04 99.02 79.22 81.37
1
Best 44.67 48.53 35.54 37.99 43.78 45.59 42.94 48.28 88.38 92.16 83.58 87.5 72.57 77.77 98.28 99.51 80.78 82.35
Last 96.71 97.55 97.35 98.04 97.11 97.55 97.6 98.28 96.86 99.26 98.09 98.53 97.2 97.55 99.85 100 97.35 97.79
2
Best 98.38 98.53 98.58 98.77 98.43 98.77 97.99 98.28 97.4 99.75 98.43 99.26 97.5 98.28 99.9 100 97.94 98.77
A Last 57.6 63.46 57.45 60.1 54.18 60.34 – – 48.51 56.97 50.53 54.57 62.62 65.87 52.16 62.98 54.28 56.73
3
Best 61.54 65.38 60.53 62.74 60.14 62.5 – – 53.89 57.69 52.39 54.57 63.16 66.35 56.3 72.6 55.91 57.93
Last 98.58 99.02 98.18 98.77 98.09 98.77 – – 85.69 90.69 88.24 90.2 96.42 97.06 99.85 100 88.78 90.2
4
Best 98.97 99.51 99.21 99.75 98.77 99.51 – – 86.52 90.69 89 91.42 97.35 98.77 99.9 100 89.81 90.69
Last 64.12 86.03 69.8 86.03 80.59 89.46 – – 14.31 42.65 33.97 39.46 4.9 4.9 98.72 100 41.42 48.04
5
Best 66.76 87.5 75.05 89.95 83.14 89.71 – – 14.46 42.65 36.52 44.85 5.05 5.15 98.77 100 46.08 50.49
(continued on next page)
Please cite this article as: Z. Zhao, T. Li, J. Wu et al., Deep learning algorithms for rotating machinery intelligent diagnosis: An open source benchmark study. ISA Transactions
(2020), https://doi.org/10.1016/j.isatra.2020.08.010.
26 Z. Zhao, T. Li, J. Wu et al. / ISA Transactions xxx (xxxx) xxx
Table A.15
SEU: Results with order split and data augmentation.
Nor Input Loc AE DAE SAE MLP CNN LeNet AlexNet ResNet18 LSTM
Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max
Last 38.81 45.95 31.81 34.76 39.9 49.52 38.05 46.67 83.24 90.71 67.05 74.76 67.62 73.1 95.53 98.33 77.76 79.76
1
Best 44.48 47.86 48.43 52.14 46.33 49.52 41.24 46.67 84.05 90.71 67.38 74.76 69.48 75.1 96.05 99.05 79.24 81.9
Last 99.24 99.76 99 100 98.71 99.52 99.38 99.76 98.48 99.76 98.24 99.52 98.62 99.05 98.38 99.52 99.48 99.76
2
Best 100 100 100 100 100 100 99.57 99.76 98.86 99.76 98.33 99.52 98.9 99.52 98.81 99.76 99.52 100
A Last 55.57 59.76 54.38 55.71 55.43 64.05 – – 51.48 58.33 50.28 61.9 57.29 60 53 64.76 53.62 55
3
Best 59.47 62.14 60.9 62.14 60.1 64.05 – – 55.1 59.05 55 61.9 59.52 63.57 54.52 68.57 55.72 60.24
Last 88 90.48 87.62 92.38 89.62 95 – – 80.62 87.38 84.05 86.19 97.08 97.86 98.81 99.29 88.62 90.24
4
Best 93.48 96.43 94.48 96.67 94.09 96.67 – – 82.71 88.33 84.95 87.62 97.86 98.57 98.91 99.29 90.62 92.38
Last 47.24 63.33 60.05 73.33 50.29 53.57 – – 13.62 27.62 34.67 41.43 5 5 94.86 98.1 30.62 38.57
5
Best 51.19 65.95 66.33 76.19 55.53 65 – – 13.76 27.62 38.91 41.67 5 5 97.1 98.57 32.57 41.43
Last 99 99.52 95.33 99.29 96.57 99.05 54.52 58.33 88.95 92.38 64.81 69.05 89.43 94.76 99.19 99.76 85.59 86.67
1
Best 99.86 100 99.9 100 100 100 56.24 62.62 91.14 93.33 67.05 70 91.19 96.9 99.43 99.76 86.61 88.1
Last 50.14 57.14 52.29 59.52 53.81 54.29 98.33 99.76 97.67 98.81 99.28 99.76 98.28 99.52 95.67 99.29 98.93 99.76
2
Best 53.95 57.14 57.38 60.95 59.48 61.43 99.48 99.76 98.29 99.29 99.52 99.76 99 99.52 96.57 99.76 99.52 99.76
B Last 91.95 92.86 89.43 93.81 88.76 92.86 – – 50.43 53.81 54.29 56.43 60.81 63.81 64.62 67.86 52.98 56.19
3
Best 95.43 96.43 94 97.62 93.38 95.71 – – 53.19 56.9 55 59.05 62.29 65.95 65.86 69.29 56.38 58.62
Last 54.76 57.86 53.81 58.33 41.14 55.24 – – 83.87 87.38 84.57 86.67 95.48 96.67 93.76 99.05 85.36 86.43
4
Best 59.9 64.52 60 63.57 46.91 61.67 – – 85.42 87.38 85.33 86.9 96.33 98.57 94.09 99.52 86.49 88.33
Last 99 99.52 95.33 99.29 96.57 99.05 – – 44.29 44.29 37.86 44.52 64.52 68.81 97.33 98.57 39.94 42.38
5
Best 99.86 100 99.9 100 100 100 – – 46.67 46.67 39.27 44.52 68.65 76.9 97.62 98.57 40.48 42.38
Last 56.52 59.29 63.24 68.1 59.19 60.95 56.29 58.57 93.52 95.24 82.95 85.48 97.24 99.29 97.29 99.29 93.29 94.76
1
Best 59.91 64.05 66 68.33 62.19 62.86 58.81 63.57 94.71 97.14 82.95 85.48 98.1 99.29 98.05 99.29 93.52 95
Last 99.52 100 99.52 99.76 99.1 99.29 99.24 99.52 99.14 99.52 99 99.52 99.38 99.76 99.09 99.76 99 99.29
2
Best 99.95 100 99.95 100 99.9 100 99.33 99.76 99.48 100 99.28 99.52 99.48 100 99.81 100 99.29 99.76
C Last 46.48 48.81 51.14 56.43 53.24 55.95 – – 56.24 58.33 54.24 55.24 62.38 63.57 62.53 69.05 57.71 62.14
3
Best 50.38 55.24 53.29 60.24 56.81 58.57 – – 59 62.62 56.14 58.81 62.95 64.76 65.48 73.57 58.76 62.14
Last 86.14 91.19 91.67 94.29 88.62 95.95 – – 83.95 91.43 84.67 86.67 95.67 97.86 99.09 99.52 88.86 91.67
4
Best 92.71 97.38 94.33 95.48 94.05 98.1 – – 86.33 91.43 86.05 88.57 97.48 98.57 99.19 99.76 91.81 93.81
Last 55.33 58.81 55 59.05 55.28 62.62 – – 22.48 53.57 42.91 48.1 31 77.14 97.95 99.52 45.28 48.1
5
Best 58.09 59.52 60.71 65.48 62.47 65.24 – – 23.29 53.57 44.29 48.81 33.09 82.14 98.52 99.76 45.76 48.1
Please cite this article as: Z. Zhao, T. Li, J. Wu et al., Deep learning algorithms for rotating machinery intelligent diagnosis: An open source benchmark study. ISA Transactions
(2020), https://doi.org/10.1016/j.isatra.2020.08.010.
Z. Zhao, T. Li, J. Wu et al. / ISA Transactions xxx (xxxx) xxx 27
Table A.16
UoC: Results with random split and data augmentation.
Nor Input Loc AE DAE SAE MLP CNN LeNet AlexNet ResNet18 LSTM
Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max
Last 27.09 29.22 27.49 29.83 25.57 28.01 26.88 28.92 41.98 48.25 31.93 33.49 11.11 11.11 66.18 76.26 33.12 34.09
1
Best 29.92 31.35 31.66 32.57 28.46 29.38 30.2 31.2 47.55 51.45 33.94 35.46 11.45 12.79 76.8 78.39 36.89 37.9
Last 92.94 94.06 90.78 91.63 92.76 93.46 92.66 93.46 68.49 71.08 79.97 81.58 70.62 75.49 88.16 89.65 79.21 80.67
2
Best 94.7 95.13 93.3 93.76 94.55 95.13 94.91 95.13 69.53 71.08 84.78 86.45 73.3 78.69 90.59 91.93 80.58 81.74
A Last 15.25 19.44 19.11 24.04 21.39 24.78 – – 24.57 31.6 34.21 38.13 11.13 11.13 62.34 63.8 37.48 40.36
3
Best 17.57 23.15 20.65 24.04 23.95 26.71 – – 26.44 32.05 36.26 40.36 11.16 11.28 67.3 69.14 38.9 42.43
Last 52.42 57.69 50.71 55.1 51.14 53.88 – – 31.32 32.42 34.19 36.99 45.91 47.34 73.52 74.43 34.4 35.77
4
Best 55.98 60.27 53 56.47 54.43 56.16 – – 34.43 35.01 37.57 39.12 48.04 50.53 79.18 80.67 37.26 37.75
Last 35.37 49.01 34.03 45.97 37.26 46.27 – – 21.92 23.59 24.35 26.03 11.11 11.11 78.51 83.56 20.12 22.53
5
Best 37.56 49.01 36.83 49.01 39.39 48.1 – – 25.69 27.85 26.88 29.07 11.11 11.11 85.93 87.82 24.99 26.64
Last 26.61 27.55 29.98 31.81 28.01 29.68 27.67 28.92 46.18 47.79 32.17 34.09 30.11 40.79 69.16 72.15 35.13 37.29
1
Best 29.47 31.2 31.81 32.72 30.14 32.12 30.44 31.96 47.34 48.86 34.19 35.01 31.29 42.01 76.84 77.78 36.71 38.81
Last 79.87 84.32 80.67 85.54 81.4 87.37 81.19 90.11 62.89 67.28 78.72 82.34 68.77 74.89 83.87 88.43 77.62 78.84
2
Best 88.05 89.19 90.02 91.63 89.53 91.63 89.5 90.87 67.91 68.65 81.19 82.34 72.33 76.56 89.8 90.87 80.7 82.19
B Last 12.85 19.73 17.86 21.81 21.19 23.15 – – 27.18 33.09 36.29 38.72 13.21 21.51 65.19 67.66 37.62 40.65
3
Best 13.5 22.4 19.97 24.18 24.6 25.67 – – 28.66 34.12 38.9 40.5 14.22 24.48 67.89 68.99 39.91 42.73
Last 47.76 54.03 44.87 47.03 48.01 55.4 – – 30.93 32.42 34.4 36.38 33 48.25 75.04 79.15 34 36.07
4
Best 52.48 56.16 48.07 52.36 52.06 56.47 – – 34.58 35.01 38.39 39.73 35.13 52.51 79.63 80.97 37.99 39.88
Last 39.7 51.45 41.95 52.97 35.59 44.9 – – 16.96 22.68 27 29.22 11.11 11.11 71.38 87.21 20.7 22.68
5
Best 40.82 52.82 44.2 53.58 39.36 47.03 – – 19.15 28.46 28.31 29.22 11.17 11.42 86.03 88.13 26.12 27.7
Last 26.27 30.14 28.1 28.46 25.15 28.16 27.67 32.12 42.98 46.42 29.86 32.57 34.95 46.58 67.09 73.52 37.63 39.73
1
Best 29.13 30.59 30.26 32.88 27.64 30.29 29.77 32.12 45.9 50.68 32.69 35.31 35.92 46.58 76.16 77.32 40.67 42.92
Last 93.12 94.22 94.09 94.82 94.49 95.59 94.12 94.98 70.23 71.84 85.87 86.45 83.68 85.69 86.91 89.35 82.41 84.47
2
Best 94.95 95.28 95.22 95.89 95.22 95.74 95.68 96.19 73.55 74.73 87.73 88.74 86.27 87.21 89.2 90.26 84.05 85.84
C Last 12.88 19.88 18.1 21.66 20.03 21.81 – – 34.69 38.87 38.6 39.17 11.13 11.13 61.45 65.58 36.94 40.5
3
Best 13.47 22.85 19.38 23.15 23.26 24.93 – – 35.99 40.5 41.25 42.43 11.22 11.42 66.08 67.95 39.97 42.58
Last 52.05 53.42 53.45 57.99 52.91 59.06 – – 30.59 31.2 36.1 40.03 45.75 51.75 75.95 77.78 37.05 38.05
4
Best 55.8 57.08 55.56 59.06 57.57 59.97 – – 34.8 35.46 39.36 41.7 47.73 52.82 80.31 81.28 39.63 40.18
Last 44.57 49.01 44.9 48.25 49.25 53.58 – – 16.29 26.64 26 27.09 11.11 11.11 71.29 88.74 22.59 23.29
5
Best 47.03 49.47 47.64 51.29 51.9 56.16 – – 18.06 28.77 27.52 29.22 11.11 11.11 87.34 88.74 27.95 30.29
Table A.17
UoC: Results with random split and without data augmentation.
Nor Input Loc AE DAE SAE MLP CNN LeNet AlexNet ResNet18 LSTM
Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max
Last 21.55 25.27 22.13 25.57 21.89 25.11 24.41 27.55 36.23 42.92 28.07 31.05 11.11 11.11 72.12 88.28 29.71 32.42
1
Best 25.39 27.7 26.7 27.7 24.47 25.11 27.37 28.46 44.08 48.55 31.96 34.4 11.26 11.87 88.04 89.5 34.43 36.38
Last 92.53 94.67 91.45 93.15 92.15 94.06 91.69 92.69 65.36 69.25 83.38 84.93 75.04 78.54 88.13 90.72 80.82 82.34
2
Best 93.94 94.67 92.95 93.61 93.59 94.37 94.07 94.52 67.03 69.25 85.33 86.76 76.29 78.69 91.9 92.39 82.5 84.02
A Last 18.22 21.51 18.22 21.36 18.37 19.44 – – 27.15 29.38 31.54 33.68 11.13 11.13 63.02 66.17 35.7 42.14
3
Best 21.72 23.29 18.66 21.81 21.87 22.85 – – 29.26 31.9 35.34 38.72 11.13 11.13 67.36 69.44 39.17 44.66
Last 55.4 57.99 51.78 54.49 53.52 58.75 – – 31.6 32.12 34.15 37.29 44.23 52.21 75.89 78.69 35.68 36.99
4
Best 58.3 59.67 54.98 56.62 55.65 59.36 – – 34.1 35.01 37.93 39.57 46.91 54.19 79.15 80.06 38.33 39.57
Last 35.43 50.23 30.23 32.57 36.38 42.92 – – 18.2 25.57 22.16 24.05 11.11 11.11 71.14 85.39 19.39 21.31
5
Best 37.75 50.84 32.12 33.33 38.69 45.97 – – 19.27 27.7 27.06 28.01 11.17 11.42 86.24 87.98 23.87 25.88
Last 22.98 23.59 24.2 28.16 23.68 24.96 24.41 26.64 39.73 43.07 27.64 31.05 36.04 42.31 81.74 87.37 28.52 30.75
1
Best 26.39 27.4 26.94 28.61 26.06 28.77 27.76 29.07 43.74 45.97 32.21 33.79 38.93 42.31 87.06 87.98 33.3 34.09
Last 78.17 84.78 68.16 78.08 77.11 82.34 73.85 78.84 64.32 67.58 81.43 83.71 78.26 81.58 89.93 91.63 80.37 82.8
2
Best 85.75 88.13 86.03 88.43 85.91 86.91 86.48 88.28 65.72 68.8 84.05 85.69 80.12 83.56 92.33 93.3 82.71 84.17
B Last 13 20.47 20.18 21.96 20.77 22.85 – – 29.94 31.01 32.17 36.2 11.13 11.13 64.01 67.51 36.86 39.76
3
Best 13.68 22.7 22.46 24.63 23.65 26.11 – – 31.75 32.94 34.95 39.91 11.51 12.31 68.25 69.14 39.35 41.99
Last 49.07 51.75 53.21 55.56 50.5 54.79 – – 30.69 31.51 32.6 34.09 39.63 48.4 67.52 76.41 35.95 37.29
4
Best 51.14 53.73 55.8 59.06 53.52 56.01 – – 34.49 35.62 37.35 39.57 42.31 49.62 79.6 81.74 39.21 40.64
Last 42.59 53.42 47.36 50.68 40.25 46.73 – – 19.66 23.74 24.99 26.79 15.19 22.98 71.63 81.28 22.04 23.59
5
Best 44.14 54.03 49.1 51.6 41.43 47.95 – – 22.04 27.09 27.55 28.31 16.47 24.96 87.12 88.13 27.37 28.01
(continued on next page)
Please cite this article as: Z. Zhao, T. Li, J. Wu et al., Deep learning algorithms for rotating machinery intelligent diagnosis: An open source benchmark study. ISA Transactions
(2020), https://doi.org/10.1016/j.isatra.2020.08.010.
28 Z. Zhao, T. Li, J. Wu et al. / ISA Transactions xxx (xxxx) xxx
Table A.18
UoC: Results with order split and data augmentation.
Nor Input Loc AE DAE SAE MLP CNN LeNet AlexNet ResNet18 LSTM
Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max
Last 25.75 27.09 26.3 29.53 24.72 26.03 25.72 27.7 28.83 32.72 29.95 31.05 11.11 11.11 38.96 43.99 34.79 35.46
1
Best 28.55 31.05 29.53 31.81 28 28.46 29.65 31.05 33.18 34.09 32.11 32.88 11.41 12.63 44.23 44.75 36.99 38.51
Last 62.74 64.54 63.23 67.88 64.02 65.3 65.36 70.02 52.24 53.27 53.06 54.79 53.36 55.86 53.82 56.32 52.48 55.56
2
Best 67.88 70.47 68.31 68.95 68.52 71.08 70.04 71.99 55.86 57.99 58.02 59.82 56.99 57.84 61.67 63.32 55.98 57.53
A Last 17.9 22.37 15.79 19.56 21.72 23.26 – – 24.89 27.7 26.43 29.48 13.04 20.74 43.41 45.63 27.23 28.3
3
Best 20.27 25.19 16.62 20.74 23.59 25.48 – – 26.81 28 29.07 30.96 13.18 21.48 49.04 50.67 29.54 31.7
Last 35.89 37.75 37.9 40.18 36.28 37.44 – – 30.65 34.55 30.93 33.18 36.47 38.81 43.8 45.66 31.38 33.33
4
Best 39.82 42.16 40.43 41.55 39.42 42.01 – – 36.32 37.6 36.29 38.2 39.12 39.73 48.46 50.23 35.74 37.44
Last 21.52 32.57 22.8 28.31 24.05 27.85 – – 18.6 26.33 24.32 26.79 11.11 11.11 34.61 39.88 21.58 23.14
5
Best 24.63 35.31 24.78 32.57 26.91 31.51 – – 20.64 27.09 29.01 30.59 11.11 11.11 41.89 45.36 25.69 26.64
Last 24.87 26.64 24.9 28.16 26.33 28.31 25.42 27.7 29.65 32.88 30.07 31.35 24.48 32.12 37.87 41.7 34.25 35.01
1
Best 27.89 28.92 27.52 28.31 28.92 30.44 28.71 29.68 33.76 37.6 32.94 35.01 26.76 33.18 43.38 44.6 36.74 37.29
Last 52.27 59.21 55.95 60.43 46.76 53.58 57.26 63.47 48.8 50.99 54.7 56.16 49.77 54.03 51.81 55.1 50.99 52.51
2
Best 62.41 64.08 65.36 67.88 62.31 64.99 65.9 68.49 53.42 56.01 57.2 58.9 53.36 57.23 59.97 61.64 54.7 55.1
B Last 13.95 20.59 14.72 20.59 19.23 21.48 – – 22.55 24.3 24.03 25.33 12.56 17.78 46.19 48.44 20.89 21.61
3
Best 14.73 23.7 16.12 23.7 22.04 23.7 – – 23.76 24.3 27.73 29.33 13.21 20.89 49.07 50.07 23.75 23.99
Last 32.94 35.77 36.01 39.73 32.88 34.55 – – 31.54 33.03 29.01 30.14 37.02 38.96 40.21 49.16 31.35 33.49
4
Best 36.71 40.33 39.18 41.7 38.23 39.57 – – 37.2 38.05 35.4 36.99 41.25 43.23 49.22 49.92 35.62 36.07
Last 26.79 28.77 19.72 27.85 26.36 29.22 – – 18.32 24.81 25.48 28.46 17.47 21.61 30.75 40.03 23.01 24.96
5
Best 31.17 35.01 23.01 31.51 30.05 32.57 – – 20.97 28.31 29.41 31.2 20.4 24.51 41.61 43.53 27.27 28.92
Last 23.62 27.25 24.47 27.09 23.93 26.48 23.93 25.11 31.2 32.88 29.68 32.72 29.96 31.81 38.63 45.66 36.44 38.2
1
Best 27.08 29.38 27.2 28.46 27.37 29.22 28.13 28.77 35.68 37.9 31.48 34.4 32.33 34.25 44.63 45.66 38.39 39.12
Last 63.18 67.28 62.97 66.97 62.59 66.21 62.59 64.84 51.29 54.03 57.9 59.21 54.46 57.69 53.36 56.16 50.9 51.75
2
Best 68.25 70.62 67.78 70.47 68.51 72.15 69.89 70.78 55.68 56.62 61.8 62.71 58.66 60.12 59.21 61.04 56.89 58.14
C Last 12.8 19.56 15.82 20.89 20.89 24.15 – – 23.53 28.44 26.49 28.89 23.11 34.52 39.73 48.59 25.96 27.26
3
Best 13.39 21.63 16.68 21.48 22.67 25.33 – – 26.16 30.67 30.28 31.11 24.62 35.11 47.32 48.89 29.63 30.81
Last 37.17 40.33 37.09 39.27 37.28 38.51 – – 31.05 33.03 30.78 32.42 28.34 35.92 43.35 49.62 30.93 32.72
4
Best 40.62 43.68 40.76 42.62 41.48 43.84 – – 35.83 36.68 36.13 37.14 31.29 37.29 49.86 52.51 35.59 36.68
Last 28.17 32.88 28.1 32.27 28.11 35.77 – – 11.11 11.11 24.29 27.55 11.11 11.11 31.84 37.6 23.65 26.03
5
Best 32.04 34.86 30.59 34.86 31.93 36.99 – – 11.96 13.09 28.71 30.44 11.14 11.26 39.64 41.25 28.31 29.38
Table A.19
XJTU-SY: Results with random split and data augmentation.
Nor Input Loc AE DAE SAE MLP CNN LeNet AlexNet ResNet18 LSTM
Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max
Last 68.59 74.22 71.98 80.99 70.62 78.12 66.72 73.7 94.64 99.74 96.09 99.22 84.12 98.18 88.39 99.74 96.82 97.92
1
Best 74.17 76.3 79.84 83.07 75.62 80.21 77.24 80.21 99.95 100 98.6 99.22 86.36 98.18 99.79 100 98.91 99.48
Last 100 100 100 100 100 100 100 100 96.82 100 90.89 100 6.51 6.51 100 100 95.83 100
2
Best 100 100 100 100 100 100 100 100 100 100 100 100 6.72 6.77 100 100 100 100
A Last 37.81 48.7 30.78 40.89 38.85 47.14 – – 71.2 73.44 65.05 71.88 6.51 6.51 74.38 92.19 67.45 70.57
3
Best 39.01 49.48 34.9 47.66 41.15 48.18 – – 75.47 76.56 75.52 76.3 7.08 7.81 91.83 92.71 72.55 74.22
Last 99.95 100 99.84 100 99.69 100 – – 98.65 99.22 98.33 99.48 99.58 100 100 100 99.17 99.74
4
Best 100 100 100 100 100 100 – – 98.75 99.22 99.48 99.74 100 100 100 100 99.79 100
Last 42.97 90.62 68.02 90.89 59.17 89.84 – – 6.51 6.51 60.16 74.48 6.51 6.51 100 100 71.41 76.56
5
Best 56.04 92.71 72.71 93.75 65.68 91.93 – – 6.72 6.77 74.58 76.82 6.72 6.77 100 100 77.97 80.99
(continued on next page)
Please cite this article as: Z. Zhao, T. Li, J. Wu et al., Deep learning algorithms for rotating machinery intelligent diagnosis: An open source benchmark study. ISA Transactions
(2020), https://doi.org/10.1016/j.isatra.2020.08.010.
Z. Zhao, T. Li, J. Wu et al. / ISA Transactions xxx (xxxx) xxx 29
Table A.20
XJTU-SY: Results with random split and without data augmentation.
Nor Input Loc AE DAE SAE MLP CNN LeNet AlexNet ResNet18 LSTM
Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max
Last 36.04 48.18 55.21 57.29 48.85 52.6 60 62.24 99.58 100 93.8 96.35 73.44 93.23 94.69 100 94.95 97.14
1
Best 56.98 58.85 57.55 59.64 54.95 57.03 61.72 63.02 99.79 100 97.4 98.7 75.16 95.83 99.95 100 97.87 98.18
Last 100 100 100 100 100 100 100 100 99.64 99.74 100 100 6.51 6.51 100 100 100 100
2
Best 100 100 100 100 100 100 100 100 99.79 100 100 100 6.72 6.77 100 100 100 100
A Last 39.64 48.96 36.2 43.75 37.92 41.41 – – 70.36 76.56 71.46 75.52 43.85 72.4 82.08 90.62 69.53 72.14
3
Best 42.14 50.52 39.43 46.88 41.51 42.97 – – 77.18 77.6 76.09 79.43 46.67 76.56 91.72 92.45 73.02 75
Last 100 100 100 100 99.9 100 – – 98.34 98.7 99.27 100 99.74 100 99.06 100 98.91 99.74
4
Best 100 100 100 100 99.9 100 – – 98.49 98.96 99.53 100 100 100 100 100 99.84 100
Last 54.43 92.19 36.93 86.72 63.28 92.97 – – 6.51 6.51 52.97 66.15 6.51 6.51 78.54 100 59.9 75.78
5
Best 54.95 92.71 38.65 87.24 63.8 93.49 – – 7.66 10.94 74.53 79.69 6.82 7.03 100 100 77.61 79.69
Last 57.87 60.16 57.24 59.11 57.97 59.9 60 61.2 95.68 100 93.91 96.61 96.88 98.18 99.79 100 95.83 98.44
1
Best 60.37 62.5 58.33 60.16 59.85 60.68 61.82 63.54 99.84 100 97.97 98.96 98.6 99.22 100 100 98.02 98.44
Last 68.18 100 70.89 84.64 62.66 100 62.66 100 99.84 100 100 100 6.51 6.51 100 100 99.95 100
2
Best 100 100 85.68 89.06 100 100 100 100 99.95 100 100 100 6.77 6.77 100 100 100 100
B Last 20.47 42.71 24.58 43.23 44.95 50.78 – – 75.78 77.6 72.97 75.26 79.89 83.85 90.52 91.93 70.26 72.14
3
Best 21.98 45.31 26.98 46.35 47.34 52.6 – – 78.28 80.21 75.21 76.56 81.93 84.38 92.14 92.71 74.38 75.78
Last 99.64 100 100 100 100 100 – – 98.39 98.96 98.44 98.96 99.48 100 100 100 96.15 100
4
Best 99.79 100 100 100 100 100 – – 98.6 99.22 99.01 99.22 100 100 100 100 100 100
Last 86.77 89.84 50.78 89.06 68.13 87.24 – – 31.93 76.56 72.19 75.78 37.5 91.15 98.91 100 77.5 80.99
5
Best 88.75 93.23 51.46 90.36 85.47 93.49 – – 35.05 78.65 76.25 80.21 38.91 92.97 100 100 80.42 82.55
Last 62.29 63.02 62.19 64.84 61.67 64.84 61.1 64.84 99.22 99.74 96.93 98.18 98.54 99.48 99.95 100 97.03 98.44
1
Best 64.16 65.62 63.75 65.62 65.42 66.15 62.66 65.1 99.48 100 98.7 99.22 99.12 99.74 100 100 98.86 99.22
Last 99.9 100 100 100 100 100 100 100 99.74 100 100 100 100 100 100 100 100 100
2
Best 100 100 100 100 100 100 100 100 99.9 100 100 100 100 100 100 100 100 100
C Last 17.76 36.46 17.97 40.1 35.78 41.15 – – 77.13 81.77 74.27 76.04 82.76 84.9 91.88 92.97 72.08 76.3
3
Best 19.58 40.62 19.01 43.49 38.23 42.71 – – 78.65 81.77 77.5 78.39 85.37 87.24 93.39 94.53 75 76.3
Last 100 100 100 100 100 100 – – 98.34 99.22 98.6 99.22 99.74 100 99.84 100 99.48 100
4
Best 100 100 100 100 100 100 – – 98.75 99.48 98.86 99.48 100 100 100 100 100 100
Last 93.96 97.66 85.52 96.35 90.26 93.75 – – 34.84 77.86 76.77 79.43 89.79 94.79 98.75 99.74 82.61 85.16
5
Best 94.9 98.7 86.3 98.18 92.34 94.79 – – 35.62 79.95 80.16 81.51 91.77 95.31 100 100 85.62 87.76
Please cite this article as: Z. Zhao, T. Li, J. Wu et al., Deep learning algorithms for rotating machinery intelligent diagnosis: An open source benchmark study. ISA Transactions
(2020), https://doi.org/10.1016/j.isatra.2020.08.010.
30 Z. Zhao, T. Li, J. Wu et al. / ISA Transactions xxx (xxxx) xxx
Table A.21
XJTU-SY: Results with order split and data augmentation.
Nor Input Loc AE DAE SAE MLP CNN LeNet AlexNet ResNet18 LSTM
Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max Mean Max
Last 62.97 73.59 68.36 76.41 60.46 71.54 68.41 75.38 98.97 99.74 96.31 96.67 91.9 95.13 89.54 96.41 96.26 97.95
1
Best 72.67 76.92 81.08 83.33 70.92 72.05 77.64 80.26 99.84 100 98.1 98.97 94.2 96.15 99.69 100 98 98.97
Last 100 100 100 100 100 100 73.33 73.33 73.02 73.33 72.2 73.33 6.67 6.67 73.33 73.33 71.38 73.33
2
Best 100 100 100 100 100 100 73.33 73.33 73.33 73.33 73.33 73.33 6.67 6.67 73.33 73.33 73.33 73.33
A Last 26.05 45.64 29.95 40.77 41.95 46.67 – – 72.3 76.15 75.28 78.21 14.72 46.92 73.8 88.46 72.77 74.62
3
Best 27.64 47.44 33.13 43.08 46.72 51.28 – – 76.77 78.21 78.26 79.74 15.03 46.92 90.51 91.79 77.38 80
Last 99.95 100 99.84 100 99.9 100 – – 97.38 98.97 97.49 99.49 99.64 100 100 100 98.15 99.23
4
Best 100 100 100 100 100 100 – – 98.26 98.97 99.08 99.74 100 100 100 100 99.59 99.74
Last 58.36 92.56 54.05 88.46 78.1 89.74 – – 19.44 70.51 53.28 75.9 6.67 6.67 95.18 99.74 71.79 78.97
5
Best 60.56 93.59 55.39 89.23 82.31 95.13 – – 21.33 79.74 74 76.92 6.67 6.67 100 100 78.05 79.49
Last 82.67 85.9 84.26 86.92 82.05 83.59 85.13 87.44 98.87 99.74 97.59 97.95 97.54 98.46 97.44 99.49 96.51 97.69
1
Best 84.41 85.9 87.03 87.44 85.03 85.9 87.54 88.21 99.79 100 98.46 99.23 98.51 99.49 99.84 100 97.9 98.46
Last 14 36.15 29.23 32.82 20.82 47.44 11.39 18.46 54.2 72.05 73.28 73.33 6.67 6.67 52.77 73.33 73.33 73.33
2
Best 42.92 58.21 52.31 63.59 32.51 47.44 33.95 38.97 73.28 73.33 73.33 73.33 6.67 6.67 73.33 73.33 73.33 73.33
B Last 42.87 47.44 33.54 47.18 45.59 47.69 – – 75.74 78.46 76.41 79.23 81.64 82.56 87.49 88.97 75.13 76.67
3
Best 45.9 50.77 36.05 48.72 49.85 52.31 – – 78.62 79.23 78.67 79.74 85.28 85.9 90.62 91.28 77.79 79.23
Last 100 100 98.51 100 99.59 100 – – 96.87 98.46 96.1 98.72 81.23 100 100 100 98.62 99.49
4
Best 100 100 98.82 100 100 100 – – 98.41 98.97 98.51 99.23 81.33 100 100 100 99.44 99.74
Last 89.33 92.56 55.28 91.79 66.67 94.1 – – 26.72 76.67 70.72 74.36 35.39 84.62 99.33 100 72.31 75.64
5
Best 91.18 93.59 56.46 93.85 72.36 94.87 – – 27.44 76.67 76.72 79.23 38.05 89.74 100 100 78.15 82.05
Last 81.44 83.85 85.38 87.44 81.08 83.59 83.9 85.13 96.67 99.74 98.05 98.97 98.82 98.97 98.92 100 97.29 98.21
1
Best 83.59 86.15 87.44 89.23 84.26 85.64 85.9 87.18 99.95 100 99.33 99.74 99.64 99.74 100 100 99.18 99.49
Last 99.95 100 100 100 100 100 73.33 73.33 73.33 73.33 73.23 73.33 73.33 73.33 73.33 73.33 73.33 73.33
2
Best 100 100 100 100 100 100 73.33 73.33 73.33 73.33 73.33 73.33 73.33 73.33 73.33 73.33 73.33 73.33
C Last 6.67 6.67 6.67 6.67 49.69 57.44 – – 75.64 76.92 76.97 81.54 83.69 85.9 90.46 91.54 75.49 76.92
3
Best 6.67 6.67 6.67 6.67 51.85 61.54 – – 78.05 80.26 78.82 81.54 85.95 88.46 93.18 93.59 78.87 80
Last 99.95 100 99.79 100 99.95 100 – – 98.26 98.97 97.59 98.46 99.95 100 100 100 98.87 99.74
4
Best 100 100 100 100 100 100 – – 98.77 98.97 98.15 98.72 100 100 100 100 99.44 99.74
Last 74.05 93.08 69.95 92.56 90.1 93.85 – – 57.95 73.85 77.38 81.28 90.51 94.62 90.67 100 79.23 82.82
5
Best 76.77 95.13 73.33 96.92 94.56 98.21 – – 64.46 82.31 81.28 82.31 93.69 97.69 100 100 85.79 88.72
Declaration of competing interest [8] Young T, Hazarika D, Poria S, Cambria E. Recent trends in deep
learning based natural language processing. IEEE Comput Intell Mag
2018;13:55–75.
The authors declare that they have no known competing finan-
[9] Feng Q, Zhao X, Fan D, Cai B, Liu Y, Ren Y. Resilience design method
cial interests or personal relationships that could have appeared based on meta-structure: A case study of offshore wind farm. Reliab Eng
to influence the work reported in this paper. Syst Saf 2019;186:232–44.
[10] Li D, Liu Y, Huang D. Development of semi-supervised multiple-output
soft-sensors with co-training and tri-training mpls and mrvm. Chemometr
Acknowledgment
Intell Lab Syst 2020;199:103970.
[11] Hinton GE, Salakhutdinov RR. Reducing the dimensionality of data with
This work was supported by Natural Science Foundation of neural networks. Science 2006;313:504–7.
China (No. 51835009, No. 51705398). [12] MIT Technology Review. 10 breakthrough technologies 2013. 2019,
https://www.technologyreview.com/lists/technologies/2013/ [Accessed on
August 2019].
Appendix. Evaluation results [13] LeCun Y, Bengio Y, Hinton G. Deep learning. Nature 2015;521:436.
[14] Riley P. Three pitfalls to avoid in machine learning. 2019.
See Tables A.1–A.21. [15] Goodfellow I, Bengio Y, Courville A. Deep learning. MIT Press; 2016.
[16] Li C, d. Oliveira JLV, Lozada MC, Cabrera D, Sanchez V, Zurita G. A
systematic review of fuzzy formalisms for bearing fault diagnosis. IEEE
References Trans Fuzzy Syst 2018.
[17] Hoang D-T, Kang H-J. A survey on deep learning based bearing fault
[1] Zhao Z, Wu S, Qiao B, Wang S, Chen X. Enhanced sparse period- diagnosis. Neurocomputing 2019;335:327–35.
group lasso for bearing fault diagnosis. IEEE Trans Ind Electron [18] Zhang S, Zhang S, Wang B, Habetler TG. Deep learning algorithms
2018;66:2143–53. for bearing fault diagnosticsx-a comprehensive review. IEEE Access
[2] Wang S, Chen X, Tong C, Zhao Z. Matching synchrosqueezing wavelet 2020;8:29857–81.
transform and application to aeroengine vibration monitoring. IEEE Trans [19] Hamadache M, Jung JH, Park J, Youn BD. A comprehensive review of
Instrum Meas 2016;66:360–72. artificial intelligence-based approaches for rolling element bearing phm:
[3] Sun C, Ma M, Zhao Z, Chen X. Sparse deep stacking network for fault shallow and deep learning. JMST Adv 2019;1:125–51.
diagnosis of motor. IEEE Trans Ind Inf 2018;14:3261–70. [20] Ali YH, Ali SM, Rahman RA, Hamzah RIR. Acoustic emission and artifi-
[4] Krizhevsky A, Sutskever I, Hinton GE. Imagenet classification with cial intelligent methods in condition monitoring of rotating machine–a
deep convolutional neural networks. In: Advances in neural information review. In: National conference for postgraduate research.
processing systems, p. 1097–105. [21] Liu R, Yang B, Zio E, Chen X. Artificial intelligence for fault diagnosis of
[5] Farabet C, Couprie C, Najman L, LeCun Y. Learning hierarchical features rotating machinery: A review. Mech Syst Signal Process 2018;108:33–47.
for scene labeling. IEEE Trans Pattern Anal Mach Intell 2012;35:1915–29. [22] Wei Y, Li Y, Xu M, Huang W. A review of early fault diagnosis approaches
[6] Hirschberg J, Manning CD. Advances in natural language processing. and their applications in rotating machinery. Entropy 2019;21:409.
Science 2015;349:261–6. [23] Zhao G, Zhang G, Ge Q, Liu X. Research advances in fault diagnosis
[7] Sun S, Luo C, Chen J. A review of natural language processing techniques and prognostic based on deep learning. In: 2016 Prognostics and system
for opinion mining systems. Inf Fusion 2017;36:10–25. health management conference. IEEE; p. 1–6.
Please cite this article as: Z. Zhao, T. Li, J. Wu et al., Deep learning algorithms for rotating machinery intelligent diagnosis: An open source benchmark study. ISA Transactions
(2020), https://doi.org/10.1016/j.isatra.2020.08.010.
Z. Zhao, T. Li, J. Wu et al. / ISA Transactions xxx (xxxx) xxx 31
[24] Duan L, Xie M, Wang J, Bai T. Deep learning enabled intelligent [52] Khodja AY, Guersi N, Saadi MN, Boutasseta N. Rolling element bear-
fault diagnosis: Overview and applications. J Intell Fuzzy Systems ing fault diagnosis for rotating machinery using vibration spectrum
2018;35:5771–84. imaging and convolutional neural networks. Int J Adv Manuf Technol
[25] Zhang W, Jia M-P, Zhu L, Yan X-A. Comprehensive overview on compu- 2020;106:1737–51.
tational intelligence techniques for machinery condition monitoring and [53] Li Y, Du X, Wan F, Wang X, Yu H. Rotating machinery fault diagnosis
fault diagnosis. Chin J Mech Eng 2017;30:782–95. based on convolutional neural network and infrared thermal imaging.
[26] Zhao R, Yan R, Chen Z, Mao K, Wang P, Gao RX. Deep learning and Chin J Aeronaut 2020;33:427–38.
its applications to machine health monitoring. Mech Syst Signal Process [54] Zhang J, Sun Y, Guo L, Gao H, Hong X, Song H. A new bearing fault
2019;115:213–37. diagnosis method based on modified convolutional neural networks. Chin
[27] Lei Y, Yang B, Jiang X, Jia F, Li N, Nandi AK. Applications of machine J Aeronaut 2020;33:439–47.
learning to machine fault diagnosis: A review and roadmap. Mech Syst [55] Zhao M, Tang B, Deng L, Pecht M. Multiple wavelet regularized deep
Signal Process 2020;138:106587. residual networks for fault diagnosis. Measurement 2020;152.
[28] Nasiri S, Khosravani MR, Weinberg K. Fracture mechanics and mechanical [56] Li X, Zhang W, Ding Q, Sun J-Q. Intelligent rotating machinery fault
fault detection by artificial intelligence methods: A review. Eng Fail Anal diagnosis based on deep learning using data augmentation. J Intell Manuf
2017;81:270–93. 2020;31:433–52.
[29] Tian Y, Guo D, Zhang K, Jia L, Qiao H, Tang H. A review of fault diagnosis [57] Tang T, Hu T, Chen M, Lin R, Chen G. A deep convolutional neural
for traction induction motor. In: 2018 37th Chinese control conference. network approach with information fusion for bearing fault diagnosis
2018, p. 5763–8. under different working conditions. Proc Inst Mech Eng C 2020.
[30] Khan S, Yairi T. A review on the application of deep learning in system [58] Xue Y, Dou D, Yang J. Multi-fault diagnosis of rotating machinery
health management. Mech Syst Signal Process 2018;107:241–65. based on deep convolution neural network and support vector machine.
[31] Stetco A, Dinmohammadi F, Zhao X, Robu V, Flynn D, Barnes M, et Measurement 2020;156.
al. Machine learning methods for wind turbine condition monitoring: A [59] Verstraete DB, Lope Droguett E, Meruane V, Modarres M, Ferrada A.
review. Renew Energy 2018. Deep semi-supervised generative adversarial fault diagnostics of rolling
[32] Ellefsen AL, Æsøy V, Ushakov S, Zhang H. A comprehensive survey element bearings. Struct Health Monitor 2020;19:390–411.
of prognostics and health management based on deep learning for [60] Zhang W, Li X, Jia X-D, Ma H, Luo Z, Li X. Machinery fault diagno-
autonomous ships. IEEE Trans Reliab 2019;68:720–40. sis with imbalanced data using deep generative adversarial networks.
[33] Ademujimi TT, Brundage MP, Prabhu VV. A review of current machine Measurement 2020;152.
learning techniques used in manufacturing diagnosis. In: IFIP inter- [61] Li T, Zhao Z, Sun C, Yan R, Chen X. Adaptive channel weighted cnn with
national conference on advances in production management systems. multi-sensor fusion for condition monitoring of helicopter transmission
Springer; p. 407–15. system. IEEE Sens J 2020.
[34] Chang C-W, Lee H-W, Liu C-H. A review of artificial intelligence
[62] Xu G, Liu M, Jiang Z, Shen W, Huang C. Online fault diagnosis method
algorithms used for smart machine tools. Inventions 2018;3:41.
based on transfer convolutional neural networks. IEEE Trans Instrum
[35] Wang J, Ma Y, Zhang L, Gao RX, Wu D. Deep learning for smart Meas 2020;69:509–20.
manufacturing: Methods and applications. J Manuf Syst 2018;48:144–56.
[63] Mao W, Ding L, Tian S, Liang X. Online detection for bearing incipient
[36] Sharp M, Ak R, Hedber Jr T. A survey of the advancing use and
fault based on deep transfer learning. Measurement 2020;152.
development of machine learning in smart manufacturing. J Manuf Syst
[64] Chen Z, Gryllias K, Li W. Intelligent fault diagnosis for rotary machinery
2018;48:170–9.
using transferable convolutional neural network. IEEE Trans Ind Inf
[37] Mao W, Chen J, Liang X, Zhang X. A new online detection approach for
2020;16:339–49.
rolling bearing incipient fault via self-adaptive deep feature matching.
[65] Li Q, Tang B, Deng L, Wu Y, Wang Y. Deep balanced domain adaptation
IEEE Trans Instrum Meas 2020;69:443–56.
neural networks for fault diagnosis of planetary gearboxes with limited
[38] Chen L, Zhang Z, Cao J, Wang X. A novel method of combining non-
labeled data. Measurement 2020;156.
linear frequency spectrum and deep learning for complex system fault
[66] Jiao J, Zhao M, Lin J. Unsupervised adversarial adaptation network for
diagnosis. Measurement 2020;151.
intelligent fault diagnosis. IEEE Trans Ind Electron 2019.
[39] Zhang Y, Li X, Gao L, Chen W, Li P. Intelligent fault diagnosis of
[67] Grezmak J, Zhang J, Wang P, Loparo KA, Gao RX. Interpretable convo-
rotating machinery using a new ensemble deep auto-encoder method.
lutional neural network through layer-wise relevance propagation for
Measurement 2020;151.
machine fault diagnosis. IEEE Sensors J 2020;20:3172–81.
[40] Kong X, Mao G, Wang Q, Ma H, Yang W. A multi-ensemble method
based on deep auto-encoders for fault diagnosis of rolling bearings. [68] Haidong S, Junsheng C, Hongkai J, Yu Y, Zhantao W. Enhanced deep gated
Measurement 2020;151. recurrent unit and complex wavelet packet energy moment entropy for
early fault prognosis of bearing. Knowl-Based Syst 2020;188.
[41] Jiang N, Hu X, Li N. Graphical temporal semi-supervised deep learning-
based principal fault localization in wind turbine systems. Proc. Inst. [69] Zhao K, Jiang H, Li X, Wang R. An optimal deep sparse autoencoder with
Mech. Eng. I 2020. gated recurrent unit for rolling bearing fault diagnosis. Meas Sci Technol
[42] Li X, Li J, Qu Y, He D. Semi-supervised gear fault diagnosis using raw 2020;31.
vibration signal based on deep learning. Chin J Aeronaut 2020;33:418–26. [70] Wu Z, Jiang H, Zhao K, Li X. An adaptive deep transfer learning method
[43] Xiong X, Jiang H, Li X, Niu M. A wasserstein gradient-penalty generative for bearing fault diagnosis. Measurement 2020;151.
adversarial network with deep auto-encoder for bearing intelligent fault [71] Ma Y, Jia X, Bai H, Wang G, Liu G, Guo C. A new fault diagnosis
diagnosis. Meas Sci Technol 2020;31. method using deep belief network and compressive sensing. J Vibroeng
[44] Zhou F, Yang S, Fujita H, Chen D, Wen C. Deep learning fault diag- 2020;22:83–97.
nosis method based on global optimization GAN for unbalanced data. [72] Yan L-P, Dong X-Z, Wang T, Gao Q, Tan C-Q, Zeng D-T, et al. A fault
Knowl-Based Syst 2020;187. diagnosis method for gas turbines based on improved data preprocessing
[45] Guo Q, Li Y, Song Y, Wang D, Chen W. Intelligent fault diagnosis method and an optimization deep belief network. Meas Sci Technol 2020;31.
based on full 1-d convolutional generative adversarial network. IEEE Trans [73] Yu K, Lin TR, Tan J. A bearing fault and severity diagnostic technique
Ind Inf 2020;16:2044–53. using adaptive deep belief networks and Dempster–Shafer theory. Struct
[46] Zhao X, Jia M, Lin M. Deep Laplacian Auto-encoder and its application Health Monitor 2020;19:240–61.
into imbalanced fault diagnosis of rotating machinery. Measurement [74] Ding Y, Ma L, Ma J, Suo M, Tao L, Cheng Y, et al. Intelligent fault
2020;152. diagnosis for rotating machinery using deep q-network based health
[47] Zhiyi H, Haidong S, Lin J, Junsheng C, Yu Y. Transfer fault diagno- state classification: A deep reinforcement learning approach. Adv Eng Inf
sis of bearing installed in different machines using enhanced deep 2019;42:100977.
auto-encoder. Measurement 2020a;152. [75] Dai W, Mo Z, Luo C, Jiang J, Zhang H, Miao Q. Fault diagnosis of
[48] Zhiyi H, Haidong S, Ping W, Lin JJ, Junsheng C, Yu Y. Deep transfer multi- rotating machinery based on deep reinforcement learning and reciprocal
wavelet auto-encoder for intelligent fault diagnosis of gearbox with few of smoothness index. IEEE Sens J 2020.
target training samples. Knowl-Based Syst 2020b;191. [76] Zhang D, Stewart E, Entezami M, Roberts C, Yu D. Intelligent acoustic-
[49] Li X, Jia X-D, Zhang W, Ma H, Luo Z, Li X. Intelligent cross-machine based fault diagnosis of roller bearings using a deep graph convolutional
fault diagnosis approach with deep auto-encoder and domain adaptation. network. Measurement 2020;156.
Neurocomputing 2020;383:235–47. [77] Rumelhart DE, Hinton GE, Williams RJ. Learning internal representations
[50] Mao W, Zhang D, Tian S, Tang J. Robust detection of bearing early fault by error propagation. Technical report, California Univ San Diego La Jolla
based on deep transfer learning. Electronics 2020;9. Inst for Cognitive Science; 1985.
[51] Zhang Y, Xing K, Bai R, Sun D, Meng Z. An enhanced convolutional neural [78] Vincent P, Larochelle H, Bengio Y, Manzagol P-A. Extracting and compos-
network for bearing fault diagnosis based on time–frequency image. ing robust features with denoising autoencoders. In: Proceedings of the
Measurement 2020;157. 25th international conference on Machine learning. ACM; p. 1096–103.
Please cite this article as: Z. Zhao, T. Li, J. Wu et al., Deep learning algorithms for rotating machinery intelligent diagnosis: An open source benchmark study. ISA Transactions
(2020), https://doi.org/10.1016/j.isatra.2020.08.010.
32 Z. Zhao, T. Li, J. Wu et al. / ISA Transactions xxx (xxxx) xxx
[79] Ranzato M, Poultney C, Chopra S, Cun YL. Efficient learning of sparse [96] Lee J, Qiu H, Yu G, Lin Ja. Bearing data set. Moffett Field, CA: IMS,
representations with an energy-based model. In: Advances in neural University of Cincinnati, NASA Ames Prognostics Data Repository, NASA
information processing systems. p. 1137–44. Ames Research Center; 2007, https://ti.arc.nasa.gov/tech/dash/groups/
[80] LeCun Y, Bengio Y. Convolutional networks for images, speech, and time pcoe/prognostic-data-repository/.
series. In: The handbook of brain theory and neural networks. vol. 3361. [97] Zhang Y, Li X, Gao L, Wang L, Wen L. Imbalanced data fault diagnosis of
1995, p. 1995. rotating machinery using synthetic oversampling and feature learning. J
[81] He K, Zhang X, Ren S, Sun J. Deep residual learning for image recognition. Manuf Syst 2018;48:34–50.
In: Proceedings of the IEEE conference on computer vision and pattern [98] Mao W, Liu Y, Ding L, Li Y. Imbalanced fault diagnosis of rolling bearing
recognition. p. 770–8. based on generative adversarial network: A comparative study. IEEE
[82] Hochreiter S, Schmidhuber J. Long short-term memory. Neural Comput Access 2019;7:9515–30.
1997;9:1735–80. [99] Buda M, Maki A, Mazurowski MA. A systematic study of the class
[83] Case western reserve university (CWRU) bearing data center. 2019, imbalance problem in convolutional neural networks. Neural Netw
[Online]. Available: https://csegroups.case.edu/bearingdatacenter/pages/ 2018;106:249–59.
download-data-file/. [Accessed September 2019]. [100] Zhao Z, Zhang Q, Yu X, Sun C, Wang S, Yan R, et al. Unsupervised
[84] Society for machinery failure prevention technology. 2019, [Online]. deep transfer learning for intelligent fault diagnosis: An open source and
Available: https://mfpt.org/fault-data-sets/. [Accessed September 2019]. comparative study. 2019, arXiv preprint arXiv:1912.12528.
[85] Lessmeier C, Kimotho JK, Zimmer D, Sextro W. KAt-datacenter, chair of [101] Zheng H, Wang R, Yang Y, Yin J, Li Y, Li Y, et al. Cross-domain
design and drive technology. Paderborn University; 2019, https://mb.uni- fault diagnosis using knowledge transfer strategy: a review. IEEE Access
paderborn.de/kat/forschung/datacenter/bearing-datacenter/ [Accessed on 2019;7:129260–90.
August 2019]. [102] Yan R, Shen F, Sun C, Chen X. Knowledge transfer for rotary machine
[86] Lessmeier C, Kimotho JK, Zimmer D, Sextro W. Condition monitoring fault diagnosis. IEEE Sens J 2019.
of bearing damage in electromechanical drive systems by using motor [103] Han T, Liu C, Yang W, Jiang D. Deep transfer network with joint
current signals of electric motors: A benchmark data set for data- distribution adaptation: A new intelligent fault diagnosis framework for
driven classification. In: Proceedings of the European conference of the industry application. ISA Trans 2019a.
prognostics and health management society. p. 05–8. [104] Han T, Liu C, Yang W, Jiang D. Learning transferable features in deep
[87] Cao P, Zhang S, Tang J. Gear fault data. 2019, [Online]. Available: https: convolutional neural networks for diagnosing unseen machine conditions.
//doi.org/10.6084/m9.figshare.6127874.v1. [Accessed September 2019]. ISA Trans 2019b;93:341–53.
[88] XJTU-SY bearing datasets. 2019, [Online]. Available:http://biaowang.tech/ [105] Rudin C. Stop explaining black box machine learning models for high
xjtu-sy-bearing-datasets/. [Accessed September 2019]. stakes decisions and use interpretable models instead. Nat Mach Intell
[89] Wang B, Lei Y, Li N, Li N. A hybrid prognostics approach for estimating 2019;1:206–15.
remaining useful life of rolling element bearings. IEEE Trans Reliab 2018. [106] Li X, Zhang W, Ding Q. Understanding and improving deep learning-based
[90] SEU gearbox datasets. 2019, [Online]. Available: https://github.com/ rolling bearing fault diagnosis with attention mechanism. Signal Process
cathysiyu/Mechanical-datasets. [Accessed September 2019]. 2019a;161:136–54.
[91] Shao S, McAleer S, Yan R, Baldi P. Highly accurate machine fault diagnosis [107] Li T, Zhao Z, Sun C, Cheng L, Chen X, Yan R, et al. Waveletkernelnet:
using deep transfer learning. IEEE Trans Ind Inf 2018;15:2446–55. An interpretable deep neural network for industrial intelligent diagnosis.
[92] Li K. School of mechanical engineering. Jiangnan University; 2019, 2019b, arXiv preprint arXiv:1911.07925.
http://mad-net.org:8765/explore.html?t=0.5831516555847212 [Accessed [108] Zeiler MD, Fergus R. Visualizing and understanding convolutional net-
on August 2019]. works. In: European conference on computer vision. Springer; p.
[93] Li K, Ping X, Wang H, Chen P, Cao Y. Sequential fuzzy diagnosis method 818–33.
for motor roller bearing in variable operating conditions based on [109] Tang M, Perazzi F, Djelouah A, Ben Ayed I, Schroers C, Boykov Y. On reg-
vibration analysis. Sensors 2013;13:8013–41. ularized losses for weakly-supervised cnn segmentation. In: Proceedings
[94] PHM IEEE 2012 data challenge. 2019, [Online]. Available: https: of the European conference on computer vision. p. 507–22.
//github.com/wkzs111/phm-ieee-2012-data-challenge-dataset. [Accessed [110] Ravanelli M, Bengio Y. Interpretable convolutional filters with sincnet.
September 2019]. 2018, arXiv preprint arXiv:1811.09725.
[95] Nectoux P, Gouriveau R, Medjaher K, Ramasso E, Chebel-Morello B, [111] Gregor K, LeCun Y. Learning fast approximations of sparse coding. In: Pro-
Zerhouni N, et al. Pronostia: An experimental platform for bearings accel- ceedings of the 27th international conference on international conference
erated degradation tests. In: IEEE international conference on prognostics on machine learning. Omnipress; p. 399–406.
and health management. IEEE Catalog Number: CPF12PHM-CDR. p. 1–8. [112] Wang Y, Yao Q, Kwok J, Ni LM. Generalizing from a few examples: A
survey on few-shot learning. 2019, arXiv:1904.05046.
[113] Elsken T, Metzen JH, Hutter F. Neural architecture search: A survey. J
Mach Learn Res 2019;20:1–21.
Please cite this article as: Z. Zhao, T. Li, J. Wu et al., Deep learning algorithms for rotating machinery intelligent diagnosis: An open source benchmark study. ISA Transactions
(2020), https://doi.org/10.1016/j.isatra.2020.08.010.