Next Article in Journal
Development of an Internet of Things-Based Ultra-Pure Water Quality Monitoring System
Previous Article in Journal
Data Reconstruction Methods in Multi-Feature Fusion CNN Model for Enhanced Human Activity Recognition
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Multi-Scale Self-Supervision Approach for Bearing Anomaly Detection Using Sensor Data Under Multiple Operating Conditions

College of Science and Technology, Ningbo University, Ningbo 315300, China
*
Author to whom correspondence should be addressed.
Submission received: 15 January 2025 / Revised: 3 February 2025 / Accepted: 13 February 2025 / Published: 15 February 2025
(This article belongs to the Section Fault Diagnosis & Sensors)

Abstract

:
Early fault detection technologies play a decisive role in preventing equipment failures in industrial production. The primary challenges in early fault detection for industrial applications include the severe imbalance of time-series data, where normal operating data vastly outnumber anomalous data, and in some cases, anomalies may be virtually absent. Additionally, the frequent changes in operational modes during machinery operation further complicate the detection process, making it difficult to effectively identify faults across varying conditions. This study proposes a bearing early anomaly detection method based on contrastive learning and reconstruction approaches to address the aforementioned issues. The raw time-domain vibration data, which were collected from sensors mounted on the bearings of the machinery, are first preprocessed using the Ricker wavelet transform to effectively remove noise and extract useful signal components. These processed signals are then fed into a BYOL-based contrastive learning network to learn more discriminative global feature representations. In addition, we design the reconstruction loss to complement contrastive learning. By reconstructing the masked original data, the reconstruction loss forces the model to learn detailed information, thereby emphasizing the preservation and restoration of local details. Our model not only eliminates the reliance on negative samples found in mainstream unsupervised methods but also captures data features more comprehensively, achieving superior fault detection accuracy under different operating conditions compared to related methods. Experiments on the widely used CWRU multi-condition-bearing fault dataset demonstrate that our method achieves an average fault detection accuracy of 96.97%. Moreover, the experimental results show that on the full-cycle IMS dataset, our method detects early faults at least 2.3 h earlier than the other unsupervised methods. Furthermore, the validation results for the full-cycle XJTU-SY dataset further demonstrate its excellent generalization ability.

1. Introduction

Bearings are critical components in many industrial machines, and even minor damage can gradually escalate, leading to significant malfunctions. By leveraging advanced early anomaly detection techniques, it becomes possible to detect the initial signs of bearing wear, cracks, or other forms of damage. This timely detection not only allows for timely maintenance but also contributes to a substantial reduction in both equipment downtime and maintenance costs [1,2,3].
In the realm of anomaly detection, methodologies are broadly bifurcated into supervised and unsupervised paradigms, predicated on the exigency for annotated data. Supervised anomaly detection techniques are contingent upon meticulously labeled datasets to discern the quintessence of normal and anomalous states for prognostic endeavors. During the pedagogical phase, these techniques harness labeled data to architect models that adjudicate nascent samples based on the assimilated patterns. A cohort of relatively orthodox supervised algorithms, such as support vector machines (SVMs) [4] and decision trees [5], are ensconced within this category. Nonetheless, these classical methodologies grapple with impediments in feature extraction. The prodigious evolution of deep learning techniques in recent epochs has ameliorated this deficiency and has been ubiquitously employed in anomaly detection tasks. Convolutional neural networks (CNNs) [6] and long short-term memory (LSTM) networks [7] epitomize two salient exemplars. CNN and LSTM epitomize quintessential neural network architectures and manifest prodigious prowess in feature extraction for spatial and sequential data, respectively. Savants have proffered pertinent methodologies [8,9,10,11] and have executed experiments on a panoply of predictable and unpredictable time-series datasets. The empirical evidence has evinced their robustness in both ephemeral and protracted time-series analyses, underscoring the efficacy of these deep learning-centric approaches in anomaly detection.
However, in the crucible of practical industrial applications, the procurement of labeled data is often beleaguered by challenges due to its exorbitant cost and temporally intensive nature [12,13]. Consequently, the deployment of supervised methods is circumscribed to scenarios where data are pre-labeled [14]. In riposte to these challenges, unsupervised anomaly detection methodologies have burgeoned as a pivotal research trajectory. These methodologies dissect data structures to pinpoint anomalies, thereby attenuating the reliance on labeled data [15]. They evince remarkable advantages in the incipient detection of bearing anomalies, especially when labeled data are sparse or arduous to procure [16].
To this end, a compendium of unsupervised methods has been cultivated. Huang et al. [17,18] introduced the first-kind tensor singular value decomposition (1K-FTSVD) and an adaptive signal decomposition technique christened second-kind tensor singular spectrum decomposition (2K-FTSSD). These innovations surmount the inherent limitations of traditional tensor SVD predicated on the n-mode product, as well as the quandaries of non-unique optimization results and non-pseudo-diagonal core tensors. They eclipse traditional methods in the incipient detection of anomalies. Kang et al. [19] propounded a method denominated DIDAD, predicated on autoencoders, which distills features from high-frequency signals and refines anomaly detection. Other unsupervised anomaly detection methods, such as USSCNN, a CNN-based model proffered by Song et al. [20], UODA, a model predicated on autoencoders and RNNs proposed by Lu et al. [21], and FDDA, a method based on deep neural networks (DNN) and LSTM posited by Lu et al. [22], all evince superlative performance in feature extraction and excel in manipulating disparate types of data.
Despite the strides made by unsupervised methods in vanquishing the challenges concomitant with the need for specialized knowledge and prior experience with labeled data, their implementation in veritable industrial milieus remains formidable. Particularly in the context of bearing anomaly detection, industrial samples often suffer from a pronounced imbalance in time-series data. Normal data are copious, while anomalous data are scarce or even nonexistent. This imbalance exacerbates the conundrum of model generalization, as the paucity of anomalous data stymies the model’s capacity to efficaciously learn the features of anomalies. Moreover, the vibration signal data of bearings under disparate operating conditions can exhibit substantial variations [23]. These variations can precipitate shifts in data distribution, which in turn undermine the model’s generalization capabilities. A model that excels under certain conditions may falter under others, rendering it intricate and challenging to directly apply extant unsupervised anomaly detection methods in authentic industrial settings.
To tackle the critical issues of data imbalance and variations in operating conditions in real industrial environments, we specifically devise a novel early anomaly detection model named ADBR (Anomaly Detection method, based on the BYOL structure and Reconstructed structures), and the main contributions of this study are as follows:
(1)
To address the issue of scarce or nonexistent abnormal samples in real industrial data, this paper presents an ADBR unsupervised model. Based on the BYOL contrastive structure, this model effectively trains the anomaly detection model while completely eliminating the reliance on negative sample pairs, which is common in other unsupervised models.
(2)
To address the frequent variations in operating conditions, the proposed ADBR model learns global features while retaining and reconstructing local details, enabling it to focus on both the overall structure and local nuances of the signals. By combining global- and local-scale learning strategies, the model comprehensively captures data features under different operating conditions, achieving superior accuracy in bearing anomaly detection compared to related methods across various conditions.
(3)
Experiments on the widely used CWRU multi-condition bearing fault dataset demonstrate that our method achieves an average fault detection accuracy of 96.97%. Moreover, the experimental results show that on the full-cycle IMS dataset, our method detects early faults at least 2.3 h earlier than the other unsupervised methods. Furthermore, the validation results for the full-cycle XJTU-SY dataset further demonstrate its excellent generalization ability.
The remainder of this paper is organized as follows. Section 2 provides a detailed description of the proposed method, and Section 3 describes the details of the experimental data used. Section 4 describes the data preprocessing method. Section 5 presents the specific experiments and analyzes the results. Section 6 summarizes the results of this study.

2. Methodology

ADBR primarily consists of three key components. First, we utilize the Ricker wavelet to process the original one-dimensional time-series data, which helps remove noise while extracting useful signal components. Next, we employ a BYOL-based network structure to extract global features. Bootstrap Your Own Latent (BYOL) model [24] trains without needing negative sample pairs, making it suitable for real-world industrial data. It focuses on bringing positive sample pairs closer together, emphasizing the relationships between samples rather than their absolute values. This strategy not only enhances robustness against noise and interference, which are common challenges in industrial settings, but also breaks the dependency on negative samples that is characteristic of other contrastive models. Additionally, the Ricker wavelet is highly sensitive to sharp changes and transient signals. Anomalies in industrial time-series data typically manifest as brief spikes. Therefore, we design a reconstructed structure to extract local features. The reconstruction loss compels the model to focus more on the detailed information of the data by comparing the original and reconstructed data. By integrating global- and local-scale learning strategies, the model can comprehensively capture data features for early anomaly detection of bearings under multiple operating conditions, thereby enhancing the precision of anomaly detection across diverse operational conditions.
Figure 1 illustrates the detailed workflow, which consists of three parts: the Ricker wavelet transform, BYOL-based network structure, and reconstruction structure. Table 1 provides detailed notes on all the symbols in this section.

2.1. Ricker Wavelet Transform

For a given segment of the normal sample, the raw time-series data D normal = [ X 1 , X 2 , , X n ] R t × n , where each X i consists of
X i = [ x 1 i , x 2 i , , x t i ] T R t ,
The Ricker wavelet transform is defined as
ψ ( t ) = 1 t 2 σ 2 e t 2 2 σ 2 .
Applying the Ricker wavelet transform to each sample X i yields the corresponding time–frequency representation P i :
P i = X i × ψ ( t ) .
Thus, we obtain set P, where P = [ P 1 , P 2 , , P n ] .

2.2. BYOL-Based Network Structure

After obtaining the time–frequency representation set P, we input it into f θ to decode it into features y and y , which are subsequently fed into the online and target network structures, respectively. As shown in Figure 1, the network structure based on the BYOL model consists of online and target networks. The online network includes a projection head g θ and prediction head q θ , whereas the target network contains a projection head z ξ , where ξ is the exponential moving average of θ :
ξ τ ξ + ( 1 τ ) θ .
In the model implementation, we use ResNet-18 as the base model for the encoder, employing pretrained weights from PyTorch. The output dimension of the encoder is 512, whereas that of the online network projection head is modified to 2 for anomaly detection. The projection and prediction heads consist of a combination of linear and batch normalization layers. For each time–frequency representation P i , the outputs of the online and target networks are denoted by q θ ( z θ ) and z ξ , respectively. Therefore, we have
L contra = 1 n i = 1 n q θ ( z θ ) z ξ 2 2 .

2.3. Reconstruction Structure

To further enhance the performance of the model, we introduce the reconstruction loss. During training, P i is simultaneously masked to generate set P i . The masked P i is input into the encoder and reconstructed back to the original time–frequency representation P ^ i by the decoder. The decoder is a simple, fully connected network that reconstructs the original image from the output of the encoder.
The reconstruction loss is calculated as follows:
L recon = 1 n i = 1 n P i P ^ i 2 2 .

2.4. Objective Function

The combination of contrastive learning and reconstruction loss utilizes two different training objectives, with the former focusing on the alignment of global features, and the latter emphasizing the recovery of local details. This dual-scale training strategy enhances the generalization and representation capabilities of the model, resulting in improved performance in various industrial scenarios. The overall objective function is given by
L = α L contra + ( 1 α ) L recon .
where α is a hyperparameter used to balance the importance of the contrastive loss and reconstruction loss. The following Algorithm 1 demonstrates a more detailed process.
Algorithm 1: ADBR
  Input: D train 70 % of D normal
  Input: X i , X j Randomly select two samples X i and X j from D train
  Output: g θ
  Data: Initialize: all parameters
1.   Training:
2.    P i = X i ψ , P j = X j ψ
3.    z 1 g θ ( f θ ( P i ) ) , z 2 g θ ( f θ ( P j ) )
4.    z 1 g ξ ( f ξ ( P j ) ) , z 2 g ξ ( f ξ ( P i ) )
5.    L contra = 1 2 q θ ( z 1 ) z 1 2 2 + q θ ( z 2 ) z 2 2 2
6.    L recon = 1 2 P i P ^ i 2 2 + P j P ^ j 2 2
7.    L = α L contra + ( 1 α ) L recon

3. Experimental Setup

In the experimental section, we select three classic bearing datasets: the CWRU [25] abnormal bearing dataset, and the IMS [26,27,28] and XJTU-SY [29] full-cycle datasets. The datasets from CWRU, IMS, and XJTU-SY all contain rolling bearing data. The CWRU dataset contains the vibration signals of bearings under different operating conditions, making it suitable for testing the anomaly detection capability of the model across multiple conditions. The IMS and XJTU-SY datasets covered the complete degradation process of the bearings from normal to failure, allowing us to verify the performance of the model for early anomaly detection. To validate the accuracy of our method for early anomaly detection in bearings, we first use the classic CWRU abnormal bearing dataset to demonstrate our method’s ability to distinguish between positive and negative samples in various conditions, and then use the full-cycle bearing data to show the effectiveness and generalization ability of our method in detecting early anomalies. Table 2 presents the configuration of our training environment. Table 3 presents the training details of our model.

3.1. Preparation of Data

3.1.1. Abnormal Bearing Dataset

The CWRU dataset [30] contains vibration signals collected from motor bearings, covering various types of bearing faults. Its working platform is shown in Figure 2. The data are acquired under different operating conditions with corresponding loads of 0 horsepower, 1 horsepower, 2 horsepower, and 3 horsepower. The vibration signals are collected using accelerometers mounted on the bearing housing with a sampling frequency of 12 kHz. The operating conditions in the CWRU dataset are categorized into four types: (1) normal state, no faults; (2) inner race fault, where defects are present in the inner race of the bearing; (3) outer race fault, where defects are present in the outer race of the bearing; and (4) rolling element fault, where issues occur in the rolling elements of the bearing. In the fault states, the sizes of the bearing defects are categorized into three types: 0.007, 0.014, and 0.021, corresponding to different levels of fault severity. Table 4 lists all abnormal types.

3.1.2. Full-Cycle IMS Bearing Dataset

The IMS dataset [26] contains vibration signal data spanning from normal operating conditions to the occurrence of bearing faults. Data are collected from four identical bearings driven by a motor. The motor speed is maintained at 2000 rpm, and a radial load of 6000 lbs is applied using a spring mechanism. The test platform is illustrated in Figure 3. In the first dataset, two high-sensitivity accelerometers are installed on each bearing housing, and 1 s of vibration signals is recorded every 10 min (the first 43 files are recorded every 5 min). This dataset includes 2156 files, with the first 1000 files representing normal operating conditions. At the end of the experiment, an inner race failure is detected in the third bearing, a rolling element failure occurs in the fourth bearing, and the remaining two bearings remain intact.
In the second dataset, only one high-sensitivity accelerometer is mounted on each bearing housing, and the recording interval is maintained at 1 s every 10 min. This dataset comprises 984 files, with the first 400 files representing normal conditions. By the end of the experiment, only the outer race of the first bearing fails, whereas the other three bearings remain healthy.
Both datasets have a sampling frequency of 20,480 Hz, with more detailed specifications provided in Table 5. This dataset offers extensive vibration signal information for evaluating the performance of models in bearing fault detection, particularly in the early fault detection and identification of different fault types.

3.1.3. Full-Cycle XJTU-SY Dearing Dataset

The full-cycle XJTU-SY bearing fault data [31] are collected using the testing platform shown in Figure 4, which includes an AC motor, a motor speed controller, a support shaft, two support bearings, two accelerometers, and a test bearing. This testing platform is capable of conducting accelerated degradation testing, providing experimental data that closely reflect real-world conditions, and comprehensively capturing the entire lifecycle of bearings during the degradation process. The dataset used in this study includes vibration signals from three types of naturally occurring bearing faults: inner race faults (IF), outer race faults (OF), and cage faults (CF). Detailed data are presented in Table 6. The vibration signals are collected via accelerometers at a sampling frequency of 25.6 kHz, with a sampling interval of 1 min and a duration of 1.28 s for each sampling event.

4. Data Preprocessing

For the CWRU dataset, normal samples under all operating conditions are divided into training, testing, and validation sets at a ratio of 7:2:1. The training set is used for model training, whereas the validation set is used to evaluate the model performance during training. Subsequently, the detection effectiveness of the model is tested using a testing set that included normal and various types of abnormal data under different operating conditions. For the full-cycle IMS and XJTU-SY, only normal sample files are used as the training set, whereas the remaining files are used for testing. Each sample uniformly contains 2048 data points.

5. Experimental Results and Discussion

5.1. Results on Abnormal Bearing Dataset

In this section, we select samples from the CWRU dataset to evaluate the capability of the proposed model for bearing anomaly detection under multiple operating conditions and compare it with several classic supervised networks. The experimental results for the multi-condition CWRU dataset are presented in Table 7.
Our proposed self-supervised model demonstrates a high accuracy rate in detecting ten abnormal states under multiple conditions, with an average accuracy surpassing that of supervised neural networks (NNs) and deep neural networks (DNNs), which utilize negative samples. Although its performance is only slightly lower than that of supervised CNN and deep convolutional neural networks (DCNNs), as a self-supervised model that does not rely on labels, achieving performance close to that of some supervised models represents a significant advancement. This indicates that the proposed model can effectively reflect system changes under multiple conditions. A learning strategy that combines global and local scales allows the model to capture data features more comprehensively for the detection of anomalies in bearings under various conditions, thus improving the precision of early anomaly detection in the full-cycle dataset.
To illustrate the effectiveness of the combined global- and local-scale learning strategy visually, we conduct an ablation experiment comparing the anomaly detection capabilities of our proposed ADBR model with those of the BYOL model alone under multiple conditions. We train both models on identical training sets and employ a uniform data preprocessing method to evaluate their performance on the same test set, allowing us to compare the contribution of the reconstruction loss. Figure 5 shows the accuracy rates of the two models for bearings with different types of losses under multiple conditions.
The results show that our method significantly outperforms the detection accuracy of various anomaly states under multiple operating conditions compared to the performance achieved by using only the BYOL network architecture. Experiments demonstrated that eliminating the reconstruction structure leads to a substantial decrease in accuracy, which more intuitively illustrates how utilizing the reconstruction loss enhances the model’s focus on local scales, thereby contributing to improved anomaly detection accuracy across multiple conditions.
To evaluate the impact of the weight coefficient α for contrastive learning loss and reconstruction loss on model performance, we design a series of experiments to analyze the effects of different α values on the model’s performance.
During the training process, we select different α values ranging from [0.1:0.1:0.9] to balance the weights of the contrastive learning loss and reconstruction loss.
As shown in Figure 6, when the loss weight coefficient α of the ADBR model is set to 0.6, the model achieves the highest accuracy in detecting various bearing anomalies, demonstrating optimal model performance. Based on the experimental results, we select α = 0.6 as the loss weight coefficient for the ADBR model, with the corresponding loss function formulated as follows:
L = 0.6 L contra + 0.4 L recon .

5.2. Results on Full-Cycle Bearing Dataset

In this section, we validate and compare the early anomaly detection capability of our model for bearings through experimental results on a full-cycle dataset, while also assessing the generalization ability of the model.

5.2.1. Results on the IMS Dataset

Each sample contains 2048 data points; therefore, a single time file may contain multiple samples. We conduct experiments on two sets of full-cycle data from the IMS dataset and compare the results for normal and abnormal bearings. The results indicate that when the number of abnormal samples in a data folder is greater than or equal to four, the bearing is considered abnormal. Figure 7 shows that in the first set of IMS data, the number of abnormal samples in all time files of the first and second bearings is below the threshold, and no anomalies are detected. An anomaly is detected in the 1206th time file of the third bearing and the 1416th time file of the fourth bearing, while the other two normal bearings show no anomalies. These experimental results are fully consistent with the final state of the bearings. Figure 8 shows that in the second set of IMS data, an anomaly is detected in the 513th time file of the first bearing, while the number of abnormal samples in all time files of the other three bearings is below the threshold, and no anomalies are detected, classifying them as normal. This outcome also aligns with the final state of the bearings in the dataset.
The detected final states of the bearings are aligned with the final states provided by the dataset. However, to assess the reliability of the early anomaly points identified by the ADBR model, the root-mean-square (RMS) method is used for verification. RMS is sensitive to changes in energy within the vibration signals because bearing abnormalities are typically associated with increases in the amplitude of the vibration signals. For instance, wear, pitting, or cracks in the rolling elements or inner and outer rings can lead to a significant increase in the energy of the vibration signal, resulting in a corresponding increase in RMS values. Therefore, RMS is effective in detecting changes in bearing conditions. We compare the anomaly detection results of these three bearings with their RMS curves as shown in Figure 9. We observe that our model successfully detects bearing anomalies in all cases before significant changes occur in the RMS curves. This demonstrates that our method is capable of detecting anomalies prior to the significant deterioration of bearing conditions (manifested as a notable increase in RMS values), indicating that its detection results are highly reliable. Such reliability is crucial for decision-making support in practical industrial applications.
To further validate the performance of our proposed ADBR model compared to other anomaly-detection methods, we conduct comparative experiments on the IMS dataset, specifically the Bearing2_1 dataset, using the early anomaly-detection methods listed in Table 8.
Among these, methods 1–3 [33,34,35] are classical anomaly detection techniques, while methods 4–12 employ a combination of feature extraction and detection. Methods 13–17 are unsupervised anomaly detection methods similar to our approach, and Figure 10 illustrates the locations of the anomaly samples detected by various anomaly detection methods on Bearing2_1 of the IMS dataset.
Based on the results in Table 8, we find that the ADBR model detects an anomaly in the 513th data file, which occurs after 85.5 h of testing. This result is at least 50 data files (approximately 8.3 h) earlier than those of classical anomaly detection methods. In comparison, methods that combine feature extraction and detection identify early bearing anomalies in at least 22 data files (approximately 3.6 h). Additionally, compared to similar popular unsupervised models, the ADBR model detects early bearing anomalies in at least 14 data files (approximately 2.3 h). These results are obtained under consistent experimental conditions by using a publicly available dataset for early bearing anomaly detection. Therefore, it can be concluded that by integrating global- and local-scale learning strategies, the model can capture data features more comprehensively in anomaly detection, leading to the superior performance of our ADBR model compared to other unsupervised models in detecting early-bearing anomalies.

5.2.2. Results on the XJTU-SY Dataset

This section further validates the generalization capability of the proposed model using another full-cycle bearing dataset, XJTU-SY, to comprehensively evaluate the performance of the model under prolonged operating conditions. For abnormal bearings, the amplitude gradually increases over time, and this significant increase in amplitude is a typical signal of equipment anomalies, indicating a sharp increase in vibration intensity at the onset of failure. Therefore, we compare the model results with the amplitude plots. As shown in Figure 10, the ADBR model detects anomalies before these changes became apparent. Therefore, in two early anomaly detection experiments conducted on two full-cycle datasets, the model successfully detects early anomalies in the bearings, which not only demonstrates its superior early anomaly detection ability compared to similar models but also verifies its generalization capability.

6. Conclusions

Based on the findings of this study, we propose a novel early anomaly detection model for bearings, named ADBR. The model combines a BYOL-based network structure with a reconstruction approach, enabling it to learn data features across multiple operating conditions at both the global and local scales, while being trained solely on positive samples. This approach effectively addresses two key challenges in industrial anomaly detection: the lack of negative samples and the frequent changes in the operating environment of the equipment. Experimental results on the multi-condition CWRU anomaly bearing dataset show that the ADBR model achieves an average accuracy of 96.97% in detecting various bearing anomaly states, demonstrating its great potential for early anomaly detection. Notably, when applied to the IMS dataset, the ADBR model detects faults at least 2.3 h earlier than traditional algorithms and other unsupervised methods, highlighting its significant advantage in early fault prediction. Finally, the model’s generalization ability is further validated using the XJTU-SY dataset. Future work will focus on exploring the impact of additional operational factors, such as lubrication and temperature variations, on the bearing fault detection model. These factors can significantly influence the operating conditions and fault characteristics of the equipment. By incorporating them, the model’s robustness can be further enhanced, enabling it to maintain high efficiency in fault detection even under complex and dynamic industrial environments. Moreover, accounting for these environmental factors will broaden the model’s applicability, allowing it to better handle bearing fault diagnosis tasks across a wider range of operational conditions, thus increasing its value and reliability in practical industrial applications.

Author Contributions

Conceptualization, Z.D. and Y.C.; methodology, Z.D. and Y.C.; software, Z.D.; validation, Z.D.; formal analysis, Y.C.; investigation, F.L.; resources, L.J.; data curation, Y.C.; writing—original draft preparation, Z.D.; writing—review and editing, Y.C. and Z.D.; visualization, Z.D.; supervision, Y.C.; project administration, Y.C. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Natural Science Foundation of China [Grant number 12304513]; and Ningbo Natural Science Foundation [Grant number 2022J138].

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The CRWU dataset, the IMS dataset, and the XJTU-SY dataset are used in this article (all three datasets are open source on the web).

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Martí, L.; Sanchez-Pi, N.; Molina, J.M.; Garcia, A.C.B. Anomaly Detection Based on Sensor Data in Petroleum Industry Applications. Sensors 2015, 15, 2774–2797. [Google Scholar] [CrossRef] [PubMed]
  2. Khan, S.; Yairi, T. A review on the application of deep learning in system health management. Mech. Syst. Signal Process. 2018, 107, 241–265. [Google Scholar] [CrossRef]
  3. Nandi, S.; Toliyat, H.A.; Li, X. Condition Monitoring and Fault Diagnosis of Electrical Motors—A Review. IEEE Trans. Energy Convers. 2005, 20, 719–729. [Google Scholar] [CrossRef]
  4. Widodo, A.; Yang, B.-S. Support vector machine in machine condition monitoring and fault diagnosis. Mech. Syst. Signal Process. 2007, 21, 2560–2574. [Google Scholar] [CrossRef]
  5. Sugumaran, V.; Muralidharan, V.; Ramachandran, K.I. Feature selection using Decision Tree and classification through Proximal Support Vector Machine for fault diagnostics of roller bearing. Mech. Syst. Signal Process. 2007, 21, 930–942. [Google Scholar] [CrossRef]
  6. Kanarachos, S.; Christopoulos, S.R.G.; Chroneos, A.; Fitzpatrick, M.E. Detecting anomalies in time series data via a deep learning algorithm combining wavelets, neural networks and Hilbert transform. Expert Syst. Appl. 2017, 85, 292–304. [Google Scholar] [CrossRef]
  7. Taylor, A.; Leblanc, S.; Japkowicz, N. Anomaly Detection in Automobile Control Network Data with Long Short-Term Memory Networks. In Proceedings of the 2016 IEEE International Conference on Data Science and Advanced Analytics (DSAA), Montreal, QC, Canada, 17–19 October 2016; pp. 130–139. [Google Scholar] [CrossRef]
  8. Wen, T.; Keyes, R. Time Series Anomaly Detection Using Convolutional Neural Networks and Transfer Learning. arXiv 2019, arXiv:1905.13628. [Google Scholar]
  9. Malhotra, P.; Ramakrishnan, A.; Anand, G.; Vig, L.; Agarwal, P.; Shroff, G. LSTM-based Encoder-Decoder for Multi-sensor Anomaly Detection. arXiv 2016, arXiv:1607.00148. [Google Scholar]
  10. Lin, S.; Clark, R.; Birke, R.; Schönborn, S.; Trigoni, N.; Roberts, S. Anomaly Detection for Time Series Using VAE-LSTM Hybrid Model. In Proceedings of the ICASSP 2020—IEEE International Conference on Acoustics, Speech and Signal Processing, Barcelona, Spain, 4–8 May 2020; pp. 4322–4326. [Google Scholar] [CrossRef]
  11. Dong, Z.; Jiang, Y.; Jiao, W.; Zhang, F.; Wang, Z.; Huang, J.; Wang, X.; Zhang, K. Double attention-guided tree-inspired grade decision network: A method for bearing fault diagnosis of unbalanced samples under strong noise conditions. Adv. Eng. Inform. 2025, 103004. [Google Scholar] [CrossRef]
  12. Jiang, W.; Hong, Y.; Zhou, B.; He, X.; Cheng, C. A GAN-Based Anomaly Detection Approach for Imbalanced Industrial Time Series. IEEE Access 2019, 7, 143608–143619. [Google Scholar] [CrossRef]
  13. Wang, D.; Shang, Y. A new active labeling method for deep learning. In Proceedings of the 2014 International Joint Conference on Neural Networks (IJCNN), Beijing, China, 6–11 July 2014; pp. 112–119. [Google Scholar] [CrossRef]
  14. Love, B.C. Comparing supervised and unsupervised category learning. Psychon. Bull. Rev. 2002, 9, 829–835. [Google Scholar] [CrossRef]
  15. Chandola, V. Anomaly Detection: A Survey. ACM Comput. Surv. 2009, 41, 1–58. [Google Scholar] [CrossRef]
  16. Pimentel, M.A.; Clifton, D.A.; Clifton, L.; Tarassenko, L. A review of novelty detection. Signal Process. 2014, 99, 215–249. [Google Scholar] [CrossRef]
  17. Huang, J.; Zhang, F.; Coombs, T.; Chu, F. The first-kind flexible tensor SVD: Innovations in multi-sensor data fusion processing. Nonlinear Dyn. 2024, 1–19. [Google Scholar] [CrossRef]
  18. Huang, J.; Zhang, F.; Safaei, B.; Qin, Z.; Chu, F. The flexible tensor singular value decomposition and its applications in multisensor signal fusion processing. Mech. Syst. Signal Process. 2024, 220, 111662. [Google Scholar] [CrossRef]
  19. Kang, Y.; Chen, G.; Wang, H.; Pan, W.; Wei, X. A New Dual-Input Deep Anomaly Detection Method for Early Faults Warning of Rolling Bearings. Sensors 2023, 23, 8013. [Google Scholar] [CrossRef] [PubMed]
  20. Song, W.; Shen, W.; Gao, L.; Li, X. An Early Fault Detection Method of Rotating Machines Based on Unsupervised Sequence Segmentation Convolutional Neural Network. IEEE Trans. Instrum. Meas. 2022, 71, 3504712. [Google Scholar] [CrossRef]
  21. Lu, W.; Cheng, Y.; Xiao, C.; Chang, S.; Huang, S.; Liang, B.; Huang, T. Unsupervised Sequential Outlier Detection with Deep Architectures. IEEE Trans. Image Process. 2017, 26, 4321–4330. [Google Scholar] [CrossRef] [PubMed]
  22. Lu, W.; Li, Y.; Cheng, Y.; Meng, D.; Liang, B.; Zhou, P. Early Fault Detection Approach with Deep Architectures. IEEE Trans. Instrum. Meas. 2018, 67, 1679–1689. [Google Scholar] [CrossRef]
  23. He, M.; He, D. Deep Learning Based Approach for Bearing Fault Diagnosis. IEEE Trans. Ind. Appl. 2017, 53, 3057–3065. [Google Scholar] [CrossRef]
  24. Grill, J.B.; Strub, F.; Altché, F.; Tallec, C.; Richemond, P.; Buchatskaya, E.; Doersch, C.; Avila Pires, B.; Guo, Z.; Gheshlaghi Azar, M.; et al. Bootstrap your own latent—A new approach to self-supervised learning. Adv. Neural Inf. Process. Syst. 2020, 33, 21271–21284. [Google Scholar]
  25. Yu, J. Local and nonlocal preserving projection for bearing defect classification and performance assessment. IEEE Trans. Ind. Electron. 2012, 59, 2363–2376. [Google Scholar] [CrossRef]
  26. Nectoux, P.; Gouriveau, R.; Medjaher, K. PRONOSTIA: A Project for Prognostics of Bearings. Int. J. Progn. Health Manag. 2012, 3, 1–6. Available online: http://www.phmsociety.org/ (accessed on 1 February 2024).
  27. Li, H.; Wang, Y. Rolling bearing reliability estimation based on logistic regression model. In Proceedings of the International Conference on Quality, Reliability, Risk, Maintenance, and Safety Engineering (QR2MSE), Chengdu, China, 15–18 July 2013; pp. 1730–1733. [Google Scholar]
  28. Tobon-Mejia, D.A.; Medjaher, K.; Zerhouni, N.; Tripot, G. A data-driven failure prognostics method based on mixture of Gaussians hidden Markov models. IEEE Trans. Reliab. 2012, 61, 491–503. [Google Scholar] [CrossRef]
  29. Yu, J. Health condition monitoring of machines based on hidden Markov model and contribution analysis. IEEE Trans. Instrum. Meas. 2012, 61, 2200–2211. [Google Scholar] [CrossRef]
  30. Case Western Reserve University Bearing Data Center. Available online: https://engineering.case.edu/bearingdatacenter (accessed on 22 December 2023).
  31. Wang, B.; Lei, Y.; Li, N.; Li, N. A hybrid prognostics approach for estimating remaining useful life of rolling element bearings. IEEE Trans. Reliab. 2018, 69, 401–412. [Google Scholar] [CrossRef]
  32. Zhang, W.; Li, X.; Ding, Q. Deep residual learning-based fault diagnosis method for rotating machinery. ISA Trans. 2019, 95, 295–305. [Google Scholar] [CrossRef] [PubMed]
  33. Guo, G.; Wang, H.; Bell, D.; Bi, Y.; Greer, K. KNN Model-Based Approach in Classification. In On the Move to Meaningful Internet Systems 2003: CoopIS, DOA, and ODBASE; Meersman, R., Tari, Z., Schmidt, D.C., Eds.; Lecture Notes in Computer Science; Springer: Berlin/Heidelberg, Germany, 2003; Volume 2888, pp. 986–996. [Google Scholar]
  34. Alghushairy, O.; Alsini, R.; Soule, T.; Ma, X. A Review of Local Outlier Factor Algorithms for Outlier Detection in Big Data Streams. BDCC 2020, 5, 1. [Google Scholar] [CrossRef]
  35. Li, K.L.; Huang, H.K.; Tian, S.F.; Xu, W. Improving one-class SVM for anomaly detection. In Proceedings of the 2003 International Conference on Machine Learning and Cybernetics, Xi’an, China, 5 November 2003; pp. 3077–3081. [Google Scholar] [CrossRef]
Figure 1. The detailed workflow: the three modules in the lower-left corner are expanded with the corresponding colors.
Figure 1. The detailed workflow: the three modules in the lower-left corner are expanded with the corresponding colors.
Sensors 25 01185 g001
Figure 2. CWRU bench.
Figure 2. CWRU bench.
Sensors 25 01185 g002
Figure 3. IMS bench.
Figure 3. IMS bench.
Sensors 25 01185 g003
Figure 4. XJTU-SY bench.
Figure 4. XJTU-SY bench.
Sensors 25 01185 g004
Figure 5. Accuracy of the two models for different anomaly detections.
Figure 5. Accuracy of the two models for different anomaly detections.
Sensors 25 01185 g005
Figure 6. Impact of α on model effectiveness.
Figure 6. Impact of α on model effectiveness.
Sensors 25 01185 g006
Figure 7. Result on dataset 1 of (a) bearing1_1, (b) bearing1_2, (c) bearing1_3, (d) bearing1_4.
Figure 7. Result on dataset 1 of (a) bearing1_1, (b) bearing1_2, (c) bearing1_3, (d) bearing1_4.
Sensors 25 01185 g007
Figure 8. Result on dataset 2 of (a) bearing2_1, (b) bearing2_2, (c) bearing2_3, (d) bearing2_4.
Figure 8. Result on dataset 2 of (a) bearing2_1, (b) bearing2_2, (c) bearing2_3, (d) bearing2_4.
Sensors 25 01185 g008
Figure 9. Comparison of the results of our method with the RMS curves for (a) bearing1_3, (b) bearing1_4, and (c) bearing2_1.
Figure 9. Comparison of the results of our method with the RMS curves for (a) bearing1_3, (b) bearing1_4, and (c) bearing2_1.
Sensors 25 01185 g009
Figure 10. Comparison of the results of our method with the amplitude diagrams of (a) bearing2_1, (b) bearing2_2, (c) bearing2_3, (d) bearing2_4, and (e) bearing2_5.
Figure 10. Comparison of the results of our method with the amplitude diagrams of (a) bearing2_1, (b) bearing2_2, (c) bearing2_3, (d) bearing2_4, and (e) bearing2_5.
Sensors 25 01185 g010
Table 1. List of symbols and their definitions.
Table 1. List of symbols and their definitions.
SymbolDefinition
D normal The raw time-series data from normal samples
X i A sample from the time-series data, where X i R t
ψ ( t ) The Ricker wavelet function, defined in Equation (1)
P i The time–frequency representation of the sample X i after the Ricker
wavelet transform
g θ The projection head of the online network
q θ The prediction head of the online network
L contra The contrastive loss, defined in Equation (4)
L recon The reconstruction loss, defined in Equation (5)
L The overall objective function, defined in Equation (6)
Table 2. Training environment configuration.
Table 2. Training environment configuration.
ComponentConfiguration
Operating SystemWindows 10
CPU12th Gen Intel(R) Core(TM) i7-12700K
GPUNVIDIA GeForce RTX 3090
RAM32 GB
StorageHS-SSD-A4000 2048G
Python VersionPython 3.8.18
Deep Learning FrameworkPyTorch 2.3.0
CUDA Version11.8
Table 3. Training parameters of the ADBR model.
Table 3. Training parameters of the ADBR model.
ParameterValue
EncoderResNet-18
LossMSE
OptimizerAdam
Learning Rate0.0003
Epoch200
Batch Size64
ShuffleTrue
Table 4. Ten types of anomaly.
Table 4. Ten types of anomaly.
TypeCondition TypeDefect Severity
0Normal0
1IRF0.007
2IRF0.014
3IRF0.021
4BF0.007
5BF0.014
6BF0.021
7ORF0.007
8ORF0.014
9ORF0.021
Table 5. Details of the IMS dataset.
Table 5. Details of the IMS dataset.
Date
Set
Number of
Files
Number of
Positive Sample Files
Abnormal
Bearings
Dataset 1  2156 1000   1_3 and 1_4
Dataset 2  984 400   2_1
Table 6. Details of the XJTU-SY dataset.
Table 6. Details of the XJTU-SY dataset.
Bearing
Dataset
Number of
Files
Number of
Positive Sample Files
Fault
Element
Bearing 2_1   491  315  Inner race
Bearing 2_2  161   30  Outer race
Bearing 2_3  533  225  Cage
Bearing 2_4  42  20  Outer race
Bearing 2_5  339  95  Outer race
Table 7. Comprehensive comparison of classification accuracy of different fault diagnosis methods.
Table 7. Comprehensive comparison of classification accuracy of different fault diagnosis methods.
MethodFault TypeAverage (%)
0 (%)1 (%)2 (%)3 (%)4 (%)5 (%)6 (%)7 (%)8 (%)9 (%)
NN10080.7565.5010082.2587.2579.7576.5099.507684.75
DNN10097.509210094.5094.5095.5092.7599.5083.2594.95
CNN10097.7590.5010099.5094.2598.2599.7510096.2597.63
DCNN [32]10010099.5010010010010010099.7510099.93
ADBR84.7610099.5810099.7995.3498.7310091.5310096.97
Table 8. Results on Bearing2_1 by various anomaly detection methods.
Table 8. Results on Bearing2_1 by various anomaly detection methods.
MethodResults     MethodResults     Method     Results
1.KNN563     7.RMS + LOF535     13.DIDAD [19]     533
2.LOF609     8.Kurtosis + LOF700     14.ANOGAN     533
3.one-class SVM609     9.SDAE + LOF700     15.USSCNN [20]     531
4.RMS + SVDD535     10.RMS + iForest535     16.UODA     531
5.Kurtosis + SVDD703     11.Kurtosis + iForest650     17.FDDA [22]     527
6.SDAE + SVDD600     12.SDAE + iForest720     18.ADBR     513
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Dai, Z.; Jiang, L.; Li, F.; Chen, Y. A Multi-Scale Self-Supervision Approach for Bearing Anomaly Detection Using Sensor Data Under Multiple Operating Conditions. Sensors 2025, 25, 1185. https://doi.org/10.3390/s25041185

AMA Style

Dai Z, Jiang L, Li F, Chen Y. A Multi-Scale Self-Supervision Approach for Bearing Anomaly Detection Using Sensor Data Under Multiple Operating Conditions. Sensors. 2025; 25(4):1185. https://doi.org/10.3390/s25041185

Chicago/Turabian Style

Dai, Zhuoheng, Lei Jiang, Feifan Li, and Yingna Chen. 2025. "A Multi-Scale Self-Supervision Approach for Bearing Anomaly Detection Using Sensor Data Under Multiple Operating Conditions" Sensors 25, no. 4: 1185. https://doi.org/10.3390/s25041185

APA Style

Dai, Z., Jiang, L., Li, F., & Chen, Y. (2025). A Multi-Scale Self-Supervision Approach for Bearing Anomaly Detection Using Sensor Data Under Multiple Operating Conditions. Sensors, 25(4), 1185. https://doi.org/10.3390/s25041185

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop