Sensors 24 06448
Sensors 24 06448
Article
KAN-HyperMP: An Enhanced Fault Diagnosis Model for
Rolling Bearings in Noisy Environments
Jun Wang 1 , Zhilin Dong 2, * and Shuang Zhang 3
1 Department of Ocean Engineering, Yantai Institute of Science and Technology, Yantai 265600, China;
[email protected]
2 School of Engineering, Zhejiang Normal University, Jinhua 321004, China
3 School of Computer Science and Technology, Anhui University, Hefei 230601, China; [email protected]
* Correspondence: [email protected]
Abstract: Rolling bearings often produce non-stationary signals that are easily obscured by noise,
particularly in high-noise environments, making fault detection a challenging task. To address
this challenge, a novel fault diagnosis approach based on the Kolmogorov–Arnold Network-based
Hypergraph Message Passing (KAN-HyperMP) model is proposed. The KAN-HyperMP model is
composed of three key components: a neighbor feature aggregation block, a feature fusion block, and
a KANLinear block. Firstly, the neighbor feature aggregation block leverages hypergraph theory to
integrate information from more distant neighbors, aiding in the reduction of noise impact, even when
nearby neighbors are severely affected. Subsequently, the feature fusion block combines the features
of these higher-order neighbors with the target node’s own features, enabling the model to capture
the complete structure of the hypergraph. Finally, the smoothness properties of B-spline functions
within the Kolmogorov–Arnold Network (KAN) are employed to extract critical diagnostic features
from noisy signals. The proposed model is trained and evaluated on the Southeast University (SEU)
and Jiangnan University (JNU) Datasets, achieving accuracy rates of 99.70% and 99.10%, respectively,
demonstrating its effectiveness in fault diagnosis under both noise-free and noisy conditions.
deep learning (ADL) [9]. The adoption of these technologies not only introduces a new
perspective on mechanical fault diagnosis but also fosters the advancement of the entire
industrial system towards greater efficiency and safety.
Graph theory models exhibit unique advantages in comprehensively describing fault
characteristic information. To effectively handle graph data, Graph Neural Networks
(GNNs) have emerged as a burgeoning field. Specifically designed for graph signal pro-
cessing, GNNs enable the precise definition of values and connections between nodes,
capturing and analyzing information from a global perspective. Recently, GNN technology
has been applied to fault diagnosis by researchers, in order to deepen their understanding
and address fault issues more effectively. GNNs enhance data extraction and inference
by aggregating information from neighbors at various depths. These networks have
been successfully applied in multiple domains, including physical models [10], chemical
structures [11], social networks [12], natural language processing [13], and image classifi-
cation [14]. For example, Li et al. [15] utilized GNNs to model and analyze graph data,
proposing three graph construction methods, exploring seven types of graph convolution
networks (GCNs), and four different graph pooling methods. They further developed an
intelligent fault diagnosis and predictive diagnosis framework based on GNNs. Addition-
ally, Zhao et al. [16] introduced a semi-supervised graph convolutional deep belief network
and applied it to electromechanical system fault diagnosis, which achieved significant
diagnostic results, even with limited labeled samples. These studies, which converted
vibration signals into graph data and utilized GNNs for fault diagnosis, demonstrate the
feasibility and advantages of GNNs in this field.
Graph-based models are becoming a prominent trend in rolling bearing fault diag-
nosis because they effectively capture the relationships between sample data. However,
traditional graph models are limited by their focus on learning pairwise correlations be-
tween adjacent samples, as each edge connects only two nodes, making them inadequate
for capturing the more complex higher-order relationships that are crucial in practical
applications [17]. For instance, during the monitoring of bearing degradation, consecutive
samples are not only interrelated but also collectively reflect the component’s gradual
deterioration. To illustrate the intricate relationships among multiple samples in fault
diagnosis, some researchers have turned to hypergraph structures to represent equipment
monitoring data. Hypergraphs connect multiple nodes through hyperedges, enabling a
more comprehensive depiction of complex relationships among multisample data. Conse-
quently, hypergraphs are used to represent intricate higher-order relationships between
vertices and model complex networks and systems with high-order interactions. Zhang
et al. [18] transformed motion current signals into a hypergraph structure and developed a
Hypergraph GCN (HGCN) to learn the higher-order relationships between nodes for fault
classification. Similarly, Shi et al. [19] transformed vibration signal samples into a hyper-
graph and mined the high-order structural information between samples using HGCN
layers. Yan et al. [20] structured the sample data into multiple hypergraph structures to
better learn the high-order data hidden among the samples. Additionally, Feng et al. [21]
introduced the Hypergraph Neural Network (HGNN), a model that naturally extends the
spectral method of GCN to hypergraphs, and designed corresponding hypergraph convolu-
tion operations. Meanwhile, Yadati et al. [22] developed the HyperGCN model, addressing
semi-supervised classification problems on hypergraphs. These advancements have pro-
moted the application of hypergraph models in fields such as computer vision [23,24],
recommendation systems [25,26], and spatiotemporal forecasting [27,28], achieving signif-
icant success. Notably, in the analysis of bearing monitoring data, utilizing hypergraph
methods to explore high-order relationships between samples offers a new perspective and
methodology for rolling bearing fault diagnosis.
To effectively capture higher-order relationships, Wang et al. [29] introduced T-spectral
convolution, a technique specifically designed for handling complex data structures, with
a particular strength in representing hypergraphs as tensors. By leveraging the multidi-
mensional characteristics of tensors, this method effectively captures complex inter-node
Sensors 2024, 24, 6448 3 of 20
2. Proposed Model
In the task of rolling bearing fault diagnosis, fault samples are unstructured, making
it challenging to construct a hypergraph that can represent the hidden structure within
sample data and across different samples. To address this issue, a hypergraph construction
method capable of capturing the data structure among fault samples is proposed, and a
corresponding neural network is developed based on the constructed hypergraph for fault
identification.
Finally, to ensure each sample is accurately classified, the samples obtained through
the above process are vertically stacked, forming a feature matrix X ∈ R N × D . The entire
process is illustrated in Figure 1.
Building on this analysis, research has introduced the hypergraph T-spectral Convolu-
tion [29], which leverages tensor representations and t-product decompositions to enable
the direct manipulation of hypergraph data in higher dimensions. This approach allows
models to handle higher-order relationships more naturally, overcoming the limitations
of traditional methods that reduce high-order hypergraphs to two-dimensional matrices.
The t-product, a powerful tool for complex algebraic operations, preserves the multidimen-
sional structure of the data, thereby capturing the deep structures and patterns within the
hypergraph. The formula is expressed as follows:
Zs = Anorm
s ∗ Xs ∗ Ws (7)
( M −2)
Here, Anorm
s is the normalized adjacency tensor, and Xs ∈ R N × D× N represents
the CNI signal tensor, defined as follows:
Given a feature (or signal) matrix X ∈ R N × D , where N is the number of nodes in the
hypergraph and D is the feature dimension for each node, the interaction of all nodes along
the d-th dimension (d = 1, . . . , D) is given by
( M −2)
CNI([x]d ) = [x]d ◦ [x]d . . . ◦ [x]d ∈ R N ×1× N (8)
| {z }
(M−1) times
where ◦ denotes the outer product (also known as the basic tensor product), and [x]d ∈ R N
represents the d-th dimensional feature vector of the nodes.
While T-spectra convoluation offers numerous theoretical advantages, such as the
ability to process high-order neighbor information, it also faces significant drawbacks,
including high computational complexity and substantial memory requirements. For in-
( M −2)
stance, in Equation (7), Xs ∈ R N × D× N describes a high-dimensional tensor. While
constructing such a tensor is feasible for small hypergraphs, it becomes impractical for
larger hypergraphs, such as those in this paper, due to computational limitations.
Figure 3. The architecture overview of our KAN-HyperMP. The raw signal is processed into the
final signal feature matrix X using techniques such as resampling and sliding window sampling.
An incidence matrix H is then constructed using the KNN algorithm, establishing a hypergraph
structure. Based on the hypergraph, the neighbor feature aggregation block extracts information from
high-order neighbor nodes. This information is then integrated with the node’s own information
through the feature fusion block. Finally, feature extraction is completed using the KANLinear block,
facilitating fault diagnosis.
Mth -order neighborhood. These concepts provide a crucial theoretical foundation for
understanding how the block processes data.
1. Mth -order neighborhood hyperedge set
In defining the Mth -order hyperedges within a hypergraph G = (V , E ), scenarios are
differentiated based on the number of nodes each hyperedge contains:
(
{ e }, if |e| = M,
eM = M M
(9)
ext (e) | |ext (e)| = M , if |e| < M
Based on this, an Mth -order neighborhood hyperedge set can be defined for each
hyperedge as follows:
n o
E M (v) = e M | e ∈ E , v ∈ e (10)
where e M \ {v} denotes the removal of the target node v from the set e M and the sort
function refers to the ordering of the remaining nodes. This structured definition
of neighborhoods offers an effective method for processing and analyzing hyper-
graph data, significantly enhancing the model’s comprehension of complex node
relationships.
For instance, consider a simple hypergraph as shown in Figure 5a, and based on
Equation (2); M = 3 is determined. According to the previously defined method,
hyperedge e1 is initially expanded to obtain ext3 (e1 ), as shown in Figure 5b. Based on
the previously defined criteria, the 3rd-order neighborhood hyperedge set for node v1
is determined, as shown in Equation (12).
(a) (b)
Figure 5. Construct the 3rd-order neighborhood hyperedge set for node v1 . (a) Hypergraph structure.
(b) Expand hyperedge.
n o
E3 (v1 ) = ext3 (e1 ), {e2 } = {{(v1 , v2 , v1 ), (v1 , v2 , v2 )}, {(v1 , v2 , v3 )}} (12)
where xv1 , xv2 , and xv3 are the feature vectors of nodes v1 , v2 , and v3 respectively, and P1 ,
P2 , P3 correspond to the combinatorial counts from sort(·). The ⊙ operation denotes the
Hadamard (element-wise) product along the feature dimension.
Step 2: Hyperedge weights.
Notably, hyperedge e1 includes two nodes, while e2 includes three. To capture the
variation among hyperedges during feature aggregation, a weight for each hyperedge (We )
is introduced, calculated as follows:
|e|
We = (18)
α
|e|
where α = ∑i=0 (−1)i (|ei |)(|e| − i ) M .
Therefore, the final 3rd-order neighborhood feature for node v1 is as follows:
FNv3 = We1 · P1 · ( xv1 ⊙ xv2 ) + We1 · P2 · ( xv2 ⊙ xv2 ) + We2 · P3 · ( xv2 ⊙ xv3 ) (19)
1
Repeating this process for all target nodes enables us to obtain neighbor features that
can be extended to the Mth -order, resulting in the final Mth -order neighbor features, as
defined in Equation (20).
where Fv ∈ R N × D represents the node’s own feature vector, and σ denotes the activation
function, with ReLU being the choice in this study. The function COMBINE is defined as
follows: h i
COMBINE( Fv , FN M (v) ) = Fv FN M (v) (22)
This method involves concatenating features along dimension D, preserving all origi-
nal feature information from the participating nodes and ensuring that both the node’s and
its neighbors’ features are clearly represented in the final feature matrix.
ϕ( x ) = wb b( x ) + ws spline( x ) (23)
where b( x ) serves as the basis function, given by b( x ) = silu( x ) = 1+xe−x ; the spline
function spline( x ) is parameterized as a linear combination of B-splines:
Owing to the smooth nature of B-spline activation functions, which possess significant
noise suppression characteristics, these functions effectively dampen random fluctuations
in input data, thereby enhancing the network’s stability and predictive accuracy in noisy
environments. In the experimental section, KAN is replaced with a traditional Multilayer
Perceptron (HyperMP-MLP), and a comparative analysis is conducted with the KAN’s
results, further affirming the method’s effectiveness. The overall architecture of the KAN-
HyperMP model is shown in Figure 6.
3. Experiments Description
In this section, the effectiveness of the constructed model is validated using two open-
source bearing fault diagnosis datasets: SEU and JNU. Experiments are conducted on a
server equipped with an Intel(R) Xeon(R) CPU and an NVIDIA L4 GPU. The network frame-
work is implemented in a PyTorch 2.3.1 and CUDA 12.1 environment. KAN-HyperMP has
a hidden dimension of 256, a combined neighbor feature aggregation block and feature
fusion block, and a single KANLinear block for final feature extraction. For constructing
the incidence matrix with the KNN algorithm, the number of nearest neighbors (K) is set to
Sensors 2024, 24, 6448 11 of 20
3, which accordingly sets the model’s M value to 3. The model training employs a negative
log-likelihood loss function and is optimized using the Adam algorithm with a learning
rate of 1 × 10−3 and a weight decay rate of 5 × 10−6 . In order to evaluate the model’s
performance, the datasets are split into training, validation, and test sets with a ratio of
60%, 20%, and 20%, respectively.
The SEU Bearing Dataset, obtained from the Dynamic Drive Simulator (DDS), is
tailored specifically for bearing fault diagnosis and learning tasks. The bearing signals
in this dataset are sampled at a frequency of 5120 Hz. Data are gathered under two
operational settings: 20 Hz-0 V and 30 Hz-2 V, encompassing normal and various faulted
conditions. These conditions are categorized into five distinct types: Normal, Ball (defects
on the rolling element), Inner (defects on the inner race), Outer (defects on the outer race),
and Combination (concurrent defects on both the inner and outer races). This dataset is
instrumental for basic bearing fault diagnostics, facilitating transfer learning across different
loading conditions and enabling the analysis of complex combined inner and outer race
faults. It effectively addresses the diverse requirements of fault diagnostics and predictive
maintenance. The SEU testbed is depicted in Figure 8.
Sensors 2024, 24, 6448 12 of 20
Table 1. Model comparison using the SEU and JNU Datasets (without noise).
SEU JNU
Model Accuracy F1-Score Accuracy F1-Score
CNN 98.60% 98.60% 99.02% 99.02%
LSTM 98.84% 98.80% 95.57% 94.80%
GCN 98.57% 98.53% 93.60% 93.70%
GAT 92.30% 92.70% 83.80% 83.80%
HGNN 98.99% 98.98% 90.40% 90.30%
HyperGCN 98.94% 98.93% 93.40% 93.20%
KAN-HyperMP 99.70% 99.70% 99.10% 99.10%
(a) (b)
Figure 9. A 2D PCA visualization of rolling bearing fault diagnosis on the SEU and JNU Datasets.
(a) SEU Dataset. (b) JNU Dataset.
Table 2. Rolling bearing fault diagnosis on the SEU Dataset at seven noise levels.
Model −6 dB −4 dB −2 dB 0 dB 2 dB 4 dB 6 dB
CNN 75.02% 79.40% 86.00% 93.20% 74.53% 72.69% 95.51%
LSTM 59.42% 57.05% 65.51% 70.05% 76.13% 83.06% 86.80%
GCN 60.10% 70.50% 76.00% 82.00% 82.70% 85.70% 87.30%
GAT 56.60% 66.60% 73.40% 72.90% 79.80% 81.60% 84.10%
HGNN 54.70% 64.80% 75.10% 78.10% 86.60% 80.00% 82.80%
HyperGCN 63.10% 67.00% 75.30% 81.90% 83.50% 89.00% 87.00%
KAN-
81.56% 86.37% 88.50% 90.47% 92.28% 93.69% 95.60%
HyperMP
Table 3. Rolling bearing fault diagnosis on the JNU Dataset at seven noise levels.
Model −6 dB −4 dB −2 dB 0 dB 2 dB 4 dB 6 dB
CNN 76.60% 85.71% 87.39% 94.00% 97.00% 98.40% 98.10%
LSTM 51.23% 64.23% 75.43% 81.36% 87.83% 81.12% 93.51%
GCN 56.20% 69.20% 74.10% 79.80% 85.80% 87.20% 88.50%
GAT 56.00% 65.60% 72.40% 77.90% 80.30% 80.40% 82.90%
HGNN 47.80% 62.10% 65.50% 70.90% 83.40% 80.20% 78.20%
HyperGCN 63.00% 66.00% 71.40% 71.90% 83.40% 85.27% 88.60%
KAN-
87.04% 91.76% 94.57% 96.54% 98.08% 98.64% 99.12%
HyperMP
(a) (b)
Figure 10. Rolling bearing fault diagnosis accuracies of compared methods at seven noise levels.
(a) Experimental results on the SEU Dataset. (b) Experimental results on the JNU Dataset.
neighbor nodes can dilute the noise’s impact with more effective information. Additionally,
the model’s final part incorporates a B-spline-based KANLinear layer, which, due to
its smoothness and local support characteristics, can handle and suppress input noise
effectively. This helps to maintain the clarity of essential information at each network layer
while filtering out unnecessary noise, as shown in Figure 11. Tables 2 and 3 show that
models using the KANLinear layer perform at 81.56% and 87.04% on the two datasets (SNR
= −6 dB), respectively.
Figure 11. Rolling bearing fault diagnosis accuracies of KAN-HyperMP at seven noise levels.
Simultaneously, we used confusion matrices to visualize the results of the third experi-
ment on two datasets (with SNR ranging from −6 dB to 0 dB). As illustrated in Figure 12a–h,
the model’s performance on the JNU Dataset is noticeably superior to that on the SEU
Dataset as the noise level increases. Specifically, within the SEU Dataset, the primary
classification errors predominantly involve samples labeled 1 and 2.
In summary, even under extreme noise conditions, KAN-HyperMP maintained higher
accuracy compared to other models, highlighting its robustness and precision.
Sensors 2024, 24, 6448 16 of 20
(a) Accuracy: 92.08% (SNR = 0 dB) (b) Accuracy: 97.49% (SNR = 0 dB)
(c) Accuracy: 88.59% (SNR = −2 dB) (d) Accuracy: 95.86% (SNR = −2 dB)
(e) Accuracy: 84.84% (SNR = −4 dB) (f) Accuracy: 92.69% (SNR = −4 dB)
(g) Accuracy: 81.94% (SNR = −6 dB) (h) Accuracy: 87.35% (SNR = −6 dB)
Figure 12. The confusion matrix of the proposed method. (1) Results (a,c,e,g) on the SEU Dataset;
(2) Results (b,d,f,h) on the JNU Dataset.
Sensors 2024, 24, 6448 17 of 20
(a) (b)
Figure 13. Fault-diagnosis accuracy of each block in the ablation experiments. (a) Experimental
results on the SEU Dataset. (b) Experimental results on the JNU Dataset.
The feature fusion block enhances the model by merging node features with those
of neighboring nodes, providing deeper insights into node interactions and introducing
non-linear processing. This helps to capture the graph’s structure and node relationships
from a broader, more “macro” perspective. However, without the feature fusion block, the
model merely adds node features to high-order neighbor features in a simplistic manner,
diminishing its ability to distinguish between noise and useful signals.
When the KANLinear block is replaced with a traditional MLP for feature extraction,
the model loses the noise suppression and smoothing capabilities of the B-spline function.
Such a change complicates the distinction between useful signals and noise in high-noise
environments, leading to a gradual degradation in performance as the noise levels increase.
In conclusion, the analysis demonstrates the effectiveness of the three components within
the overall model.
out overfitting and losing its fault diagnosis capabilities in new data. The corresponding
experimental results are depicted in Figure 14.
As shown in Figure 14a, model accuracy gradually decreases as the number of layers
increases, with the optimal number being 1. At this stage, the fault diagnosis accuracy
for the SEU and JNU datasets reaches 95.60% and 99.12%, respectively, though increasing
the layers to 4 reduces accuracy to 88.54% and 87.56%. While adding layers is expected to
deepen the model’s capacity to capture complex data features, in some hypergraph neural
network architectures, aggregating information from more neighbors with each additional
layer may dilute useful information, making node feature representations more similar and
reducing the distinction between nodes, particularly when processing graph data.
Additionally, as depicted in Figure 14b, the model achieves its highest accuracy when
M is set to 3. However, as M increases to 9, the accuracy decreases to 92.42% and 93.21%
on the SEU and JNU datasets, respectively. This decline in performance with larger M
values can be attributed to nodes aggregating features from more distant neighbors, which
may have weaker relevance to the current node, thus introducing more noise into the data.
Particularly in noisy environments, this information from distant neighbors may not only
be unhelpful but could actually disrupt the correct interpretation of the current node’s state.
Relative to the first two hyperparameters, variations in the hidden dimensions exert a less
pronounced impact on accuracy. However, it is observed that the model attains its highest
accuracy levels on the SEU and JNU datasets when the hidden dimensions are set to 256,
as illustrated in Figure 14c.
(a) (b)
(c)
Figure 14. Parameter analysis on the classification performance of the proposed method. (a) The
number of layers. (b) The maximum edge cardinality. (c) The hidden dimension.
Sensors 2024, 24, 6448 19 of 20
4. Conclusions
In this paper, an innovative rolling bearing fault diagnosis method called KAN-
HyperMP is developed. This method utilizes hypergraph theory to effectively identify
and aggregate high-order neighbor node features. By applying B-spline functions within
KAN, the smoothness of data processing is enhanced, thereby improving the accuracy
of fault diagnosis and the stability of the model in noisy environments. Experimental
results demonstrate that KAN-HyperMP exhibits exceptional fault detection capabilities
and robustness, even under conditions of high noise, effectively addressing the challenges
of complex fault diagnosis.
Although the proposed model has demonstrated commendable performance under
extreme noise conditions, there is potential for further improvement in its accuracy. Conse-
quently, future research will focus on enhancing the model’s robustness. Advanced noise
filtering technologies and data augmentation strategies are planned to be incorporated
to bolster performance in complex environments. Additionally, multimodal data fusion
techniques will be explored to enrich the sources of information for fault diagnosis. These
enhancements are expected to improve the model’s accuracy and applicability, better meet-
ing the demands of industrial applications. Through these efforts, further optimization of
the model is aimed to be achieved, ensuring its reliability in challenging conditions.
Author Contributions: Methodology, J.W.; software, J.W.; validation, Z.D. and S.Z.; writing—original
draft preparation, J.W.; writing—review and editing Z.D. and S.Z. All authors have read and agreed
to the published version of the manuscript.
Funding: The authors are grateful for the support from the General Project of the Zhejiang Provincial
Department of Education (Application No. Y202455248) and the Zhejiang Provincial Youth Fund
(Application No. QN25E050040).
Institutional Review Board Statement: Not applicable.
Informed Consent Statement: Not applicable.
Data Availability Statement: The SEU and JNU Datasets provided in this study can be found in the
following repository: https://github.com/Tan-Qiyu/Mechanical_Fault_Diagnosis_Dataset (accessed
on 1 October 2024).
Conflicts of Interest: The authors declare no conflicts of interest.
References
1. Huo, C.; Jiang, Q.; Shen, Y.; Lin, X.; Zhu, Q.; Zhang, Q. A class-level matching unsupervised transfer learning network for rolling
bearing fault diagnosis under various working conditions. Appl. Soft Comput. 2023, 146, 110739.
2. Dong, Z.; Zhao, D.; Cui, L. Rotating machinery fault classification based on one-dimensional residual network with attention
mechanism and bidirectional gated recurrent unit. Meas. Sci. Technol. 2024, 35, 086001.
3. Wang, M.; Wang, W.; Zhang, X.; Iu, H.H.C. A new fault diagnosis of rolling bearing based on Markov transition field and CNN.
Entropy 2022, 24, 751.
4. Zhang, W.; Li, C.; Peng, G.; Chen, Y.; Zhang, Z. A deep convolutional neural network with new training methods for bearing
fault diagnosis under noisy environment and different working load. Mech. Syst. Signal Process. 2018, 100, 439–453.
5. Cui, L.; Dong, Z.; Xu H. Triplet attention-enhanced residual tree-inspired decision network: A hierarchical fault diagnosis model
for unbalanced bearing datasets Adv. Eng. Inform. 2024, 59, 102322.
6. Dong, Z.; Zhao, D.; Cui, L. An intelligent bearing fault diagnosis framework: one-dimensional improved self-attention-enhanced
CNN and empirical wavelet transform. Nonlinear Dyn. 2024, 112, 6439–6459.
7. Yang, S.; Kong, X.; Wang, Q.; Li, Z.; Cheng, H.; Xu, K. Deep multiple auto-encoder with attention mechanism network: A
dynamic domain adaptation method for rotary machine fault diagnosis under different working conditions. Knowl.-Based Syst.
2022, 249, 108639.
8. Wang, Z.; Wang, J.; Wang, Y. An intelligent diagnosis scheme based on generative adversarial learning deep neural networks and
its application to planetary gearbox fault pattern recognition. Neurocomputing 2018, 310, 213–222.
9. Yao, J.; Chang, Z.; Han, T.; Tian, J. Semi-supervised adversarial deep learning for capacity estimation of battery energy storage
systems. Energy 2024, 294, 130882.
10. Han, J.; Huang, W.; Ma, H.; Li, J.; Tenenbaum, J.; Gan, C. Learning physical dynamics with subequivariant graph neural networks.
Adv. Neural Inf. Process. Syst. 2022, 35, 26256–26268.
Sensors 2024, 24, 6448 20 of 20
11. Zhang, S.; Jin, Y.; Liu, T.; Wang, Q.; Zhang, Z.; Zhao, S.; Shan, B. SS-GNN: A simple-structured graph neural network for affinity
prediction. ACS Omega 2023, 8, 22496–22507.
12. Li, X.; Sun, L.; Ling, M.; Peng, Y. A survey of graph neural network based recommendation in social networks. Neurocomputing
2023, 549, 126441.
13. Wu, L.; Chen, Y.; Ji, H.; Liu, B. Deep learning on graphs for natural language processing. In Proceedings of the 44th International
ACM SIGIR Conference on Research and Development in Information Retrieval, Online, 11–15 July 2021; pp. 2651–2653.
14. Zhou, X.; Zhang, Y.; Wei, Q. Few-shot fine-grained image classification via GNN. Sensors 2022, 22, 7640.
15. Li, T.; Zhou, Z.; Li, S.; Sun, C.; Yan, R.; Chen, X. The emerging graph neural networks for intelligent fault diagnostics and
prognostics: A guideline and a benchmark study. Mech. Syst. Signal Process. 2022, 168, 108653.
16. Li, T.; Zhao, Z.; Sun, C.; Yan, R.; Chen, X. Multireceptive field graph convolutional networks for machine fault diagnosis. IEEE
Trans. Ind. Electron. 2020, 68, 12739–12749.
17. Li, C.; Mo, L.; Yan, R. Rolling bearing fault diagnosis based on horizontal visibility graph and graph neural networks. In
Proceedings of the 2020 International Conference on Sensing, Measurement & Data Analytics in the Era of Artificial Intelligence
(ICSMD), Xi’an, China, 15–17 October 2020; pp. 275–279.
18. Zhang, K.; Li, H.; Cao, S.; Yang, C.; Sun, F.; Wang, Z. Motor current signal analysis using hypergraph neural networks for fault
diagnosis of electromechanical system. Measurement 2022, 201, 111697.
19. Shi, M.; Ding, C.; Wang, R.; Song, Q.; Shen, C.; Huang, W.; Zhu, Z. Deep hypergraph autoencoder embedding: An efficient
intelligent approach for rotating machinery fault diagnosis. Knowl.-Based Syst. 2023, 260, 110172.
20. Yan, X.; Liu, Y.; Zhang, C.A. Multiresolution hypergraph neural network for intelligent fault diagnosis. IEEE Trans. Instrum.
Meas. 2022, 71, 1–10.
21. Feng, Y.; You, H.; Zhang, Z.; Ji, R.; Gao, Y. Hypergraph neural networks. In Proceedings of the AAAI Conference on Artificial
Intelligence, Honolulu, HI, USA, 27 January 2019–1 February 2019; Volume 33, pp. 3558–3565.
22. Yadati, N.; Nimishakavi, M.; Yadav, P.; Nitin, V.; Louis, A.; Talukdar, P. Hypergcn: A new method for training graph convolutional
networks on hypergraphs. Adv. Neural Inf. Process. Syst. 2019, 32, 1509–1520.
23. Ma, Z.; Jiang, Z.; Zhang, H. Hyperspectral image classification using feature fusion hypergraph convolution neural network.
IEEE Trans. Geosci. Remote Sens. 2021, 60, 1–14.
24. Sellami, A.; Farah, M.; Dalla Mura, M. SHCNet: A semi-supervised hypergraph convolutional networks based on relevant feature
selection for hyperspectral image classification. Pattern Recognit. Lett. 2023, 165, 98–106.
25. Gharahighehi, A.; Vens, C.; Pliakos, K. Fair multi-stakeholder news recommender system with hypergraph ranking. Inf. Process.
Manag. 2021, 58, 102663.
26. Sun, Y.; Zhu, D.; Du, H.; Tian, Z. Motifs-based recommender system via hypergraph convolution and contrastive learning.
Neurocomputing 2022, 512, 323–338.
27. Sun, Y.; Jiang, X.; Hu, Y.; Duan, F.; Guo, K.; Wang, B.; Gao, J.; Yin, B. Dual dynamic spatial-temporal graph convolution network
for traffic prediction. IEEE Trans. Intell. Transp. Syst. 2022, 23, 23680–23693.
28. Wu, J.; He, D.; Jin, Z.; Li, X.; Li, Q.; Xiang, W. Learning spatial–temporal pairwise and high-order relationships for short-term
passenger flow prediction in urban rail transit. Expert Syst. Appl. 2024, 245, 123091.
29. Wang, F.; Pena-Pena, K.; Qian, W.; Arce, G.R. T-HyperGNNs: Hypergraph neural networks via tensor representations. IEEE
Trans. Neural Netw. Learn. Syst. 2024.
30. Liu, Z.; Wang, Y.; Vaidya, S.; Ruehle, F.; Halverson, J.; Soljačić, M.; Hou, T.Y.; Tegmark, M. Kan: Kolmogorov-arnold networks.
arXiv 2024, arXiv:2404.19756.
31. Li, C.; Mo, L.; Yan, R. Fault diagnosis of rolling bearing based on WHVG and GCN. IEEE Trans. Instrum. Meas. 2021, 70, 1–11.
32. Kipf, T.N.; Welling, M. Semi-supervised classification with graph convolutional networks. arXiv 2016, arXiv:1609.02907.
33. Veličković, P.; Cucurull, G.; Casanova, A.; Romero, A.; Lio, P.; Bengio, Y. Graph attention networks. arXiv 2017, arXiv:1710.10903.
34. Wen, L.; Li, X.; Gao, L.; Zhang, Y. A new convolutional neural network-based data-driven fault diagnosis method. IEEE Trans.
Ind. Electron. 2017, 65, 5990–5998.
35. Staudemeyer, R.C.; Morris, E.R. Understanding LSTM—A tutorial into long short-term memory recurrent neural networks. arXiv
2019, arXiv:1909.09586.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual
author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to
people or property resulting from any ideas, methods, instructions or products referred to in the content.