0% found this document useful (0 votes)
5 views

Review of Image Classification Algorithms Based On

This paper reviews image classification algorithms utilizing Graph Convolutional Networks (GCNs), highlighting their advantages over traditional Convolutional Neural Networks (CNNs) which are limited to 2D grid structures. It discusses the innovative approaches of classical CNN models, the challenges faced by GCNs such as over-smoothing, and potential solutions. The paper also outlines future directions for research in GCNs and their applications in image classification tasks.

Uploaded by

decktm457
Copyright
© © All Rights Reserved
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
5 views

Review of Image Classification Algorithms Based On

This paper reviews image classification algorithms utilizing Graph Convolutional Networks (GCNs), highlighting their advantages over traditional Convolutional Neural Networks (CNNs) which are limited to 2D grid structures. It discusses the innovative approaches of classical CNN models, the challenges faced by GCNs such as over-smoothing, and potential solutions. The paper also outlines future directions for research in GCNs and their applications in image classification tasks.

Uploaded by

decktm457
Copyright
© © All Rights Reserved
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 10

EAI Endorsed Transactions

on AI and Robotics Research Article

Review of Image Classification Algorithms Based on


Graph Convolutional Networks
Wenhao Tang1,*

1
School of Computer Science and Technology, Henan Polytechnic University, Jiaozuo, Henan 454000, PR China

Abstract

In recent years, graph convolutional networks (GCNs) have gained widespread attention and applications in image
classification tasks. While traditional convolutional neural networks (CNNs) usually represent images as a two-dimensional
grid of pixels when processing image data, the classical model of graph neural networks (GNNs), GCNs, can effectively
handle data with the graph structure, such as social networks, recommender systems, and molecular structures. This paper
summarizes the classical convolutional neural network models, highlighting their innovative approaches. And it will
introduce the problems that graph convolutional networks have had, such as over-smoothing, and the ways to solve them,
and suggest some possible future directions.

Keywords: Graph Convolutional Networks, Convolutional Neural Networks, Graph Neural Networks, over-smoothing.

Received on 17 June 2023, accepted on 02 July 2023, published on 06 July 2023

Copyright © 2023 Tang et al., licensed to EAI. This is an open access article distributed under the terms of the Creative Commons
Attribution license, which permits unlimited use, distribution and reproduction in any medium so long as the original work is properly
cited.

doi: 10.4108/airo.2%y.3462

Traditional image classification methods include the k-


nearest neighbor algorithm [5], support vector machine,
1. Introduction rule-based models [6], etc. The k-nearest neighbor
algorithm overly relies on the choice of distance metric
1.1. Traditional Algorithms function and k-value [7], is computationally intensive,
requires large memory, is not very interpretable, and has a
Image classification [1] is a task in computer vision that slow prediction speed. The training goal of traditional
aims to classify the input images into a predefined class or support vector machines [8] is to have the smallest error
classes. The main processes of image classification are data rate for the learned classifier obtained from the training set
preprocessing, feature extraction and selection, model to approximate the distribution of the training set [9],
training, evaluation, and application [2]. Preprocessing which is achieved by finding the optimal classification
includes operations such as scaling, cropping, grayscale, hyperplane that maximizes the data interval of the training
and normalization for subsequent processing and analysis set. However, in some practical cases, choosing the correct
[3]. The feature extractor extracts meaningful features from kernel function to avoid overfitting is a difficult task when
the image and filters the extracted features to remove the number of features is much larger than the number of
redundant or irrelevant features and keep the most samples. Therefore k-nearest neighbor algorithm and SVM
representative ones [4]. After the model is trained, the are suitable for classifying simple images [10]. And for
performance needs to be evaluated with a test set to some images that are heavily affected by noise or
calculate the accuracy, recall, precision, and other metrics incomplete, the accuracy of classifiers based on them is
of the classifier. Finally, the trained classifier is applied to significantly reduced [11].
the new image data for classification prediction.

Corresponding author. Email: [email protected]


*

EAI Endorsed Transactions on


AI and Robotics
1
W. Tang

1.2. Application Of Graph Convolutional of the paper concludes and shows the future trends of graph
Networks In Image Classification convolutional networks.

One of the main advantages of deep learning models over


traditional machine learning methods with hand-crafted
2. Graph Convolutional Networks
features is the ability to efficiently extract high-level
semantic information and rich detailed information [12]. 2.1. Classic Convolutional Neural Networks
The important backbone models of deep learning are
convolutional neural networks, residual neural networks, Convolutional neural networks (CNN) are a class of feed-
attention mechanism, etc. forward artificial neural networks that use convolutional
Graph Convolutional Network (GCN) is a deep learning computation to extract features, as shown in Figure 1,
model based on graph theory that operates directly on the mainly consisting of a convolutional layer, an activation
graph structure [13]. In image classification tasks, GCN layer, a pooling layer, and a fully-connected layer. The
can transform an image into an undirected graph and convolutional layer is responsible for extracting local
compute the feature vectors of nodes directly using the features of the image, the activation layer introduces non-
adjacency matrix [14] and the trainable weight matrix [15]. linearity, the pooling layer reduces the dimensionality of
Then, the values of the nodes and their connections are the features and the fully connected layer achieves the final
updated by a series of convolution operations [16]. Finally, classification task.
the final classification results are obtained after a fully CNNs have a long history of development, and this
connected layer. paper reviews many classic models, starting with AlexNet,
The graph convolution of the GCNs would mix node which won first place in the ImageNet Large Scale Visual
features and their nearby neighbors [17], that is, by using Recognition Challenge.
the local neighborhood information of the graph to perform AlexNet [18] made an important breakthrough in the
feature aggregation for each node to learn the node history of CNNs by implementing the first deep CNN
representations. structure on a large-scale image dataset. AlexNet consists
In image classification tasks, GCNs can be used for of five convolutional layers, three pooling layers, and three
semi-supervised classification tasks, where the core idea is fully connected layers. Compared to LeNet, AlexNet has a
to update the representation of nodes by propagating deeper network structure and a larger parameter size. In
information between them. GCNs can be modeled by addition, it introduces techniques such as the ReLU
representing images as hypergraphs. In this case, each node activation function, Dropout techniques, and data
represents a pixel in the image, and the edges represent the augmentation.
similarity between pixels. This approach can capture both Convolutional layers
global and local information in an image. By applying Pooling layers
GCNs, the local features of each node can be combined
with information from surrounding nodes to obtain a more Images
discriminative feature representation.
The graph convolutional network has many advantages
in image classification tasks. First, it can capture both local
and global structural information in an image, which leads
to a better understanding of the semantics of the image.
Second, GCNs can learn the weights of each node
adaptively, thus reducing the need for manually designed
features. In addition, GCNs can handle changing graph
structures with strong robustness and generalization ability.
Fully-connected layers

1.3. Thesis Outline


Section 2 of this paper first summarizes classical
convolutional neural networks in chronological order, Output: Classes with their probability
highlighting their innovative points and advantages. Then,
the concept of graph neural networks is elaborated and the
classical models are summarized. Finally, the object of Figure 1. The architecture of CNN.
study, graph convolutional networks, is introduced and its
concepts are elaborated. Section 3 of the paper first shows
the common datasets of GCNs. Then it summarizes the Simonyan and Zisserman (2014) proposed VGGNet
advantages and disadvantages of recent graph with a higher number of layers, increasing the number of
convolutional network models with good performance and layers to 19. The main feature of VGGNet is the use of
presents the problems they have not yet solved. Section 4 consecutive convolutional kernels of size 3 × 3 instead of

EAI Endorsed Transactions on


AI and Robotics
2
Review of Image Classification Algorithms Based on Graph Convolutional Networks

larger ones, which reduces the number of parameters and achieve effective hierarchical accumulation without the
increases computational efficiency. The authors' degradation problem encountered when training deep
experiments further confirm the conjecture that the higher convolutional networks in a single pass. In the concrete
the number of layers, the more accurate the prediction. The implementation, the authors added a batch normalization
top-1 validation error dropped to 25.5% and the top-5 error (BN) layer after each convolutional layer and before the
dropped to 8.0% by layer 19. activation layer instead of the dropout method to reduce
The architecture of the Inception network [20] (also overfitting. Due to the excellent stackability of Resnet, the
known as GoogLeNet) differs from the original preference authors incrementally increase the network level from 18
for hierarchical stacking to a more flexible network model. to 152 layers and obtain better results than 18 layers on the
The main innovation of Inception is the introduction of the ImageNet test set.
Inception module, a structure for parallel stacking of
convolutional kernels at different scales. Unlike the X
stacking of a single network of residuals or VGG networks,
GoogLe-net solves the problem of model degradation at weight layer

high levels in a different way. And a new network


F (X ) ReLU
architecture with bifurcation is proposed. One of the X
identity
reasons why the authors propose an optimization of a single weight layer
network model like the VGG model is that as the layers are
stacked, the useless weights at the shallow level cause a F (X ) + X
significant computational loss for the stacked weights at +
the deep level. Therefore, the authors propose a bifurcated
network architecture to reduce the impact of this ReLU
magnifying glass effect.
The bold innovation of the network structure of
GoogLe-Net v1[20] also enabled the invention to beat
VGG-net in the same year, winning first place in the 2014 Figure 2. Diagram of skip connection.
Image-Net competition by a narrow margin of 0.01%
accuracy improvement.
GoogLe-Net has evolved with its unique and convergent Huang, et al. (2017) et al. proposed the densely
network structure, making good use of the BN layer in connected network (DenseNet). The core idea of DenseNet
Inception-v2 [21] and replacing the convolutional kernel of is to connect the output of each layer to all subsequent
size 5 × 5 with two convolutional kernels of size 3 × 3 to layers to form a densely connected structure. This
improve performance. Chollet (2017) replaced the connection enhances feature propagation, improves the
Inception module with a depthwise separable convolution parameter utilization of the network, and reduces the
based on Inception-v3 (where the order of separable training cost.
convolution is reversed from mobilnet, with pointwise EfficientNet [25] is a CNN structure optimized based on
convolution followed by depthwise convolution, and then neural network search techniques (NAS). The main
combined with ResNet's On ImageNet, Xception is slightly contribution of EfficientNet is the introduction of a
more accurate than Inception-v3, while the number of balanced network expansion strategy that improves
parameters decreases, and the inclusion of a ResNet-like performance by adjusting the depth, width, and resolution
residual connectivity mechanism in the network of the network. EfficientNet achieves state-of-the-art
significantly speeds up its convergence process and performance on several image classification tasks while
achieves significantly higher accuracy. having a low number of parameters and computational
The key innovation of ResNet [23] is the introduction of costs.
the skip connection, as shown in Figure 2, which allows Although the classical models mentioned above and
direct information transfer between network layers. This many current CNNs perform better on image classification
structure effectively solves the problem of gradient tasks, there are some drawbacks compared to GCNs:
disappearance in deep networks, allowing the network to
reach very large depths. The whole idea of residual • CNNs are mainly designed for 2D grid structures
networks treats the action between convolutional layers as (e.g., images) and cannot directly deal with non-
a mapping function 𝐻𝐻(𝑥𝑥), and the authors believe that the Euclidean structured data [26]. Because discretized
degradation problem arises because the training of 𝐻𝐻(𝑥𝑥) convolutions are only defined for regular domains
becomes extremely difficult as the layers are [27]. However, the graph convolution of GCN can
superimposed. If this part of the work can be superimposed handle data with a non-Euclidean structure [28].
in steps, that is, fitting a residual function 𝐹𝐹(𝑥𝑥) = • CNNs ignore the relationships between nodes. When
𝐻𝐻(𝑥𝑥) − 𝑥𝑥, then the objective function can be derived as processing images usually only focus on local
𝐻𝐻(𝑥𝑥) = 𝐹𝐹(𝑥𝑥) + 𝑥𝑥. Since F(x) and x are the output and relationships between neighboring nodes and ignore
input respectively, if 𝐹𝐹(𝑥𝑥) can be found correctly, then the connections between more distant nodes. Thus, for
𝐹𝐹(𝑥𝑥) can be brought in as 𝑥𝑥 ′ to find the deeper 𝐹𝐹 ′ (𝑥𝑥) to

EAI Endorsed Transactions on


AI and Robotics
3
W. Tang

images with globally related nodes, CNNs may not be GraphSAGE uses a multi-layer aggregation function,
suitable for processing. where each layer aggregates the information of nodes and
• In CNNs, convolution operations are usually used to their neighbors to obtain the feature vector of the next
aggregate spatial information. However, the layer. GraphSAGE employs the neighborhood information
convolution operation has high computational of nodes and does not depend on the global graph structure.
complexity. When the depth of the network increases, An innovative improvement of GraphSAGE is to use node
the computation and parameters of convolution features to learn an embedding function that enables
become difficult to control [29]. invisible nodes to generate embeddings.
• Although CNNs excel in dealing with two- GraphSAGE has good performance in both
dimensional grid structures such as image data, GCNs unsupervised and supervised learning.
can better capture the relationships [30] between GAT [35] introduces what is essentially a single-layer pre-
nodes and local features for more general graph- feedback neural network with an attention mechanism that
structured data, especially when the node allows the model to learn itself. This mechanism is
relationships and attribute features are more complex, performed by adding a model learnable coefficient to each
providing stronger modeling and representation edge and performing node feature fusion with an attention
capabilities. coefficient 𝛼𝛼 . This allows the process of convolutional
fusion feature to adjust the model parameters according to
the task and become adaptive for better results. The
2.2. Graph Neural Networks formula for the Attention mechanism is defined as
For datasets containing images, traditional machine �⃗𝑖𝑖 ||𝑊𝑊ℎ
�⃗𝑗𝑗 ���
𝑒𝑒𝑒𝑒𝑒𝑒 �𝐿𝐿𝐿𝐿𝐿𝐿𝐿𝐿𝐿𝐿𝐿𝐿𝐿𝐿𝐿𝐿𝐿𝐿�a�⃗𝑇𝑇 �𝑊𝑊ℎ
learning methods first preprocess the graph structure data 𝛼𝛼𝑖𝑖𝑖𝑖 = , �1.�
[31], mapping the graph structure information to a simple �⃗𝑖𝑖 ||𝑊𝑊ℎ
∑𝑘𝑘∈𝑁𝑁𝑖𝑖 𝑒𝑒𝑒𝑒𝑒𝑒 �𝐿𝐿𝐿𝐿𝐿𝐿𝐿𝐿𝐿𝐿𝐿𝐿𝐿𝐿𝐿𝐿𝐿𝐿�a�⃗𝑇𝑇 �𝑊𝑊ℎ �⃗𝑘𝑘 ���
representation such as a high or low-dimensional feature
vector. This pre-processing stage may add to the image
where a is the weight vector of the attention mechanism,
noise by losing the topological dependency of the
information of the graph nodes [32]. is the attention coefficient between the 𝑖𝑖-th node and
Graph neural Networks (GNN) [32] are based on an the 𝑗𝑗-th node, .𝑇𝑇 is the transposition, || is the concatenation
information diffusion mechanism. Its appearance extends operation, 𝑊𝑊 is the weight matrix, and 𝑁𝑁𝑖𝑖 is some neighbor
existing neural network methods for processing the data nodes of 𝑖𝑖-th node. After getting the of each edge, the
represented in graph domains. GNN pre-processing differs node feature of the 𝑖𝑖-th point after the fusion of attention
from traditional deep learning models in that it requires the can be expressed as the following formula, which is
conversion of images into node and edge representations. essentially a weighted feature summation process, except
For large-size images or images with deep structure, that the weight coefficients in each fusion are learned with
certain GNN models face three problems: neighbor the model training, and finally after a nonlinear activation
explosion, node dependency, and over-smoothing [33]. In function.
addition to the graph itself, these problems are attributed to
the design of a multilayer GNNs framework [33]. To
address these problems, [33] proposed the Ripple Walk ���⃗ �⃗𝑗𝑗 � .
ℎ′ 𝑖𝑖 = 𝜎𝜎 � � 𝛼𝛼𝑖𝑖𝑖𝑖 𝑊𝑊ℎ �2.�
Training (RWT) method for deep and large GNNs. general
𝑗𝑗∈𝑁𝑁𝑖𝑖
subgraph-based training framework RWT does not train
directly on the whole picture but takes subgraphs from the
whole picture and constructs small batches for training. On this basis, to make the attention mechanism more
Their proposed complete GNN is based on small-batch scalable, the defined multi-head attention mechanism is
gradient updates. By computing small batches of gradients used to calculate the attention weights and then contact
within subgraphs, subgraphs with acceptable sizes can them to get the node feature. When at the last layer of the
avoid neighbor explosion altogether. Also, the gradients do network, the dimensionality of the node features output
not depend on nodes outside the subgraph, which solves the after contact is too large. To make the data reasonable, it
node dependency at the subgraph level. Aggregation was switched to accumulate and then average, and then
between subgraphs usually occurs accidentally. However, output to do specific tasks. Experiments on the PPI dataset
propagation-aggregation occurs in the subgraphs, so the show that GAT has better performance than GraphSAGE.
GNNs do not perform well on training sets with class imbalance.
over-smoothing problem can be solved.
This is because, in the class imbalance node classification, Song,
The classical models of GNN are GCN, GraphSAGE, et al. (2022) found that compensating for sub-nodes that deviate
and GAT. from the class connectivity pattern can easily lead to sub-node
GraphSAGE [34] is a graph neural network algorithm false positives. Juan, et al. (2023) point out that increasing the
proposed in 2017, which solves the limitations of GCN participation of a few nodes in the propagation process is an
networks: GCN training requires the adjacency matrix of effective solution. Their INS-GNN contains Self-supervised pre-
the entire graph, depends on the specific graph structure, training, Self-training, and Self-supervised edge augmentation.
and can generally only be used in transductive learning. self-supervised pre-training focuses on the contribution of a few

EAI Endorsed Transactions on


AI and Robotics
4
Review of Image Classification Algorithms Based on Graph Convolutional Networks

nodes, allowing the contribution to model learning, the model extracted after feature aggregation [32]. Before the
does not favor the majority of nodes. This is done by randomly improvement of the equation,
sampling the edges of the graph and masking a few nodes. Self-
training aims to reduce the cost of semi-supervised learning. It 1 1
enables unlabeled nodes to have labels by creating pseudo-labels, � −2 𝐴𝐴̂𝐷𝐷
𝐴𝐴̃ = 𝐷𝐷 � −2 + 𝐼𝐼𝑁𝑁 , �4.�
which facilitates the performance of two different numbers of
nodes. Self-supervised edge augmentation aims to involve a few where 𝐴𝐴̂ = 𝐴𝐴. The eigenvalues of the above equation are in
nodes more in information transfer so that the majority of nodes
a small range. Therefore, when used in deep neural network
have less influence on the model. However, Self-training may
introduce noise while expanding the dataset. However, Self-
models, repeated application of this operator can lead to
supervised learning may have an advancing effect on solving the numerical instability, and gradient explosion or
problem of unstable performance of GNNs in settings with too disappearance [41]. To alleviate this problem, Kipf and
few labeled nodes [38-40]. Welling (2016) let 𝐴𝐴̂ = 𝐴𝐴 + 𝐼𝐼𝑁𝑁 , and then
1 1
2.3. Graph Convolutional Networks � −2 𝐴𝐴̂𝐷𝐷
𝐴𝐴̃ = 𝐷𝐷 � −2 . �5.�

In terms of the convolution method, graph convolution


networks can be divided into spectral-based and spatial- 3. Image Classification Based on Graph
based. Spectral-based GCN is a convolutional method Convolutional Networks
based on spectral theory and convolution theorem, which
converts data from the spatial domain to the spectral
domain for processing and has a solid theoretical 3.1. Common datasets
foundation. In contrast, spatial-based GCN is a method that
does not rely on the spectral convolution theory but defines Citeseer, Cora, and Pubmed [43, 44] are commonly used
the convolution operation directly on the space, which is datasets for GCN image classification. The datasets contain
very flexible. Compared with spatial-based graph sparse bag-of-words feature vectors for each document and
convolution methods, spectral-based graph convolution is a list of citation links between documents [41]. The
currently less commonly used. Kipf and Welling (2016) following is a description of these datasets (Table 1):
proposed spectral convolution, which is to transfer the filter Table 1. Description of the three common datasets.
of the convolution network to the Fourier domain
Dataset Number Dimension Classes Number
simultaneously with the graph signal before processing.
of nodes of node of of edges
While Niepert, et al. (2016) proposed spatial-based
features nodes
convolution which allows the nodes in the graph to connect
Cora 2708 1433 7 5429
and build hierarchical structures in the spatial domain and
then convolve. Citeseer 3312 3703 6 4732
Suppose the undirected graph 𝐺𝐺 = {𝑉𝑉, 𝐸𝐸, 𝑋𝑋}. v is the set of Pubmed 19717 500 3 44338
𝑖𝑖=1 . 𝐸𝐸 is the set of edges of the
N graph nodes, 𝑉𝑉 = {𝑣𝑣𝑖𝑖 }𝑁𝑁
nodes. 𝑋𝑋𝑖𝑖 denotes the matrix of features of node v_i. If 𝑋𝑋𝑖𝑖 Citeseer. Nodes in the CiteSeer dataset denote papers
and edges denote citation relationships between papers. So,
is a d-dimensional feature ( 𝑋𝑋𝑖𝑖 ∈ ℝ𝑑𝑑 ), then 𝑋𝑋 ∈ ℝ𝑁𝑁×𝑑𝑑
this dataset contains 3312 papers in 6 categories in the field
denotes the matrix of all node features.
The propagation equation of the improved GCN [41] is of computer science, represented by 3312 nodes. The
citation relationships between them are represented by
4732 edges. These nodes can all be classified into 6
𝐻𝐻(𝑙𝑙+1) = 𝜎𝜎�𝐴𝐴̃𝐻𝐻(𝑙𝑙 ) 𝑊𝑊 (𝑙𝑙 ) �. �3.�
categories and are all represented by 3703-dimensional
feature descriptors.
Here, 𝐻𝐻𝑙𝑙 is the node feature of the GCN layer 𝑙𝑙 and 𝜎𝜎 Cora. This dataset contains 2708 papers on the field of
refers to the activation function of each graph convolution machine learning in 7 categories, which can also be
layer. 𝐷𝐷 is the diagonal matrix representing the degree of represented by 2708, nodes in one of the 7 categories.
nodes and 𝐷𝐷𝑖𝑖𝑖𝑖 = ∑𝑗𝑗 𝐴𝐴𝑖𝑖𝑖𝑖 . 𝐴𝐴𝑖𝑖𝑖𝑖 refers to the edge-forming These nodes have 5429 edges. Each node has a 1433-
relationship of nodes 𝑖𝑖 and 𝑗𝑗. If there is an edge between 𝑖𝑖 dimensional feature descriptor.
and 𝑗𝑗, then 𝐴𝐴𝑖𝑖𝑖𝑖 = 1, otherwise 𝐴𝐴𝑖𝑖𝑖𝑖 = 0. 𝐴𝐴̂ = �𝐴𝐴𝑖𝑖𝑖𝑖 � ∈ ℝ𝑁𝑁×𝑁𝑁 Pubmed. This dataset contains 19717 papers on the field
is a node-self-connected adjacency matrix, 𝐴𝐴̂ = 𝐴𝐴 + 𝐼𝐼𝑁𝑁 , of biomedicine in 3 categories. Each paper is represented
1 1 by a bag-of-words model, and the edges between nodes
and 𝐼𝐼𝑁𝑁 is the unit matrix. 𝐴𝐴̃ = 𝐷𝐷 � −2 𝐴𝐴̂𝐷𝐷
� −2 is a sparse matrix.
indicate the co-citation relationships between papers. It can
𝑊𝑊 𝑙𝑙 refers to the learnable parameters. The graph also be said that the dataset network contains 19717 nodes
convolution layer has two stages: feature fusion and feature and 44338 edges, each represented by a 500-dimensional
extraction. 𝐴𝐴̃𝐻𝐻(𝑙𝑙 ) is the aggregation phase, which serves to feature descriptor.
control the node properties of local neighbors to make them
similar. 𝐻𝐻(𝑙𝑙 ) 𝑊𝑊 (𝑙𝑙 ) is the feature extraction phase, where
common features between neighboring nodes can be 3.2. Graph Convolutional Networks Models

EAI Endorsed Transactions on


AI and Robotics
5
W. Tang

However, there are still some challenges and limitations of Pan, et al. (2022) proposed that no deep GCN model has
the original graph convolutional networks in image been used for medical diagnosis because of the problem of
classification. One of them is the problem of computational over-smoothing. To overcome the over-smoothing
efficiency because GCNs need to perform aggregation problem, they used a snowball GCN module to build a
operations on the whole graph, resulting in high multiscale convolutional module. the snowball GCN [53]
computational complexity [45]. In addition, GCNs may is a densely connected graph network that can connect
encounter the problem of memory limitation when dealing multiscale features. This graph network can superimpose
with large-scale image data because of the need to store and all the learned features as input to subsequent hidden
process a large number of nodes and edges [46]. Further, layers. This network also overcomes the gradient vanishing
the original GCN can capture information only about problem and reduces the number of parameters, etc. The
immediate neighbors with one layer of convolution. when key to solving the over-smoothing problem in Luan, et al.
multiple GCN layers are stacked, information about larger (2019) is that they define the graph convolution of a
neighborhoods is integrated [47]. This is because the graph spectral filter as the product of a block Krylov matrix and
convolution layers of GCNs can be considered low-pass a specific form of learnable parameter matrix. The formula
filters [48], and this property causes the signal to become 𝐾𝐾𝑚𝑚 (𝐴𝐴, 𝐵𝐵) ≔ [𝐵𝐵, 𝐴𝐴𝐴𝐴, … , 𝐴𝐴𝑚𝑚−1 𝐵𝐵] for the block Krylov
smoother, which is an inherent advantage of GCNs; matrix comes from the transformation of S-span of 𝕊𝕊-span
however, if the number of GCN layers is large, performing 𝑚𝑚
of {𝑋𝑋𝑘𝑘 }𝑘𝑘=1 and 𝐾𝐾𝑚𝑚𝕊𝕊 (𝐴𝐴, 𝐵𝐵). They stated that it is difficult to
this signal smoothing operation multiple times will make apply the block Krylov method directly to the GCN, so they
the signal converge, which loses the diversity of node developed the snowball and the truncated Krylov. For
features, which is the over-smoothing problem [49]. 𝑜𝑜𝑢𝑢𝑢𝑢𝑢𝑢𝑢𝑢𝑢𝑢 = 𝑠𝑠𝑠𝑠𝑠𝑠𝑠𝑠𝑠𝑠𝑠𝑠𝑠𝑠(𝐿𝐿𝑝𝑝 ∁𝑊𝑊𝐶𝐶 ), if 𝑝𝑝 = 1 and 𝐿𝐿𝑝𝑝 = 𝐿𝐿 , then
Too many graph convolution layers may cause the over- the snowball implementation maps back to the Fourier
smoothing problem, however, there are image basis of the graph, thus achieving a "snowball "The
classification studies that solve this problem to some extent Adaptive multi-channel fusion GCN implemented by Pan,
[50-52], allowing the model to extract deep-level features. et al. (2022) also contains the channel common convolution
SelfSAGCN [50] overcomes the overfitting problem module, which solves the problem of extracting the nodes
and the over-smoothing problem by combining feature in a particular channel embedding and the common
aggregation and semantic alignment. SelfSAGCN first information shared between channels. As shown in Figure
applies feature aggregation to extract semantic information 3, the output of the channel common convolution module
from the labeled nodes layer by layer, which does not suffer is represented by the equation
from the over-smoothing phenomenon. The core idea of
selfSAGCN is that the node features obtained from the 𝐻𝐻𝑐𝑐 = 𝛼𝛼𝐻𝐻𝑐𝑐 + 𝛽𝛽𝐻𝐻𝑐𝑐𝑐𝑐𝐻𝐻0 + 𝛾𝛾𝐻𝐻𝑐𝑐𝑐𝑐 , �6.�
semantic and graph structure should be consistent when the
categories are the same. This is not affected by the over- where, 𝛼𝛼, 𝛽𝛽, 𝛾𝛾 are hyperparameters that measure the
smoothing phenomenon. The unlabeled node features importance of the common convolution, respectively. The
obtained from graph aggregation are aligned with semantic multi-channel attention they introduce can fuse the outputs
features by semantic alignment techniques to find of different channels and the features of each channel to
additional supervisory information. This improves the integrate the learned embeddings for prediction. The
performance of the model and enhances the identifiability ablation experiments demonstrate that their proposed
of node features. The semantic alignment of selfSAGCN is MAMF-GCN has strong robustness and high accuracy.
based on the central similarity of homogeneous class
information, which enables the model to transfer relevant
features to unlabeled data after learning semantics from the
labeled data. Yang, et al. (2021) additionally utilized
central similarity optimization to align node features
obtained from semantic and graph structure aspects,
respectively features are aligned, which has a significant
effect on mitigating over-smoothing. Also, the central
similarity of labeled and unlabeled nodes can provide
additional supervised information, which further improves
the classification accuracy of unlabeled nodes. Moreover,
they use the pseudo-labeling trick for unlabeled data and
also suppress the noise using the practice of updating the
centers. It is experimentally confirmed that selfSAGCN
has better performance on different datasets even when the
labeled nodes are severely limited. This indicates that the
overfitting problem does not affect it too much. Even if the
number of layers is increased to 16, selfSAGCN can
mitigate the over-smoothing problem.

EAI Endorsed Transactions on


AI and Robotics
6
Review of Image Classification Algorithms Based on Graph Convolutional Networks

Multi-Scale Deep Convolution showed that it is not feasible to down-sample the original
Functional
Knn Graph Snowball HF image to the right size and compress the input graph data
and GNN model to reduce the memory consumption of the
Cd model. They state that this is because of two ways of down-
Functional sampling, one, neighborhood sampling will result in

Multi-Channel Attention
Knn Graph
Snowball H CF "neighborhood explosion" when the number of graph
convolution layers increases, i.e., many neighborhoods are

Prediction
WC( ) WC( )
l Parameter l +1
Sharing Cu H making it difficult to train. Second, although graph
Phenotypic
Graph sampling can avoid "neighborhood explosion," it does not
Snowball H CP guarantee that each node will be sampled once throughout
Channel Common Convolution the training or inference process. For compressed input
Cd images, it is difficult to manipulate because the original
graph is small. For compressed GNN models, the
Multi-Scale Deep Convolution
relationship between the compression rate and the accuracy
Phenotypic
Graph Snowball
HP of the GNN model needs to be carefully designed to ensure
that both nodes in the high-dimensional semantic space and
nodes in the low-dimensional feature space can be
characterized. However, this approach is more difficult.
The core idea of Bi-GCN is to binarize the network
Figure 3. The structure of adaptive multi-channel parameters (e.g., weights) and input node features in the
fusion GCN. feature extraction phase, while not operating in this way in
the feature aggregation phase. In addition, the original
matrix multiplication of the graph convolution operation is
The prominent innovation of NSCGCN [52] is to overcome modified to a binary addition operation. Another core idea
the over-smoothing problem using feature fusion based on is to binarize the node features by separating them and
the node-self-convolution algorithm and to preserve the assigning attention weights to each node. By deploying
spatial structure information of the original feature graph additional parameters, the model can remain effective to
using the feature reconstruction algorithm. The innovative learn more useful information. In theory, Bi-GCN can
point of the node-self-convolution algorithm is that the reduce the memory consumption of network parameters
input undirected graph 𝐺𝐺 𝑙𝑙 retains only the node-self- and input data by an average of about 30 times and increase
connected degree matrix 𝐼𝐼. The result of the convolution of the inference speed by an average of about 47 times.
the input graph node features 𝑋𝑋 is given by However, Bi-GCN had not experimented based on solving
the over-smoothing problem, i.e., Bi-GCN does not
1 1 perform well at deeper layers.
𝑍𝑍 (𝑙𝑙+1) = 𝐷𝐷 −2 𝐼𝐼𝐷𝐷 −2 𝑋𝑋 (𝑙𝑙 ) 𝑊𝑊 (𝑙𝑙 ) , �7.� Similar to CNNs, GCNs have a multilayer structure [13],
where each layer extracts higher-level features by
where 𝑊𝑊 (𝑙𝑙 ) is the learnable parameter. A new graph aggregating features from neighboring nodes and applying
structure 𝐺𝐺 𝑙𝑙+1 is then obtained by combining the original a nonlinear activation function. This allows GCNs to take
adjacency matrix of the undirected graph 𝐺𝐺 𝑙𝑙 with 𝑍𝑍 (𝑙𝑙+1) to full advantage of the topology of the graph and thus better
regain a new graph structure 𝐺𝐺 (𝑙𝑙+1) . The feature capture the relationships between nodes. However,
reconstruction algorithm is based on the image applying GCNs may not be able to adaptively extract the
neighborhood structure, which converts the image into most relevant information between topologies or node
graphically structured data and has better classification features. Xu, et al. (2021) proposed a multiscale skeleton
performance than the down-sampling approach. Although adaptive weighted graph convolution network (MS-
NSCGCN also has outstanding performance, it also suffers AWGCN) for skeleton-based human action recognition in
from overfitting. And the feature reconstruction stage is not IoT. MS-AWGCN solves the problem of learning potential
adaptive and does not achieve the best structure. For the graph topology in an adaptive extraction style by an
overfitting phenomenon due to an insufficient amount of adaptive information aggregation strategy to weight
data, semantic alignment techniques can be considered to information from different sampling strategies more
allow labeled nodes to guide unlabeled nodes [50]. efficiently. The adaptive weighted graph convolution block
For the problem of large memory consumption of graph formulation of MS-AWGCN is as follows,
convolution operations, Bi-GCN [54] provides a solution.
𝐾𝐾
Bi-GCN has two innovative approaches, one is to perform 𝑓𝑓𝑜𝑜𝑜𝑜𝑜𝑜 = � 𝛼𝛼 𝑊𝑊𝑘𝑘 𝑓𝑓𝑖𝑖𝑖𝑖 (𝐴𝐴𝑘𝑘 + 𝐿𝐿𝑘𝑘 ). �8.�
binarized network parameters and input node feature 𝑘𝑘
operations in the feature extraction phase, which can
theoretically reduce memory consumption by up to 30 where 𝑊𝑊𝑘𝑘 denotes the weight matrix, 𝐿𝐿𝑘𝑘 is learnable, and
times and improve inference speed by up to 47 times, and 𝑓𝑓𝑖𝑖𝑖𝑖 is the node feature of the input.
the other is to design a new backpropagation method to Relative to the formulation of [56] graph self-learning
accommodate binarized weights. Wang, et al. (2021) module

EAI Endorsed Transactions on


AI and Robotics
7
W. Tang

Science and Technology, vol. 34, no. 9, Sep 2023, Art


𝐾𝐾 no. 095703.
𝑓𝑓𝑜𝑜𝑜𝑜𝑜𝑜 = � 𝑊𝑊𝑘𝑘 𝑓𝑓𝑖𝑖𝑖𝑖 (𝐿𝐿𝑘𝑘 ) , �9.� [2] M. Castillo-Cara, R. Talla-Chumpitaz, R. Garcia-
𝑘𝑘
Castro, and L. Orozco-Barbosa, "TINTO: Converting
Tidy Data into image for classification with 2-
MS-AWGCN introduces weights α initialized to 1 that Dimensional Convolutional Neural Networks,"
control the importance of different sampling strategies and Softwarex, vol. 22, May 2023, Art no. 101391.
𝛼𝛼 normalized adjacency matrix 𝐴𝐴𝑘𝑘 to make learning graph [3] A. Husain and V. P. Vishvakarma, "RES-KELM fusion
topology more efficient.MS-AWGCN also contains an model based on non-iterative deterministic learning
attention mechanism, and the algorithm for the mechanism classifier for classification of Covid19 chest X-ray
is SE-Net. images," Journal of Intelligent Systems, vol. 32, no. 1,
The core idea of the SE-Net [57] algorithm is to learn to Jun 2023, Art no. 20220235.
[4] M. Hida et al., "Development of Hallux Valgus
use global information to selectively emphasize
Classification Using Digital Foot Images with Machine
information-rich features and suppress less useful features. Learning," Life-Basel, vol. 13, no. 5, May 2023, Art no.
The algorithm allows the network to recalibrate features. 1146.
Since the SE block is relatively simple, it has been applied [5] Y. Zhang, "Smart detection on abnormal breasts in
to many CNNs to improve the performance of the model digital mammography based on contrast-limited
[58-61]. Although the SE block is currently less used on adaptive histogram equalization and chaotic adaptive
GCNs, it may be an option to consider from the perspective real-coded biogeography-based optimization,"
of incorporating attention mechanisms without adding too Simulation, vol. 92, no. 9, pp. 873-885, September 12,
much to the model complexity. 2016 2016.
[6] Y. Zhang, "A Rule-Based Model for Bankruptcy
Prediction Based on an Improved Genetic Ant Colony
Algorithm," (in English), Math. Probl. Eng., Article
vol. 2013, 2013, Art no. 753251.
4. Conclusion [7] Q. H. Wang, J. W. Liu, and L. L. Zhang, "Study on the
Classification of K-Nearest Neighbor Algorithm,"
This paper focuses on graph convolutional networks for Journal of Xi'an Technological University.
image classification. Firstly, some classical models of [8] S. Wang, "Detection of Dendritic Spines using Wavelet
convolutional neural networks are discussed, and their Packet Entropy and Fuzzy Support Vector Machine,"
advantages and limitations on image classification tasks are CNS & Neurological Disorders - Drug Targets, vol. 16,
pointed out. Then, graph neural networks are introduced, no. 2, pp. 116-121, 2017.
summarizing the current status of graph neural networks [9] D. Zhao, W. Zou, and G. M. Sun, "A fast image
classification algorithm using Support Vector
that perform poorly on training sets with class imbalance.
Machine," in 2010 2nd International Conference on
And the solution of introducing self-supervised learning Computer Technology and Development.
techniques is summarized. Finally, graph convolutional [10] Y. Zhang, "Pathological brain detection in MRI
networks are introduced. I summarize some graph scanning by wavelet packet Tsallis entropy and fuzzy
convolutional network models that solve the over- support vector machine," SpringerPlus, vol. 4, no. 1,
smoothing problem and explain the reasons for solving the 2015, Art no. 716.
problem. The model that solves the high memory [11] X. Zheng and R. S. Cloutier, "A Review of Image
consumption of graph convolution operation is also Classification Algorithms in IoT," EAI Endorsed
summarized, and its advantages and disadvantages are Transactions on Internet of Things, vol. 7, no. 28, pp.
1-11, 2022.
analyzed. And based on the analysis of successful cases,
[12] Y. Zhang et al., "Unsupervised domain selective graph
the self-attentive module is suggested to be introduced into convolutional network for preoperative prediction of
graph convolutional networks. This paper does not analyze lymph node metastasis in gastric cancer," Medical
the over-smoothing problem and the high memory Image Analysis, vol. 79, p. 102467, 2022/07/01/ 2022.
consumption problem of graph convolutional networks in [13] J. Zhou, J. X. Huang, Q. V. Hu, and L. He, "SK-GCN:
more detail. In future work, I will analyze more literature Modeling Syntax and Knowledge via Graph
and summarize more methods. Convolutional Network for aspect-level sentiment
Graph convolutional networks have potential and wide classification," Knowledge-Based Systems, vol. 205, p.
application prospects in image classification tasks. With 106292, 2020/10/12/ 2020.
[14] M. Ghayekhloo and A. Nickabadi, "CLP-GCN:
the continuous development and improvement of
Confidence and label propagation applied to Graph
technology, it is believed that GCNs will play a greater role Convolutional Networks," Applied Soft Computing,
in image classification and other graph data fields. vol. 132, p. 109850, 2023/01/01/ 2023.
[15] Y. Li et al., "Identification of Mild Cognitive
Impairment based on quadruple GCN model
References constructed with Multiple features from higher-order
brain connectivity," Expert Systems with Applications,
[1] J. M. Herzog and V. Sick, "Design of a line-of-sight p. 120575, 2023/06/02/ 2023.
fluorescence-based imaging diagnostic for [16] S. Zheng, L. Qiu, and F. Lan, "TSO-GCN: A Graph
classification of microbe species," Measurement Convolutional Network approach for real-time and

EAI Endorsed Transactions on


AI and Robotics
8
Review of Image Classification Algorithms Based on Graph Convolutional Networks

generalizable truss structural optimization," Applied [34] W. L. Hamilton, R. Ying, and J. Leskovec, "Inductive
Soft Computing, vol. 134, p. 110015, 2023/02/01/ 2023. Representation Learning on Large Graphs," 2017.
[17] X. Zhu, C. Li, J. Guo, and S. Dietze, "CNIM-GCN: [35] P. Velikovi, G. Cucurull, A. Casanova, A. Romero, P.
Consensus Neighbor Interaction-based Multi-channel Liò, and Y. Bengio, "Graph Attention Networks," 2017.
Graph Convolutional Networks," Expert Systems with [36] J. Song, J. Park, and E. Yang, "TAM: Topology-Aware
Applications, vol. 226, p. 120178, 2023/09/15/ 2023. Margin Loss for Class-Imbalanced Node
[18] A. Krizhevsky, I. Sutskever, and G. Hinton, "ImageNet Classification," presented at the Proceedings of the 39th
Classification with Deep Convolutional Neural International Conference on Machine Learning,
Networks," Advances in neural information processing Proceedings of Machine Learning Research, 2022.
systems, vol. 25, no. 2, 2012. [37] X. Juan, F. Zhou, W. Wang, W. Jin, J. Tang, and X.
[19] K. Simonyan and A. Zisserman, "Very Deep Wang, "INS-GNN: Improving graph imbalance
Convolutional Networks for Large-Scale Image learning with self-supervision," Information Sciences,
Recognition," Computer Science, 2014. vol. 637, p. 118935, 2023/08/01/ 2023.
[20] C. Szegedy, W. Liu, Y. Jia, P. Sermanet, and A. [38] Q. Q. Zhao, H. F. Ma, L. J. Guo, and Z. X. Li,
Rabinovich, "Going Deeper with Convolutions," IEEE "Hierarchical attention network for attributed
Computer Society, 2014. community detection of joint representation," NEURAL
[21] S. Ioffe and C. Szegedy, "Batch Normalization: COMPUTING & APPLICATIONS, vol. 34, no. 7, pp.
Accelerating Deep Network Training by Reducing 5587-5601, APR 2022.
Internal Covariate Shift," JMLR.org, 2015. [39] Z. Yang, Y. Yan, H. Gan, J. Zhao, and Z. Ye, "A safe
[22] F. Chollet, "Xception: Deep Learning with Depthwise semi-supervised graph convolution network,"
Separable Convolutions," in 2017 IEEE Conference on Mathematical Biosciences and Engineering, vol. 19,
Computer Vision and Pattern Recognition (CVPR), no. 12, pp. 12677-12692, 2022.
2017. [40] Z. Zhou, J. Shi, S. Zhang, Z. Huang, and Q. Li,
[23] K. He, X. Zhang, S. Ren, and J. Sun, "Deep Residual "Effective stabilized self-training on few-labeled graph
Learning for Image Recognition," IEEE, 2016. data," Information Sciences, vol. 631, pp. 369-384,
[24] G. Huang, Z. Liu, L. Maaten, and K. Q. Weinberger, 2023/06/01/ 2023.
"Densely Connected Convolutional Networks," in [41] T. N. Kipf and M. Welling, "Semi-Supervised
IEEE Conference on Computer Vision and Pattern Classification with Graph Convolutional Networks,"
Recognition, 2017. 2016.
[25] M. Tan and Q. V. Le, "EfficientNet: Rethinking Model [42] M. Niepert, M. Ahmed, and K. Kutzkov, "Learning
Scaling for Convolutional Neural Networks," 2019. Convolutional Neural Networks for Graphs,"
[26] X. Liu, Z. You, Y. He, S. Bi, and J. Wang, "Symmetry- JMLR.org, 2016.
Driven hyper feature GCN for skeleton-based gait [43] B. Jiang, Z. Zhang, D. Lin, J. Tang, and B. Luo, "Semi-
recognition," Pattern Recognition, vol. 125, p. 108520, Supervised Learning With Graph Learning-
2022/05/01/ 2022. Convolutional Networks," in 2019 IEEE/CVF
[27] H. Jiang, P. Cao, M. Xu, J. Yang, and O. Zaiane, "Hi- Conference on Computer Vision and Pattern
GCN: A hierarchical graph convolution network for Recognition (CVPR), 2020.
graph embedding learning of brain network and brain [44] Prithviraj et al., "Collective Classification in Network
disorders prediction," Computers in Biology and Data," Ai Magazine, 2008.
Medicine, vol. 127, p. 104096, 2020/12/01/ 2020. [45] T. Altaf, X. Wang, W. Ni, R. P. Liu, and R. Braun, "NE-
[28] Y. Zhang, S. Satapathy, D. Guttery, J. Gorriz, and S. GConv: A lightweight node edge graph convolutional
Wang, "Improved Breast Cancer Classification network for intrusion detection," Computers &
Through Combining Graph Convolutional Network Security, vol. 130, Jul 2023, Art no. 103285.
and Convolutional Neural Network," Information [46] R. Corrias, M. Gjoreski, and M. Langheinrich,
Processing and Management, vol. 58, p. 102439, 01/18 "Exploring Transformer and Graph Convolutional
2021. Networks for Human Mobility Modeling," Sensors,
[29] Y. Zang, D. Yang, T. Liu, H. Li, S. Zhao, and Q. Liu, vol. 23, no. 10, May 2023, Art no. 4803.
"SparseShift-GCN: High precision skeleton-based [47] L. Yao, C. Mao, and Y. Luo, "Graph Convolutional
action recognition," Pattern Recognition Letters, vol. Networks for Text Classification," 2019, pp. 7370-
153, pp. 136-143, 2022/01/01/ 2022. 7377.
[30] S. Wang, V. Govindaraj, J. Gorriz, X. Zhang, and Y. [48] N. T. Hoang and T. Maehara, "Revisiting Graph Neural
Zhang, "Explainable diagnosis of secondary pulmonary Networks: All We Have is Low-Pass Filters," 2019.
tuberculosis by graph rank-based average pooling [49] Q. Li, Z. Han, and X. M. Wu, "Deeper Insights into
neural network," Journal of Ambient Intelligence and Graph Convolutional Networks for Semi-Supervised
Humanized Computing, 03/11 2021. Learning," 2018.
[31] S. Wang, "Advances in data preprocessing for [50] X. Yang, C. Deng, Z. Dang, K. Wei, and J. Yan, "Self-
biomedical data fusion: an overview of the methods, SAGCN: Self-Supervised Semantic Alignment for
challenges, and prospects," Information Fusion, vol. Graph Convolution Network," in Computer Vision and
76, pp. 376-421, 2021. Pattern Recognition, 2021.
[32] F. Scarselli, M. Gori, A. C. Tsoi, M. Hagenbuchner, and [51] J. Pan, H. Lin, Y. Dong, Y. Wang, and Y. Ji, "MAMF-
G. Monfardini, "The Graph Neural Network Model," GCN: Multi-scale adaptive multi-channel fusion deep
IEEE Transactions on Neural Networks, vol. 20, no. 1, graph convolutional network for predicting mental
p. 61, 2009. disorder," Computers in Biology and Medicine, vol.
[33] Y. Liu, M. Zhang, C. Ma, B. Bai, and G. Yan, "Graph 148, p. 105823, 2022/09/01/ 2022.
neural network," 2020. [52] C. Tang, C. Hu, J. Sun, S.-H. Wang, and Y.-D. Zhang,
"NSCGCN: A novel deep GCN model to diagnosis

EAI Endorsed Transactions on


AI and Robotics
9
W. Tang

COVID-19," Computers in Biology and Medicine, vol.


150, p. 106151, 2022/11/01/ 2022.
[53] S. Luan, M. Zhao, X. W. Chang, and D. Precup, "Break
the Ceiling: Stronger Multi-scale Deep Graph
Convolutional Networks," ed, 2019.
[54] J. Wang, Y. Wang, Z. Yang, L. Yang, and Y. Guo, "Bi-
GCN: Binary Graph Convolutional Network," in
Computer Vision and Pattern Recognition, 2021.
[55] W. Xu, M. Wu, J. Zhu, and M. Zhao, "Multi-scale
skeleton adaptive weighted GCN for skeleton-based
human action recognition in IoT," Applied Soft
Computing, vol. 104, p. 107236, 2021/06/01/ 2021.
[56] G. Zhu, L. Zhang, H. Li, P. Shen, and M. Bennamoun,
"Topology-learnable Graph Convolution for Skeleton-
based Action Recognition," Pattern Recognition
Letters, vol. 135, 2020.
[57] J. Hu, L. Shen, S. Albanie, G. Sun, and E. Wu,
"Squeeze-and-Excitation Networks," IEEE
transactions on pattern analysis and machine
intelligence, vol. 42, no. 8, pp. 2011-2023, 2020.
[58] N. Ettehadi et al., "Automated Multiclass Artifact
Detection in Diffusion MRI Volumes via 3D Residual
Squeeze-and-Excitation Convolutional Neural
Networks," (in English), Frontiers in Human
Neuroscience, Original Research vol. 16, 2022-March-
30 2022.
[59] Y. Chen, Z. Zhang, L. Zhong, T. Chen, J. Chen, and Y.
Yu, "Three-Stream Convolutional Neural Network
with Squeeze-and-Excitation Block for Near-Infrared
Facial Expression Recognition," Electronics, vol. 8, no.
4, p. 385, 2019.
[60] J. Chen et al., "An Efficient Memristor-Based Circuit
Implementation of Squeeze-and-Excitation Fully
Convolutional Neural Networks," IEEE Transactions
on Neural Networks and Learning Systems, vol. 33, no.
4, pp. 1779-1790, 2022.
[61] G. A. Altuwaijri, G. Muhammad, H. Altaheri, and M.
Alsulaiman, "A Multi-Branch Convolutional Neural
Network with Squeeze-and-Excitation Attention
Blocks for EEG-Based Motor Imagery Signals
Classification," Diagnostics, vol. 12, no. 4, p. 995,
2022.

EAI Endorsed Transactions on


AI and Robotics
10

You might also like