0% found this document useful (0 votes)
1 views16 pages

Integrated Syntactic and Semantic Tree For Targeted Sentiment Classification Using Dual-Channel Graph Convolutional Network

This document presents a novel approach for targeted sentiment classification using an Integrated Syntactic and Semantic tree (IS2 tree) and a Dual-Channel Graph Convolutional Network (DCGCN). The IS2 tree enhances the representation of semantic relations while addressing parsing errors in informal contexts, and the DCGCN effectively couples syntactic and semantic information to improve classification performance. Experimental results demonstrate that this method outperforms existing state-of-the-art techniques on multiple datasets.

Uploaded by

venomroy9363
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
1 views16 pages

Integrated Syntactic and Semantic Tree For Targeted Sentiment Classification Using Dual-Channel Graph Convolutional Network

This document presents a novel approach for targeted sentiment classification using an Integrated Syntactic and Semantic tree (IS2 tree) and a Dual-Channel Graph Convolutional Network (DCGCN). The IS2 tree enhances the representation of semantic relations while addressing parsing errors in informal contexts, and the DCGCN effectively couples syntactic and semantic information to improve classification performance. Experimental results demonstrate that this method outperforms existing state-of-the-art techniques on multiple datasets.

Uploaded by

venomroy9363
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 16

IEEE/ACM TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL.

32, 2024 1109

Integrated Syntactic and Semantic Tree for Targeted


Sentiment Classification Using Dual-Channel
Graph Convolutional Network
Puning Zhang , Rongjian Zhao , Boran Yang , Member, IEEE, Yuexian Li, and Zhigang Yang , Member, IEEE

Abstract—Targeted sentiment analysis aims to identify the sen- I. INTRODUCTION


timent polarity of specific target mentions in a sentence. Existing
ENTIMENT analysis, also known as opinion mining, aims
methods employ neural networks to extract the relations between
target mentions and their contexts. Recent approaches based on
graph convolutional networks can model the syntactic relations
S to analyze the emotional attitudes exhibited by users. Early
sentence-level sentiment analysis methods were mainly based
extracted by an external parser into adjacency matrices. However, on lexicons and rules [1], [2], [3], [4]. Therefore, they could
online reviews are informal and complex, the syntactic structures
provided by the parser can be incorrect in these syntax-insensitive
hardly predict the sentiment polarity only from the general terms
scenarios. To remedy this defect, we design a novel integrated of a sentence and there are often different sentiment polarities
syntactic and semantic tree (IS2 tree) by labeling semantic relations towards various target mentions in the same sentence. Targeted
between the target mention and contexts in a syntactic dependency sentiment analysis [5], [6], [7], [8] is a fine-grained task to model
tree. Furthermore, a dual-channel graph convolutional network the relationship between the target mention and corresponding
(DCGCN) is proposed to encode the contextual information as-
sociated with the target mention by dynamic semantic pruning opinion expressions. Specifically, a targeted sentiment analysis
mechanisms and to also retain the syntactic relations. Experimental task can identify the sentiment polarity (i.e., positive, neutral,
results demonstrate that the IS2 tree has a favorable generalization negative) of a specific target (also called aspect) appearing in a
capability comparing to the state-of-the-art baselines on four public sentence. For example, given sentence “The falafel was rather
datasets. overcooked and dried but the chicken was fine.”, the sentiment
Index Terms—Targeted Sentiment analysis, graph convolutional polarities of two target mentions “falafel” and “chicken” are
network, syntactic dependency tree, attention mechanism. negative and positive, respectively.
Dominant targeted sentiment analysis methods [9], [10], [11],
[12] mainly employed deep neural networks to encode input
sentences and leveraged attention mechanisms to capture the
semantic relations between the target mention and contexts.
Manuscript received 17 December 2021; revised 30 August 2022, 30 March However, these methods suffer from the inherent defects of at-
2023, and 27 August 2023; accepted 30 December 2023. Date of publication
12 January 2024; date of current version 19 January 2024. This work was tention mechanisms, which make the targeted sentiment analysis
supported in part by the National Natural Science Foundation of China un- vulnerable to the noise interference of words in sentences [13],
der Grants 61901071, 61871062, 61771082, and U20A20157, in part by the [14], [15]. In the above example, for target mention “chicken”,
Science and Natural Science Foundation of Chongqing, China under Grant
cstc2020jcyj-zdxmX0024, in part by the University Innovation Research Group opinion word “dried” receives higher attention than “fine”, but
of Chongqing under Grant CXQT20017, and in part by the Program for “dried” is more relevant to “falafel”, which affects the classi-
Innovation Team Building at Institutions of Higher Education in Chongqing fication performance by establishing the incorrect relationship
under Grant CXTDX201601020. The associate editor coordinating the review
of this manuscript and approving it for publication was Dr. Preslav Nakov. between “chicken” and the irrelevant “dried”. Therefore, how to
(Corresponding author: Rongjian Zhao.) introduce the semantic relations without the noise interference of
Puning Zhang, Rongjian Zhao, Yuexian Li, and Zhigang Yang are with the words in a sentence becomes a challenge for targeted sentiment
School of Communication and Information Engineering, Chongqing University
of Posts and Telecommunications, Chongqing 400065, China, and also with analysis.
the Advanced Network and Intelligent Connection Technology Key Labora- With the developments of neural encoders [16], [17], the
tory of Chongqing Education Commission of China, Chongqing Key Lab- knowledge of syntactic structures provided by external parsers
oratory of Ubiquitous Sensing and Networking, Chongqing 400065, China
(e-mail: [email protected]; [email protected]; s200131265 plays an important role in targeted sentiment analysis, which can
@stu.cqupt.edu.cn; [email protected]). help to establish connections between the target and contexts
Boran Yang is with the School of Communication and Information Engi- in a sentence to learn syntax-aware representations of target
neering, Chongqing University of Posts and Telecommunications, Chongqing
400065, China, also with the Advanced Network and Intelligent Connec- mentions. However, traditional neural encoders lacked effective-
tion Technology Key Laboratory of Chongqing Education Commission of ness in dealing with such graph-structured data as in syntactic
China, Chongqing Key Laboratory of Ubiquitous Sensing and Networking, dependency trees. Consequently, recent approaches [18], [19]
Chongqing 400065, China, and also with the School of Artificial Intelli-
gence, Chongqing University of Technology, Chongqing 400065, China (e-mail: leverage graph neural networks, such as graph convolutional net-
[email protected]). work (GCN) [20], graph attentional network [21], to encode the
Digital Object Identifier 10.1109/TASLP.2024.3350877 syntactic structures of input sentences and model dependency

2329-9290 © 2024 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission.
See https://www.ieee.org/publications/rights/index.html for more information.

Authorized licensed use limited to: Mepco Schlenk Engineering College. Downloaded on February 24,2025 at 08:22:52 UTC from IEEE Xplore. Restrictions apply.
1110 IEEE/ACM TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 32, 2024

The contributions of our work are summarized as follows.


r A novel IS2 tree is designed to integrate semantic infor-
mation with the syntactic dependency tree and overcome
the parsing errors in syntax-insensitive scenarios by in-
troducing semantic dependency labels and enhancing the
dependency relations by target-specific reshaping.
r A novel DCGCN is proposed to simultaneously cou-
ple the syntactic and semantic relations in IS2 trees.
Specifically, DCGCN exploits a dynamic pruning mech-
anism (DPM) based on the semantic dependency labels
to avoid the interference caused by irrelevant opinion
words in IS2 trees. In this way, DCGCN can not only
pay attention to semantic information, but also reduce
the redundancy and complexity in the syntax-insensitive
Fig. 1. Example syntactic dependency tree of a sentence provided by an scenarios.
external parser. r The experimental results on four public datasets demon-
strate that DCGCN is a competitive alternative comparing
to the state-of-the-art baselines and verify the generaliza-
trees into adjacency matrices. As shown in Fig. 1, each word tion capability of the IS2 tree. Our code will be released at
in the dependency tree is regarded as a node in the graph, and https://github.com/lnfszrj/DCGCN.
the syntactic dependency relations between words correspond
to an edge in the graph. In particular, the fine-grained syntactic II. RELATED WORK
dependency labels in the graph can distinguish the contextual
From the perspective of information extraction, existing
information that is related to the target mention and achieve
works of targeted sentiment analysis can be divided into two
better performance [22].
categories: methods based on semantic information and methods
However, due to informal and complex reviews, the inaccurate
based on syntactic structures [22], [26], [27].
results of the syntactic dependency trees provided by an external
parser are inevitable on syntax-insensitive datasets [23], [24],
[25], [26]. Moreover, the GCN based models perform poorly in A. Semantic-Based Methods
modeling the long-distance association or disconnected words Targeted sentiment analysis methods based on semantic in-
in syntactic dependency trees [27]. As shown in Fig. 1, target formation extract features from two parts of a given sentence,
mention “mocha” and decisive opinion word “happy” are dis- including the target mention and contexts, and predict the
tant (at least 7-hops) in the syntactic dependency tree due to the emotional polarity of the target mention by combining these
complex syntactic structures, resulting in a weak dependency features. For example, Zhang et al. [23] exploited Recurrent
relation between them. Therefore, it is crucial to alleviate the Neural Network (RNN) to extract features from the target and
errors introduced by the external parser in syntax-insensitive contexts, and then employed a gate mechanism to further inte-
scenarios and appropriately enhance the relations between the grate these features. Tang et al. [24] designed two Long-Short
target mention and its related yet long-distance opinion words Term Memory (LSTM) networks to encode contexts and the
in syntactic dependency trees. target mention, and aggregated the output features of these
To address the above problems, we first design an Integrated two LSTMs to capture the global contextual information for
Syntactic and Semantic tree (IS2 tree) by embedding semantic classification. With the developments of neural networks, the
dependency labels into syntactic dependency trees to better convolutional neural network (CNN) is leveraged by sentiment
represent the semantic similarity between the target mention and analysis tasks to encode input sentences. Huang et al. [29]
contexts. In this way, the model can avoid the inaccurate parsing integrated parameter filters in the model, which could obtain
of syntactic dependency trees by external parsers in syntax- the semantic relations between the target mention and rele-
insensitive scenarios. Through the target-specific reshaping, the vant opinion words. Xue et al. [30] introduced gate mech-
IS2 tree can enhance the dependency relations between the target anisms in CNN to encode sentences for better information
mention and long-distance opinion words. In addition, we design integration.
a novel Dual-Channel Graph Convolutional Network (DCGCN) Attention mechanisms [31] are recently proposed to capture
to couple the semantic and syntactic labels in IS2 trees through the semantic interaction between targets and contexts. Wang
the designed dual channel mechanism in parallel, which makes et al. [11] employed LSTM and attention mechanisms to model
our model more lightweight and achieve optimal performance the interactive relationship between the target and the whole
in the syntax-insensitive scenarios. Besides, to introduce the sentence. Ma et al. [10] designed interactive attention networks
semantic information without incurring the noise interference to leverage the semantic similarity between the target men-
of words from sentences, our proposed DCGCN leverages a tion and contexts in the linear interaction form of sentences.
dynamic pruning mechanism, which can acquire the semantic However, models based on attention mechanisms have inherent
information relevant to the target mention in the IS2 tree. defects, which make the targeted sentiment analysis vulnerable

Authorized licensed use limited to: Mepco Schlenk Engineering College. Downloaded on February 24,2025 at 08:22:52 UTC from IEEE Xplore. Restrictions apply.
ZHANG et al.: INTEGRATED SYNTACTIC AND SEMANTIC TREE FOR TARGETED SENTIMENT CLASSIFICATION USING DCGCN 1111

to the noise interference of words in sentences. In light of this, However, it not only takes a lot of time to parse the sentence with
Fan et al. [13] introduced a multi-grained attention network to multiple parsers but also causes the problem of error accumu-
capture the semantic interaction information of different gran- lations of multiple parsers. Furthermore, Li et al. [54] proposed
ularities to correct the sentiment polarity of the target mention. a DualGCN structure to obtain the feature representation of the
Zeng et al. [14] explored the local context focus mechanism, target by designing two GCNs that extract syntactic and semantic
leveraging the local context words that were close to the target information respectively. However, it still achieved limited per-
to model the semantic relations and eliminate the interference formance in reducing the paring errors of the dependency tree,
caused by distant irrelevant opinion words. Lin et al. [15] de- which accordingly brought more redundancy and complexity to
signed a selective attention mechanism to make full use of the the model. Similarly, Dai et al. [55] employed the syntax-guided
inter-target information. relationship provided by the pre-trained BERT and RoBERTa to
Similar to the attention mechanisms, memory networks are build a dependency tree, whereas the pre-trained model would
also employed to model the interaction between the target men- receive interference from the initial learning data and provide
tion and opinion words. Chen et al. [32] combined memory limited semantic information. Sun et al. [56] built complex
networks and attention mechanisms to obtain the long-term con- three-graph structures to capture the interaction between doc-
textual information. Wang et al. [33] proposed target-sensitive uments for document-level relationship extraction, which was
memory networks to preserve the contextual information that not suitable for sentiment analysis. However, we only used the
was relevant to the target. However, the contextual information concise IS2 tree to capture the structural information, semantic
of the sentence could be forgotten with the stacking of the information, and syntactic information of the sentence at the
memory networks. same time, and it was suitable for sentiment analysis, which
In addition, BERT [34], ALBERT [35], and other large-scale could capture the interaction between the target and the context
pre-trained language models were published to further improve words. Lan et al. [57] proposed to use syntactic information and
the performance of NLP tasks and targeted sentiment classifica- dependency information between target mentions. However, in
tion tasks. Since the knowledge provided by BERT is beneficial syntactic-insensitive scenarios, the lack of semantic information
for downstream tasks [36], we combine it with the contextual in the syntactic dependency tree would lead to wrong word in-
encoder in our model to embed the sentence. teractions, so it could cause error propagation when the sentence
contains multiple target mentions. Dai et al. [58] used GCN to
learn the semantic information in the syntactic structure of the
B. Syntactic-Based Methods sentence, but it was based on the original syntactic dependency
Methods based on syntactic structures leverage the syntac- tree without the semantic relationship, which not only intro-
tic knowledge of sentences to model and obtain the feature duced two redundant GCNs to obtain the semantic and syntactic
representations of targets. Early studies [37], [38] exploited information respectively but also increased the parameters of the
manually-extracted syntactic rules. Subsequently, the external adjacency matrix. Moreover, Chen et al. [43] combined syntactic
parser and neural network-based approaches were introduced dependency trees with latent graph structures to obtain the
into the targeted sentiment classification task. Dong et al. [39] target representation. Zhang et al. [44] designed a hierarchical
used an adaptive recursive neural network to encode targets into syntactic and lexical graph based on the syntactic dependency
syntactic dependency trees. Furthermore, Nguyen et al. [40] tree, leveraging the co-occurrence characteristics of words in a
developed phase-level recursive neural networks, employing sentence to improve the classification performance. However,
both the dependency tree and combination tree as the model the syntactic dependency tree provided by external parsers
input to encode the syntactic knowledge of the target mention may negatively affect the model performance because of the
and contexts simultaneously. He et al. [41] established the con- inaccurate dependency parsing and the informal expressions in
nections between the target and contexts based on the syntactic syntax-insensitive scenarios.
relative distance between them in the syntactic dependency tree Recent research found that semantic relations between target
to distinguish the importance of contexts. mentions and contexts could be exploited to enhance the encod-
To further improve the ability to process graph-structure data ing of syntax-insensitive and informal reviews [23], [24], [25].
(e.g., dependency trees), recent studies introduce graph neu- Zhang et al. [19] proposed aspect-specific GCN to introduce
ral networks [20], [21] to model the relations between targets the semantic attention based on target mentions to allocate the
and contexts in dependency trees into adjacency matrices. Sun attention weights of opinion words and to obtain the represen-
et al. [18] proposed to leverage GCN for encoding syntactic tations of syntactic structures. Similarly, Tang et al. [26] devel-
features of sentences to exchange contextual information in oped a graph-dependent enhanced double transformer network
the model. Wang et al. [42] reconstructed the target-specific to combine the semantic representations of contexts and the
dependency tree of the sentence to emphasize the importance of syntactic structures extracted by GCN. Bai et al. [22] designed
the contextual information adjacent to the target mention. Fur- an attention mechanism using syntactic dependency labels and
thermore, recent studies combine several different types of graph proved the importance of fine-grained dependency labels in im-
neural networks to perform targeted sentiment analysis tasks. proving classification performance. However, existing methods
For example, Hou et al. [53] designed an ensemble learning based on attention mechanisms fail to address the interference
method to fuse syntactic dependency trees obtained by different caused by irrelevant opinion words with the opposite sentiment
parsers to enhance the syntactic representation of the target. polarity.

Authorized licensed use limited to: Mepco Schlenk Engineering College. Downloaded on February 24,2025 at 08:22:52 UTC from IEEE Xplore. Restrictions apply.
1112 IEEE/ACM TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 32, 2024

Fig. 2. (a) A syntactic dependency tree of an example sentence with two targets of opposite sentiment polarities. The color of each word represents the attention
weight. (b) A syntactic dependency tree of an example sentence with one target mention.

In contrast to existing targeted sentiment classification meth- and they have the identical syntactic relative distance (both
ods, we propose a novel IS2 tree to integrate semantic depen- 2-hops) with target “falafel”, but only “dried” is related to
dency labels onto the syntactic dependency tree provided by “falafel”. Fortunately, the same dependency label “conj” can be
parsers. In this way, the IS2 tree can address the inaccurate distinguished according to the semantic similarity with different
dependency parsing by encoding the semantic dependency la- targets “falafel” and “chicken”. In addition, important opinion
bels. Meanwhile, the dependency relations between the target words are often neglected by the syntactic relative distance in
mention and contexts can be enhanced by the target-specific the dependency tree and not directly associated with the target.
reshaping of IS2 trees. Furthermore, we propose a DCGCN As shown in Fig. 2(b), although “hypes” is not rooted in target
to couple the syntactic and semantic relations in IS2 tree in “average cake”, it plays a decisive role in predicting the senti-
parallel while implementing network updates at each layer based ment polarity. These results indicate that the noise interference
on the dual-channel mechanism. In this way, only a single of words in a sentence cannot be overcome by the syntactic
GCN network is employed to couple syntactic and semantic dependency tree with attention mechanisms and the necessity
information in the case of insensitive syntax, which greatly of enhancing the dependency relations between the target and
reduces the redundancy of the model. Considering the impact of relevant opinion words.
some important fine-grained dependency labels on classification
performance [22], a dynamic pruning mechanism is proposed to
obtain the contextual information that is relevant to the target. B. Labels of Semantic Dependency Relations
In contrast to previous research [42], [43], we propose to in-
III. INTEGRATED SYNTACTIC AND SEMANTIC TREE tegrate semantic information with the syntactic structures using
the IS2 tree to avoid inaccurate dependency parsing and enhance
A. Limitations of Syntactic Dependency Trees
the dependency relations between the target and contexts. For
Generally, the syntactic structure of sentences is obtained each input sentence, a parser is employed to obtain its syntactic
from an external parser. The relationship between words is dependency tree T , and rij represents the syntactic dependency
represented as a directed edge between a pair of nodes and label between word i and j. As shown in Algorithm 1, there are
a specific dependency label. Existing studies can leverage the two main steps to establish an IS2 tree: reshaping (lines 4–6)
syntactic information (e.g., syntactic dependency label [22], and labeling (lines 7–9).
syntactic relative distance [19], [42]) in the dependency tree Reshaping The target mention in a sentence serves as the
provided by an parser to design attention mechanisms. How- root of an IS2 tree. Specially, a multiple-word target mention
ever, due to the informal and complex reviews, these attention- should be regarded as a whole, whose internal dependency label
based methods may not work well and even lead to classi- can be ignored [42]. In addition, the other nodes in a syntactic
fication mistakes. We illustrate two possible defects of the dependency are reshaped as leaf nodes of the IS2 tree, and
syntactic dependency tree and methods based on attention their original syntactic dependency labels rij are preserved. In
mechanisms. this way, the dependency relations between the target mention
As shown in Fig. 2(a), the score under each word represents and contexts in the original syntactic dependency tree can be
the attention weight obtained by the attention-based LSTM [11]. enhanced.
For target “chicken”, the model mistakenly assigns high atten- Labeling The semantic dependency relations between other
tion weights to “but” and “dried”, because they are close to context words and the target mention in sentence s are labeled
the target. Even given the syntactic knowledge, the syntactic as Sim : wjs , which represents the semantic similarity between
dependency labels for “fine” and “dried” are both “conj” syntactic dependency label rij of context word wjs and the

Authorized licensed use limited to: Mepco Schlenk Engineering College. Downloaded on February 24,2025 at 08:22:52 UTC from IEEE Xplore. Restrictions apply.
ZHANG et al.: INTEGRATED SYNTACTIC AND SEMANTIC TREE FOR TARGETED SENTIMENT CLASSIFICATION USING DCGCN 1113

IV. METHODOLOGY
2
An IS tree contains both syntactic and semantic relations,
and therefore we leverage DCGCN to couple the syntactic and
semantic dependency relations of the IS2 tree and dynamic
pruning to select the contextual information relevant to the target
mention. Formally, we denote each sample in the dataset as a
triplet T, S, G, where T = {wi , wi+1 , . . ., wi+k−1 } is a target
mention sequence and S = {w1 , w2 , . . ., wi , . . ., wi+k , . . ., wn }
is a sentence sequence. The lengths of T and S are k and n,
respectively. G = (V, A, R) represents a graph, i.e., the IS2 tree
of sentence S, where V represents the set of all nodes (words)
in the graph, A ∈ Rn×n is an adjacent matrix, and Rij denotes
the dependency label (i.e., rij or Sim : wjs ) between word i and
Fig. 3. Structure of an IS2 tree reshaped and labeled from a syntactic depen- j in the IS2 tree when Aij = 1, otherwise Rij = N one.
dency tree. The overall architecture of our proposed DCGCN model is
shown in Fig. 4, which is mainly composed of four compo-
nents: the contextual encoder, DCGCN encoder, gate unit, and
Algorithm 1: Integrated Syntactic and Semantic Tree. classifier. Specifically, the contextual encoder (e.g., GloVe [59],
BERT) is responsible for vectorizing the input sentences and
Input:Target mention a = {wia , wi+1 a a
, . . ., wi+m }, obtaining the contextual information. The DCGCN encoder
sentence s = {w1 , w2 , . . ., wn }, syntactic dependency
s s s
processes both the syntactic and semantic dependency relations
tree T , and syntactic dependency labels r. in the IS2 tree and selects the contextual information using
Output:Integrated syntactic and semantic tree T̃ . a dynamic pruning mechanism (DPM) based on the semantic
1: Construct the center of the target a as the root R̃ for T̃ ; dependency labels. Finally, the output of the contextual encoder
2: for i → m do and the output of the DCGCN encoder are dynamically aggre-
3: for j = 1 → n do gated by the gate unit and sent to the classifier to predict the
4: if wjs →rji wia or wjs ←rij wia then sentiment polarity (i.e., positive, neutral, negative). The four
5: preserving the original syntactic structures and components are detailed as follows.
dependency labels rij by wjs ←rij R̃
6: else
7: labeling the semantic relations as A. Contextual Encoder
zij = Sim : wjs
In this paper, we consider two sentence encoding structures:
8: embedding semantic dependency labels into the
the first is the bidirectional long short-term memory (Bi-LSTM)
IS2 tree by wjs ←zij R̃
network, which is widely used for contextual encoding of input
9: end if
sentences in sentiment analysis tasks [14], [19]; the second is
10: end for
a large-scale pre-training model named BERT [34]. Previous
11: end for
studies [36] have proved that BERT can significantly improve
12: return T̃
the classification performance of sentiment analysis tasks. To
compare our proposed DCGCN model with baseline models,
we also employ BERT as the contextual encoding layer.
Bi-LSTM Encoder models the bi-directional contextual infor-
target mention i. The purpose of introducing a new dependency
mation of sentences. We employ GloVe [59] embeddings vi ∈
label Sim : wjs is to avoid the incorrect parsing of syntactic
Rdv , lexical tags ti ∈ Rdt , and position embeddings pi ∈ Rdp ,
dependency trees in syntax-insensitive scenarios.
where dv , dt , and dp denote the dimension of the word, the
The structure of an IS2 tree reshaped from a syntactic de-
lexical tag, and position embeddings, respectively. Therefore,
pendency tree is illustrated in Fig. 3. The IS2 tree contains not
the representation of word wi in sentence S can be denoted by
only the syntactic dependency relations but also the semantic
a concatenation of vi , ti and pi as ei = [vi ; ti ; pi ].
information between the target mention and context words us-
Given a word embedding sequence e = {e1 , e2 , . . ., en }, the
ing the semantic dependency labels. If the sentence contains −−−−→ →

more than one target mention, we construct a unique tree for forward LST M is leveraged to generate hidden state h =

→ − → −
→ ←−−−−
each target mention. In summary, our proposed IS2 structure {h1 , h2 , . . ., hn } and the backward LST M to generate hidden
←− ←− ← − ←

has at least two advantages. First, the semantic information is state h = {h1 , h2 , . . ., hn }. Finally, the output hidden state
integrated into the IS2 tree to overcome inaccurate dependency vector H can be obtained as:
parsing in syntax-insensitive scenarios. Second, the dependency
between target mentions and opinion words in the IS2 tree can be −
→ −−−−→
h = LST M (ec ), c ∈ [1, n] (1)
enhanced by using the semantic dependency labels to represent

− ←−−−−
the semantic similarity. h = LST M (ec ), c ∈ [1, n] (2)

Authorized licensed use limited to: Mepco Schlenk Engineering College. Downloaded on February 24,2025 at 08:22:52 UTC from IEEE Xplore. Restrictions apply.
1114 IEEE/ACM TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 32, 2024

Fig. 4. Flowchart of our proposed DCGCN for targeted sentiment classification. The Bi-LSTM encoder can be replaced by BERT. The right half shows the details
of the DPM, where the top K=2 context words with the largest attention weights are selected for aggregation in each iteration based on semantic dependency labels
Sim : wjs .

→ ←
− − on graph data structures [13]. GCN can directly perform con-
H = [h; h] (3)
volutional operations on adjacent nodes to encode information,
BERT Encoder is a pre-trained language model based on the which is then passed through a multi-layer neural network so that
transformer mechanisms. The introduction of BERT in targeted each node in the graph can learn more contextual information.
sentiment analysis tasks can significantly improve the classi- Given a graph with n nodes, its adjacency matrix A ∈ Rn×n as
fication performance [30]. To compare the proposed DCGCN the discretized output of a dependency parser and each word with
model with baselines, we also employ BERT to generate the an adjacency node j ∈ N (i), GCN updates the representations
contextual embeddings of words. To fine-tune the BERT model, of words with the multi-head attention mechanisms [31] at the
each sentence sequence is reconstructed into “[CLS] + target lth layer by:
mention + [SEP] + sentence + [SEP]” as: ⎛ ⎞

e = {e0 , e1 , . . ., ek , ek+1 , ek+2 , . . ., ek+1+n , ek+2+n } (4) Hil = ||M
m=1 σ
⎝ Alm
ij W
lm l−1 ⎠
Hj (6)
j∈N (i)
where, e0 and ek+1 are the vector representations of “[CLS]” 
and “[SEP]”, respectively. The output of BERT is a different 1 , if j ∈ N (i)
ij =
Alm
0 , if j∈/ N (i)
(7)
sequence of words with the same length as the input:
h = {h0 , h1 , . . ., hk , hk+1 , hk+2 , . . ., hk+1+n , hk+2+n } (5) where || represents the concatenation of m vectors, W lm ∈
R M ×d is the parameter matrix in layer l, N (i) refers to the
d

where h0 is the pooling vector of the BERT output, h1 , . . ., hk


set of nodes in A(i) = 1, and σ is the ReLU activation function.
are the contextual feature vector representations of the target
2) Dual-Channel Graph Convolutional Network: GCN ex-
mention, hk+2 , . . ., hk+1+n represent the embedding vectors
ploits the states of neighbor nodes to update the representation of
of sentence sequences, and these embedding vectors output by
target mentions and their syntactic dependency relations. How-
BERT serve as the input of the DCGCN encoder.
ever, the original GCN cannot encode the semantic dependency
Compared with the Bi-LSTM encoder, BERT does not involve
relations introduced by the IS2 tree. In addition, the interference
the lexical, and positional tag embeddings, because BERT can
of the irrelevant opinion words to the target in the IS2 tree should
learn these features from the sentence sequence autonomously.
be pruned for better classification performance.
To solve these above problems, we leverage DCGCN with
B. DCGCN Encoder syntactic and semantic channels to encode the IS2 tree output
DCGCN takes the IS2 tree reshaped from the syntactic de- by Algorithm 1. Different from the DualGCN [54], which en-
pendency tree as the input and further prunes the IS2 tree based codes the syntactic and semantic information between words via
on the semantic dependency relations to acquire the contextual building two GCN networks separately, we only employ a single
information relevant to the target mention. DCGCN network to encode the IS2 tree based on the syntactic
1) Graph Convolutional Network: Graph convolutional net- and semantic channels in parallel. Our model is more lightweight
work is a CNN variant that provide efficient implementations and achieves optimal performance. Furthermore, we introduce

Authorized licensed use limited to: Mepco Schlenk Engineering College. Downloaded on February 24,2025 at 08:22:52 UTC from IEEE Xplore. Restrictions apply.
ZHANG et al.: INTEGRATED SYNTACTIC AND SEMANTIC TREE FOR TARGETED SENTIMENT CLASSIFICATION USING DCGCN 1115

a DPM in the semantic channels of DCGCN to acquire the Algorithm 2: DCGCN with Dynamic Pruning Mechanism.
opinion words relevant to the target mention based on semantic 1: Encode the input sentence H
dependency labels in the IS2 tree, as shown in Fig. 4. In this way, 2: for l = 1 → L do
only the contextual information related to the target mention is 3: for j ∈ N (i) do
encoded. 4: map syntactic dependency label rij into vector γij
Syntactic Channel: Specifically, syntactic dependency re- 5: lm
calculate syntactic attention weight βij by
lations rij are directly mapped into vectors γij ∈ Rdr . The (9)–(10)
syntactic information is extracted from the two-layer neural 6: obtain syntactic representation hl+1
Syni by (8)
networks and the contexts located in the neighbors of the target 7: end for
are calculated to update the syntactic representation hl+1 Syni of 8: for ĵ ∈ D(i) do
the target mention i by 9: calculate semantic attention weight ϕlz
ij by
 (12)–(13)
hl+1
Syni = M
m=1
lm
βij Wzl hlj (8)
10: sum semantic attention ϕlij by (14) and select the
j∈N (i)
top K attention to update the set D(i) by (15)
lm
gij = σ(relu(γij Wr1 + br1 )Wr2 + br2 ) (9) 11: obtain semantic representation hl+1
Simi by (11)
12: end for
lm
exp(gij ) 13: calculate output feature H T by concatenating
lm
βij = N (i) (10)
j=1 exp(gij )
lm syntactic representation hl+1
Syni and semantic
representation hl+1
Semi
where Wr1 and Wr2 are trainable parameter matrices, br1 and
14: end for
br2 denote the bias vectors, σ represents the sigmod activation
15: return D(i) and H T
function, N (i) refers to the set of nodes rooted in the target
lm
mention in the original syntactic dependency tree, and βij
represents the mth syntactic attention weight at layer l.
Semantic Channel with DPM: To couple the semantic depen- where D(i) denotes the set of opinion words related to the target
dency relations in the IS2 tree, we extract semantic information mention i in the sentence processed by DPM. For example, when
between the target mention and contexts based on the semantic K = 2, the sequence output by DP M2 ({0.6, 0.1, 0.3, 0.2}) is
dependency label Sim : wjs . Furthermore, DPM can eliminate {0.6, 0.3}, and the corresponding opinion words are added to
the interference of the semantic information irrelevant to the D(i). After each iteration of the DPM, the relevant contexts in
target in the IS2 tree. The semantic representation hl+1 D(i) are updated according to the adjacency relation between
Simi of the
target mention i can be updated as: the target mention and contexts.
 Finally, the DCGCN encoder concatenates syntactic repre-
Semi =m=1
hl+1 M
ϕlm l l
ij Wm hj (11) sentation hl+1 l+1
Syni and semantic representation hSimi filtered out
j∈D(i) by DPM into:
Sim : wjs H̃ l+1 = hl+1
Syni  hSemi
l+1
(16)
ij =
ϕlm  (12)
Sim : wĵs
ĵ ∈
/N (i) H l+1 = ReLU (H̃ l+1 W + b) (17)


k where H l+1 is the output of the DCGCN encoder at layer l + 1.
Sim : wjs = Similarity(target, rij ) = hc Ws · γij (13)
c=1 C. Gate Unit
where Ws represents the trainable parameter matrix, hc is the To learn a composite representation containing both contex-
inner word vector of a multi-word target mention, γij refers to tual and target-specific information and to control the infor-
the vector mapped from the original syntactic dependency label mation fusion ratio, we introduce a fine-grained feature fusion
of word wjs , and ϕlmij represents the attention weight provided mechanism. This feature fusion mechanism leverages a gate unit
by the semantic dependency label Sim : wjs . In this way, the to aggregate target representation H L output by the last DCGCN
semantic information relevant to the target mention hc contained encoder and global contextual representation H T of the target
in γij can be obtained. In addition, the contextual information mention output by a mask mechanism [16]. Because the target
relevant to the target mention is encoded by a DPM based on the mention is viewed as a single whole term in the IS2 tree, the
semantic dependency labels. Specifically, M attention weights averaging pooling operation is unnecessary. The feature fusion
ij of the context words not in N (i) are sorted, where the first
ϕlm process is:
K attention weights are selected and recorded in D(i):
Hf ull = g ◦ H L + (1 − g) ◦ H T (18)

M
ϕlij = ϕlm
ij (14) where ◦ refers to the element-wise product operation and g
m=1 denotes the fusion rate:
l
ψij = DP MK ({ϕlij |j ∈
/ N (i)}), j ∈ D(i) (15) g = sigmod(Wg [H L ||H T ] + bg ) (19)

Authorized licensed use limited to: Mepco Schlenk Engineering College. Downloaded on February 24,2025 at 08:22:52 UTC from IEEE Xplore. Restrictions apply.
1116 IEEE/ACM TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 32, 2024

TABLE I TABLE II
STATISTICS OF DATASETS INPUT/OUTPUT DIMENSION OF MODEL SUB-MODULES

are 300-dimensional GloVe [59] embedding vectors as the initial


vector representations of the words in the sentence and the values
where || represents the vector concatenation, Wg and bg are of the vectors can be fine-tuned by the training. The dimensions
trainable parameters. of the lexical tag embeddings, location tag embeddings, synactic
and semantic dependency label embeddings are set to 20. The
D. Classifier L2 regularization term is λ = 10−5 . The learning rate of the
Adamax optimizer [47] is set to 10−3 . Each channel of DCGCN
The classifier is a fully connected layer that calculates the
contains M = 5 vector representations to ensure the stability
probability of each sentiment polarity:
of the model. The input/output dimension of each module is
Classif ier(Hf ull ) = sof tmax(Wp Hf ull + bp ) (20) listed in Table II. All experimental results were averaged over
ten runs. In addition, the tool we use to obtain semantic labels
where Hf ull is the fusion vector, Wp is the trainable param- in IS2 tree is the Biaffine parser [48]. The method of obtaining
eter, and bp is a bias vector. DCGCN is trained by a gradient semantic labels can be divided into three steps. First, we record
descent algorithm, using the cross-entropy loss function and the the syntactic dependency labels of the words that are not rooted
regularization term λ: at the target mention in the syntactic dependency tree obtained by

C the parser. Second, we artificially concatenate the word and the
Loss(θ) = − gi log(yi ) + λ  θ2 (21) string “Sim” into a new semantic dependency label representing
i=1 semantic similarity to the target mention and embed it in the IS2
where C denotes the sentiment label classes, gi is the one-hot tree. Finally, the semantic information related to the target in the
ground truth vector of the label, yi is the output vector calculated syntactic dependency label of the word is mined by the semantic
by (20), and θ contains all trainable parameters. channel of DCGCN.

E. Training B. Baselines
The training process consists of two stages in each iteration Our proposed DCGCN is compared with state-of-the-art tar-
as in Algorithm 2. During the first step, DCGCN encodes the geted sentiment classification systems, which are divided into
syntactic dependency labels for the syntactic representation two categories: baselines based on semantic information and
of the target mention (lines 3–7). In the second step, DPM baselines based on syntactic information.
selects semantic dependency labels from the IS2 tree to obtain 1) Semantic Baselines:
the semantic representation of the target mention (lines 8–12). r IAN [10] uses two LSTM networks and attention mecha-
Finally, we concatenate these two contextual representations to nisms for the fine-grained interaction between targets and
determine the sentiment polarity (lines 13–14). l ∈ [1, 2, . . ., L] context words.
layers of DCGCN encoders can be stacked to form deeper r MGAN [13] employs the Bi-LSTM to capture contextual
networks. information and a multi-grained attention mechanism to
obtain the relationships between targets and context words.
V. EXPERIMENTS r AOA [49] models the interaction between target men-
tions and context words by introducing the Attention-over-
In this paper, experiments are conducted on four public
Attention module.
datasets including SemEval 2014 Restaurant, SemEval 2014 r DSMN [15] designs a selective attention mechanism to
Laptop [45], ACL14 Twitter [23], and MAMS [46]. Each sen-
make full use of the inter-target information.
tence in each dataset contains the target mention and its cor- r AEN [50] uses an attentional encoder network to enhance
responding sentiment polarity label including positive, neutral,
the feature representation of target mentions.
and negative [20]. The statistics of the samples in each dataset r KGCapsAN [51] utilizes syntactical and n-gram structures
are shown in Table I.
to guide the capsule attention network.
r DC-GCN [57] considers both syntactic structure informa-
A. Settings
tion and multi-aspect sentiment dependencies in sentences
Two types of contextual encoders are employed: Bi-LSTM and employs GCNs to learn its node information represen-
encoders and BERT encoders. Given a Bi-LSTM encoder, there tation.

Authorized licensed use limited to: Mepco Schlenk Engineering College. Downloaded on February 24,2025 at 08:22:52 UTC from IEEE Xplore. Restrictions apply.
ZHANG et al.: INTEGRATED SYNTACTIC AND SEMANTIC TREE FOR TARGETED SENTIMENT CLASSIFICATION USING DCGCN 1117

TABLE III
PERFORMANCE COMPARISONS OF DIFFERENT MODELS ON BENCHMARK DATASETS

2) Syntactic Baselines: r RGAT [22] adds fine-grained syntactic dependency labels


r AdaRNN [40] learns the representation of target mentions and introduces the node attention to form the final attention
in sentences by RNNs on dependency trees. distribution.
r CDT [24] integrates the dependency trees with GCN to r kumaGCN [43] proposes a latent graph mechanism to
learn the target vector using syntactic information. obtain the semantic information and learn the syntactic
r BiGCN [44] employs two GCNs to model syntactic infor- dependency between the targets and contexts.
mation and word co-occurrence information of sentences r DGEDT [26] designs a dual-transformer structure to com-
separately and combines them to obtain the target repre- bine the syntactic and semantic representations of target
sentations for sentiment analysis. mentions.
r TD-GAT [27] uses GAT to capture the syntactic structure r DSS-GCN [58] propose a human cognition-based method
of sentences and model the cross-layer relationships of the for ABSA, which establishes the learning from word se-
network. mantics to sentence syntax.
r MergeGraph [53] designs a graph ensemble technique to
make use of the predictions from different parsers, then
C. Results
GNNs are applied over the resulting graph.
r DualGCN [54] proposes dual graph convolutional net- Table III shows the comparison between DCGCN and base-
works to consider the complementarity of syntax structures lines tagged by their categories: semantic models (w Sem.) and
and semantic correlations simultaneously. syntactic models (w Syn.). Compared with the semantic models
r RoBERTa+ASGCN/RGAT [55] utilizes the pre-trained except DC-GCN, DCGCN achieves a significant improvement
fine-tuned RoBERTa and perturbed masking method to in the classification accuracy on all four datasets.
induce trees in the corresponding datasets. Compared to DC-GCN, although our research predates DC-
r R-GAT [42] reconstructs the dependency tree provided by GCN, our model still achieves comparable or even better per-
the parser based on syntactic relative distance and then formance on the laptop and Twitter datasets. This is because
GAT is utilized to encode it. DCGCN incorporates semantic dependency labels and target-
r ASGCN [19] extracts the contextual information from the specific reconstruction to enhance dependency relationships,
LSTM network as the input to GCN and introduces se- combining semantic information with syntactic dependency
mantic attention and masking mechanisms to obtain the trees, and overcoming parsing errors in syntax-insensitive sce-
representation of target mentions. narios. In addition, the dynamic pruning mechanism based

Authorized licensed use limited to: Mepco Schlenk Engineering College. Downloaded on February 24,2025 at 08:22:52 UTC from IEEE Xplore. Restrictions apply.
1118 IEEE/ACM TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 32, 2024

on semantic dependency labels allows DCGCN to focus on 77.55, and 84.01 on Restaurant, Laptop, Twitter, and MAMS,
semantic information while reducing redundancy and complex- respectively.
ity in syntax-insensitive scenarios. In contrast, DC-GCN does
not consider syntax-insensitive scenarios. By calculating the
model complexity, we found that the training time cost and total VI. ANALYSIS
model parameters of DC-GCN on the Restaurant dataset are
A. Ablation Study
higher than those of DCGCN by 3.43 s and 0.21 M, respectively.
Although DC-GCN performs well on the Restaurant dataset, To verify the effectiveness of the IS2 tree and DCGCN,
using two different GCNs to encode syntactic dependency trees we conduct ablation experiments on all datasets. As shown in
and multi-aspect sentiment graphs increases model complexity Table IV, we consider five baselines for comparison purposes:
and leads to severe overfitting on syntax-insensitive datasets. 1) DCGCN-Attention using the traditional attention mechanisms
For syntactic models, R-GAT calculates the contribution of to take the place of the semantic channels; 2) DCGCN w/o
contextual words based on the syntactic relative distance be- IS2 tree substituing the syntactic dependency tree provided
tween them and the target in the original dependency tree. How- by an external parser for the IS2 tree, i.e., replacing semantic
ever, the information provided by syntactic relative distances is relations Sim : wjs with syntactic relations rij ; 3) DCGCN w/o
limited due to parsing errors in the case of syntactic insensitivity. DPM removing the DPM and directly introducing the semantic
Therefore, it is necessary to label the semantic relations between dependency relations of all context words in the IS2 tree; 4)
words and embed them into the IS2 tree as a supplement to DCGCN w/o Syn-channel taking out syntactic representation
the syntactic information. RoBERTa+ASGCN/RGAT utilizes hl+1 l+1
Syni but reserving semantic representation hSemi ; 5) DCGCN
RoBERTa as a guide to build the dependency tree. However, w/o Sem-channel retaining only syntactic representation hl+1 Syni .
RoBERTa is a pre-trained model, it will receive interference Impacts of Attention Mechanisms
from the initial learning data and can only provide limited Because the attention mechanisms cannot appropriately
semantic information. DCGCN achieves the accuracy improve- handle the noise interference of irrelevant opinion words in
ment of 2.58 on the syntax-insensitive Twitter dataset with the sentences, the accuracy of DCGCN-Attention and DCGCN-
help of IS2 tree, which directly embeds the displayed semantic BERT-Attention decrease by 1.68 and 1.75 percent on average,
relations. DualGCN employs two GCN structures to extract respectively, compared with DCGCN. The average decrease
syntactic and semantic information respectively. However, the in accuracy is 1.47 and 1.75 percent on the syntax-insensitive
original dependency tree is utilized as input to the model, which Twitter and MAMS, which further demonstrates that the
can introduce parsing errors and will make the model more integration of semantic information with syntactic information
redundant and complex. DCGCN encodes IS2 tree by syntactic in the IS2 tree is beneficial in encoding informal reviews.
and semantic channels in parallel, rather than constructing dual Effectiveness of the IS2 tree
GCN structures. It is lightweight by integrating dual-channel The purpose of constructing the IS2 tree is to overcome the
mechanism in a single GCN. In addition, dynamic pruning parsing errors in the original dependency tree and introduce
for IS2 tree can further reduce the complexity of DCGCN. semantic labels to supplement the syntactic relations in the de-
Compared with DualGCN, the accuracy and macro F1 scores pendency tree. As shown in Table IV, when replacing the seman-
of DCGCN are improved by the highest 1.41 and 2.81 on all tic relations with the syntactic relations, we compare DCGCN
datasets. GraphMerge fuses the parsing results provided by w/o IS2 tree to DCGCN, the accuracy and macro F1 scores
different parsers into a syntactic dependency tree to supplement decrease by 2.26 and 2.55 percent on average on all datasets,
syntactic information. However, it not only requires a lot of respectively. Furthermore, the accuracy and macro F1 scores
preparation but also raises an issue of error accumulations across of DCGCN-BERT w/o IS2 tree compared to DCGCN-BERT
multiple parsers. DCGCN only employs one parser to obtain the decrease by 2.35 and 2.65 percent, respectively, which further
original dependency tree and then performs simple pruning and indicates that the semantic relations benefits targeted sentiment
semantic embedding as an effective supplement to syntactic in- classification and the proposed IS2 tree is effective.
formation. DSS-GCN establishes a dual-channel semantic GCN Effectiveness of the DPM
to obtain word semantics from general semantic information and To exclude the interference of contextual information irrele-
structural semantic information. However, in syntax-insensitive vant to the target, DCGCN adopts the DPM when encoding IS2
scenarios, the information obtained by GCN is limited, resulting tree, which calculates its contribution to the sentiment analysis
in poorer performance of DSS-GCN on the Twitter dataset. On by the semantic similarity between context words and the target.
the other hand, DCGCN overcomes parsing errors not only At the same time, with the help of the DPM, our model can
in the Restaurant and Laptop datasets but also in the syntax- avoid invalid connections in the dependency tree, and reduce the
insensitive Twitter dataset by introducing semantic dependency complexity while ensuring the classification accuracy. DCGCN
labels and target-specific reconstruction to enhance dependency w/o DPM lacks a pruning mechanism for the IS2 tree, which
relationships. Moreover, these results also demonstrate the ne- results in the impact of contextual information irrelevant to
cessity of integrating semantics with syntactic dependency trees. the target mention on the classification performance and sig-
With the help of the IS2 tree, DCGCN achieves better results nificantly decreases the accuracy and macro F1 scores by 1.34
than all the baselines, giving macro-F1 scores of 83.97, 80.96, and 1.37 percent compared to DCGCN. The performance of

Authorized licensed use limited to: Mepco Schlenk Engineering College. Downloaded on February 24,2025 at 08:22:52 UTC from IEEE Xplore. Restrictions apply.
ZHANG et al.: INTEGRATED SYNTACTIC AND SEMANTIC TREE FOR TARGETED SENTIMENT CLASSIFICATION USING DCGCN 1119

TABLE IV
EXPERIMENTAL RESULTS OF ABLATION STUDY

BERT-based models is also lower than that of DCGCN, which the IS2 tree, so they cannot beat DualGCN. It is the collaboration
demonstrates the importance of introducing the DPM. of several new modules that makes our model achieve the best
Syn-Channel vs. Sem-Channel performance.
We design a dual-channel structure of syntax and semantics
within DCGCN for two reasons: on the one hand, DCGCN needs
B. Qualitative Case Study
to deal with the syntactic and semantic relations in the IS2 tree,
respectively; on the other hand, the dual-channel structure can As shown in Table V, we extract one example sentence from
help avoid redundant and complex network structures. In this each dataset and compare the attention weights offered by the
way, DCGCN can encode these two relations in parallel while strong baselines (DGEDT, ASGCN, RGAT) and our model
updating each layer of the network. It is also the reason why our DCGCN. We also give the original syntactic dependency tree
model is more lightweight than DualGCN. Comparing DCGCN and the ground truth label for each example. The shades of
to DCGCN w/o syn-channel and DCGCN w/o Sem-channel, the words represent the corresponding attention weights. These
we find that these two ablation models have essentially the same cases indicate the necessity of adding semantic dependency
degree of performance degradation on all datasets. For example, relations about the target mention into targeted sentiment clas-
the accuracy and macro F1 scores of DCGCN are higher than sification task again.
those of DCGCN w/o Syn-channel up to 0.96 and 1.07, and The first example: “This is literally a hot spot when it comes
higher than those of DCGCN w/o Sem-channel up to 1.09 and to the food.” with the target “food”. In the original syntactic
1.13 percent, respectively. These results show that the syntactic dependency tree, the dependency between the opinion term “hot
information of sentences is favorable but the semantic informa- spot” and the target “food” is weakened by the syntactic relative
tion related to the target mention cannot be neglected as well. distance, which leads to ASGCN and RGAT using syntactic
Specifically, the syntactic dependency relations preserved in the structures to make a wrong prediction. However, our DCGCN
IS2 tree are as important as the semantic dependency relations. focuses on the opinion term through the semantic channels
However, the proposed model cannot beat DualGCN (the combined with the semantic dependency labels in the IS2 tree.
most similar model) without any module. First, it is reason- DGEDT may correctly make a prediction with the help of dual
able that the module w/o IS2 Tree cannot beat DualGCN, attention mechanisms.
because once DCGCN uses the original syntactic dependency The second one: “If you are a Tequila fan you will not be
tree without labeling semantic information, the performance will disappointed.” with the double negatives “not be disappointed”.
decrease in the syntax-insensitive scenarios. Second, DCGCN It is difficult to model such a structure using the conventional
cannot accurately identify semantic dependency labels in the methods. For example, the negative expression “not” is easily
IS2 Tree that are related to the target without DPM. There- recognized by traditional methods, while the implicit opinion
fore, it results in a decline in sentiment classification perfor- word “disappointed” is often ignored. DCGCN can make pos-
mance. Third, syn-channel and sem-channel in DCGCN are itive prediction due to the dependency labels in the IS2 tree
respectively responsible for processing the syntactic depen- representing the semantic similarity of the target mention, so
dency labels and semantic dependency labels in the IS2 tree. that make the model consider both “not” and “disappointed”.
If we remove any channel, it means that the model is missing In the third example: “I never tried any external mics with that
some important information provided by the IS2 tree. Similarly, iMac.”, both ASGCN and RGAT are affected by “never tried”
the module DCGCN-Attention doesn’t use the IS2 tree to encode adjacent to the target mention “external mics” in the syntactic
the semantic relationship between words offline, it encodes dependency tree, leading to the negative prediction. The dual
the semantic information between words online during training attention mechanisms in DGEGT also fail to eliminate the noise
DCGCN based on the attention mechanism. Therefore, the above interference of “never”, which leads to the incorrect prediction.
benchmark modules are missing all or part of the information in However, DCGCN utilizes the DPM in the semantic channels

Authorized licensed use limited to: Mepco Schlenk Engineering College. Downloaded on February 24,2025 at 08:22:52 UTC from IEEE Xplore. Restrictions apply.
1120 IEEE/ACM TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 32, 2024

TABLE V
CASE VISUALIZATION FOR ATTENTION WEIGHTS OF DGEDT, ASGCN, RGAT, AND DCGCN

to exclude opinion word “never” that is not related to the target C. Impacts of Different Parsers
mention. The classification accuracy shows the impact of different
Finally, we give a negative example: “Last Kiss by Taylor external parsers on the classification performance. Specifically,
Swift if like the saddest song I have heard.”. There is a long-range
the following baselines are set: 1) Random using the dependency
words distance between the opinion expression and the target tree provided by the Biaffine parser [48], but randomly adjusting
mention, which make the traditional methods hard to detect the syntactic dependency labels corresponding to the words; 2)
implicit semantics. For example, ASGCN and RGAT focus
Stanford employing the syntactic dependency labels constructed
on the irrelevant term “Last Kiss” towards the target mention by Stanford Transition-based Parser [52]; 3) Biaffine applying a
“Taylor Swift”. DCGCN can handle such a sample. It utilizes
Deep Biaffine Parser [48] to obtain the syntactic dependency la-
the semantic dependency structures in the IS2 tree to enhance the
bels. The performances of the Stanford Parser and Deep Biaffine
relation between the target mention and the related opinion term Parser on the Treebank are given in Table VI.
“saddest song”, and employs the DPM to exclude the irrelevant
The impact of syntactic dependency trees provided by differ-
“Last Kiss”.
ent external parsers on the classification accuracy is shown in

Authorized licensed use limited to: Mepco Schlenk Engineering College. Downloaded on February 24,2025 at 08:22:52 UTC from IEEE Xplore. Restrictions apply.
ZHANG et al.: INTEGRATED SYNTACTIC AND SEMANTIC TREE FOR TARGETED SENTIMENT CLASSIFICATION USING DCGCN 1121

TABLE VI
CLASSIFICATION ACCURACY BASED ON DIFFERENT EXTERNAL PARSERS

TABLE VII
ACCURACY WHEN USING DIFFERENT PARSERS AS THE INPUT
Fig. 6. Impacts of iteration number L on all datasets. (a) Accuracy-L curves.
(b) F1-L curves.

hyper-parameters K of the DPM and the classification perfor-


mance. On Restaurant and Laptop, DCGCN achieves the highest
accuracy when K = 2 and K = 3. On MAMS and Twitter,
DCGCN has the best classification performance when K = 4
and K = 5. Since users’ online reviews are generally informal,
more attention should be paid to the contextual information
related to the target mention. The experimental results show
that a small K cannot preserve important (task-relevant) con-
textual information and lead to a decrease in the classification
performance, whereas a large K retains too much contextual
information unrelated to the target mention and also affects the
classification performance.

E. Impacts of Iteration Number


Fig. 5. Impacts of different hyper-parameter K on all datasets. As shown in Fig. 6, we conduct numerous experiments to
(a) Accuracy-K curves. (b) F1-K curves. figure out a proper iteration number (i.e., the number of stacked
DCGCN layers). After each iteration, the DPM reselects the
opinion words to retain and update the representation of the
Table VII. We find that the classification accuracy of DCGCN target mention. DCGCN achieves the best classification per-
using Biaffine and Stanford parsers is not significantly different formance when L = 3 on Laptop, when L = 4 on Twitter and
on all datasets, with a maximum difference of 0.35 percent MAMS. The more DCGCN iterations signify the more complex
on Twitter, although Biaffine has achieved higher UAS and syntactic structures of sentences and the more related semantic
LAS than those of Stanford. The reason is that the IS2 tree information introduced by the semantic channel in DCGCN.
leverages semantic dependency labels to replace some of the However, the model will suffer from over-fitting and redundant
original syntactic dependency labels to avoid the inaccurate information passing, which results in the performance drop when
parsing results introduced by different parsers. These results the iteration number is too large.
also demonstrate the generalization capability of the IS2 tree to
cooperate with different parsers. F. Computational Complexity Analysis
When compared with the Stanford parser and Biaffine parser,
As shown in Table VIII, we compare the total training param-
the random parser has the lower classification accuracy due
eters and time spent for DCGCN with MGAN for the syntax-free
to random syntactic dependency labels. This result indicates
system and DualGCN for the syntax-based system, respectively.
that the accuracy of the syntactic dependency relations pro-
To validate the time complexity, we train DCGCN for five epochs
vided by a parser is also important because they influence the
and report the average time cost. It can be seen from Table VIII
representation of words provided by the syntactic channel in
that the total parameters and training time of DCGCN are slightly
DCGCN.
higher than MGAN because it takes time to encode the IS2 tree.
However, the total parameters of DCGCN are significantly lower
D. Impacts of Hyper-Parameter K than that of DualGCN, and the training time of each epoch is also
Because the noise interference of the contextual informa- shorter. Different from the complex structure that contains two
tion irrelevant to the target mention in the IS2 tree cannot GCN networks in DualGCN, we have pre-embedded semantic
be addressed by the attention mechanism, it is necessary to labels in the IS2 tree. Therefore, only a single DCGCN is re-
introduce the DPM. Fig. 5 shows the relationship between the quired to couple and encode the syntax and semantic information

Authorized licensed use limited to: Mepco Schlenk Engineering College. Downloaded on February 24,2025 at 08:22:52 UTC from IEEE Xplore. Restrictions apply.
1122 IEEE/ACM TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 32, 2024

TABLE VIII
COMPARISON OF MODEL PARAMETERS AND TRAINING TIME COST, 1 M = 1e6

capture accurate users’ sentiment and eliminate semantic noise.


(3) As shown in Fig. 7(c) and 7(d), on the dataset in strong
professional fields or dataset containing many irregular user
comments, DualGCN is less adaptable to the IS2 tree than
DCGCN. On the one hand, the model cannot understand the
semantic information from the professional field. On the other
hand, the irregular user comments will bring some semantic
information irrelevant to the target mention, which requires a
module like DPM in DCGCN to filter the semantic noise in the
IS2 tree.

H. Error Analysis
To analyze the limitation of the proposed model, we further
summarize the prediction errors of DCGCN. There are several
reasons for explaining the findings. First, DCGCN may not
always generate an accurate prediction that reflects what people
really think in ironic sentences. For example, the correct label
for the sentence “The scene hunky waiters dub dinners darling
and it sounds like they mean it.” is “negative”. Our method
tends to classify the sentence as “positive”, since DCGCN cannot
Fig. 7. Transfer Study of IS2 Tree. (a) Twitter. (b) Restaurant. (c) Laptop. understand the literally positive words “hunky” and “darling”,
(d) MAMS. which are actually ironic. Second, our method fails to recognize
the potential implications of the sentences such as “Premium
price for the OS more than anything else”. The correct label
in the IS2 tree while implementing network updates at each layer of the sentence is “positive”, but DCGCN may classify it as
based on the dual-channel mechanism. In addition, DCGCN “negative”. This is because the model cannot realize that the
adopts the DPM to prune irrelevant context words in IS2 tree, more expensive price often means the better OS performance,
which avoids invalid connections in the dependency tree and and the model only cares about “Premium price” and “more”.
reduces the complexity of the model. The above results show
that our algorithm makes DCGCN lightweight while fusing syn-
tactic and semantic information and ensuring the classification VII. CONCLUSION
accuracy.
In this paper, we propose a novel IS2 tree to integrate syntactic
and semantic information. By integrating semantic dependency
G. Transfer Study labels into the syntactic dependency trees, our proposed IS2 tree
To verify the transferability of the IS2 tree, we decouple the can overcome the inaccurate parsing results introduced by the
2
IS tree from DCGCN, and migrate it to DualGCN [54]. As external parser in syntax-insensitive scenarios. Furthermore, we
shown in Fig. 7, we can draw the following conclusions: (1) design DCGCN to encode both syntactic and semantic depen-
As shown in Fig. 7(a), in the syntactic-insensitive dataset, the dency relations in the IS2 tree and select the task-related contex-
IS2 Tree has a significant effect on enhancing the sentiment tual information using the DPM. The experiment results show
classification performance of the model. In this case, the model that DCGCN demonstrate the necessity of integrating semantic
needs the semantic dependency information contained in the IS2 relations with the syntactic structure and the effectiveness of
Tree. (2) As shown in Fig. 7(b), compared with the models with DCGCN in encoding both syntactic and semantic information.
GloVe embeddings, the IS2 Tree can give the models with BERT Worth noting is that the IS2 tree has an excellent generalization
embeddings higher performance gain. BERT can provide the capability to cooperate with different parsers and DCGCN im-
IS2 tree with a large amount of sentence structure information proves the accuracy of targeted sentiment classification tasks by
and semantic information of words, which can help the model large margins.

Authorized licensed use limited to: Mepco Schlenk Engineering College. Downloaded on February 24,2025 at 08:22:52 UTC from IEEE Xplore. Restrictions apply.
ZHANG et al.: INTEGRATED SYNTACTIC AND SEMANTIC TREE FOR TARGETED SENTIMENT CLASSIFICATION USING DCGCN 1123

REFERENCES [25] D. Bahdanau, K. Cho, and Y. Bengio, “Neural machine translation by


jointly learning to align and translate,” in Proc. Int. Conf. Learn. Repre-
[1] P. D. Turney et al., “Thumbs up or thumbs down? Semantic orientation sentations, 2015, pp. 3104–3112.
applied to unsupervised classification of reviews,” in Proc. Assoc. Comput. [26] H. Tang, D. Ji, C. Li, and Q. Zhou, “Dependency graph enhanced dual-
Linguistics, 2002, pp. 417–424. transformer structure for aspect-based sentiment classification,” in Proc.
[2] B. Pang, L. Lee, and S. Vaithyanathan, “Thumbs up? Sentiment classifica- Assoc. Comput. Linguistics, 2020, pp. 6578–6588.
tion using machine learning techniques,” in Proc. Conf. Empirical Methods [27] B. Huang and K. Carley, “Syntax-aware aspect level sentiment classifi-
Natural Lang. Process., 2002, pp. 79–86. cation with graph attention networks,” in Proc. Conf. Empirical Methods
[3] A. Yessenalina, Y. Yue, and C. Cardie, “Multi-level structured models Natural Lang. Process., 2019, pp. 5469–5477.
for document-level sentiment classification,” in Proc. Conf. Empirical [28] D. Tang, B. Qin, and T. Liu, “Aspect level sentiment classification with
Methods Natural Lang. Process., 2010, pp. 1046–1056. deep memory network,” in Proc. Conf. Empirical Methods Natural Lang.
[4] A. Meena and T. V. Prabhakar, “Sentence level sentiment analysis in the Process., 2016, pp. 214–224.
presence of conjuncts using linguistic analysis,” in Proc. Adv. Inf. Retrieval, [29] B. Huang and K. Carley, “Parameterized convolutional neural networks for
2007, pp. 573–580. aspect level sentiment classification,” in Proc. Conf. Empirical Methods
[5] K. Schouten et al., “Survey on aspect-level sentiment analysis,” IEEE Natural Lang. Process., 2018, pp. 1091–1096.
Trans. Knowl. Data Eng., vol. 28, no. 3, pp. 813–830, Mar. 2016. [30] W. Xue and T. Li, “Aspect based sentiment analysis with gated
[6] A. Nazir et al., “Issues and challenges of aspect-based sentiment analysis: convolutional networks,” in Proc. Assoc. Comput. Linguistics, 2018,
A comprehensive survey,” IEEE Trans. Affect. Comput., vol. 13, no. 2, pp. 2514–2523.
pp. 845–863, Apr.–Jun, 2022, doi: 10.1109/TAFFC.2020.2970399. [31] A. Vaswani et al., “Attention is all you need,” in Proc. Adv. Neural Inf.
[7] L. Jiang, M. Yu, M. Zhou, X. Liu, and T. Zhao, “Target-dependent twit- Process. Syst, 2017, pp. 5998–6008.
ter sentiment classification,” in Proc. Assoc. Comput. Linguistics, 2011, [32] P. Chen, Z. Sun, L. Bing, and W. Yang, “Recurrent attention network on
pp. 151–160. memory for aspect sentiment analysis,” in Proc. Conf. Empirical Methods
[8] J. Zhou, J. X. Huang, Q. Chen, Q.V. Hu, T. Wang, and L. He, “Deep learning Natural Lang. Process., 2017, pp. 452–461.
for aspect-level sentiment classification: Survey, vision, and challenges,” [33] S. Wang, S. Mazumder, B. Liu, M. Zhou, and Y. Chang, “Target-sensitive
IEEE Access, vol. 7, pp. 78454–78483, 2019. memory networks for aspect sentiment classification,” in Proc. Assoc.
[9] Y. Wang, M. Huang, X. Zhu, and L. Zhao, “Attention-based LSTM for Comput. Linguistics, 2018, pp. 957–967.
aspect-level sentiment classification,” in Proc. Conf. Empirical Methods [34] J. Devlin, M. W. Chang, K. Lee, and K. Toutanova, “BERT: Pretraining
Natural Lang. Process., 2016, pp. 606–615. of deep bidirectional transformer for language understanding,” in Proc.
[10] D. Ma, S. Li, X. Zhang, and H. Wang, “Interactive attention networks for North Amer. Chapter Assoc. Comput. Linguistics: Hum. Lang. Technol.,
aspect-level sentiment classification,” in Proc. Int. Joint Conf. Artif. Intell., 2019, pp. 4171–4186.
2017, pp. 4068–4074. [35] Z. Lan, M. Chen, S. Goodman, K. Gimpel, P. Sharma, and R. Soricut,
[11] Y. Wang, M. Huang, and L. Zhao, “Attention-based LSTM for aspect-level “ALBERT: A lite BERT for self-supervised learning of language repre-
sentiment classification,” in Proc. Conf. Empirical Methods Natural Lang. sentations,” in Proc. Int. Conf. Learn. Representations, 2020.
Process., 2016, pp. 606–615. [36] H. Xu, B. Liu, L. Shu, and P. Yu, “BERT post-training for review reading
[12] J. Tang et al., “Progressive self-supervised attention learning for aspect comprehension and aspect-based sentiment analysis,” in Proc. Conf. North
level sentiment analysis,” in Proc. Assoc. Comput. Linguistics, 2019, Amer. Chapter Assoc. Comput. Linguistics: Hum. Lang. Technol., 2019,
pp. 557–566. pp. 2324–2335.
[13] F. Fan et al., “Multi-grained attention network for aspect-level sentiment [37] G. Qiu, B. Liu, J. Bu, and C. Chen, “Opinion word expansion and target
classification,” in Proc. Conf. Empirical Methods Natural Lang. Process., extraction through double propagation,” Comput. linguistics, vol. 37, no. 1,
2018, pp. 3433–3442. pp. 9–27, 2011.
[14] B. Zeng et al., “LCF: A local context focus mechanism for aspect-based [38] K. Liu, H. Xu, Y. Liu, and J. Zhao, “Opinion target extraction using
sentiment classification,” Appl. Sci., vol. 16, 2019, Art. no. 3389. partially-supervised word alignment model,” in Proc. Int. Joint Conf. Artif.
[15] P. Lin, M. Yang, and J. Lai, “Deep selective memory network with Intell., 2013, pp. 2134–2140.
selective attention and inter-aspect modeling for aspect level sentiment [39] L. Dong et al., “Adaptive recursive neural network for target-dependent
classification,” IEEE/ACM Trans. Audio, Speech, Lang. Process., vol. 29, twitter sentiment classification,” in Proc. Assoc. Comput. Linguistics,
pp. 1093–1106, 2021. 2014, pp. 49–54.
[16] K. S. Tai, R. Socher, and C. D. Manning, “Improved semantic represen- [40] T. H. Nguyen and K. Shirai, “Phrasernn: Phrase recursive neural network
tations from tree-structured long short-term memory networks,” in Proc. for aspect-based sentiment analysis,” in Proc. Conf. Empirical Methods
Assoc. Comput. Linguistics, 2015, pp. 1556–1566. Natural Lang. Process., 2015, pp. 2509–2514.
[17] X. Li, L. Bing, W. Lam, and B. Shi, “Transformation networks for target [41] R. He, W. S. Lee, H. T. Ng, and D. Dahlmeier, “Effective attention
oriented sentimnet classification,” in Proc. Assoc. Comput. Linguistics, modeling for aspect-level sentiment classification,” in Proc. 27th Int. Conf.
2018, pp. 946–956. Comput. Linguistics, 2018, pp. 1121–1131.
[18] K. Sun, R. Zhang, S. Mensah, Y. Mao, and X. Liu, “Aspect-level sentiment [42] K. Wang et al., “Relational graph attention network for aspect-based senti-
analysis via convolution over dependency tree,” in Proc. Conf. Empirical ment analysis,” in Proc. Assoc. Comput. Linguistics, 2020, pp. 3229–3238.
Methods Natural Lang. Process., 2019, pp. 5679–5688. [43] C. Chen, Z. Teng, and Y. Zhang, “Inducing target-specific latent structures
[19] C. Zhang, Q. Li, and D. Song, “Aspect-based sentiment classification with for aspect sentiment classification,” in Proc. Conf. Empirical Methods
aspect-specific graph convolutional networks,” in Proc. Conf. Empirical Natural Lang. Process., 2020, pp. 5596–5607.
Methods Natural Lang. Process., 2019, pp. 4568–4578. [44] M. Zhang and T. Qian, “Convolution over hierarchical syntactic and lexical
[20] T. N. Kipf and M. Welling, “Semi-supervised classification with graph graphs for aspect level sentiment analysis,” in Proc. Conf. Empirical
convolutional networks,” in Proc. Int. Conf. Learn. Representations, Methods Natural Lang. Process., 2020, pp. 3540–3549.
2017. [45] M. Pontiki, D. Galanis, J. Pavlopoulos, H. Papageorgiou, I. Androutsopou-
[21] P. Velickovic, G. CucurullA. Casanova, A. Romero, P. Liò, and Y. Bengio, los, and S. Manandhar, “SemEval-2014 task 4: Aspect based sentiment
“Graph attention networks,” in Proc. Int. Conf. Learn. Representations, analysis,” in Proc. SemEval, 2014, pp. 27–35.
2018. [46] Q. Jiang, L. Chen, R. Xu, X. Ao, and M. Yang, “A challenge dataset
[22] X. Bai, P. Liu, and Y. Zhang, “Investigating typed syntactic dependencies and effective models for aspect-based sentiment analysis,” in Proc. Conf.
for targeted sentiment classification using graph attention neural network,” Empirical Methods Natural Lang. Process., 2019, pp. 6280–6285.
IEEE/ACM Trans. Audio, Speech, Lang. Process., vol. 29, pp. 503–514, [47] D. P. Kingma and J. Ba, “Adam: A method for stochastic optimization,”
2021. in Proc. Int. Conf. Learn. Representations, 2015.
[23] M. Zhang, Y. Zhang, and D. T. Vo, “Gated neural networks for targeted [48] T. Dozat and D. C. Manning, “Deep biaffine attention for neural depen-
sentiment analysis,” in Proc. AAAI Conf. Artif. Intell., 2016, pp. 3087– dency parsing,” in Proc. Int. Conf. Learn. Representations, 2017, pp. 1–8.
3093. [49] B. Huang, Y. Ou, and K. M. Carley, “Aspect level sentiment classifica-
[24] D. Tang, B. Qin, X. Feng, and T. Liu, “Effective LSTMs for tion with attention-over-attention neural networks,” in Proc. Int. Conf.
target-dependent sentiment classification,” in Proc. COLING, 2016, Social Comput., Behav.-Cultural Model. Prediction Behav. Representation
pp. 3298–3307. Model. Simul., 2018, pp. 197–206.

Authorized licensed use limited to: Mepco Schlenk Engineering College. Downloaded on February 24,2025 at 08:22:52 UTC from IEEE Xplore. Restrictions apply.
1124 IEEE/ACM TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 32, 2024

[50] Y. Song et al., “Attentional encoder network for targeted sentiment classi- Boran Yang (Member, IEEE) received the Ph.D.
fication,” 2019, arXiv:1902.09314. degree from the Chongqing University of Posts and
[51] B. Zhang, X. Li, X. Xu, K.-C. Leung, Z. Chen, and Y. Ye, “Knowledge Telecommunications and joined the Chongqing Uni-
guided capsule attention network for aspect-based sentiment analysis,” versity of Technology, Chongqing, China as a faculty
IEEE/ACM Trans. Audio, Speech, Lang. Process., vol. 28, pp. 2538–2551, Member and research Secretary with the School of
2020. Artificial Intelligence. His research interests include
[52] D. Chen and C. Manning, “A fast and accurate dependency parser us- edge computing, edge resource sharing, and network
ing neural networks,” in Proc. Conf. Empirical Methods Natural Lang. security. He has authored or coauthored more than
Process., 2014, pp. 740–750. 20 technical papers in top journals such as IEEE
[53] X. Hou et al., “Graph ensemble learning over multiple dependency trees for TRANSACTIONS ON GREEN COMMUNICATIONS AND
aspect-level sentiment classification,” in Proc. Assoc. Comput. Linguistics, NETWORKING, IEEE TRANSACTIONS ON VEHICULAR
2021, pp. 2884–2894. TECHNOLOGY, IEEE TRANSACTIONS ON MULTIMEDIA, IEEE INTERNET OF
[54] R. Li et al., “Dual graph convolutional networks for aspect-based sentiment THINGS JOURNAL, and GLOBECOM. He was the co-recipient of the Best Paper
analysis,” in Proc. Assoc. Comput. Linguistics, 2021, pp. 6319–6329. Awards from IEEE MSN 2020 and IEEE GreenCom 2019. He was the Technical
[55] J. Dai et al., “Does syntax matter? A strong baseline for Aspect-based Editor/Guest Editor of Digital Communications and Networks, Sensors, and
Sentiment Analysis with RoBERTa,” in Proc. Assoc. Comput. Linguistics, World Electric Vehicle Journal and reviewer for IEEE TRANSACTIONS ON
2021, pp. 1816–1829. COMMUNICATIONS and IEEE INTERNET OF THINGS JOURNAL. He was also a
[56] Q. Sun et al., “Dual-channel and hierarchical graph convolutional net- TPC Member of IEEE Healthcom 2023.
works for document-level relation extraction,” Expert Syst. Appl., vol. 205,
pp. 117678–117688, 2022.
[57] Z. Lan et al., “Dual-channel interactive graph convolutional networks
for aspect-level sentiment analysis,” in Mathematics, vol. 10, no. 18,
pp. 3317–3331, 2022.
[58] A. Dai et al., “Learning from word semantics to sentence syntax by graph
convolutional networks for aspect-based sentiment analysis,” Mathemat- Yuexian Li is currently working toward the mas-
ics, vol. 14, no. 1, pp. 17–26, 2022. ter’s degree with the Chongqing University of Posts
[59] J. Pennington et al., “Glove: Global vectors for word representa- and Telecommunications, Chongqing, China. Her re-
tion,” in Proc. Conf. Empirical Methods Natural Lang. Process., 2014, search interests include sentiment analysis, machine
pp. 1532–1543. translation, and natural language processing.

Puning Zhang received the Ph.D. degree from the


Beijing University of Posts and Telecommunications,
Beijing, China, in 2017. He is currently an Associate
Professor with the School of Communication and
Information Engineering, Chongqing University of
Posts and Telecommunications, Chongqing, China.
Zhigang Yang (Member, IEEE) received the M.S. de-
His research interests include Internet of Things
search and sentiment analysis. gree in 2006 from the Chongqing University of Posts
and Telecommunications, Chongqing, China, where
he is currently working toward the Ph.D. degree. He is
an Associate Professor with the Chongqing Univer-
sity of Arts and Sciences, Chongqing. His research
Rongjian Zhao is currently working toward the mas- interests include edge computing, network security,
ter’s degree with the Chongqing University of Posts and privacy.
and Telecommunications, Chongqing, China. His re-
search interests include sentiment analysis, machine
translation, and natural language processing.

Authorized licensed use limited to: Mepco Schlenk Engineering College. Downloaded on February 24,2025 at 08:22:52 UTC from IEEE Xplore. Restrictions apply.

You might also like