Tailieu RGAT
Tailieu RGAT
7360
Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing, pages 7360–7370,
November 16–20, 2020. c 2020 Association for Computational Linguistics
into account; it is inspired by relational graph an effective representation of other positional vari-
convolutional networks (RGCN) (Schlichtkrull ations with absolute or relative position encodings.
et al., 2018) and graph attention networks (GAT)
(Veličković et al., 2017). This method takes into 2 Related Work
account the conversational context by using a di-
Emotion Recognition in Conversation Several
rected graph, where the nodes denote individual
studies have tackled the ERC task. Hazarika
utterances, the edges represent relationships be-
et al. (2018a,b) used memory networks for rec-
tween pairs of nodes (utterances), and the labels
ognizing humans emotion in conversation, where
of the edges represent the types of relationships.
two distinct memory networks consider the inter-
However, graph-based neural networks do not take
speaker interaction. DialogueRNN (Majumder
sequential information contained in utterances into
et al., 2019) employs an attention mechanism
account. Table 1 also represents the importance of
for grasping the relevant utterance from the en-
the sequential information. B’s emotional change
tire conversation. More related to our method
at utterance ]4 is caused by utterance ]3 rather than
is the DialogueGCN model proposed by Ghosal
]2 or ]1. In this way, human emotions may depend
et al. (2019), in which RGAT is used for model-
on more immediate utterances in the temporal or-
ing both self-dependency and inter-speaker depen-
der, and thus it is essential to take the sequence of
dency. This model has achieved state-of-the-art
utterances into account.
performance on several conversational datasets.
A common response to this issue is to en- On the other hand, as a way of considering contex-
code information about absolute position features tual information, Luo and Wang (2019) proposed
(Vaswani et al., 2017) or relative position fea- to propagate each of the utterances into an embed-
tures (Shaw et al., 2018), where these encodings ded vector. Likewise, a pre-trained BERT model
are added to nodes (utterances) or edges (relation- (Devlin et al., 2018) has been used for generating
ships). However, in order to account for self- and dialogue features to combine several utterances by
inter-speaker dependency, our model focuses on inserting separate tokens (Yang et al., 2019).
relation types rather than nodes (utterances) and
edges (relationships); thus, our position encoding Graph Neural Network Graph-based neural
also focuses on relation types. networks are used in various tasks. The fun-
In this paper, we propose novel position encod- damental model is the graph convolutional net-
ings (relational position encodings) that provide work (GCN) (Kipf and Welling, 2016), which
the RGAT model with sequential information re- uses a fixed adjacency matrix as the edge weight.
flecting relation types. By using the relational po- Our method is based on RGCN (Schlichtkrull
sition encodings, our RGAT model can capture et al., 2018) and GAT (Veličković et al., 2017).
both the speaker dependency and the sequential in- The RGCN model prepares a different structure
formation. Experiments on four ERC benchmark for each relation type and hence considers self-
datasets showed that our relational position en- dependency and inter-speaker dependency sepa-
coding outperformed baselines and state-of-the-art rately. The GAT model uses an attention mech-
methods. In addition, our method outperformed anism to attend to the neighborhood’s representa-
both the absolute and relative position encodings. tions of the utterances.
In summary, our contributions are as follows: Position Encodings In our work, positional in-
(1) For the first time, we apply position encod- formation is added to the graphical structure. Sev-
ings to RGAT to account for sequential informa- eral studies add position encodings to several
tion. (2) We propose relational position encodings structures, such as self-attention networks (SANs)
for the relational graph structure to reflect both se- and GCN. SANs (Vaswani et al., 2017) perform
quential information contained in utterances and the attention operation under the position-unaware
speaker dependency in conversations. (3) We con- assumption, in which the positions of the input are
duct extensive experiments demonstrating that the ignored. In response to this issue, the absolute po-
graphical model with relational position encod- sition (Vaswani et al., 2017) or relative position
ings is beneficial and that our method outperforms (Shaw et al., 2018), or structure position (Wang
state-of-the-art methods on four ERC datasets. (4) et al., 2019) are used to capture the sequential or-
We also empirically demonstrate that our model is der of the input. Similarly, graph-based neural net-
7361
Contextual Embedding Emotion Classification
BERT
Speaker Dependency Modeling
BERT
RGAT
BERT
BERT FFNN
Labels
BERT
Position Encodings
BERT
-2 -1 -1 0 -1 -1 -2
BERT
Concatenation
Figure 1: Our entire framework. First, we obtain a contextual embedding for each utterance by using BERT.
Then, we modify this embedding by using RGAT to consider speaker dependency. The position encodings in the
RGAT structure take sequential information into account. Finally, after concatenating the contextual embedding
to the output embedding through RGAT, we classify the concatenated vector into emotion labels by using a fully
connected feed-forward network.
works do not take sequential information. In the logueGCN model does not.
design of proteins, the relative spatial structure be-
tween proteins is modeled in order to account for 3.1 Contextual Utterance Embedding
the complex dependencies in the protein sequence We generate contextual utterance features from the
and is applied to the edges of the graph represen- tokens by following the method in (Luo and Wang,
tations (Ingraham et al., 2019). 2019). First, every utterance u1 , u2 , · · · , uN is
tokenized by the BPE tokenizer (Sennrich et al.,
3 Method 2015), i.e., ui = (ui,1 , ui,2 , · · · , ui,Ti ), where Ti
denotes the number of tokens. The tokens are
First, we define the problem of the ERC task. The
embedded through WordPiece embeddings (Wu
task is to recognize emotion labels (Happy, Sad,
et al., 2016). The pre-trained uncased BERT-Base1
Neutral, Angry, Excited, and Frustrated) of utter-
model converts the token embeddings into con-
ances u1 , u2 , · · · , uN , where N denotes the num-
textualized token representations, which can be
ber of utterances in a conversation. Let sm for
converted to the vector representations via max
m = 1, · · · , M be a collection of speakers in a
pooling, so that they are regarded as the contex-
given conversational dataset, where M denotes the (0)
number of speakers. The utterance ui is uttered by tual utterance embeddings hi ∈ RDm for i =
speaker sm , where m is the correspondence be- 1, · · · , M , where Dm denotes the dimension of
tween the utterance and its speaker. the utterance embeddings. This BERT model is
Our framework consists of three components - fine-tuned through a training process.
contextual utterance embedding, speaker depen- 3.2 Speaker Dependency Modeling with
dency modeling with position encodings and emo- Position Encodings
tion classification. The entire model architecture
is shown in Figure 1. Although our method is Graph-based neural networks are used to cap-
based on the DialogueGCN (Ghosal et al., 2019) ture the speaker dependency features of conver-
model, it considers the positional information con- sations. We design relational graph attention net-
tained in utterances in a sequential conversation 1
See https://github.com/google-research/bert for
as described in Section 3.2.3, whereas the Dia- details.
7362
works to capture both self-dependency and inter- contrast to the 8 types used by DialogueGCN2 .
speaker dependency of utterances. In addition, we In addition, the window sizes p and f repre-
introduce an attention mechanism to attend to the sent the number of past or future utterances from
neighborhood’s representations of the utterances. a target utterance in a neighborhood where each
Furthermore, novel position encodings (relational utterance ui has an edge with the p utterances
position encodings) are added to the graph to ac- (i.e. ui−1 , ui−2 , · · · , ui−p ), the f utterances (i.e.
count for the sequential information contained in ui+1 , ui+2 , · · · , ui+f ), and itself. An appropriate
utterances. window size has to be determined because a small
window makes each utterance connect to too small
3.2.1 Graphical Structure a neighborhood while an immense window size
makes the calculation very expensive. Although
We introduce the following notation: we denote
the window size can be different for each type, we
directed and multi-graphs as G = (V, E, R) with
determine the same window size for each relation.
a node (utterance) vi ∈ V and a labeled edge (re-
lation) (vi , r, vj ) ∈ E, where r ∈ R is a relation 3.2.2 Edge Weight
type. We introduce an edge weight by using an attention
mechanism. Although our attention mechanism is
Nodes Representation Each utterance in a con- based on the GAT (Veličković et al., 2017) model,
versation is represented as a node vi ∈ V. Each it is independent for each relational type r:
node vi is initialized with the contextual utterance
(0)
embeddings hi . Through a stack of graphical
layers, this embedding is modified by aggregating αijr = softmaxi LRL aTr [Wr hi ||Wr hj ]
their neighborhood’s representations, described as (1)
(L)
hi , where L denotes the number of graphical where αijr denotes the edge weight from a tar-
layers. get utterance i to its neighborhood j under rela-
tional type r, Wr denotes a parametrized weight
Labeled Edges Representation Following the matrix for the attention mechanism, ar denotes
state-of-the-art method (Ghosal et al., 2019), the a parametrized weight vector, and ·T represents
labeled edges depend on two aspects: (a) speak- transposition. After applying LeakyReLU nonlin-
ers dependency - this depends upon both self- earity (LRL), a softmax function is used to obtain
dependency and inter-speaker dependency. In the incoming edges whose sum total weight is 1.
detail, the former indicates how utterance ui of 3.2.3 Position Encodings
speaker sm influences sm ’s other utterances (in-
We propose relational position encodings for the
cluding itself). On the other hand, the latter de-
relational graph attention networks. Our position
scribes how utterance ui of speaker sm influences
encodings are based on the relative position since
the other speaker sk6=m ’s utterances; (b) temporal
it is appropriate for graph-based neural networks.
dependency - this also depends on temporal turns
The target utterance feature is connected to its
in conversation. Namely, it relies upon whether
neighborhood by an edge in the graph. There-
one utterance uj is uttered in the past or future
fore, in order to account for the sequential infor-
of the target utterance ui . While the future de-
mation between them, we need to consider the dis-
pendencies are not used in on-going conversation,
tance from the target to its neighborhood, which
the ERC task is an offline system. Furthermore,
is undoubtedly the relative distance between ut-
as past utterances plausibly influence future ut-
terances. Furthermore, we follow the speaker de-
terances, the converse may help the model fill in
pendency modeling described in 3.2.1 and use re-
some missing information like the speaker’s back-
lational graph attention networks. It is necessary
ground. For these reasons, we take the converse
that the sequential information depends on the re-
influence into account, referring to (Ghosal et al.,
lation type r. In summary, we use a different rel-
2019).
ative distance for each relation type, which is re-
Accordingly, there are four relation types of
2
edges: (1) self - past type, (2) inter - past type, The type of DialogueGCN depends on 2 distinct speak-
ers and therefore implies 2 × 4 distinct relation types, which
(3) self - future type, and (4) inter - future type, indicates that both the speaker dependency and the temporal
described as (r1 , r2 , r3 , r4 ). Note that this is in dependency are prepared for each distinct speaker.
7363
weight in (1) as
αijr = softmaxi LRL aTr [Wr hi ||Wr hj ] + PEijr
(3)
Figure 2: Example of relational positions. The rela- To add position encodings to the edge weight,
tional position depends on each relational type, and the our relational position has the same scalar dimen-
background color represents the relational type from sion as the edge weight. Because it is a scalar
the target utterance h4 . These positions, which are value, it may have limited ability to express po-
based on the relative distance, are different for each re- sitional information. In future studies, we will in-
lation.
crease the dimension of the position encodings.
3.2.4 RGAT
A graphical propagation module modifies the rep-
(l)
resentation of a node hi by aggregating represen-
tations of its neighborhood N r (i), and an attention
mechanism is used to attend to the neighborhood’s
(l−1)
representations. The features hir under relation
r are summed to compose the output embedding
(l)
of a node hi . Through a stack of graphical lay-
PE PE PE PE ers l, the representation of a node changes within
its l-hop neighborhood. We define the propagation
Figure 3: Illustration of relational position encodings. module as follows:
The encodings, which are composed of four represen-
tations, are added to the edges in a graph for each rela-
(l−1) (l−1) (l−1)
X
tion. “PE” denotes the position encodings. hir = αijr Wr(l−1) hj (4)
j∈N r (i)
R
(l) (l−1)
X
ferred to as relational position encodings. Figure 2 hi = hir (5)
illustrates the idea of relational positions. r=1
We compare two types of relational position en- where Wr
(l−1)
denotes a learnable weight matrix
coding, i.e., a fixed function and a learned repre- for each relation r. In addition, We apply multi-
sentation (Gehring et al., 2017). As the fixed posi- head attention to the aggregation module in (4)
tional function, we define its representation as and concatenate its outputs. After this propagation
module in (5), we use layer normalization with
learnable affine transform parameters.
max(−p, min(p, j − i))
r = 1, where j ∈ N 1 (i)
N 2 (i)
max(−p, min(p, j − i)) r = 2, where j ∈ 3.3 Emotion Classification
P Eijr =
max(−f, min(f, j − i)) r = 3, where j ∈ N 3 (i)
(L)
N 4 (i) After obtaining the representations hi of each
max(−f, min(f, j − i)) r = 4, where j ∈
(2)
node through the speaker dependency modeling
where P Eijr denotes the relational distance from
with relational position encodings, we concatenate
a target utterance i to its neighborhood j under (0)
the contextual utterance embeddings hi and the
relational type r. The maximum relational po- (L)
sition is clipped to a size of p or f , which de- representation of hi . The concatenated vector is
notes the window size of past or future utterances. classified by using a fully connected feed-forward
N r (i) denotes the neighborhood of the target i un- network, which consists of two linear transforma-
der relation type r. As the learned representations, tions with a ReLu activation between them:
we use one-layer feed-forward neural networks for
positional embeddings, whose argument is the re- Classifier(x) = max(0, xW1 + b1 )W2 + b2
lational fixed function. (6)
Our relational position is based on the relative where W1 and W2 denote learnable weight ma-
position; thus, it can be added to the edge weight, trixes, and b1 and b2 denote learnable bias vec-
as illustrated in Figure 3. We redefine the attention tors.
7364
Conversations Utterances
Datasets Classes Evaluation Metrics
train validation test train validation test
IEMOCAP 108 12 31 5320 490 1623 6 Weighted-F1
MELD 1038 114 280 9989 1109 2610 7 Weighted-F1
EmoryNLP 713 99 85 9934 1344 1328 7 Weighted-F1
DailyDialog 11118 1000 1000 87170 8069 7740 7 Micro-F1
7365
Models IEMOCAP MELD EmoryNLP DailyDialog
CNN 48.18 55.86 32.59 49.34
CNN+cLSTM 54.95 56.87 32.89 50.24
BERT BASE 53.31 56.21 33.15 53.12
KET 59.56 58.18 34.39 53.37
DialogueRNN 62.75 57.03 31.70 50.65
DialogueGCN 64.18 58.10 - -
Ours 65.22 60.91 34.42 54.31
Table 3: Performance of our method, baseline, and state-of-the-art methods on the three test sets (the values in the
table are in terms of the evaluation metrics listed in Table 2). Bold font denotes the best performance. “-” signifies
that no results were reported for the given dataset. “Ours” denotes our methods, which are composed of a BERT
model and RGAT with relational position encodings. The position representations were learned.
initial learning rates of 4e-5 in the BERT struc- more, it achieved a weighted average F1 score
ture and 2e-3 in the RGAT structure and used the of 60.91% on the MELD dataset, outperform-
Adam optimizer (Kingma and Ba, 2014) under ing DialogueGCN by more than 2 points. For
the scheduled learning rate with a batch size of 1. EmoryNLP, it achieved a weighted average F1
The number of dimensions of the contextual em- score of 34.42%. It achieved a micro-averaged
beddings and utterance representations was set to F1 score of 54.31% on the DailyDialog dataset,
768, and the size of the internal hidden layer in improving recognition performance over the base-
the emotion classification module was set to 384. lines and KET model by around 1 point. From
We used 8-head attention for calculating the edge these results, we can see that adding our posi-
weight of RGAT and set 0.1 as the dropout rate in tion encodings caused an improvement over the
the BERT structure. We also carried out experi- baselines, KET, and DialogueGCN on all datasets.
ments with different contextual past window sizes Further, it is obvious that our approach is robust
p and future window sizes f , (1, 1), (2, 2), (3, 3), across datasets having varying training-data sizes,
(10, 10), (all, all), and RGAT layers, 1, 2, 3. We conversation lengths, and numbers of speakers.
selected either a concatenated function or a sum-
mation function as a mixing operation in the emo- 5.2 Analysis of the Experimental Results
tion classification module, as described in 3.3. We Let us investigate the importance of our model
chose the hyper-parameter that achieved the best components by analyzing the predicted emotional
score on each dataset by using development data. labels, as shown in Table 4. The results of the
All of the presented results are averages of 5 runs. model using BERT without speaker dependency
We conducted all experiments on a CentOS server modeling are listed on row ]0, while the results of
using Xeon(R) Gold 6246 CPU with 512GB of DialogueRNN, as described in Section 4.3, are on
memory, and we used Quadro RTX 8000 GPU row ]1. The results of DialogueGCN, as described
with 48GB of memory. in Section 4.3, are reported in ]2. The results of the
BERT and RGAT model without position encod-
5 Results and Discussion ings are on row ]3, and those of our model are on
]4. Note that DialogueGCN’s RGAT differs from
5.1 Comparison with Baselines and
our model in terms of its graphical structure and
State-of-the-Art
relational types.
We compared the performance of our approach As shown in the table, our method did not
with those of the baselines and state-of-the-art achieve the best score for almost all labels. How-
methods listed in Table 3. We have quoted the ever, interestingly, it achieved a state-of-the-art av-
results for the baselines and state-of-the-art re- erage F1 score, which is the target metric on the
sults reported in (Zhong et al., 2019; Ghosal et al., dataset. A possible reason for this performance is
2019), except for the results of BERT BASE on that our method consists of effective components.
IEMOCAP. Each component of BERT and RGAT with posi-
For IEMOCAP, our model obtained a weighted tion encodings worked well for each label. As a re-
average F1 score of 65.22%, outperforming Di- sult, these components led to a strong average per-
alogueGCN by more than 1 point. Further- formance. Each effective component is explained
7366
Background Components
Contextual Speaker
] Models Happy Sad Neutral Angry Excited Frustrated Average
Utterance Dependency
Embedding Modeling
0 BERT BASE BERT × 37.09 59.53 51.73 54.33 54.26 55.83 53.31
1 DialogueRNN CNN, GRU 33.18 78.80 59.21 65.28 71.86 58.91 62.75
2 DialogueGCN CNN, GRU RGAT 42.75 84.54 63.54 64.19 63.08 66.99 64.18
3 Ours(without PE) BERT RGAT 50.69 76.78 65.85 59.66 64.04 62.37 64.36
4 Ours BERT RGAT with PE 51.62 77.32 65.42 63.01 67.95 61.23 65.22
Table 4: Weighted average F1 scores of ours (with or without PE), baseline, and state-of-the-art methods for each
label in the IEMOCAP dataset. Bold font denotes the best performance. “Average” denotes the weighted average
F1 score. The variations of their background components are shown in the third and fourth columns.
7367
] Position Encodings (PE) Type Average experiment by increasing the past and future win-
0 - - 64.36 dow sizes [(1,1), (3,3), (5,5), (7,7), (9,9), (11,11),
1 fixed 63.95
Node-based PE (20,20), (30,30), and (40,40)] on the IEMOCAP
2 learn 64.95
3 fixed 63.97 dataset and compared the results with those of the
Edge-based PE
4 learn 64.59 baseline model using BERT and RGAT without
5
Relational PE
fixed 63.99 positional information. The experimental results
6 learn 65.22 are illustrated in Figure 4.
Table 5: Impact of various position encodings compo- As an illustration, it is clear that both models
nents on the IEMOCAP dataset. The base model using perform better with a window size around 3, 5, 7.
BERT and RGAT without position encodings is shown On the other hand, long utterance information may
in ]0. “fixed” and “learn” denote a fixed function and a obstruct efficient recognition (see the results for
learned representation respectively. a window size around 30, 40). Although it is re-
quired to select a small window size, too small
a size results in poor performance, no better than
choosing a size of 1.
Furthermore, the proposed position encoding
method is robust to a varying window size. As
the window size increased, the baseline model’s
F1 score decreased, while our model maintained
its performance even with a large window. One
possible reason is that, as our position encodings
clearly distinguish between immediate and far ut-
terances, it can reduce the influence of these dis-
Figure 4: Effect of different window sizes on the tant utterances.
weighted average F1 score of our method (Ours) and
the baseline model (Base) on the IEMOCAP dataset.
We plotted the scores by using a marker with a con-
6 Conclusion
fidence interval of 95%, which was estimated using a
We proposed relational position encodings for
bootstrap.
RGAT to recognize human emotions in textual
conversation. We incorporated the relational po-
tion encoding, i.e., a fixed function and a learned sition encodings in the RGAT structure to cap-
representation. ture both speaker dependency and the sequen-
The baseline model using BERT and RGAT tial order of utterances. On four ERC datasets,
without position encodings (]0) had a recogni- our model improved recognition performance over
tion performance of 64.36%. We added various those of the baselines and existing state-of-the-art
position encodings to the baseline model and se- methods. Additional experimental studies demon-
lected fixed functions or learned representations strated that the relational position encoding ap-
as the position representation (from ]1 to ]6). The proach outperformed the other position encodings
model using the relational position encodings with and showed that it is robust to changes in window
learned representations had a recognition perfor- size.
mance of 65.22%, the best score and outperform- In future studies, we plan to increase the number
ing the base model by around 1 point. Our rela- of dimensions of the relational position encodings,
tional position encodings were more effective than since a scalar value may not be able to express po-
the other position encodings. sitional information adequately.
We also found that the fixed functions in various
positions resulted in a score lower than that of the Acknowledgements
baseline model. We can conclude that it is required
to learn a position representation. We would like to thank Dr. Ichiro Yamada, Dr.
Rei Endo, and Hideya Mino for the valuable dis-
5.4 Effect of Varying the Window Size cussions. We also thank the anonymous reviewers
We conducted another experiment to evaluate the for their helpful comments.
key aspects of our framework. We carried out an
7368
References Yoon Kim. 2014. Convolutional neural net-
works for sentence classification. arXiv preprint
James Bradbury, Stephen Merity, Caiming Xiong, and arXiv:1408.5882.
Richard Socher. 2016. Quasi-recurrent neural net-
works. arXiv preprint arXiv:1611.01576. Diederik P Kingma and Jimmy Ba. 2014. Adam: A
method for stochastic optimization. arXiv preprint
Carlos Busso, Murtaza Bulut, Chi-Chun Lee, Abe arXiv:1412.6980.
Kazemzadeh, Emily Mower, Samuel Kim, Jean-
nette N Chang, Sungbok Lee, and Shrikanth S Thomas N Kipf and Max Welling. 2016. Semi-
Narayanan. 2008. IEMOCAP: Interactive emo- supervised classification with graph convolutional
tional dyadic motion capture database. Language networks. arXiv preprint arXiv:1609.02907.
resources and evaluation, 42(4):335.
Jieun Lee and Ilyoo B Hong. 2016. Predicting positive
Junyoung Chung, Caglar Gulcehre, KyungHyun Cho, user responses to social media advertising: The roles
and Yoshua Bengio. 2014. Empirical evaluation of of emotional appeal, informativeness, and creativity.
gated recurrent neural networks on sequence model- International Journal of Information Management,
ing. arXiv preprint arXiv:1412.3555. 36(3):360–373.
Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Yanran Li, Hui Su, Xiaoyu Shen, Wenjie Li, Ziqiang
Kristina Toutanova. 2018. BERT: Pre-training of Cao, and Shuzi Niu. 2017. DailyDialog: A man-
deep bidirectional transformers for language under- ually labelled multi-turn dialogue dataset. arXiv
standing. arXiv preprint arXiv:1810.04805. preprint arXiv:1710.03957.
Jonas Gehring, Michael Auli, David Grangier, Denis Pengfei Liu, Xipeng Qiu, and Xuanjing Huang.
Yarats, and Yann N Dauphin. 2017. Convolutional 2016. Recurrent neural network for text classi-
sequence to sequence learning. In Proceedings fication with multi-task learning. arXiv preprint
of the 34th International Conference on Machine arXiv:1605.05101.
Learning-Volume 70, pages 1243–1252. JMLR. org.
Ilya Loshchilov and Frank Hutter. 2016. SGDR:
Deepanway Ghosal, Navonil Majumder, Soujanya Po- Stochastic gradient descent with warm restarts.
ria, Niyati Chhaya, and Alexander Gelbukh. 2019. arXiv preprint arXiv:1608.03983.
DialogueGCN: A graph convolutional neural net-
work for emotion recognition in conversation. arXiv Linkai Luo and Yue Wang. 2019. EmotionX-HSU:
preprint arXiv:1908.11540. Adopting pre-trained bert for emotion classification.
arXiv preprint arXiv:1907.09669.
Chuan Guo, Juan Cao, Xueyao Zhang, Kai Shu, and
Huan Liu. 2019. DEAN: Learning dual emotion for Navonil Majumder, Soujanya Poria, Devamanyu Haz-
fake news detection on social media. arXiv preprint arika, Rada Mihalcea, Alexander Gelbukh, and Erik
arXiv:1903.01728. Cambria. 2019. DialogueRNN: An attentive rnn for
emotion detection in conversations. In Proceedings
Devamanyu Hazarika, Soujanya Poria, Rada Mihal- of the AAAI Conference on Artificial Intelligence,
cea, Erik Cambria, and Roger Zimmermann. 2018a. volume 33, pages 6818–6825.
ICON: Interactive conversational memory network
for multimodal emotion detection. In Proceedings Rosalind W Picard. 2010. Affective computing: From
of the 2018 Conference on Empirical Methods in laughter to ieee. IEEE Transactions on Affective
Natural Language Processing, pages 2594–2604. Computing, 1(1):11–17.
Devamanyu Hazarika, Soujanya Poria, Amir Zadeh, Soujanya Poria, Erik Cambria, Devamanyu Hazarika,
Erik Cambria, Louis-Philippe Morency, and Roger Navonil Majumder, Amir Zadeh, and Louis-Philippe
Zimmermann. 2018b. Conversational memory net- Morency. 2017. Context-dependent sentiment anal-
work for emotion recognition in dyadic dialogue ysis in user-generated videos. In Proceedings of the
videos. In Proceedings of the 2018 Conference 55th annual meeting of the association for compu-
of the North American Chapter of the Association tational linguistics (volume 1: Long papers), pages
for Computational Linguistics: Human Language 873–883.
Technologies, Volume 1 (Long Papers), pages 2122–
2132. Soujanya Poria, Devamanyu Hazarika, Navonil Ma-
jumder, Gautam Naik, Erik Cambria, and Rada Mi-
Sepp Hochreiter and Jürgen Schmidhuber. 1997. halcea. 2018. MELD: A multimodal multi-party
Long short-term memory. Neural computation, dataset for emotion recognition in conversations.
9(8):1735–1780. arXiv preprint arXiv:1810.02508.
John Ingraham, Vikas Garg, Regina Barzilay, and Michael Schlichtkrull, Thomas N Kipf, Peter Bloem,
Tommi Jaakkola. 2019. Generative models for Rianne Van Den Berg, Ivan Titov, and Max Welling.
graph-based protein design. In Advances in Neu- 2018. Modeling relational data with graph convolu-
ral Information Processing Systems, pages 15794– tional networks. In European Semantic Web Confer-
15805. ence, pages 593–607. Springer.
7369
Rico Sennrich, Barry Haddow, and Alexandra Birch.
2015. Neural machine translation of rare words with
subword units. arXiv preprint arXiv:1508.07909.
Peter Shaw, Jakob Uszkoreit, and Ashish Vaswani.
2018. Self-attention with relative position represen-
tations. arXiv preprint arXiv:1803.02155.
Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob
Uszkoreit, Llion Jones, Aidan N Gomez, Łukasz
Kaiser, and Illia Polosukhin. 2017. Attention is all
you need. In Advances in neural information pro-
cessing systems, pages 5998–6008.
Petar Veličković, Guillem Cucurull, Arantxa Casanova,
Adriana Romero, Pietro Lio, and Yoshua Bengio.
2017. Graph attention networks. arXiv preprint
arXiv:1710.10903.
Xing Wang, Zhaopeng Tu, Longyue Wang, and Shum-
ing Shi. 2019. Self-attention with structural position
representations. arXiv preprint arXiv:1909.00383.
Yonghui Wu, Mike Schuster, Zhifeng Chen, Quoc V
Le, Mohammad Norouzi, Wolfgang Macherey,
Maxim Krikun, Yuan Cao, Qin Gao, Klaus
Macherey, et al. 2016. Google’s neural ma-
chine translation system: Bridging the gap between
human and machine translation. arXiv preprint
arXiv:1609.08144.
Kisu Yang, Dongyub Lee, Taesun Whang, Seolhwa
Lee, and Heuiseok Lim. 2019. EmotionX-KU: Bert-
max based contextual emotion classifier. arXiv
preprint arXiv:1906.11565.
Sayyed M Zahiri and Jinho D Choi. 2018. Emotion de-
tection on tv show transcripts with sequence-based
convolutional neural networks. In Workshops at the
Thirty-Second AAAI Conference on Artificial Intelli-
gence.
Peixiang Zhong, Di Wang, and Chunyan Miao. 2019.
Knowledge-enriched transformer for emotion de-
tection in textual conversations. arXiv preprint
arXiv:1909.10681.
7370