CASCADE: Contextual Sarcasm Detection in Online Discussion Forums
CASCADE: Contextual Sarcasm Detection in Online Discussion Forums
Abstract
arXiv:1805.06413v1 [cs.CL] 16 May 2018
The literature in automated sarcasm detection has mainly focused on lexical, syntactic and
semantic-level analysis of text. However, a sarcastic sentence can be expressed with contextual
presumptions, background and commonsense knowledge. In this paper, we propose CASCADE (a
ContextuAl SarCasm DEtector) that adopts a hybrid approach of both content and context-driven
modeling for sarcasm detection in online social media discussions. For the latter, CASCADE
aims at extracting contextual information from the discourse of a discussion thread. Also, since
the sarcastic nature and form of expression can vary from person to person, CASCADE utilizes
user embeddings that encode stylometric and personality features of the users. When used along
with content-based feature extractors such as Convolutional Neural Networks (CNNs), we see a
significant boost in the classification performance on a large Reddit corpus.
1 Introduction
Sarcasm is a linguistic tool that uses irony to express contempt. Its figurative nature poses a great challenge
for affective systems performing sentiment analysis. Previous research in automated sarcasm detection has
primarily focused on lexical, pragmatic cues found in sentences (Kreuz and Caucci, 2007). Interjections,
punctuations, sentimental shifts, etc., have been considered as major indicators of sarcasm (Joshi et al.,
2017). When such lexical cues are present in sentences, sarcasm detection can achieve high accuracy.
However, sarcasm is also expressed implicitly, i.e., without the use of any explicit lexical cues. Such use
of sarcasm also relies on the context which involves the presumption of commonsense and background
knowledge of an event. When it comes to detecting sarcasm in a discussion forum, it may not only
require understanding the context of the previous comments but also need necessary external background
knowledge about the topic of discussion. The usage of slangs and informal language also diminishes the
reliance on lexical cues. This particular type of sarcasm is tough to detect (Poria et al., 2016).
Contextual dependencies for sarcasm can take many forms. As an example, a sarcastic post from
Reddit1 , “I’m sure Hillary would’ve done that, lmao.” requires background knowledge about the event,
i.e., Hillary Clinton’s action at the time the post was made. Similarly, sarcastic posts like “But atheism,
yeah *that’s* a religion!” requires the knowledge that topics like atheism often contain argumentative
discussions and are more prone towards sarcasm.
In this work, we attempt the task of sarcasm detection in online discussion forums. Particularly, we
propose a hybrid network, named CASCADE, that utilizes both content and contextual-information
required for sarcasm detection. It starts by processing contextual information in two ways. First, it
performs user profiling to create user embeddings that capture indicative behavioral traits for sarcasm.
Recent findings suggest that such modeling of the user and their preferences, is highly effective for
the given task (Amir et al., 2016). It makes use of users’ historical posts to model their writing style
(stylometry) and personality indicators, which are then fused into comprehensive user embeddings using
a multi-view fusion approach, Canonical Correlation Analysis (CCA). Second, it extracts contextual
1
https://www.reddit.com/
information from the discourse of comments in the discussion forums. This is done by document modeling
of these consolidated comments belonging to the same forum. We hypothesize that these discourse
features would give the important contextual information, background cues along with topical information
required for detecting sarcasm.
After the contextual modeling phase, CASCADE is provided with a comment for sarcasm detection. It
performs content-modeling using a Convolutional Neural Network (CNN) to extract its syntactic features.
This CNN representation is then concatenated with the relevant user embedding and discourse features to
get the final representation which is used for classification. The overall contribution of this work can be
summarized as:
• We propose a novel hybrid sarcasm detector, CASCADE that models content and contextual information.
• We model stylometric and personality details of users along with discourse features of discussion
forums to learn informative contextual representations. Experiments on a large Reddit corpus, SARC,
demonstrate significant performance improvement over state-of-the-art automated sarcasm detectors.
In the remaining paper, Section 2 compares our model to related works; Section 3 provides the task
description and proposed approach; here, Section 3.3 explains the process of learning contextual features
comprising user embeddings and discourse features; Section 3.6 presents the hybrid prediction model
followed by experimentation details and result analysis in Section 4; finally, Section 5 draws conclusion.
2 Related Work
Automated sarcasm detection is a relatively recent field of research. The previous works in the literature
can be largely classified into two categories, content and context-based sarcasm detection models.
Content-based: These networks model the problem of sarcasm detection as a standard classification
task and try to find lexical and pragmatic indicators to identify sarcasm. Numerous works have taken
this path and presented innovative ways to unearth interesting cues for sarcasm. Tepperman et al. (2006)
investigate sarcasm detection in spoken dialogue systems using prosodic and spectral cues. Carvalho
et al. (2009) use linguistic features like positive predicates, interjections and gestural clues such as
emoticons, quotation marks, etc. Davidov et al. (2010), Tsur et al. (2010) use syntactic patterns to
construct classifiers. González-Ibánez et al. (2011) also study the use of emoticons, mainly amongst
tweets. Riloff et al. (2013) assert sarcasm to be a contrast to positive sentiment words and negative
situations. Joshi et al. (2015) use multiple features comprising lexical, pragmatics, implicit and explicit
context incongruity. In the explicit case, they include relevant features to detect thwarted sentimental
expectations in the sentence. For implicit incongruity, they generalize Riloff et al. (2013)’s work in
identifying verb-noun phrases containing contrast in both polarities.
Context-based: Usage of contextual sarcasm has increased in the recent past, especially in online
platforms. Texts found in microblogs, discussion forums, social media, etc., are plagued by grammatical
inaccuracies and contain information which is highly temporal and contextual. In such scenarios, mining
linguistic information becomes relatively inefficient and need arises for additional clues (Carvalho et al.,
2009). Wallace et al. (2014) demonstrate this need by showing how traditional classifiers fail in instances
where humans require additional context. They also indicate the importance of speaker and/or topical
information associated to a text to gather such context. Poria et al. (2016) use additional information
by sentiment, emotional and personality representations of the input text. Previous works have mainly
used historical posts of users to understand sarcastic tendencies (Rajadesingan et al., 2015; Zhang et
al., 2016). Khattri et al. (2015) try to find users’ sentiments towards entities in their histories to find
contrasting evidence. Wallace et al. (2015) utilize sentiments and noun phrases used within a forum to
gather context typical to that forum. Such forum based modeling simulates user-communities. Our work
follows similar motivation where we explore context provided by user profiling and the topical knowledge
embedded in the discourse of comments in discussion-forums (subreddits 2 ). Amir et al. (2016) perform
user modeling by learning embeddings that capture homophily. This work is closest to our approach given
2
https://www.reddit.com/reddits/
the fact that we too learn user embeddings to acquire context. However, we take a different approach that
involve stylometric and personality description of the users. Empirical evidence shows that these proposed
features are better than previous user modeling approaches. Moreover, we learn discourse features which
has not been explored before in the context of this task.
3 Method
3.1 Task Definition
The task involves detection of sarcasm for comments made in online discussion forums, i.e., Reddit.
Let us denote the set U = {u1 , ..., uNu } for Nu -users, where each user participates across a subset of
Nt -discussion forums (subreddits). For a comment Cij made by the ith user ui in the j th discussion
forum tj , the objective is to predict whether the comment posted is sarcastic or not.
1 ni −ks
∑ log p(wt ∣di , wt−ks , ..., wt+ks ) (1)
ni t=ks
To predict a word within a window, we take the average of all the neighboring context word vectors
along with the document vector d⃗i and use a neural network with softmax prediction:
ey⃗wt
p(wt ∣di , wt−ks , ..., wt+ks ) = (2)
∑i ey⃗i
Here, y⃗ = [y1 , ..., y∣V ∣ ] is the output of the neural network, i.e.,
y⃗ = Ud h(d⃗i , w
⃗ t−ks , ..., w
⃗ t+ks ; D, Ws ) + b
⃗d (3)
⃗d ∈ R∣V ∣ , Ud ∈ R∣V ∣×ds are parameters and h(⋅) represents the average of vectors d⃗i , w
b ⃗ t−ks , ..., w
⃗ t+ks
taken from D and Ws . Hierarchical softmax is used for faster training (Morin and Bengio, 2005). Finally,
after training, D learns the users’ document vectors which represent their stylometric features.
3.4.2 Personality features
Discovering personality from text has numerous NLP applications such as product recognition, mental
health diagnosis, etc. (Majumder et al., 2017). Described as a combination of multiple characteristics,
personality detection helps in identifying behavior, thought patterns of an individual. To model the
dependencies of users’ personality with their sarcastic nature, we include personality features in the user
embeddings. Previously, Poria et al. (2016) also utilize personality features in sentences. However, we
take a different and more-involved approach of extracting the personality features of a user instead.
For user ui , we iterate over all the vi -comments {Su1i , ..., Suvii } written by them. For each Suj i , we
provide the comment as an input to a pre-trained Convolutional Neural Network (CNN) which has been
trained on a multi-label personality detection task. Specifically, the CNN is pre-trained on a benchmark
corpus developed by Matthews and Gilliland (1999) which contains 2, 400 essays and is labeled with
the Big-Five personality traits, i.e., Openness, Conscientiousness, Extraversion, Agreeableness, and
Neuroticism (OCEAN). After the training, this CNN model is used to infer the personality traits present
in each comment. This is done by extracting the activations of the CNN’s last hidden layer vector which
we call as the personality vector p⃗jui . The expectation over the personality vectors for all vi -comments
made by the user is then defined as the overall personality feature vector p⃗i of user ui :
1 vi j
p⃗i = Ej∈[vi ] [p⃗jui ] = ∑ p⃗ (4)
vi j=1 ui
CNN: Here, we describe the CNN that generates the personality vectors. Given a user’s comment,
which is a text S = [w1 , ..., wn ] composed of n words, each word wi is represented as a word embedding
w⃗ i ∈ Rdem using the pre-trained FastText embeddings (Bojanowski et al., 2016). A single-layered CNN is
then modeled on this input sequence S (Kim, 2014). First, a convolutional layer is applied having three
filters F[1,2,3] ∈ Rdem ×h[1,2,3] of heights h[1,2,3] , respectively. For each k ∈ {1, 2, 3}, filter Fk slides across
Personality CNN
Stylometric
Personality
Post 1
u1 d 1⃗ d N⃗ u p N⃗ u p 1⃗ u1
… post 1 … <END>
… post 2 … <END> Average
Multiview Fusion
CCA Personality CNN Post v 1
ParagraphVector
… post 1 … <END>
User embeddings
Figure 1: The figure describes the process of user profiling. Stylometric and Personality embeddings are generated and then
fused in a multi-view setting using CCA to get the user embeddings.
⃗ k of size R∣S∣−hk +1 ,
S and extracts hk -gram features at each instance. This creates a feature map vector m
whose each entry mk,j is obtained as:
ŷ = σ( W2 q⃗ + b
⃗2 ) , where q⃗ = α( W1 o⃗ + b
⃗1 ) (6)
3.4.3 Fusion
We take a multi-view learning approach to combine both stylometric and personality features into a
comprehensive embedding for each user. We use Canonical Correlation Analysis (CCA) (Hotelling, 1936)
to perform this fusion. CCA captures maximal information between two views and creates a combined
representation (Hardoon et al., 2004; Benton et al., 2016). In the event of having more than two views,
fusion can be performed using an extension of CCA called Generalized CCA (see Supplementary).
Canonical Correlation Analysis: Let us consider the learnt stylometric embedding matrix D ∈ Rds ×Nu
and personality embedding matrix P ∈ Rdp ×Nu containing the respective embedding vectors of user ui
in their ith columns. The matrices are then mean-centered and standardized across all user columns.
We call these new matrices as X1 and X2 , respectively. Let the correlation matrix for X1 be R11 =
X1 X1 T ∈ Rds ×ds , for X2 be R22 = X2 X2 T ∈ Rdp ×dp and the cross-correlation matrix between them be
R12 = X1 X2 T ∈ Rds ×dp . For each user ui , the objective of CCA is to find the linear projections of both
embedding vectors that have a maximum correlation. We create K such projections, i.e., K-canonical
variate pairs such that each pair of projection is orthogonal with respect to the previous pairs. This is done
by constructing:
W = X1T A1 and Z = X2T A2 (7)
where, A1 ∈ Rds ×K , A2 ∈ Rdp ×K and W T W = Z T Z = I. To maximize correlation between W and Z,
optimal A1 and A2 are calculated by performing singular value decomposition as:
−1 −1 −1 −1
R112 R12 R222 = AΛB ⊺ , where A1 = R112 A and A2 = R222 B (8)
Content Modeling Context Modeling Classification
c i,⃗ j
u i⃗
Comment Ci, j
User Embedding
Reddit
is
so
liberal
and
progressive !
Discourse feature
vector t j⃗
Input embedding Convolution with multiple Max-pooling
sequence of sentence filter widths and feature over time
maps
Figure 2: This figure presents the overall hybrid network CASCADE. For the comment Ci,j , its content-based sentential
representation c⃗i,j is extracted using a CNN and appended with context vectors u
⃗ i and t⃗j .
−1 N 2
Loss = ∑ ∑ yi,j log2 (ŷi,j ) , where ŷ = sof tmax(Wo ĉi,j + b
⃗o ) (12)
N i=1j=1
Here, N is the number of comments in the training set, yi is the one-hot vector ground truth of the ith
comment and ŷi,j is its predicted probability of belonging to class j.
4 Experimental Results
4.1 Dataset
We perform our experiments on a large-scale self-annotated corpus for sarcasm, SARC3 (Khodak et al.,
2017). This dataset contains more than a million examples of sarcastic/non-sarcastic statements made
in the social media site Reddit. Reddit comprises of topic-specific discussion forums, also known as
subreddits, each titled by a post. In each forum, users communicate either by commenting to the titled
post or other’s comments, resulting in a tree-like conversation structure. This structure can be unraveled
to a linear format, thus creating a discourse of the comments by keeping the topological constraints intact.
Each comment is accompanied with its author details and parent comments (if any) which is subsequently
used for our contextual processing. It is important to note that almost all comments in the SARC dataset
are composed of a single sentence. We consider three variants of the SARC dataset in our experiments.
• Main balanced: This is the primary dataset which contains a balanced distribution of both sarcastic
and non-sarcastic comments. The dataset contains comments from 1246058 users (118940 in training
and 56118 in testing set) distributed across 6534 forums (3868 in training and 2666 in testing set).
• Main imbalanced: To emulate real-world scenarios where the sarcastic comments are typically lesser
than non-sarcastic ones, we use an imbalanced version of the Main dataset. Specifically, we maintain a
20 ∶ 80 ratio (approx.) between the sarcastic and non-sarcastic comments in both training/testing sets.
• Pol: To further test the effectiveness of our user embeddings, we perform experiments on a subset
of Main, comprising of forums associated with the topic of politics. Table 1 provides the comment
distribution of all the dataset variants mentioned.
The choice of using SARC for our experiments comes with multiple reasons. First, this corpus is the
first of its kind that was purposely developed to investigate the necessity of contextual information in
sarcasm classification. This characteristic aligns well with the main goal of this paper. Second, the large
size of the corpus allows for statistically-relevant analyses. Third, the dataset annotations contain a small
false-positive rate for sarcastic labels thus providing reliable annotations. Also, its self-annotation scheme
rules out the annotation errors induced by third-party annotators. Finally, the corpus structure provides
meta-data (e.g., user information) for its comments, which is useful for contextual modeling.
• Bag-of-Words: This model uses a comment’s word-counts as features in a vector. The size of the
vector is the vocabulary size of the training dataset.
• CNN: We compare our model with this individual CNN version. This CNN is capable of modeling only
the content of a comment. The architecture is similar to the CNN used in CASCADE (see Section 3.2).
• CNN-SVM: This model proposed by Poria et al. (2016) consists of a CNN for content modeling and
other pre-trained CNNs for extracting sentiment, emotion and personality features from the given
comment. All the features are concatenated and fed into an SVM for classification.
• CUE-CNN: This method proposed by Amir et al. (2016) also models user embeddings with a method
akin to ParagraphVector. Their embeddings are then combined with a CNN thus forming the CUE-CNN
model. We compare with this model to analyze the efficiency of our embeddings as opposed to theirs.
Released software4 is used to produce results on the SARC dataset.
4.4 Results
Table 2 presents the performance results on the SARC datasets. CASCADE manages to achieve major
improvement across all datasets with statistical significance. The lowest performance is obtained by the
Bag-of-words approach whereas all neural architectures outperform it. Amongst the neural networks,
the CNN baseline receives the least performance. CASCADE comfortably beats the state-of-the-art
neural models CNN-SVM and CUE-CNN. Its improved performance on the Main imbalanced dataset
also reflects its robustness towards class imbalance and establishes it as a real-world deployable network.
We further compare our proposed user-profiling method with that of CUE-CNN, with absolute dif-
ferences shown in the bottom row of Table 2. Since CUE-CNN generates its user embeddings using a
method similar to the ParagraphVector, we test the importance of personality features being included in
our user profiling. As seen in the table, CASCADE without personality features drops in performance to a
range similar to CUE-CNN. This suggests that the combination of stylometric and personality features are
indeed crucial for the improved performance of CASCADE.
• For the comment Whew, I feel much better now!, its sarcasm is evident only when its previous comment
is seen So all of the US presidents are terrorists for the last 5 years.
• The comment The part where Obama signed it. doesn’t seem to be sarcastic until looked upon as a
remark to its previous comment What part of this would be unconstitutional?.
Such observations indicate the impact of discourse features. However, sometimes contextual cues from
the previous comments are not enough and misclassifications are observed due to lack of necessary
commonsense and background knowledge about the topic of discussion. There are also other cases where
our model fails despite the presence of contextual information from the previous comments. During
exploration, this is primarily observed for contextual comments which are very long. Thus, sequential
discourse modeling using RNNs may be better suited for such cases. Also, in the case of user embeddings,
sarcastic
non-sarcastic
Table 3: Comparison with variants of the proposed CASCADE Figure 3: 2D-Scatterplot of the user embeddings visualized
network. All combinations use content-based CNN. using t-SNE (Maaten and Hinton, 2008).
misclassifications were common for users with lesser historical posts. In such scenarios, potential solutions
would be to create user networks and derive information from similar users within the network. These are
some of the issues which we plan to address in future work.
5 Conclusion
In this paper we introduce Contextual Sarcasm Detector called as CASCADE which leverages both
content and contextual information for the classification. For contextual details, we perform user profiling
along with discourse modeling from comments in discussion threads. When this information is used
jointly with a CNN-based textual model, we obtain state-of-the-art performance on a large-scale Reddit
corpus. Our results show that discourse features along with user embeddings play a crucial role in the
performance of sarcasm detection.
References
Silvio Amir, Byron C Wallace, Hao Lyu, and Paula Carvalho Mário J Silva. 2016. Modelling context with user
embeddings for sarcasm detection in social media. arXiv preprint arXiv:1607.00976.
Adrian Benton, Raman Arora, and Mark Dredze. 2016. Learning multiview embeddings of twitter users. In
Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 2: Short
Papers), volume 2, pages 14–19.
James Bergstra and Yoshua Bengio. 2012. Random search for hyper-parameter optimization. Journal of Machine
Learning Research, 13(Feb):281–305.
Piotr Bojanowski, Edouard Grave, Armand Joulin, and Tomas Mikolov. 2016. Enriching word vectors with
subword information. arXiv preprint arXiv:1607.04606.
Paula Carvalho, Luı́s Sarmento, Mário J Silva, and Eugénio De Oliveira. 2009. Clues for detecting irony in
user-generated contents: oh...!! it’s so easy;-. In Proceedings of the 1st international CIKM workshop on
Topic-sentiment analysis for mass opinion, pages 53–56. ACM.
Na Cheng, Rajarathnam Chandramouli, and KP Subbalakshmi. 2011. Author gender identification from text.
Digital Investigation, 8(1):78–88.
Dmitry Davidov, Oren Tsur, and Ari Rappoport. 2010. Semi-supervised recognition of sarcastic sentences in
twitter and amazon. In Proceedings of the fourteenth conference on computational natural language learning,
pages 107–116. Association for Computational Linguistics.
Roberto González-Ibánez, Smaranda Muresan, and Nina Wacholder. 2011. Identifying sarcasm in twitter: a closer
look. In Proceedings of the 49th Annual Meeting of the Association for Computational Linguistics: Human
Language Technologies: Short Papers-Volume 2, pages 581–586. Association for Computational Linguistics.
David R Hardoon, Sandor Szedmak, and John Shawe-Taylor. 2004. Canonical correlation analysis: An overview
with application to learning methods. Neural computation, 16(12):2639–2664.
Harold Hotelling. 1936. Relations between two sets of variates. Biometrika, 28(3/4):321–377.
Aditya Joshi, Vinita Sharma, and Pushpak Bhattacharyya. 2015. Harnessing context incongruity for sarcasm
detection. In Proceedings of the 53rd Annual Meeting of the Association for Computational Linguistics and the
7th International Joint Conference on Natural Language Processing (Volume 2: Short Papers), volume 2, pages
757–762.
Aditya Joshi, Pushpak Bhattacharyya, and Mark J Carman. 2017. Automatic sarcasm detection: A survey. ACM
Computing Surveys (CSUR), 50(5):73.
Anupam Khattri, Aditya Joshi, Pushpak Bhattacharyya, and Mark Carman. 2015. Your sentiment precedes you:
Using an author’s historical tweets to predict sarcasm. In Proceedings of the 6th Workshop on Computational
Approaches to Subjectivity, Sentiment and Social Media Analysis, pages 25–30.
Mikhail Khodak, Nikunj Saunshi, and Kiran Vodrahalli. 2017. A large self-annotated corpus for sarcasm. arXiv
preprint arXiv:1704.05579.
Yoon Kim. 2014. Convolutional neural networks for sentence classification. In Proceedings of the 2014 Confer-
ence on Empirical Methods in Natural Language Processing (EMNLP), pages 1746–1751.
Diederik P Kingma and Jimmy Ba. 2014. Adam: A method for stochastic optimization. arXiv preprint
arXiv:1412.6980.
Roger J Kreuz and Gina M Caucci. 2007. Lexical influences on the perception of sarcasm. In Proceedings of
the Workshop on computational approaches to Figurative Language, pages 1–4. Association for Computational
Linguistics.
Quoc Le and Tomas Mikolov. 2014. Distributed representations of sentences and documents. In Proceedings of
the 31st International Conference on Machine Learning (ICML-14), pages 1188–1196.
Laurens van der Maaten and Geoffrey Hinton. 2008. Visualizing data using t-sne. Journal of machine learning
research, 9(Nov):2579–2605.
Navonil Majumder, Soujanya Poria, Alexander Gelbukh, and Erik Cambria. 2017. Deep learning-based document
modeling for personality detection from text. IEEE Intelligent Systems, 32(2):74–79.
Gerald Matthews and Kirby Gilliland. 1999. The personality theories of hj eysenck and ja gray: A comparative
review. Personality and Individual differences, 26(4):583–626.
Gerald Matthews, Ian J Deary, and Martha C Whiteman. 2003. Personality traits. Cambridge University Press.
Tomas Mikolov, Ilya Sutskever, Kai Chen, Greg S Corrado, and Jeff Dean. 2013. Distributed representations of
words and phrases and their compositionality. In Advances in neural information processing systems, pages
3111–3119.
Frederic Morin and Yoshua Bengio. 2005. Hierarchical probabilistic neural network language model. In Aistats,
volume 5, pages 246–252. Citeseer.
Soujanya Poria, Erik Cambria, Devamanyu Hazarika, and Prateek Vij. 2016. A deeper look into sarcastic tweets
using deep convolutional neural networks. arXiv preprint arXiv:1610.08815.
Ashwin Rajadesingan, Reza Zafarani, and Huan Liu. 2015. Sarcasm detection on twitter: A behavioral modeling
approach. In Proceedings of the Eighth ACM International Conference on Web Search and Data Mining, pages
97–106. ACM.
Ellen Riloff, Ashequl Qadir, Prafulla Surve, Lalindra De Silva, Nathan Gilbert, and Ruihong Huang. 2013. Sar-
casm as contrast between a positive sentiment and negative situation. In Proceedings of the 2013 Conference
on Empirical Methods in Natural Language Processing, pages 704–714.
Efstathios Stamatatos. 2009. A survey of modern authorship attribution methods. Journal of the Association for
Information Science and Technology, 60(3):538–556.
Joseph Tepperman, David Traum, and Shrikanth Narayanan. 2006. ” yeah right”: Sarcasm recognition for spoken
dialogue systems. In Ninth International Conference on Spoken Language Processing.
Oren Tsur, Dmitry Davidov, and Ari Rappoport. 2010. Icwsm-a great catchy name: Semi-supervised recognition
of sarcastic sentences in online product reviews. In ICWSM, pages 162–169.
Byron C Wallace, Laura Kertz, Eugene Charniak, et al. 2014. Humans require context to infer ironic intent (so
computers probably do, too). In Proceedings of the 52nd Annual Meeting of the Association for Computational
Linguistics (Volume 2: Short Papers), volume 2, pages 512–516.
Byron C Wallace, Eugene Charniak, et al. 2015. Sparse, contextually informed models for irony detection: Ex-
ploiting user communities, entities and sentiment. In Proceedings of the 53rd Annual Meeting of the Association
for Computational Linguistics and the 7th International Joint Conference on Natural Language Processing (Vol-
ume 1: Long Papers), volume 1, pages 1035–1044.
Meishan Zhang, Yue Zhang, and Guohong Fu. 2016. Tweet sarcasm detection using deep neural network. In
Proceedings of COLING 2016, the 26th International Conference on Computational Linguistics: Technical
Papers, pages 2449–2460.