4 TH
4 TH
74
Type Precision Recall F1 score Type Precision Recall F1 score
BERT(CR) 0.672 0.673 0.671 BERT(C) 0.587 0.589 0.585
BERT(C) 0.695 0.701 0.693 BERT(CR) 0.493 0.492 0.477
BERT(PCRW) 0.704 0.705 0.703 BERT(R) 0.679 0.679 0.679
BERT(PR) 0.638 0.638 0.637
BERT(PCW) 0.703 0.703 0.703
LR(CR) 0.526 0.526 0.526
BERT(PC1RW) 0.677 0.678 0.677
LR(R) 0.563 0.564 0.563
BERT(PC1W) 0.689 0.690 0.689
NB(R) 0.557 0.557 0.557
RNN-LSTM(CR) 0.361 0.361 0.361 SVC(R) 0.551 0.551 0.550
BERT(R) 0.722 0.722 0.722 XGB(R) 0.539 0.543 0.528
BERT(PC2R) 0.658 0.685 0.645 SVC(CR) 0.516 0.516 0.516
BERT(PR) 0.706 0.706 0.706 XGB(CR) 0.544 0.544 0.544
SVM(CR) 0.646 0.647 0.646
Table 4: Results for Reddit Dataset
NB(CR) 0.672 0.672 0.672
NB(R) 0.632 0.632 0.632
LR(R) 0.642 0.643 0.642 0.8 BERT(R)
0.6 LR(R)
Table 3: Results for Twitter Dataset 0.4 NB(R)
0.2 SVC(R)
0 XGB(CR)
4 Results Precision Recall F1 score
We have evaluated the test data of Twitter and Figure 4: Results analysis for Reddit Dataset
Reddit dataset which is shared by Figurative
Language processing 2020 shared task organizers. PC1RW represents the processed isolated first
The performance is evaluated by using the metrics context of meaningful words and response, PC1W
as precision, recall and F1 score. We have chosen represents the isolated first context of meaningful
the classifiers to predict the test data based on the words without response, R represents the
performance of the cross validation of training response, PC1R represents the processed second
data. We have performed to predict the test data context with response, PR represents the
by using various combinations of Conversion processed response. The results of the approaches
context and response that are CR represents the are presented in the Table 3 shows the response
combined context of sentences with response, C text from conversion dialogue by using BERT
represents the combined full context of sentences have higher performance than others for the
without response, PCRW represents the processed shared task of the Twitter dataset and the Table 4
combined context of meaningful words and shows BERT response text from conversion
response, PCW represents the combined full dialogue thread performs well for the shared task
context of meaningful words without response, of the Reddit dataset. The best results have
obtained by using BERT model with the isolated
0.8 response(R) text for both the Twitter and Reddit
RNN-LSTM(CR)
dataset respectively. We have noticed that the
0.6 BERT(R) BERT performs well in continuous conversion
0.4 SVM(CR) dialogues or continuous sentences with previous
NB(CR) dialogues compared with the meaningful words
0.2 from conversion context. In both the dataset, the
LR(R)
0
RNN-LSTM performs poor than the SVM, NB
Precision Recall F1 score and LR because of the smaller dataset. The
machine learning approach performs better with
Figure 3: Results analysis for Twitter Dataset the smaller dataset. But the BERT model performs
75
well for the response text of both the Twitter and International Workshop on Semantic Evaluations ,
Reddit dataset with the non-grammatical pages 519–535, Association for Computational
sentences even the data size is small. Figure 3 Linguistics.
shows the chart representations of the Raj Kumar Gupta, and Yinping Yang. 2017.
performance analysis of the different methods in CrystalNest at SemEval-2017 Task 4: Using
the Twitter data. Figure 4 shows the chart Sarcasm Detection for Enhancing Sentiment
representations of the performance analysis of the Classification and Quantification, ACM.
different methods in the Reddit data. David Bamman and Noah A. Smith. 2016.
Contextualized Sarcasm Detection on Twitter,
5 Conclusion Association for the Advancement of Artificial
Intelligence (www.aaai.org).
We have implemented traditional machine Mondher Bouazizi And Tomoaki Otsuki (Ohtsuki),.
learning, deep learning approach and BERT 2016. A Pattern-Based Approach for Sarcasm
model for identifying the sarcasm from Detection on Twitter, IEEE. Translations and
Conversion dialogue thread and to detecting content mining, Digital Object Identifier
sarcasm from social media. The approaches are 10.1109/ACCESS.2016.2594194
evaluated on Figurative Language 2020 dataset. Kalaivani A and Thenmozhi D. 2019. Sentimental
The given utterance of combined text and isolated Analysis using Deep Learning Techniques,
text are preprocessed and vectorized using word International journal of recent technology and
embeddings in deep learning models. We have engineering, ISSN: 2277-3878.
employed RNN-LSTM to build the model for Meishan Zhang, Yue Zhang, and Guohong Fu,. 2016.
both the datasets. The instances are vectorized Tweet Sarcasm Detection Using Deep Neural
using Doc2Vec and TFIDF score for traditional Network, Proceedings of COLING 2016, the 26th
machine learning models. The classifiers namely International Conference on Computational
Logistic Regression (LR), Random Forest Linguistics: Technical Papers, pages 2449–2460.
Classifier (RF), XGBoost Classifier (XGB), Malave N., and Dhage S.N. 2020. Sarcasm Detection
Linear Support vector machine (SVC), Gaussian on Twitter: User Behavior Approach. In: Thampi S.
Naïve Binomial (NB) were employed to build the et al. (eds) Intelligent Systems, Technologies and
models for both the Twitter and Reddit datasets. Applications. Advances in Intelligent Systems and
BERT uncased model with isolated response Computing, vol 910. Springer, Singapore. DOI
https://doi.org/10.1007/978-981-13-6095-4_5.
context gives better results for both the datasets
respectively. The performance may be improved Yitao Cai, Huiyu Cai and Xiaojun Wan. 2019. Multi-
further by using larger datasets. Modal Sarcasm Detection in Twitter with
Hierarchical Fusion Model, Proceedings of the
References 57th Annual Meeting of the Association for
Computational Linguistics, pages 2506–2515
Joshi, A., Bhattacharyya, P., and Carman, M. J. 2017. Association for Computational Linguistics.
Automatic sarcasm detection: A survey. ACM
Computing Surveys (CSUR), 50(5), 73. Jacob Devlin, Ming-Wei Chang, Kenton Lee, and
Kristina Toutanova. 2018. Bert: Pre-training of
Ghosh, D., Fabbri, A. R., and Muresan, S. 2018. deep bidirectional transformers for language
Sarcasm analysis using conversation context. understanding. arXiv preprint arXiv:1810.04805.
Computational Linguistics, 44(4), 755-792.
Khodak, M., Saunshi, N., and Vodrahalli, K. 2017. A
large self-annotated corpus for sarcasm. arXiv
preprint arXiv:1704.05579.
Aniruddha Ghosh, and Tony Veale. 2016. Fracking
Sarcasm using Neural Network”, research gate
publication, Conference Paper. DOI:
10.13140/RG.2.2.16560.15363.
Keith Cortis, Andre Freitas, Tobias Daudert, Manuela
Hurlimann, Manel Zarrouk, Siegfried Handschuh,
and Brian Davis. 2017. SemEval-2017 Task 5:
Fine-Grained Sentiment Analysis on Financial
Microblogs and News”, Proceedings of the 11th
76