Rumor Identification in Twitter Data For 2020 US Presidential Election Using BERT Model
Rumor Identification in Twitter Data For 2020 US Presidential Election Using BERT Model
Journal QR Code:
Article QR:
Copyright
Information: This article is open access and is distributed under the terms of
Creative Commons Attribution 4.0 International License
A publication of the
Dr Hasan Murad School of Management
University of Management and Technology, Lahore, Pakistan
Rumor Identification in Twitter Data for 2020 US
Presidential Election using BERT Model
Abdul Rahim 1*
ABSTRACT: Social media validation of the information shared.
platforms provide rich resources to Also, once established, such models
their users to connect, share and help find the behavior of rumors and
search for the information of their patterns in American politics.
interest. These platforms are even KEYWORDS:
more significant for governmental BERT Model, rumor detection,
issues and political campaigns. As social media, US elections
information spreads within seconds,
it is incredibly challenging to I. INTRODUCTION
control and monitor the authenticity Social media and microblogging
of the information. Many attempts
platforms are great examples of the
have been made in this regard. This
paper briefly overviews some latest technologies becoming part of
significant efforts and discusses the our daily lives. These
patterns of rumors and fake news communication advancements are
using the latest machine learning used in personal and professional
techniques. For this purpose, we domains, although they are used
extracted the tweets, specifically primarily in journalism and for
with the hash tag Donald Trump,
extending political influences.
during the high time of the 2020 US
presidential election in order to test Nevertheless, such communication
their authenticity. Similar data was platforms allow their users to
extracted from the FactCheck experience more information flow
websites Snopes.com, quickly, easily, and cost-free.
factcheck.org, and politifact.org. These are undeniable advantages,
We applied the already established although they have given rise to
BERT model to train the data and
unnecessary competitiveness,
tested one million tweets. We found
the model as reliably accurate and which, in turn, has resulted in
proposed that once all the truthful inappropriate use of these exciting
information is saved and pretrained new technologies. Indeed, their dark
in the model, it can auto-identify the side is becoming more prominent
with time due to the spread of false focus on the 2020 presidential
propaganda, fake news, and election.
tampered information, creating a The current research aims to review
battleground for hate speech [1], [2]. the previously utilized techniques
All of this has resulted in and then evaluate the BERT model
irreversible damage in various to classify the rumors based on a
forms, that is, damage to mental small set of information. Efforts are
health and repute and social being made to control incorrect
shaming [3], [4]. Such negative and information flow; indeed,
unethical use of social platforms remarkable efforts have been made
remains in the limelight for quite by Twitter in this regard [10].
some time, and this fact challenges However, fake news is still a norm
the reliability and survival of these in the glamour world when it comes
modern communication platforms. to journalism and politics. Keeping
It also creates prospects for machine in view the said consideration, this
learning techniques to provide study investigates the circulation of
solutions. rumors related to politicians during
With the rise of online and fact- elections as it provides us with
checking platforms and machine selected patterns for rumor
learning techniques, control over the propagation.
spread of incorrect information is The rest of the paper is organized as
being improved, though the need for follows. In Section 2, we briefly
a reliable autodetection system is review the existing studies related to
still there [5], [6]. This study focuses rumors in general by applying
on rumors and disinformation machine learning models,
propagated during elections, particularly in US elections. We
suggesting how these can be tackled proceed with the paper by
using machine learning techniques. presenting Section 3, including data
Specifically, we were inspired by preparation, the modeling aspect,
the studies on rumors regarding the and the results derived. Later, in
2016 US presidential election [7], Section 4, we discuss the results and
[8], [9]. In this work, we mainly
the limitations and prospects of the Similarly, Harrag et al. [14] carried
current work. out another study that employed the
II. LITERATURE REVIEW BERT model with GPT2 to
recognize and classify the
Many studies have been carried information (in this case, tweets) as
out to identify and classify rumors either human-generated or machine-
and fake news by employing generated. They specifically
machine learning methods; targeted the Arabic language tweets
however, the recent trend has and compared their predictions with
shifted towards advanced deep hybrid models, such as RNN,
learning and hybrid approaches LSTM, GRU, and their variants.
[10], [11], [12]. New tools and They reported 98% accuracy for
technologies are emerging with their data.
advancements in the machine Indeed, much work has been carried
learning domain. We utilized out with considerable accuracy by
complex techniques by eschewing employing the BERT model, though
the details to achieve the goals more lack of resources, absence of
meaningfully. This section context, and unavailability of
discusses some prominent and latest standard corpora for fake or
studies explicitly aimed at rumor propaganda news are the challenges
detection using advanced machine faced in this domain [13]. Da San
learning approaches. Extensive Martino et al [13] proposed a dataset
work has been carried out by of annotated news articles with 14
leveraging BERT and its variants as fake approaches to address this
the foundation model. For instance, issue. This study proposed a BERT-
in the study of [13], the authors based model adopted by Patil,
proposed a combination of Singh, and Agarwal [14] for
Convolutional Neural Network SemEval 2020-Task 11. The
(CNN) with BERT. Adding the proposed approach consists of two
CNN layer was to enhance the aspects: identification of
word's semantic representation with propaganda and classification of the
varying sentence lengths. In doing techniques used to disseminate it
so, the said authors achieved results (among 14 classes), such as
with 98.9% accuracy.
Dr Hasan Murad School of Management
47
Volume 1 Issue 1, Spring 2021
Rumor Identification in Twitter Data …
making it easier for us to know if the BERT model. The maximum length
data requires modification in of the sentence in the dataset was 55
annotations. Moreover, we (as shown in Fig. 3), and we set it to
manually cross-validated the results 128 for training purposes.
based on the information provided Consequently, it required
by FactCheck and people voter significantly less effort to develop
websites. the calibrated model. We fine-tuned
This practice enabled us to prepare the top fully connected layer with
reliable data corresponding to the the word embedding vector. Also,
desired objectives and the we opted for a broad-match strategy
identification and reliable that used a minimum number of
classification of rumors. keywords to predict as many tweets
b. Model Preparation that can be classified as rumors.
Col_0 Col_1
Row_0 66 40
Row_1 9 95
With these specs, we could run the Therefore, the proposed model
model provided that if we required achieved significantly accurate
more RAM for computing data, we results with minimal data loss and
minor information loss. Further, we
used local hardware and saved the
compare the train and test data. This
file. Moreover, if we needed GPU comparison contains
for computing, we used Kaggle a smaller set of train data and
hardware resources. It took around 3 extensively tested data with the
hours to predict the results. We same context. It can be established
made pandas data frames for these that bidirectional, pre-trained, word
predictions and saved them into embedding BERT leads to faster
training of model and lower cross-
CSV.
entropy loss.
The model displayed a precision
score of 77% accuracy (as given in V. CONCLUSION