0% found this document useful (0 votes)
3 views6 pages

Fake News Detection Using Machine Learning Algorithms

The document discusses a machine learning approach for detecting fake news using neural networks and linguistic processing techniques. It highlights the challenges posed by the proliferation of misinformation on social media and presents a system that classifies news articles as real or fake with a reported accuracy of 94%. The methodology includes data preprocessing, feature extraction, and the use of classifiers like Random Forest and Logistic Regression to improve detection accuracy.

Uploaded by

rv3912790
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
3 views6 pages

Fake News Detection Using Machine Learning Algorithms

The document discusses a machine learning approach for detecting fake news using neural networks and linguistic processing techniques. It highlights the challenges posed by the proliferation of misinformation on social media and presents a system that classifies news articles as real or fake with a reported accuracy of 94%. The methodology includes data preprocessing, feature extraction, and the use of classifiers like Random Forest and Logistic Regression to improve detection accuracy.

Uploaded by

rv3912790
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 6

Fake News Detection Using Machine Learning

2024 Third International Conference on Electrical, Electronics, Information and Communication Technologies (ICEEICT) | 979-8-3503-6908-3/24/$31.00 ©2024 IEEE | DOI: 10.1109/ICEEICT61591.2024.10718469

Algorithms
E. Bharat Babu a, Kaki Archana a, J Rakesh Goud a, Kadapa Dileep Hussain a, Santhosh Kumar Veeramalla b
a
Electronics and Communication Engineering, B V Raju Institute of Technology, Narsapur, India
b
Electronics and Communication Engineering, BVRIT Hyderabad College of Engineering for Women, Hyderabad, India

[email protected], [email protected], [email protected], [email protected], [email protected]

Abstract—Fake news detection is a news which is not real. Now events of the interest in online format. However, enormous
a days, with the internet everywhere, everyone gets their news challenges also accompany immense possibilities. Since the
from several online sources. As more people used online media mass media has a noticeable influence on society, some people
sites like Facebook, Twitter, and others. Wide-range of false news try to profit from the fact that this influence exists. The media
dissemination include the formation of prejudiced viewpoints and
the manipulation of election results to favor candidates.
may occasionally alter knowledge in a few ways to achieve
Additionally, spammers utilize attention-grabbing news headlines certain objectives. This causes the news stories to be written
as clickbait to monetize their ads. In this research, we use that are either entirely fake or partially true. Even now, there are
principles as neural networks as well as linguistic processing a lot of websites that only provide bogus news. The main
concepts for the binary classification of different types of online objectives Websites that distribute false information typically
news stories that categorize news articles of various kinds that are aim to sway the general population on various of individuals,
available online using binary classification. In addition, apart including but not limited to issues related to affairs of state.
from enabling the consumer to ascertain the legitimacy of the Numerous other countries, including China, Germany, the
headlines, we also want to provide them with the option to flag United States of America, and the Ukraine, also have examples
news as fake or real. In this case, as an attribute harvesting
strategy, we as a species deployed a technique called term
of these websites available [4]. Thus, its feasible that the spread
frequency of an inverse document as frequency spanning an of false information poses a worldwide menace as well as a
assortment word pool plus an N gram and Radom Forest (RF) as planetary assignment. plenty academics claim that their issue of
a classifier. The outcomes obtained from analyzing the real-world incorrect information could be solved with the use of machine
data demonstrate that our algorithm is 94% accurate than the learning alongside synthetic intelligence [5]. The possibilities
baselines in identifying bogus news a few minutes after it spreads. for solving the issue are briefly summarized and approved by
the writers in [6]. According on Commentary regarding the
Keywords—Random Forest Classifier, TF-IDF, Natural particular stories posted in the microblogs, the bloggers
Language Processing, Data Set, Data preprocessing, Machine approach for detecting the bug news is described in [7]. The
learning.
accuracy of the system detection is approximately 70%. In the
I. INTRODUCTION current piece, the ignorant Bayes classification algorithm,
randomly generated forest, and Logical Regression predictive
As more and more people decide to use social media algorithms are used to illustrate an easy approach for
platforms for news consumption and search instead of recognizing misleading information. Provided a news that
traditional news sources [1]. Additionally, research has shown investigate the information set, the study endeavours to
studies have demonstrated whether blogging is presently better discover how these particular strategies contribute to this
instead of television as a significant news source [2]. Stories on particular matter that has been manually labeled and to either
social media are of a lower calibre than those published by validate or refute the notion of using neural networks to detect
established news organizations, despite the advantages that the fraudulent information [8]. These articles different from the
social media offers. Yet, there is a lot of fake news on the others on related subjects in contained within that document,
internet, which is news articles with intentionally false material Logistic Regression that worked particularly by the detection of
produced for several purposes, such as political or commercial misleading information. Additionally, another chance to assess
gain, because journalism can be supplied digitally for a the recently built mechanism's functionality on present
relatively affordable rate, along with because people on social information had been provided by checking it on an inventory
networks can share it swiftly and effortlessly and also quickly of statistics that was legitimately contemporary. Their evidence
[2]. Building up techniques to recognize the bogus news on can often be questionable [9].
social media broadcasts automatically is crucial [3]. Access to
news information is now a lot more convenient and easy thanks II. LITERATURE SURVEY
to the internet and social media [2]. Considering the growing
number of mobile devices and Internet user can easily pursue Mykhailo Granik et al. describe the straightforward
criteria to detect the bug information by Random Forest

Authorized licensed use limited to: Zhejiang University. Downloaded on March 23,2025 at 15:53:44 UTC from IEEE Xplore. Restrictions apply.
classifier in the study [3]. Using a set of Facebook news data
assessed. Sources included three large major political news
outlets and two on left as well as one on the right—that
collected them. They obtained an accuracy of almost 74% in
categorization. The accuracy of classification for bogus news is
somewhat lower. Given that just 4.9% of the dataset is bogus
news, this could be the result of dataset skewness [8] offered a
framework built on a range of machine learning approaches that
tackles several problems, including long processing times, time
lag (BotMaker), and accuracy inadequacies. First, they have
gathered 400,000 tweets from the dataset known as HSpam14.
They continue by providing further details about the 160,000of
spam misleading information and also characterized the 92% of
bug information.
As the first unique idea for a machine learning fake news
detection system came from Marco L. Della Vedova and
Fig 1. Design of the system
colleagues. [10]. It combines social context and news content
elements to improve upon previous approaches seen in the Processing that training data with feature extraction.
literature and achieve a precision of 78.8%. Subsequently, they Features are characteristics or attributes of the data the model
integrated the technique into the Facebook Messenger chatbot will use to learn how to classify new examples, features might
and verified it by recognizing misleading information. Their include the length of the article, the number of exclamation
goal is to ascertain the authenticity or fraudulent Ness of the points, or the sentiment of the language used. Then training the
news article. They started by outlining the datasets they utilized, classifier with extraction. Potentially incorporate user input,
then presented their content-based strategy and suggested a a train the model on the features and labels. Evaluate its
method to include it into the current literature-based social- performance and finally use it to make predictions with
based approach. The most predictive features for crowdsourced probability scores for new, unseen examples it shows the
and journalistic accuracy evaluations are found by feature results.
analysis. The relevance of this work is limited to the collection
of popular tweets because they rely on finding properties such
as given that most tweets.

III. METHODOLOGY
The system is developed in three components, which are
explained in this paper. After eliminating extraneous words and
punctuation. First the data set is classified by using the classifier
of machine learning taxonomy. After studying the training of
the version with three distinct taxonomy, we selected most
effective classifier by precision, recall and accuracy. The next
portion is dynamic research which uses the user’s phase or text
to identify the internet information whether the information is
true or false. And the last one third section which confirms the
URL where buyer entered is legitimate [12]. To train a
classification model, for detection of misleading information, Fig. 2. System Architecture
we are using LIAR dataset, this it encompasses the acquisition
of knowledge that the model will be trained on. Before the IV. IMPLEMENTATION
training process begins, the data in the training set may need to
be pre-processed. This might help us to swab the given data and A. Data collection and analysis
to detach irrelevant information and keep the data in to step by There are several places to find online news, such as news
step format. agency homepages, social networking platforms, fact-checking
websites, and search engines. Online, there are a few publicly
accessible databases for the categorization of fake news,
including BS Detector, Buzzfeed News, LIAR [14], and others
[15]. These databases have been extensively utilized by
numerous research publications to evaluate the veracity of news
reports. Nevertheless, benchmark datasets for the task of
detecting fake news are still nonexistent. The acquired data

Authorized licensed use limited to: Zhejiang University. Downloaded on March 23,2025 at 15:53:44 UTC from IEEE Xplore. Restrictions apply.
must be pre-processed, which entails cleaning, converting, and c) Get rid of stop words-Common keywords like "stop
integrating it, prior to beginning the training procedure. words" are certain to display any literature. Since it doesn’t
LIAR: PolitiFact, a fact-checking website, supplied the provide more information about the data, eliminate them. Lead,
dataset via its API. It consists of 12,836 human-labeled brief for example, acceptable to me -> lead, fine, silver.
utterances extracted from press releases, political speeches, TV
and radio interviews, and other venues. News veracity is d) Converging-By reducing a word to its stem form,
categorized using a fine-grained set of classes, including mostly stemming aids in word reduction. Treating related words
2. untrue, almost real, partially true, largely true, pants-fire, as similarly makes sense most of the time. Using a straightforward
well as true. The set of new Benchmark for the detection of bug rule-based methodology, it does this by getting rid of suffices
news, dubbed by “Liar, Liar Pants on Fire” William Yang Wang. like ing‖, ly‖, s‖, and so on. The word corpus is growing, yet
In this a brief paper will be presented by the association for words are still frequently disregarded. A good example would
computation linguistics (ACL 2017) at the fifth biannual be entitled -> entitled. Bear in mind that certain search engines
symposium took place from the thirty-first of July to the fourth categorize words that share a stem with synonyms.
of August 2018, in the Canadian city of Vancouver.Below is a
summary of the columns that were utilized to create the three ii) Creation of Features
datasets that were utilized in this investigation: From text of data many properties can be described as word
of count, large word frequency of words, N grams. Computers
● in the first row (written material or title for an may be able to comprehend text and carry out operations like
assertion of information). grouping and categorization as we can create a specimen of
● Descriptor in Row 2 (adherence, inaccurate make up words which involves its meanings, relationships of semantic
the description category). and the condition of different types were used.
The train.csv, test.csv, and valid.csv datasets were utilized iii)Vectorizing Data
for this research and were in the CSV format. By encoding text in integer, or numeric, form, the
FAKE OR REAL? CSV: This dataset was used to create a vectorizing process creates feature vectors that our data may be
passive-aggressive classifier. Three columns are there in it: interpreted by machine learning algorithms.
1. Text or keyword, 2. Statement, 3. False or True Label. iv)Data Vectorization:
B. Termination And Explanation Bag-Of-Words Count Word presence in the text data is
characterized by Vectorizer or Bag of Words (Bow). If it is
i)Data Preparation present in the sentence, the outcome is 1; if not, it is 0. As such,
The great bulk of the content on social media is it creates a word bag for every text document based on the
unstructured, informal communication riddled with slang, document matrix count.
typos, bad language, and other mistakes. It is now crucial to v) N-Grams for Vectorizing Data
develop resource usage strategies in order to make informed
judgements due to the requirement to increase performance and All potential pairings of neighboring N grams are phrases
reliability. Before being utilized for predictive modelling, the or characters with an extension of n that may display in our
data needs to be cleaned in order to yield better insights. For source material. When(n=1) a unigram is a N gram. In the same
this reason, basic preprocessing can be completed by the data way, you may also use trigrams (n=3), bigrams (n=2), and so
of information training. The data cleaning was included in this forth. Generally speaking, bigrams and trigrams are more
phase. When we read, the content is either organized or informative than unigrams N grams basic concept is to predict
unstructured. Unstructured data is devoid of a suitable structure, the alphabet or syllable that's going to most likely happen after
whereas structured formats follow a well-defined structure. The the supplied word. More background must be dealt with in a
semi structured format classified between the two categories longer n-gram (higher n).
such as in contrast to an unstructured format, it is more vi)Data Vectorization
structured and the machine learning needs the characterize as The "comparable prevalence" of an expression among the
the system to recognize the input text must be cleaned. paper is calculated using TF-IDF, which compares the word's
Preprocessing, sometimes referred to as cleaning, typically prevalence in all papers. A term's TF-IDF weight indicates how
entails numerous steps: important it is in relation to other terms in the document and the
a) Remove-punctuation-Punctuation can improve our entire corpus [16]. Used for summarization of text, documents
understanding by providing grammatical context for a sentence. groupings and search engine scoring.
However, our vectorizer adds little value because it only counts 𝑇𝐹(𝑡, 𝑑) = (𝐷𝑜𝑐𝑢𝑚𝑒𝑛𝑡 ⬚′𝑑 ′ 𝑡𝑜𝑡𝑎𝑙 𝑤𝑜𝑟𝑑𝑠 𝑐𝑜𝑢𝑛𝑡) /
words without considering context, therefore Every special (𝑇ℎ𝑒 𝑑𝑜𝑐𝑢𝑚𝑒𝑛𝑡 ⬚′𝑑 ′ 𝑜𝑐𝑐𝑢𝑟𝑒𝑑 𝑖𝑛 𝑛𝑢𝑚𝑏𝑒𝑟 𝑜𝑓 𝑡𝑖𝑚𝑒𝑠 ′ 𝑡′)(1)
character is eliminated. For example: How are you feeling? -> The IDF is abbreviation. If a word appears in every paper,
How are things going for you? it is not highly valuable. Some nouns, such as "a," "an," "the,"
b) Tokenization-Written word is separated into words or "on," "of," etc., are usually used yet have little meaning in texts.
phrases by tokenization. It organizes previously disorganized IDF makes rare phrases more important while making typical
text. Plata o Plomo, for instance, becomes Plomo, 'o', and Plata'. phrases less relevant. The more unique a phrase is, the greater
its IDF value.

Authorized licensed use limited to: Zhejiang University. Downloaded on March 23,2025 at 15:53:44 UTC from IEEE Xplore. Restrictions apply.
𝑇𝐷𝐹(𝑡, 𝑑) = (𝐴𝑚𝑜𝑢𝑛𝑡 𝑜𝑓 𝑑𝑜𝑐𝑢𝑚𝑒𝑛𝑡𝑠 𝑜𝑣𝑒𝑟𝑎𝑙𝑙) / ln(p/(1-p)) = ln(odds) (4)
(𝑇𝑜𝑡𝑎𝑙 𝑛𝑢𝑚𝑏𝑒𝑟 𝑜𝑓 𝑑𝑜𝑐𝑢𝑚𝑒𝑛𝑡𝑠 𝑤𝑖𝑡ℎ 𝑝ℎ𝑟𝑎𝑠𝑒 𝑡 𝑖𝑛 𝑖𝑡) (2) 4. Classifier with Passive Aggressive Mode - The design of the
When TF IDF have relevance to sentence of body then algorithm of passive aggressive is for internet classification of
document of the matrix is recorded by every phrase of every large data streams, such as Twitter data. It quickly learns from
text and sentence to the relevant of the count. Matrix sparseness examples, discards them, and remains passive for correct
the result of vectorizers. A sparse matrix is one where many of classifications but becomes aggressive in case of errors,
the elements is zeros. The classifier's training is covered in this continuously updating and adjusting without convergence. Its
section. We looked at a number of classifiers to predict the text's aim is to make updates that correct losses with minimal change
class. To be precise, we examined Naïve bayes, Logistic to the weight vector's norm.
Regression, passive aggressive classifier are the three machine D. Evaluation Matrices
learning algorithms. The Python software Sci-Kit Learn was
used to implement these classifiers. Use a range of evaluation measures to assess how well the
algorithms performed in resolving the false news detection
C. Short Overview Of Algorithm challenge. In this piece, we examine the most popular methods
1. Unsupervised Gaussian Estimator - As the Bayes theorem, for identifying false information. The majority of current
which asserts that a features membership. Any of the other strategies view the issue of fake news as one of classification,
feature is independent as existence of the class serves as the one that determines the veracity of a news report:
foundation for this classification technique. It provides an Authentic Positive (TP): While bogus news reports that were
algorithm to compute the probability of the backend. assumed to be accurate reveal out to be inaccurate be, that is
an authentic positive (TP);
𝑝(𝑥) =
𝑝(𝑐|𝑥)∗𝑝(𝑐)
(3) factual Negative (TN): stories are supposed to be factual news
𝑝(𝑥) are inadvertently mislabeled as such.
p(c/x) = backend plausibility regarding the specified category False Negative (FN): labeling news articles that are expected to
forecast. be factual as false news;
p(c) = preparatory plausibility. False Positive (FP): publications containing phony news that
p(x/c) = predictor given the forecast. were assumed to be true but turn out not to be.
p(x) = preparatory plausibility.
1) Confusion Matrix- The performance of confusion matrix is a
2.Random Forest - Leveraging the results of numerous decision classification of model which is used for classifier as shown in
trees in the random forest methodology is an effective the below table. When confusion matrix is applied to the test of
communal learning method. Each choices decision is built data set whose real and moral values are known. The
independently, using different subsets of data created through performance of algorithms is visualized through the matrix. To
bootstrapping. The results of these individual trees are then inaccurate the projection of the class to count the values which
aggregated through methods like simple averaging or majority are employed to describe the accurate of the radiation. All of
voting. When making predictions for new input data, the the predictions made for a particular classification task are
Random Forest algorithm provides an average of predictions summarized in a confusion of the matrix. The illustration of the
from all the classifiers. Random Forest has demonstrated its Confusion Matrix is to confusion that arises during prediction-
ability to enhance performance and reduce variance compared making in your classification model. It provides information
to using a single classifier. It also exhibits robustness when about the kinds of mistakes a classifier is making, which is more
dealing with outliers and missing values. These qualities make significant than just counting errors.
Random Forest a widely adopted algorithm across various
TABLE I. CONFUSION OF MATRIX
applications. The process of bagging minimum modification to
Total Phase 1(Awaited) Phase 2(Awaited)
the training set might yield noticeably different tree designs.
Features Uncertainty is when splitting a node in a conventional Phase1
decision tree we consider all the features and select one by one (Real) True positive False negative
as for the result. Phase 2
3.Logistic Regression – This is a categorization, not an (Real) False positive False negative
algorithm. Given a set of independent variables(s), that are
estimated of method for creating an exponential connection
amongst several uncorrelated variables and the dependent of By defining taking into account this categorization task,
variable. Single-variate linear analysis is the term used when the subsequent criteria may be determined:
there is just one autonomous attribute and multiple of attributes.
The equation derived from Linear Regression results logically, |𝑇 𝑃|
then, linear combination predictor variables would be the 1. Preciseness = |𝑇 (5)
𝑃|+|𝐹 𝑃|
representation of the result's log probability. Odds ar equal to |𝑇 𝑃|
p/(1-p) or the chance of an event occurring as opposed to its not 2. Revocation = |𝑇 (6)
𝑃|+|𝐹 𝑁|
occurring.

Authorized licensed use limited to: Zhejiang University. Downloaded on March 23,2025 at 15:53:44 UTC from IEEE Xplore. Restrictions apply.
|𝑇 𝑃|+|𝑇 𝑁|
3. Exactness= |𝑇 𝑃|+|𝑇 𝑁|+|𝐹 𝑃|+|𝐹 𝑁| (7)
𝑃𝑟𝑒𝑐𝑖𝑐𝑖𝑜𝑛∗𝑅𝑒𝑐𝑎𝑙𝑙
4. F1 Score = 2 ∗ 𝑃𝑟𝑒𝑐𝑖𝑠𝑖𝑜𝑛+𝑅𝑒𝑐𝑎𝑙𝑙 (8)

The above equations will help us to determine the


preciseness, Revocation, Exactness and F1 score of the
classifier. By calculating all those three we come to know that
Which one is better, and which one gives more accuracy to the
user.
V. RESULT
In this research, we investigate Assessing the efficacy of
algorithms for neural networks predicated on facts SVM, Linear Fig. 4. Comparison of revocation
Regression, Random Forest using real time data collected from
the social media by using vector features and count of vectors From Fig 4 we can observe that Naïve Bayes is predicted
and with vectors of TF-IDF with validation cross technique of with recall of 0.75, For Logistic Regression 0.81, whereas for
k fold to upgrade the productiveness. After analyzing various Random Forest 0.84. As result Random Forest is having highest
extracted features from the Logistic Regression, SVM, Naïve Recall.
Bayes, passive Aggressive classifier, Random Forest. For
predicting sets their confusion matrix shown below.

TABLE II. COMPARISION OF THE THREE CLASSIFIERS.


Preciseness Revocation F1- Exactness
Analysis Score
Naïve 0.78 0.75 0.56 73.5%
Bayes
Linear 0.53 0.81 0.62 93.7%
Regression

Random 0.61 0.84 0.66 94.5%


Forest

Fig. 5. Comparison of f1 score

From Fig 5 we can observe that Naïve Bayes is predicted


with F1 score of 0.56, For Logistic Regression 0.62, whereas
for Random Forest 0.66. As result Random Forest is having
highest F1 score.

Fig. 3. Comparison of preciseness

From Fig 3 we can observe that Naïve Bayes is predicted


with precision of 0.78, For Logistic Regression 0.53, whereas
for Random Forest 0.61. As result Random Forest has the
highest precision.
Fig. 6. Comparison of accuracy

Authorized licensed use limited to: Zhejiang University. Downloaded on March 23,2025 at 15:53:44 UTC from IEEE Xplore. Restrictions apply.
As we can see from above results, Naïve Bayes having [5] Conroy, N., Rubin, V. and Chen, Y. (2015)―Automatic deception
detection: Methods for finding fake news‖ at Proceedings of the
accuracy of 73%, Logistic Regression having accuracy of 93%, Association for Information Science and Technology, 52(1), pp.1-4.
Random Forest with an accuracy of 94%. As a result, Random [6] Markines, B., Cattuto, C., & Menczer, F. (2009, April). ―Social spam
Forest turned out to be our best model. To boost Random Forest detection‖. In Proceedings of the 5th International Workshop on
Adversarial.
execution, it becomes evident that the Random Forest [7] Nair, D. Deva Hema, ―Fake News Detection using Machine Learning
algorithm outperforms all the others. and Natural Language Processing,‖ International Journal of Recent
Technology and Engineering (IJRTE) ISSN: 2277-3878, Volume-7,
VI. CONCLUSION Issue-6, March 2019.
[8] H. Gupta, M. S. Jamal, S. Madisetty and M. S. Desarkar, "A
In our study, we analyzed the results of a few models, such framework for real-time spam detection in Twitter," 2018 10th
as Naïve bayes, Logistic Regression, passive aggressive International Conference on Communication Systems & Networks
(COMSNETS), Bengaluru, 2018, pp. 380-383.
classifier. The performance of these algorithms is accessed by
[9] M. L. Della Vedova, E. Tacchini, S. Moret, G. Ballarin, M. DiPierro
calculating the precision and recall with F1 score. After and L. de Alfaro, "Automatic Online Fake News Detection
analyzing the outcomes obtained from these different Combining Content and Social Signals," 2018 22nd Conference of
algorithms. We arrived at a clear conclusion that Random Forest Open Innovations Association (FRUCT), Jyvaskyla, 2018, pp. 272-
279.
algorithm consistently delivered the highest precision and recall
[10] C. Buntain and J. Golbeck, "Automatically Identifying Fake News
to detect the fake news. It is effectiveness in providing the in Popular Twitter Threads," 2017 IEEE International Conference on
detection of fake news whether the information is real or fake. Smart Cloud (SmartCloud), New York, NY, 2017, pp. 208-215.
The fake news detection mainly helps in consuming the time of [11] S. B. Parikh and P. K. Atrey, "Media-Rich Fake News Detection: A
Survey," 2018 IEEE Conference on Multimedia Information
the user to detect the information. After analyzing all the
Processing and Retrieval (MIPR), Miami, FL, 2018, pp. 436-441
classifiers Random Forest is the best for detection of misleading [12] Dataset- Fake News detection William Yang Wang. " liar, liar pants
information. on _re": A new benchmark dataset for fake news detection. arXiv
preprint arXiv:1705.00648, 2017.
REFERENCES [13] Shankar M. Patil, Dr. Praveen Kumar, ―Data mining model for
effective data analysis of higher education students using
[1] Kai Shu, Amy Sliva, Suhang Wang, Jiliang Tang, and Huan Liu,
MapReduce‖ IJERMT, April 2017 (Volume-6, Issue-4).
―Fake News Detection on Social Media: A Data Mining Perspective‖
arXiv:1708.01967v3 [cs.SI], 3 Sep 2017. [14] Aayush Ranjan, ― Fake News Detection Using Machine Learning‖,
[2] M. Granik and V. Mesyura, "Fake news detection using naive Bayes Department Of Computer Science & Engineering Delhi
classifier," 2017 IEEE First Ukraine Conference on Electrical and Technological University, July 2018.
Computer Engineering (UKRCON),Kiev, 2017, pp. 900-903. [15] Patil S.M., Malik A.K. (2019) Correlation Based Real-Time Data
Analysis of Graduate Students Behaviour. In: Santosh K., Hegadi R.
[3] Fake news websites. (n.d.) Wikipedia. [Online]. Available:
(eds) Recent Trends in Image Processing and Pattern Recognition.
https://en.wikipedia.org/wiki/Fake_news_website. Accessed Feb. 6,
RTIP2R 2018. Communications in Computer and Information
2017.
Science, vol 1037. Springer, Singapore.
[4] Cade Metz. (2016, Dec. 16). The bittersweet sweepstakes to build an
AI that destroys fake news.

Authorized licensed use limited to: Zhejiang University. Downloaded on March 23,2025 at 15:53:44 UTC from IEEE Xplore. Restrictions apply.

You might also like