Use_of_Supervised_Machine_Learning_Class
Use_of_Supervised_Machine_Learning_Class
Use of Supervised Machine Learning Classifiers for Online Fake Review Detection
Maysara Mazin Badr Alsaad1*, Prof. Dr. Hiren Joshi2
1
PhD Research Scholar, Department of Computer Science, Rollwala Computer Centre, Gujarat
University, Navarangpura, Ahmedabad 380009, Gujarat, India.
[email protected]
2
Professor, Department of Computer Science, Rollwala Computer Centre,
Gujarat University, Navarangpura, Ahmedabad 380009, Gujarat, India.
[email protected]
Abstract
Social media and e-commerce sites have prompted online communities to use reviews to
provide input on goods, products, and services, as well as to support people to analyze
customer input for buying choices, and corporations to improve manufacturing quality.
Internet shoppers support or degrade the reputation of competitive brands. However, the
dissemination of fake reviews fools people, making these reviews a worrying problem. This
study proposes a guided learning online textual content fraudulent review detecting method.
The work splits bogus data using machine learning classifiers and honest reviews.
Experimental findings are compared to assessment measures. Planned system performance is
compared to the baseline. The research comes to the conclusion that supervised machine
learning techniques may be useful in identifying fraudulent reviews, but how well these
techniques work is largely reliant on the characteristics that are chosen. In terms of accuracy,
AUC, and other performance metrics, the SVM classifier with N-Gram feature extraction
and CV feature selection performs better than other classifiers and feature selection
techniques. This is shown by the examination of various feature extraction and selection
techniques. According to the research, N-Gram feature extraction and CV feature selection
may be helpful in spotting fraudulent reviews on e-commerce platforms. This would assist
customers in making wise selections and increase the reliability of online reviews.
Keywords: Social Media, E-commerce, Fake Reviews, Spam Detection Machine Learning,
ML Classification, Naïve Bayes, SVM.
are used in testing for optimizing results. features to handle Arabic social media
Pre-prepared vectors are used to increase material. The idea suggested the system
future preparation knowledge. Mewada et al. employs a novel supervised method mostly
(2022) presented a Rating and Review related to language: Arabic. Interesting
Processing Method to identify the general findings with 91.73% precision for
spam score rates for a review as supportive imbalanced datasets are accomplished. Li et
or unaccommodating based on review al. (2023) introduced CNN.
ratings. The best findings rely on Amazon- Qayyum et al. (2023) find text depiction
scraped product survey data using Python for fake review recognition. The authors
scraping tools. Different studies are to be tried with a record using 2 categorization
introduced later. Jáñez-Martino et al. (2022) methods. This was analyzed using feature
group material for crossover arranging as aggregations of Decision Tree and neural
spam and ham in opinion spam detection. A network algorithms. Implementations
weighting scheme prioritizes spam. demonstrate that sentence-weighted neural
Performance metrics like Accuracy, networks are more useful than other
Precision, and Recall analyze the model’s network-dependent approaches. The location
performance. The shortcoming of the may be calculated to prolong the job weight
approach is a limited set of skills and feature of each phrase using a memory network-
selection is manual. based approach. Bali proposes ML and N-
Aljabri et al. (2023) used several gram study (Bali et al., 2019) for fake news
supervised learning classification systems to estimation. Classifiers Random Forest and
identify spam reviews. Their system reached Naïve Bayes are utilized in feature
85.63% accuracy. The research may be extraction. With 92% accuracy, the outcome
expanded by studying unsupervised machine is excellent. This method works on other
learning and semi-supervised machine recent datasets. Active learning was used by
learning methods. Li, J., et al. (2023) Yan et al. (2023) to identify misleading and
suggested an unsupervised online honest feedback. Only authorized users can
recognition approach spams, an approach review. Productive, negative, and impartial
which is now a density-dependent anomaly reviews are categorized utilizing Natural
estimation problem. Language Processing and text mining
The suggested research comprises many approaches.
phases: (1) Evaluate viewpoint rating count, Martis et al. (2023) categorize J48
(2) Aspect rating-dependent local outlier classifiers reviews of films by spectators.
factor technique (3) Viewpoint positioning TP, TN, and Accuracy are contrasted
for spam reviews. The results proved the between J48 and Random Forest algorithms.
model is persuasive and outperforms current Eshtehardian et al. (2022) suggested a
methods. Kaddoura et al. (2023) developed a continual fake detection method, which
spam detection algorithm using chosen surveys, and gathers spammers
1
(TFI 84.0
SVM 858.00 84 8.38 7.57 85 83 84 83 85 84
DF) 5
84.0
L.R 0.29 84 8.35 7.58 85 83 84 83 85 84
7
73.9
D.T 10.10 74 13.06 13.03 74 74 74 74 74 74
1
81.2
A.B 53.60 81 9.69 9.03 82 81 81 81 82 81
8
85.3
N.B 0.04 85 6.74 7.88 85 87 86 86 84 85
7
84.2
R.F 348.00 84 8.57 7.15 85 83 84 83 86 84
CV 8
n=10 86.5
SVM 27931.00 87 6.75 6.74 87 87 87 86 86 86
0K 1
86.9
L.R 0.56 87 6.49 6.53 87 87 87 87 87 87
(N- 8
Gram) 75.9
D.T 18.60 76 12.97 11.07 77 74 76 75 78 76
5
84.0
A.B 56.20 84 8.14 7.79 84 84 84 84 84 84
7
This table shows the results of several - Accuracy (%): the percentage of
machine learning models trained to classify correctly classified reviews.
reviews as either truthful (Class == 1) or - AUC (%): the area under the Receiver
fake (Class == 0). The models were trained Operating Characteristic (ROC) curve
on feature vectors extracted using different that measures the trade-off between the
methods, namely Word Vectors (WV) and True Positive Rate (TPR) and the False
Count Vectors (CV), with different Positive Rate (FPR) of the classifier.
vocabulary sizes (n=20K and n=100K), and - Type I Error (%): the % of falsely
using different classifiers (NB, RF, SVM, categorized fake reviews.
LR, DT, and Ada Boost). The testing result - Type II Error (%): the % of falsely
reports several performance metrics for each categorized truthful reviews.
combination of feature extraction method, - Precision (%): the % of rightly
classifier, and vocabulary size. These categorized reviews among all reviews
metrics include: categorized as truthful (Class == 1).
- Learning Time(S): the time it took to - Recall (%): the % of correctly
train the model. categorized reviews among all truthful
reviews (Class == 1).
- F-score (%): the harmonic mean of not hold in high-dimensional feature spaces,
precision and recall, which balances the which could limit its performance. The
trade-off between them. Decision Tree classifier generally performed
poorly compared to the other classifiers,
The results also showed that the best- with lower accuracy and higher type I and
performing models achieved an accuracy of type II error rates.
around 87% and an AUC of around 87%, Decision trees are simple and
indicating that they are able to distinguish interpretable models that recursively split
between truthful and fake reviews with a feature space into regions based on
high degree of accuracy. The SVM and thresholds, but they may suffer from
Logistic Regression classifiers generally overfitting and instability, especially in
performed better than the other classifiers, high-dimensional spaces. The N-Gram
while the Naive Bayes classifier had the method generally performed better than the
shortest learning time. The N-Gram method Word Vector method, which may be
generally performed better than the Word attributed to its ability to capture local
Vector method. Increasing the vocabulary dependencies between words in the text,
size from 20K to 100K generally improved which are important for detecting patterns of
performance. However, the Decision Tree deception and sentiment. However, the
classifier generally performed poorly, with Word Vector method may be more suitable
low accuracy and high type I and type II for capturing global semantic relationships
error rates. between words and phrases. Increasing the
The results in table 7 above show that vocabulary size from 20K to 100K generally
the SVM and Logistic Regression classifiers improved performance, indicating that a
generally performed the best across most larger vocabulary can capture more fine-
combinations of feature extraction methods grained distinctions between words and
and vocabulary size, achieving accuracy and phrases, thus enhancing the accuracy of the
AUC scores of up to 87%. These classifiers model. However, this comes at the cost of
are known for their ability to handle high- increased computational complexity and
dimensional data and to learn complex memory requirements. Overall, the results
decision boundaries, which may explain suggest that machine learning models can be
their superior performance in this task. The effective at detecting fake reviews, but the
Naive Bayes classifier generally had the choice of feature extraction approach,
shortest learning time, but its performance vocabulary size, and classifier can have a
was slightly lower than that of the SVM and greater impact on performance. It is essential
Logistic Regression classifiers. Naive Bayes to rigorously assess and compare various
is a simple but effective probabilistic models on a representative dataset to select
classifier that assumes independence the best-performing one for a given task.
between features, but this assumption may
results, including increased accuracy, recall, to reduce noise before providing text to the
precision, and F-measure. The suggested ML classifier. SVM outperforms other ML
strategy outperforms baseline techniques, as classifiers such as XGBoost, KNN, Random
shown by the acquired results. Forest, Na¨ıve Bayes, DT, and LR in fake
Through our proposed work, the review classification in experimental results.
following list of opinion mining jobs is As compared to other classifiers, K-Nearest
recommended to assist businesses and Neighborhood has the weakest performance.
merchants in gathering and evaluating a
significant volume of consumer reviews: Limitations of work
a) Sentiment classification, which 1. 1.The imbalanced dataset in this work
indicates if a viewpoint is neutral, led to a low performance of ML
positive (ham), or negative (spam). classifiers.
b) Learning about the attributes of an 2. The dataset is separated into testing and
entity that has been reviewed and training using the random splitting
obtaining the reviewer's viewpoint approach.
regarding a specific item. 3. This work uses just TF-IDF feather
c) Comparative language and the ability engineering.4. The low dataset size
to find relationships between one (5573) in this work impacts classifier
thing and several related objects. outcomes, indicating the necessity to
d) Supervised machine-learning expand the dataset for enhanced
techniques surpassed human outcomes.
judgment in distinguishing between
genuine and false opinions, Future Paths
classifying consumer opinions with A balanced dataset improves ML classifier
the best accuracy between the two. performance.
e) False opinions affect customers in
two ways: 1) they influence them to Acknowledgment
make poor choices when making a The experimental work described in this
purchase, and 2) they cause them to research was conducted in the lab of the
lose faith in online product reviews. Department of Computer Science, Rollwala
Computer Centre, Gujarat University.
4. Conclusion
This work uses supervised ML techniques, References
such as SVM with specified parameters, to Aljabri, M., Zagrouba, R., Shaahid, A.,
categorize content into spam and non-spam Alnasser, F., Saleh, A., & Alomari, D.
reviews. We also tested different ML M. (2023). Machine learning-based
classifiers and reviewed their outcomes. social media bot detection: a
Different preprocessing approaches are used comprehensive literature review. Social
Pramanik, S. (2022). Utilizing Machine Martis, E., Deo, R., Rastogi, S., Chhaparia,
Learning and Deep Learning in K., & Biwalkar, A. (2023). A proposed
Cybesecurity: An Innovative system for understanding the consumer
Approach. Wiley eBooks, 271– opinion of a product using sentiment
293. https://doi.org/10.1002/978111979 analysis. In Advances in intelligent
5667.ch12 systems and computing (pp. 555–
Khanh, P. T., Ngoc, T. T. H., & Pramanik, 568). https://doi.org/10.1007/978-981-
S. (2023). Future of smart agriculture 19-5443-6_42
techniques and applications. Meslie, Y., Enbeyle, W., Pandey, B. K.,
In Advances in environmental Pramanik, S., Pandey, D., Dadeech, P.,
engineering and green technologies Belay, A., & Saini, A. K. (2021).
book series (pp. 365– Machine Intelligence-Based Trend
378). https://doi.org/10.4018/978-1- Analysis of COVID-19 for total daily
6684-9231-4.ch021 confirmed cases in Asia and Africa.
Li, J., Hu, J., Zhang, P., & Yang, L. (2023). In Advances in systems analysis,
Exposing collaborative spammer groups software engineering, and high
through the review-response performance computing book
graph. Multimedia Tools and series (pp. 164–
Applications, 82(14), 21687– 185). https://doi.org/10.4018/978-1-
21700. https://doi.org/10.1007/s11042- 7998-7701-1.ch009
023-14650-4 Mewada, A., & Dewang, R. K. (2022). A
Liu, S., & Lee, I. (2019). Extracting comprehensive survey of various
features with medical sentiment lexicon methods in opinion spam
and position encoding for drug detection. Multimedia Tools and
reviews. Health Information Science Applications, 82(9), 13199–
and 13239. https://doi.org/10.1007/s11042-
Systems, 7(1). https://doi.org/10.1007/s 022-13702-5
13755-019-0072-6 Ngoc, T. T. H., Khanh, P. T., & Pramanik,
Mandal, A., Dutta, S., & Pramanik, S. S. (2023). Smart Agriculture using a
(2021). Machine intelligence of PI from soil monitoring system. In Advances in
geometrical figures with variable environmental engineering and green
parameters using SCILab. In Advances technologies book series (pp. 200–
in systems analysis, software 220). https://doi.org/10.4018/978-1-
engineering, and high performance 6684-9231-4.ch011
computing book series (pp. 38– Padminivalli, S. J. R. K., V., Rao, M. V. P.
63). https://doi.org/10.4018/978-1- C. S., & Narne, N. S. R. (2023).
7998-7701-1.ch003 Sentiment based emotion classification
in unstructured textual data using dual
susceptibility mapping. Bulletin of Zhao, P., Ma, Z., Gill, T., & Ranaweera, C.
Engineering Geology and the (2023). Social media sentiment
Environment, 81(4). https://doi.org/10.1 polarization and its impact on product
007/s10064-022-02615-0 adoption. Marketing Letters, 34(3),
Zhan, P., Qin, X., Zhang, Q., & Sun, Y. 497–
(2023). Output-Only modal 512. https://doi.org/10.1007/s11002-
identification based on auto-regressive 023-09664-9
Spectrum-Guided symplectic geometry
mode decomposition. Journal of
Vibration Engineering &
Technologies. https://doi.org/10.1007/s
42417-022-00832-1