A Comparative Study of Using Boosting-Based Machine Learning Algorithms For Iot Network Intrusion Detection
A Comparative Study of Using Boosting-Based Machine Learning Algorithms For Iot Network Intrusion Detection
A Comparative Study of Using Boosting-Based Machine Learning Algorithms For Iot Network Intrusion Detection
https://doi.org/10.1007/s44196-023-00355-x
RESEARCH ARTICLE
Abstract
The Internet-of-Things (IoT) environment has revolutionized the quality of living standards by enabling seamless connec-
tivity and automation. However, the widespread adoption of IoT has also brought forth significant security challenges for
manufacturers and consumers alike. Detecting network intrusions in IoT networks using machine learning techniques shows
promising potential. However, selecting an appropriate machine learning algorithm for intrusion detection poses a consid-
erable challenge. Improper algorithm selection can lead to reduced detection accuracy, increased risk of network infection,
and compromised network security. This article provides a comparative evaluation to six state-of-the-art boosting-based algo-
rithms for detecting intrusions in IoT. The methodology overview involves benchmarking the performance of the selected
boosting-based algorithms in multi-class classification. The evaluation includes a comprehensive classification performance
analysis includes accuracy, precision, detection rate, F1 score, as well as a temporal performance analysis includes training
and testing times.
Keywords Internet-of-Things · Machine learning · Cyber security · Intrusion detection · Extreme boosting · Light boosting ·
Categorical boosting · Supervised learning
patches that adds a difficulty to an already challenging envi- algorithm for IoT network intrusion detection, this research
ronment [2]. This situation makes it difficult to address can guide the development of robust and efficient intrusion
known vulnerabilities and apply security fixes promptly, detection systems tailored to the unique characteristics and
leaving devices exposed to known attacks. constraints of IoT environments. Furthermore, the insights
IoT standards and protocols are still evolving [3], resulting gained from this study can inform the design of proactive
in a lack of uniform security practices across different IoT security measures to mitigate the risks associated with IoT
devices and ecosystems. Inconsistent security implementa- network intrusions.
tions can create vulnerabilities, as attackers can exploit weak The main contributions of this paper are four folds:
links in the network. Especially, with considering that IoT
devices are often deployed in physically exposed and uncon- 1. Examining the literature of using boosting-based ML
trolled environments, such as industrial settings or public algorithms in IoT network intrusion detection.
infrastructure [4]. This physical exposure increases the risk 2. Conducting an exploratory data analysis (EDA) to N-
of physical tampering, unauthorized access, and device com- BaIoT data set [8] to analyze and summarize their main
promise. characteristics and features.
Securing IoT is the only solution for supporting its spread- 3. Investigating the potential of boosting-based methods
ing or decaying is the alternative. Detecting and mitigating for detecting IoT botnet attacks through an experimental
intrusions in IoT networks is of paramount importance to performance evaluation of six boosting-based ML algo-
safeguard sensitive data, ensure privacy, and maintain the rithms representing boosting technique-based algorithms
integrity of IoT systems. Network intrusion detection plays ADB, GDB, XGB, CAB, HGB, and LGB.
a crucial role in providing real-time protection for IoT envi- 4. Benchmarking the six models through a computational
ronment. It is used to monitor network traffic and distinguish analysis to gain more insight into how light they are to
between normal and abnormal network behaviors. Tradi- an IoT environment.
tional network intrusion detection systems (NIDS) may not
be well-suited to address the unique characteristics and The remaining sections of this paper are structured as fol-
challenges presented by IoT networks [5]. The scale, hetero- lows. Section 2 surveys the related work. Section 3 presents
geneity, and resource constraints of IoT devices necessitate a background for the boosting-based ML algorithms. Sec-
innovative approaches to effectively detect and respond to tion 4 demonstrates the evaluation scheme. It describes the
network intrusions. used data set, shows the data set preprocessing and the evalua-
Incorporating Machine Learning (ML) into the defense tion metrics. Section 5 introduces the experimental results for
architecture has shown promise in this domain. It contributes model performance evaluation. Section 6 provides the con-
in achieving higher detecting accuracy rates in addition to the clusion of this work and provides possible future research
capability of detecting zero day infections [6]. Boosting is an directions.
ensemble modeling technique that refers to improving ML
algorithms predictive accuracy through combining weak or
base learning models into strong predictive model [7]. Its core 2 Related Work
idea is to iteratively train the base models, and then combine
their predictions for the sake of improving the accuracy of A number of studies have been proposed for the sake
the overall ensemble model. of detecting network intrusions in IoT environment. This
section investigates papers applying boosting-based ML
The objective of this paper is to conduct a compara- algorithms for detecting intrusions in IoT environments.
tive study on the effectiveness of boosting-based machine A quantitative systematic review approach is followed
learning algorithms for IoT network intrusion detection. It to select relevant studies. An extensive search was con-
conducts a comprehensive comparative study of multiple ducted using scientific electronic search engines on scientific
boosting-based models, i.e., Adaptive Boosting (ADB), Gra- databases including IEEE Xplore, Science Direct, Scopus,
dient Descent Boosting (GDB), Extreme Gradient Boosting and Research Gate. The search is limited to publications
(XGB), Categorical Boosting (CAB), Hist Gradient Boosting written in English and published in scientific journals, con-
(HGB), and Light Gradient Boosting (LGB). This study aims ferences, or theses. All combinations of “machine learning”,
to evaluate their efficacy within the context of IoT network “boosting”, “intrusion detection” and “IoT” were used in the
security and identify the most suitable algorithm for accurate title, abstract, and keywords over the period from 2017 to
and efficient intrusion detection. The findings of this compar- 2023. The focus was only on published work during that
ative study will contribute to the existing body of knowledge period due to the fact that the trigger for this research field was
on IoT network security and intrusion detection. By iden- the reported botnet malware (Mirai) in 2016. US Computer
tifying the most effective boosting-based machine learning Emergency Readiness Team (US-CERT) reported a botnet
123
International Journal of Computational Intelligence Systems (2023) 16:177 Page 3 of 15 177
malware that had disrupted the services of a major US Inter- updated global weights after averaging. They compared using
net provider. It caused a disruption of multiple major websites of ANN and XGB models through BoT–IoT data set. The
via a series of massive distributed denial of service (DDoS) results show that ANN has better performance of 99.99 rather
attacks. It spread quickly and infected thousands of malicious than XGB of 98.96.
endpoints. ML umbrella covers several learning techniques. Khan et al. [18] proposed a proactive interpretable predic-
Boosting algorithms have been around for years and yet it’s tion model to detect different types of security attacks using
only recently when they have become mainstream in the ML the log data generated by heating, ventilation, and air condi-
community. This section surveys and discusses the litera- tioning (HVAC) attacks. Several ML algorithms were used,
ture of network intrusion detection in IoT environment for such as DT, RF, GDB, ADB, LGB, XGB, and CAB. They
boosting-based related work. reported that the XGB classifier has produced the best result
Kumar et al. [9] used a two-step process for identifying with 99.98% accuracy. Their study was performed using the
peer to peer P2P bots which are detection step and analyz- Elnour et al. [19] HVAC systems data set.
ing step. For the classification step, tenfold cross-validation Alissa et al. [20] proposed a DT, an XGB model, and a
is used on Random Forest (RF), Decision Tree (DT) and logistic regression (LR) model. They used UNSW-NB15 data
XGB. Their approach achieved detection rate of 99.88%. set with applying features correlation technique resulting in
They trained the model for P2P botnet detection using traffic discarding nine features. They reported that the DT outper-
from three botnets namely Waledac, Vinchuca, and Zeus. formed with 94% test accuracy with slight higher accuracy
Liu et al. [10] studied eleven ML algorithms for detect- that XGB while LR achieved the worst accuracy.
ing intrusions in Contiki-NG-Based IoT Networks. They Al-Haija et al. [21] proposed an ensemble learning model
reported that XGB achieved the best performance with 97% for botnet attack detection in IoT. Their approach is to
accuracy using the NSL–KDD data set. applying the voting-based probability to ensemble three
Alqahtani et al. [11] proposed Fisher-score for reducing ML classifiers, i.e., ADB, Random under sampling boosting
the number of features for an IoT botnet attack detection. model (RUS), and bagged model. The individual perfor-
Their approach used a genetic-based extreme gradient boost- mance for the selected classifiers was 97.30, 97.70, and
ing (GXGB) model. Fishers score allows them to select only 96.20, respectively. While the performance of the proposed
three out of 115 data features of the N-BaIoT data set [8]. ensemble model was 99.60%.
Their approach achieved an accuracy of 99.96%. Dash et al. Garg et al. [22] compared the performance of boosting
[12] proposed a multi-class Adaptive Boost model (ADB) for techniques with the non boosting ensemble-based tech-
predicting the anomaly type. They used IoT security data set niques. They identified two types of attacks: IoT attacks and
from DS2OS [13] for the model evaluation. This data set cov- DDoS attacks as binary class and multiclass output, respec-
ers eight types of anomalies those are data probing, denial of tively. Three data sets were used to for evaluation BoT–IoT,
service (DoS), malicious control, malicious operation, scan, IoT-23 and CIC–DDoS-2019. Two boosting methods were
spying and wrong setup. They reported an anomaly detection used, i.e., XGB and LGB. LGB achieved the best perfor-
accuracy of 95%. mance with an accuracy of 94.79%.
Krishna et al. [14] proposed hybrid approach based on Bhoi et al. [23] proposed an LGB-based model for anoma-
ML and feature selection. The NSL–KDD [8] and NBaIoT lies detection in IoT environment. They used Gravitational
data sets [14] are used with applying feature extraction using Search-based optimization (GSO) for optimizing LGB hyper
Recursive feature elimination (RFE). They reported an accu- parameters and compared with the Particle swarm optimiza-
racy of 99.98%. They compared it with GDB classifier which tion (PSO). They used a simulated IoT sensors data set called
achieved an accuracy of 99.30%. Hazman et al. [15] proposed IoT data set that is cited in [24]. They reported an optimal
an approach for intrusion detection in IoT-based smart envi- accuracy of 100%.
ronments with Ensemble Learning called IDS–SIoEL. Their Awotunde et al. [25] proposed a boosting-based model for
approach uses ADB and combining different feature selec- intrusion detection in industrial Internet-of-Things networks.
tion techniques Boruta, mutual information and correlation They investigated the detection of various ensemble classi-
furthermore. They evaluated their approach on IoT-23, BoT– fiers, such as XGB, Bagging, extra trees (ET), RF, and ADB.
IoT [16], and Edge-IIoT data sets. They reported a detection They utilized the Telemetry data of the TON_IoT data sets.
accuracy of 99.90%. The results indicated that XGB showed the highest accu-
Ashraf et al. [17] proposed a federated intrusion detection racy in detecting and classifying IIoT attacks. Rani et al.
system for blockchain enabled IoT healthcare applications. [26] compared several algorithms for intrusion detection in
Their approach is based on using lightweight artificial neu- IoT environments, i.e., LR, RF, XGB, and LGB. They uti-
ral networks in a federated learning way. In addition, it lized DS2OS data set [27] and reported that XGB and LGBM
uses blockchain technology to provide a distributed ledger achieved almost equal accuracy of 99.92%.
for aggregating the local weights and then broadcasting the
123
177 Page 4 of 15 International Journal of Computational Intelligence Systems (2023) 16:177
Table 1 presents a comparative analysis for the previous mented in scikit class sklearn.ensemble.AdaBoostClassifier
related work in tabular form. It lists the surveyed papers in [30].
adopting boosting-based approaches in detection IoT net- The mathematical architecture of the ADB model training
work intrusions and their characteristics. The papers are can be concluded as follows [31]:
ordered chronologically, and their characteristics in terms of
the objective, the employed boosting algorithm(s), the evalu- 1. Initialize the weights of the training samples:
ation data set, the number of classes, the number of features,
and the reported accuracy. wi = 1/N , (1)
1. Initialize the training data set and assign equal weights 3. Output the final boosted model: H (x) = sign(t αt ∗
to each training instance. h t (x)).
2. Train a base learner on the weighted data set.
3. Adjust the weights of misclassified instances to give them ADB is known for its ability to handle complex data sets and
higher importance. achieve high accuracy. It focuses on mis-classified samples,
4. Repeat steps 2 and 3 for a specified number of iterations giving them higher weights in subsequent iterations, lead-
(or until a stopping criterion is met). ing to improved performance. It is resistant to over-fitting
5. Combine the predictions of all weak learners using a and can work well with weak classifiers. However, It can
weighted voting or averaging scheme to obtain the final be sensitive to noisy data and outliers, which can negatively
prediction. impact its performance. It may struggle with data sets that
have imbalanced class distributions.
3.1 Adaptive Gradient Boosting
3.2 Gradient Descent Boosting
Adaptive Gradient Boosting algorithm (AdaBoost or ADB)
of Freund and Schapire was the first practical boosting algo- Gradient descent boosting (GDB) is an extension of boosting
rithm [29]. The algorithm begins by fitting a classifier on the technique where the process of additively generating weak
original data set, and then fits additional copies of the classi- models is formalized as a gradient descent algorithm. The
fier on the same data set. It assigns higher weights to the incor- final prediction is a weighted sum of all of the tree pre-
rectly classified classes and lower weights to the correctly dictions. All its weak learners are decision trees. The idea
classified classes to focus more on difficult cases. The exact behind is to take a weak hypothesis or weak learning algo-
process repeats until the best possible result is achieved and rithm and make a series of tweaks to it, that will improve the
the algorithm has used all the instances in the data. It is imple- strength of the hypothesis/learner. This type of Hypothesis
123
Table 1 Comparative analysis for the related work
Boosting Algorithm Author Ref Year Data set Objective No of classes No of features Accuracy (%)
NBaIoT
EL ADB Hazman [15] 2022 IoT-23, BoT-IoT, NIDS for Smart cities 2 30 99.90
Edge-IIoT IoT
XGB Ashraf [17] 2022 CIC-IDS2018, NIDS for Blockchain 2 10 98.96
N-BaIoT, KDD enabled IoT
(2023) 16:177
Cup 99 Healthcare
Applications
XGB Khan [18] 2022 Elnour et al. HVAC Attack detection for 2 24 99.98
data set HVAC
XGB/DT Alissa [20] 2022 UNSW-NB15 Botnet attack detection 2 40 94.00
in IoT
ADB RUS ELBA Al-Haija [21] 2022 N-BaIoT Botnet attack detection 3 10 97.30 97.70 99.60
in IoT
XGB LGB Garg [22] 2022 BoT-IoT, IoT-23, Attacks Identification: 2 35 94.49 94.79
CIC-DDoS-19 IoT attacks and DDoS
attacks
LGB PSO-LGB Bhoi [23] 2022 IoT data set Identify cation of 2 13 99.99 100.0 100.0
GSA-LGB Malicious Access in
IoT Network
XGB Awotunde [25] 2023 ToN-IoT NIDS for IIoT 7 17 99.73
XGB LGB Rani [26] 2023 DS2OS NIDS for Smart home 8 13 99.92
Page 5 of 15
IoT
177
123
177 Page 6 of 15 International Journal of Computational Intelligence Systems (2023) 16:177
Boosting is based on the idea of Probability Approximately 1. Initialize the model’s predictions:
Correct Learning (PAC). Gradient boosting classifiers are
the Ada-Boosting method combined with weighted mini- F0 (x) = argmin c i L(yi , c), (10)
mization, after which the classifiers and weighted inputs are
recalculated. The objective of Gradient Boosting classifiers where L is the loss function.
is to minimize the loss. It is implemented in scikit class
sklearn.ensemble.GradientBoostingClasifier [30]. 2. For each boosting iteration t = 1 to T :
The mathematical architecture of the GDB model training
can be concluded as follows [32]:
• Compute the negative gradient of the loss function:
1. Initialize the model’s predictions: git = −[δL(yi , F(xi ))/δ F(xi )] F(xi )=Ft−1 (xi ) .
(11)
F0 (x) = argmin c i L(yi , c), (6)
• Compute the second derivative approximation of the
where L is the loss function, and c is the predicted value. loss function:
2. For each boosting iteration t = 1 to T :
• Compute the negative gradient of the loss function: h it = [δ 2 L(yi , F(xi ))/δ F(xi )2 ] F(xi )=Ft−1 (xi ) .
(12)
rit = −[δL(yi , F(xi ))/δ F(xi )] F(xi )=Ft−1 (xi ) .
(7) • Fit a weak learner to the negative gradient and second
derivative:
• Fit a weak learner to the negative gradient:
h t (x) = argmin h i [git ∗ h(xi )
h t (x) = argmin h i L(rit , h(xi )) (8) +0.5 ∗ h it ∗ h(xi )2 ] + (h), (13)
Ft (x) = Ft−1 (x) + η ∗ h t (x) (9) Ft (x) = Ft−1 (x) + η ∗ h t (x), (14)
GDB builds models sequentially, minimizing the loss XGB excels in both speed and performance. It supports
function by gradient descent, resulting in improved perfor- parallel processing and has a comprehensive set of hyper-
mance. However, it can be computationally expensive and parameters for fine-tuning. However, it is sensitive to hyper-
may require careful tuning of hyper-parameters. It is more parameter settings. Selecting the optimal combination of
prone to over-fitting compared to other algorithms. hyper-parameters can be time-consuming and computation-
ally expensive. Additionally, the interpretability of XGB
3.3 Extreme Gradient Boosting models can be challenging due to their complexity.
Extreme Gradient Boosting (XGB) is simply an improved 3.4 Light Gradient Boosting
version of the GBM algorithm. It implements machine learn-
ing algorithms under the Gradient Boosting framework. Its Light Gradient Boosting algorithm (LGB) is an ensemble
working procedure is the same as GBM, except that XGB learning method. It is an implementation of Gradient Boosted
implements parallel pre-processing at the node level which Decision Trees (GBDT) similar to random forest [35]. It
makes it generally over ten times faster than GBM [33]. combines multiple decision trees to obtain a better predic-
XGB also includes a variety of regularization techniques that tion. LGB algorithm is an implementation of GBD [35]. It
reduce over fitting and improve overall performance. The uses boosting to eliminate the residual error. LGB is able to
mathematical architecture of the XGB model training can be handle huge amounts of data with ease. It does not perform
concluded as follows [34]: well with a small number of data points. The trees in LGB
123
International Journal of Computational Intelligence Systems (2023) 16:177 Page 7 of 15 177
have a leafwise growth, rather than a levelwise growth. After 3.5 Categorical Boosting
the first split, the next split is done only on the leaf node
that has a higher delta loss. To speed up the training pro- Categorical Boosting is an algorithm for gradient boosting
cess, LGB uses a histogram-based method for selecting the on decision trees (also know as CatBoost or CAB) [37].
best split. Observing the high training time requirement for It is a special version of GBDT. It solves problems with
gradient boosting decision trees (GBD), Ke et al. [28] pro- ordered features while also supporting categorical features.
posed two novel techniques to overcome the challenge based It shuffles the data randomly and mean is calculated for every
on Gradient-based One-Side Sampling (GOSS) and Exclu- object only on its historical data. It constructs combinations
sive Feature Bundling (EFB). This new implementation was in a greedy way. It incorporates an ordered boosting with
named LGB, and it improved training and inference time of a permutation driven alternative to the conventional gradient
GBD by 20%. boosting. Such permutations decrease the final model predic-
The mathematical architecture of the LGB model training tions’ variance compared to the general boosting algorithm
can be concluded as follows [36]: [38].
The mathematical architecture of the CAB model training
1. Initialize the model’s predictions: can be concluded as follows [39]:
• Compute the second derivative approximation of the • Fit a weak learner to the pseudo-residuals and the
loss function: categorical features.
• Update the model’s predictions:
h it = [δ 2 L(yi , F(xi ))/δ F(xi )2 ] F(xi )=Ft−1 (xi ) .
(17) Ft (x) = Ft−1 (x) + η ∗ h t (x). (21)
3. Output the final boosted model: H (x) = FT (x). 3.6 Hist Gradient Boosting
LGB utilizes a leafwise tree growth strategy and gradient- Histogram-based Gradient Boosting Classification Tree (HGB)
based optimization, resulting in faster training times and is much faster than Gradient Boosting Classifier for big
lower memory usage. However, it may not perform well data sets. Its implementation is inspired by LGB. Dur-
when dealing with smaller data sets. It is more sensitive to ing training based on the potential gain, the tree grower
over-fitting and may require careful regularization. The inter- learns at each split point whether samples with missing
pretability of LGB models can be challenging due to their values should go to the left or right child. When pre-
complex nature. dicting, samples with missing values are assigned to the
123
177 Page 8 of 15 International Journal of Computational Intelligence Systems (2023) 16:177
Features
left or right child consequently. If no missing values were
encountered for a given feature during training, then sam-
35
15
35
ples with missing values are mapped to whichever child
5
5
5
concluded as follows [40]:
7
1. Initialize the model’s predictions:
Stream aggregation designation Stream aggrega- Weight Mean Variance/standard deviation Magnitude Radius Covariance Correlation Coefficient
F0 (x) = argmin c i L(yi , c), (22)
X
X
X
git = −[δL(yi , F(xi ))/δ F(xi )] F(xi )=Ft−1 (xi ) .
(23)
X
X
X
• Compute the second derivative approximation of the
loss function:
X
X
X
h it = [δ 2 L(yi , F(xi ))/δ F(xi )2 ] F(xi )=Ft−1 (xi ) .
(24)
X
X
X
• Construct a histogram of the feature values and their
corresponding gradients and second derivatives.
• Find the best split points in the histogram using a
greedy algorithm.
• Compute the leaf values for the histogram bins.
• Update the model’s predictions:
Var
Std
Std
where η is the learning rate.
Host to Host
Host to Host
channel
tion description
Host source IP
destination IP)
channel jitter
4 Evaluation Scheme
(IP+Socket)
channel
Table 2 Data set attributes information
MAC
HpHp
123
International Journal of Computational Intelligence Systems (2023) 16:177 Page 9 of 15 177
and a variety of attack traffic. It was suggested by Meidan et set attributes information. The attacks executed by botnets
al. [8] through gathering traffic of nine commercially avail- include: Scan that can discover vulnerable devices; flooding
able IoT devices authentically infected by Mirai and Bashlite that makes use of SYN, ACK, UDP and TCP flooding; and
malware. The devices were two smart doorbells, one smart combo attacks to open connections and send junk data.
thermostat, one smart baby monitor, four security cameras Our study uses Median’s data set in Naveed organized
and one webcam. Traffic was captured when the devices formats. Figure 2 shows the data exploration for the data
were in normal execution and after infection with malware. set collected by three labeled types, i.e., benign, Mirai and
The traffic was captured through network sniffing utility into Gafgyt. Figure 3 shows the data set individual distribution of
raw network traffic packet capture format (PCAP). It can the 10 malware classes in addition to the benign traffic.
be achieved through using port mirroring. Five features are
extracted from the network traffic as abstracted in Table 2.
Three or more statistical measures are computed for each of 4.2 Data Set Preprocessing
these five features for data aggregation, resulting in a total
of 23 features. These 23 distinct features are computed over Data preprocessing is the process of preparing the data set for
five separate time-windows (100 milliseconds (ms); 500 ms; analysis. It is an essential step in ML as it helps to ensure that
1.5 seconds (s); 10 s; and 1 minute) as demonstrated in Fig. 1. the data is appropriate and correct for feeding into the model.
Using time windows makes this data set appropriate for state- As demonstrated during data set exploration in Sect. 4.1, the
ful IDS and resulting in total of 115 features. Naveed et al. data set is imbalanced and diversified into many files based
[41] organized this data set in an easier file structure and on the attack type, as shown in Fig. 3.
made it available at Kaggle. To integrate the data set files, the data set files are inte-
The data set contains instances of network traffic data grated together into three main categories, i.e., Benign, Mirai,
divided into three categories: normal traffic (Benign data), and Gafgyt. The Bengin category contains all normal traf-
Bashlite infected traffic, and Mirai infected traffic. Each fic records represented in Light Green color in Fig. 3. Mirai
data instance consists of 115 features represented by 23 category includes all Mirai related attacks, i.e., Mirai_Ack,
different traffic characteristics in 5 different time frames. Mirai_Scan, Mirai_Syn, Mirai_Udp, Mirai_Udpplain repre-
Table 2 presents an abstracted demonstration for the data sented in Blue color in Fig. 3. The data set file of this category
is called “All_Mirai”. The third category is Gafgyt and
123
177 Page 10 of 15 International Journal of Computational Intelligence Systems (2023) 16:177
123
International Journal of Computational Intelligence Systems (2023) 16:177 Page 11 of 15 177
123
177 Page 12 of 15 International Journal of Computational Intelligence Systems (2023) 16:177
123
International Journal of Computational Intelligence Systems (2023) 16:177 Page 13 of 15 177
123
177 Page 14 of 15 International Journal of Computational Intelligence Systems (2023) 16:177
123
International Journal of Computational Intelligence Systems (2023) 16:177 Page 15 of 15 177
38. Prokhorenkova, L., Gusev, G., Vorobev, A., Dorogush, A.V., Gulin, 41. Naveed, K., Wu, H., Abusaq, A.: Dytokinesis: A cytokinesis-
A.: Catboost: unbiased boosting with categorical features. Adv. inspired anomaly detection technique for IoT devices. In: 2020
Neural. Inf. Process. Syst. 4, 1–11 (2018) IEEE 45th Conference on Local Computer Networks (LCN), pp.
39. Guo, C., Berkhahn, F.: Entity embeddings of categorical variables. 373–376 (2020)
arXiv preprint arXiv:1604.06737 (2016) 42. Saied, M., Guirguis, S., Madbouly, M.: Review of artificial intel-
40. Guryanov, A.: Histogram-based algorithm for building gradient ligence for enhancing intrusion detection in the internet of things.
boosting ensembles of piecewise linear decision trees. In: Aalst, Engineering Applications of Artificial Intelligence 127, 107231
W.M.P., Batagelj, V., Ignatov, D.I., Khachay, M., Kuskova, V., (2024)
Kutuzov, A., Kuznetsov, S.O., Lomazova, I.A., Loukachevitch,
N., Napoli, A., Pardalos, P.M., Pelillo, M., Savchenko, A.V.,
Tutubalina, E. (eds.) Analysis of Images, Social Networks and
Texts, pp. 39–50. Springer, Cham (2019)
123